Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 844 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 12 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 248k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
50,672 | 21,302,367,260 | IssuesEvent | 2022-04-15 06:09:36 | PreMiD/Presences | https://api.github.com/repos/PreMiD/Presences | opened | I need help developing an app | Service Request | ### Website name
livelinkbio
### Website URL
http://livelinkbio.com/
### Website logo
https://cdn.discordapp.com/attachments/957097221210390640/960382757102551100/PnzJfuE3_400x400.png
### Prerequisites
- [ ] It is a paid service
- [ ] It displays NSFW content
- [ ] It is region restricted
### Description
basically I want it so that it shows up on the side of Discord to see how long you actually been on the website and the logo | 1.0 | I need help developing an app - ### Website name
livelinkbio
### Website URL
http://livelinkbio.com/
### Website logo
https://cdn.discordapp.com/attachments/957097221210390640/960382757102551100/PnzJfuE3_400x400.png
### Prerequisites
- [ ] It is a paid service
- [ ] It displays NSFW content
- [ ] It is region restricted
### Description
basically I want it so that it shows up on the side of Discord to see how long you actually been on the website and the logo | non_priority | i need help developing an app website name livelinkbio website url website logo prerequisites it is a paid service it displays nsfw content it is region restricted description basically i want it so that it shows up on the side of discord to see how long you actually been on the website and the logo | 0 |
69,438 | 7,134,235,527 | IssuesEvent | 2018-01-22 20:08:44 | servo/servo | https://api.github.com/repos/servo/servo | opened | Enable css-paint-api WPT tests and remove forked duplicates | A-testing E-easy I-cleanup | https://github.com/servo/servo/tree/master/tests/wpt/mozilla/tests/mozilla/css-paint-api is a fork we previously made of https://github.com/w3c/web-platform-tests/tree/master/css/css-paint-api. Since we can now use the vendored copy from upstream (https://github.com/servo/servo/tree/master/tests/wpt/web-platform-tests/css/css-paint-api), we should remove any of the duplicated tests in our copy and [enable](https://github.com/servo/servo/blob/master/tests/wpt/include.ini) the originals.
Use `./mach test-wpt tests/wpt/web-platform-tests/css/css-paint-api` and `./mach test-wpt tests/wpt/mozilla/tests/mozilla/css-paint-api` to run the tests in the two directories. | 1.0 | Enable css-paint-api WPT tests and remove forked duplicates - https://github.com/servo/servo/tree/master/tests/wpt/mozilla/tests/mozilla/css-paint-api is a fork we previously made of https://github.com/w3c/web-platform-tests/tree/master/css/css-paint-api. Since we can now use the vendored copy from upstream (https://github.com/servo/servo/tree/master/tests/wpt/web-platform-tests/css/css-paint-api), we should remove any of the duplicated tests in our copy and [enable](https://github.com/servo/servo/blob/master/tests/wpt/include.ini) the originals.
Use `./mach test-wpt tests/wpt/web-platform-tests/css/css-paint-api` and `./mach test-wpt tests/wpt/mozilla/tests/mozilla/css-paint-api` to run the tests in the two directories. | non_priority | enable css paint api wpt tests and remove forked duplicates is a fork we previously made of since we can now use the vendored copy from upstream we should remove any of the duplicated tests in our copy and the originals use mach test wpt tests wpt web platform tests css css paint api and mach test wpt tests wpt mozilla tests mozilla css paint api to run the tests in the two directories | 0 |
71,167 | 30,824,981,284 | IssuesEvent | 2023-08-01 19:17:14 | hashicorp/terraform-provider-azurerm | https://api.github.com/repos/hashicorp/terraform-provider-azurerm | closed | false positive config change with Azure VNet Subnet service endpoints config | bug service/virtual-networks |
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Issue Description
Terraform reports false positive config change report if we update the same list of Subnet service endpoints but in different order. This is false positive config change report is impacting the infrastructure as a code design.
### Terraform (and AzureRM Provider) Version
0.13.4
### Affected Resource(s)
azurerm_subnet
### Terraform Configuration Files
```hcl
resource "azurerm_subnet" "Subnet001" {
name = "Subnet001"
resource_group_name = azurerm_resource_group.rgterraform.name
virtual_network_name = azurerm_virtual_network.eastasiavnet01.name
address_prefixes = ["192.168.12.0/26"]
service_endpoints = ["Microsoft.ServiceBus", "Microsoft.Sql", "Microsoft.Storage", "Microsoft.Web"]
}
```
### Debug Output
**Current Service Endpoint Config:**
"serviceEndpoints": [
{
"locations": [
"*"
],
"provisioningState": "Succeeded",
"service": "Microsoft.Web"
},
{
"locations": [
"eastasia",
"southeastasia"
],
"provisioningState": "Succeeded",
"service": "Microsoft.Storage"
},
{
"locations": [
"eastasia"
],
"provisioningState": "Succeeded",
"service": "Microsoft.Sql"
},
Apply terraform with the same service endpoints but in different order.
resource "azurerm_subnet" "Subnet001" {
name = "Subnet001"
resource_group_name = azurerm_resource_group.rgterraform.name
virtual_network_name = azurerm_virtual_network.eastasiavnet01.name
address_prefixes = ["192.168.12.0/26"]
service_endpoints = ["Microsoft.ServiceBus", "Microsoft.Sql", "Microsoft.Storage", "Microsoft.Web"]
}
### Panic Output
### Expected Behaviour
there should be no change on the service endpoints.
### Actual Behaviour
# azurerm_subnet.Subnet001 will be updated in-place
~ resource "azurerm_subnet" "Subnet001" {
address_prefix = "192.168.12.0/26"
address_prefixes = [
"192.168.12.0/26",
]
enforce_private_link_endpoint_network_policies = false
enforce_private_link_service_network_policies = false
id = "/subscriptions/7718cc33-171b-4e89-93b7-8edfe4ee2c6f/resourceGroups/RG_19831029/providers/Microsoft.Network/virtualNetworks/19831029-vnet01/subnets/Subnet001"
name = "Subnet001"
resource_group_name = "RG_19831029"
~ service_endpoints = [
**- "Microsoft.Web",
- "Microsoft.Storage",
- "Microsoft.Sql",
"Microsoft.ServiceBus",
+ "Microsoft.Sql",
+ "Microsoft.Storage",
+ "Microsoft.Web",**
]
virtual_network_name = "19831029-vnet01"
}
Plan: 0 to add, **2 to change**, 0 to destroy.
### Steps to Reproduce
1. `terraform apply`
### Important Factoids
terraform apply out.plan
Acquiring state lock. This may take a few moments...
azurerm_subnet.Subnet001: Modifying... [id=/subscriptions/xxxxxxxxxxxxxxxx/resourceGroups/yyyyyyyyyyy/providers/Microsoft.Network/virtualNetworks/19831029-vnet01/subnets/Subnet001]
azurerm_network_security_group.main: Modifying... [id=/subscriptions/xxxxxxxxxx/resourceGroups/yyyyyyyyyyyyyy/providers/Microsoft.Network/networkSecurityGroups/19831029-nsg]
azurerm_subnet.Subnet001: Modifications complete after 2s [id=/subscriptions/xxxxxxxxxxxxx/resourceGroups/yyyyyyyyyyyyy/providers/Microsoft.Network/virtualNetworks/19831029-vnet01/subnets/Subnet001]
azurerm_network_security_group.main: Modifications complete after 5s [id=/subscriptions/xxxxxxxxxxxxx/resourceGroups/yyyyyyyyyyyyy/providers/Microsoft.Network/networkSecurityGroups/19831029-nsg]
Apply complete! Resources: 0 added**, 2 changed,** 0 destroyed.
### References
* #0000
| 1.0 | false positive config change with Azure VNet Subnet service endpoints config -
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Issue Description
Terraform reports false positive config change report if we update the same list of Subnet service endpoints but in different order. This is false positive config change report is impacting the infrastructure as a code design.
### Terraform (and AzureRM Provider) Version
0.13.4
### Affected Resource(s)
azurerm_subnet
### Terraform Configuration Files
```hcl
resource "azurerm_subnet" "Subnet001" {
name = "Subnet001"
resource_group_name = azurerm_resource_group.rgterraform.name
virtual_network_name = azurerm_virtual_network.eastasiavnet01.name
address_prefixes = ["192.168.12.0/26"]
service_endpoints = ["Microsoft.ServiceBus", "Microsoft.Sql", "Microsoft.Storage", "Microsoft.Web"]
}
```
### Debug Output
**Current Service Endpoint Config:**
"serviceEndpoints": [
{
"locations": [
"*"
],
"provisioningState": "Succeeded",
"service": "Microsoft.Web"
},
{
"locations": [
"eastasia",
"southeastasia"
],
"provisioningState": "Succeeded",
"service": "Microsoft.Storage"
},
{
"locations": [
"eastasia"
],
"provisioningState": "Succeeded",
"service": "Microsoft.Sql"
},
Apply terraform with the same service endpoints but in different order.
resource "azurerm_subnet" "Subnet001" {
name = "Subnet001"
resource_group_name = azurerm_resource_group.rgterraform.name
virtual_network_name = azurerm_virtual_network.eastasiavnet01.name
address_prefixes = ["192.168.12.0/26"]
service_endpoints = ["Microsoft.ServiceBus", "Microsoft.Sql", "Microsoft.Storage", "Microsoft.Web"]
}
### Panic Output
### Expected Behaviour
there should be no change on the service endpoints.
### Actual Behaviour
# azurerm_subnet.Subnet001 will be updated in-place
~ resource "azurerm_subnet" "Subnet001" {
address_prefix = "192.168.12.0/26"
address_prefixes = [
"192.168.12.0/26",
]
enforce_private_link_endpoint_network_policies = false
enforce_private_link_service_network_policies = false
id = "/subscriptions/7718cc33-171b-4e89-93b7-8edfe4ee2c6f/resourceGroups/RG_19831029/providers/Microsoft.Network/virtualNetworks/19831029-vnet01/subnets/Subnet001"
name = "Subnet001"
resource_group_name = "RG_19831029"
~ service_endpoints = [
**- "Microsoft.Web",
- "Microsoft.Storage",
- "Microsoft.Sql",
"Microsoft.ServiceBus",
+ "Microsoft.Sql",
+ "Microsoft.Storage",
+ "Microsoft.Web",**
]
virtual_network_name = "19831029-vnet01"
}
Plan: 0 to add, **2 to change**, 0 to destroy.
### Steps to Reproduce
1. `terraform apply`
### Important Factoids
terraform apply out.plan
Acquiring state lock. This may take a few moments...
azurerm_subnet.Subnet001: Modifying... [id=/subscriptions/xxxxxxxxxxxxxxxx/resourceGroups/yyyyyyyyyyy/providers/Microsoft.Network/virtualNetworks/19831029-vnet01/subnets/Subnet001]
azurerm_network_security_group.main: Modifying... [id=/subscriptions/xxxxxxxxxx/resourceGroups/yyyyyyyyyyyyyy/providers/Microsoft.Network/networkSecurityGroups/19831029-nsg]
azurerm_subnet.Subnet001: Modifications complete after 2s [id=/subscriptions/xxxxxxxxxxxxx/resourceGroups/yyyyyyyyyyyyy/providers/Microsoft.Network/virtualNetworks/19831029-vnet01/subnets/Subnet001]
azurerm_network_security_group.main: Modifications complete after 5s [id=/subscriptions/xxxxxxxxxxxxx/resourceGroups/yyyyyyyyyyyyy/providers/Microsoft.Network/networkSecurityGroups/19831029-nsg]
Apply complete! Resources: 0 added**, 2 changed,** 0 destroyed.
### References
* #0000
| non_priority | false positive config change with azure vnet subnet service endpoints config community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment issue description terraform reports false positive config change report if we update the same list of subnet service endpoints but in different order this is false positive config change report is impacting the infrastructure as a code design terraform and azurerm provider version affected resource s azurerm subnet terraform configuration files hcl resource azurerm subnet name resource group name azurerm resource group rgterraform name virtual network name azurerm virtual network name address prefixes service endpoints debug output current service endpoint config serviceendpoints locations provisioningstate succeeded service microsoft web locations eastasia southeastasia provisioningstate succeeded service microsoft storage locations eastasia provisioningstate succeeded service microsoft sql apply terraform with the same service endpoints but in different order resource azurerm subnet name resource group name azurerm resource group rgterraform name virtual network name azurerm virtual network name address prefixes service endpoints panic output expected behaviour there should be no change on the service endpoints actual behaviour azurerm subnet will be updated in place resource azurerm subnet address prefix address prefixes enforce private link endpoint network policies false enforce private link service network policies false id subscriptions resourcegroups rg providers microsoft network virtualnetworks subnets name resource group name rg service endpoints microsoft web microsoft storage microsoft sql microsoft servicebus microsoft sql microsoft storage microsoft web virtual network name plan to add to change to destroy steps to reproduce terraform apply important factoids terraform apply out plan acquiring state lock this may take a few moments azurerm subnet modifying azurerm network security group main modifying azurerm subnet modifications complete after azurerm network security group main modifications complete after apply complete resources added changed destroyed references | 0 |
11,157 | 7,458,495,614 | IssuesEvent | 2018-03-30 10:37:32 | AngleSharp/AngleSharp | https://api.github.com/repos/AngleSharp/AngleSharp | closed | The ConditionalWeakTable becomes a bottleneck with parallel html processing | performance up-for-grabs | We've tried master and devel with production work-loads and found that we aren't able to utilize more than ~50% of the CPU.
Our use-case is parsing and running CSS selectors on millions of HTML documents each day. We noticed that no amount of parallelism would allow the majority of the CPU to be utilised.
I decided to try and track down the limitation, the attached code spins up some threads that each parse and query the same html document.
When I profile the application I see that the hot and contentious parts of the code all involve the ConditionalWeakTable used by Node.cs
I'm unable to think of an optimization that would either reduce contention or improve performance.
[cpu_perf_test.txt](https://github.com/AngleSharp/AngleSharp/files/1634853/cpu_perf_test.txt)
| True | The ConditionalWeakTable becomes a bottleneck with parallel html processing - We've tried master and devel with production work-loads and found that we aren't able to utilize more than ~50% of the CPU.
Our use-case is parsing and running CSS selectors on millions of HTML documents each day. We noticed that no amount of parallelism would allow the majority of the CPU to be utilised.
I decided to try and track down the limitation, the attached code spins up some threads that each parse and query the same html document.
When I profile the application I see that the hot and contentious parts of the code all involve the ConditionalWeakTable used by Node.cs
I'm unable to think of an optimization that would either reduce contention or improve performance.
[cpu_perf_test.txt](https://github.com/AngleSharp/AngleSharp/files/1634853/cpu_perf_test.txt)
| non_priority | the conditionalweaktable becomes a bottleneck with parallel html processing we ve tried master and devel with production work loads and found that we aren t able to utilize more than of the cpu our use case is parsing and running css selectors on millions of html documents each day we noticed that no amount of parallelism would allow the majority of the cpu to be utilised i decided to try and track down the limitation the attached code spins up some threads that each parse and query the same html document when i profile the application i see that the hot and contentious parts of the code all involve the conditionalweaktable used by node cs i m unable to think of an optimization that would either reduce contention or improve performance | 0 |
27,482 | 4,057,186,925 | IssuesEvent | 2016-05-24 21:12:18 | LearnersGuild/rocketchat-lg-slash-commands | https://api.github.com/repos/LearnersGuild/rocketchat-lg-slash-commands | closed | UX: which flex-panels in chat should be permanent and which should be ephemeral | game-design ux | Specifically:
- which commands deserve a permanent "tab button" on the right-hand side
- what icons should be used
- do we want to _remove_ any of the "standard" Rocket.Chat tab buttons
Related: [learning-os-software/#43](https://github.com/LearnersGuild/learning-os-software/issues/43) | 1.0 | UX: which flex-panels in chat should be permanent and which should be ephemeral - Specifically:
- which commands deserve a permanent "tab button" on the right-hand side
- what icons should be used
- do we want to _remove_ any of the "standard" Rocket.Chat tab buttons
Related: [learning-os-software/#43](https://github.com/LearnersGuild/learning-os-software/issues/43) | non_priority | ux which flex panels in chat should be permanent and which should be ephemeral specifically which commands deserve a permanent tab button on the right hand side what icons should be used do we want to remove any of the standard rocket chat tab buttons related | 0 |
179,092 | 30,116,461,301 | IssuesEvent | 2023-06-30 11:58:02 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | [web] TextFormField does not get blurred when tapped outside of the widget. | a: text input severe: regression framework f: material design platform-web f: focus has reproducible steps P3 found in release: 1.22 | <!-- Thank you for using Flutter!
If you are looking for support, please check out our documentation
or consider asking a question on Stack Overflow:
* https://flutter.dev/
* https://api.flutter.dev/
* https://stackoverflow.com/questions/tagged/flutter?sort=frequent
If you have found a bug or if our documentation doesn't have an answer
to what you're looking for, then fill out the template below. Please read
our guide to filing a bug first: https://flutter.dev/docs/resources/bug-reports
-->
## Steps to Reproduce
<!-- You must include full steps to reproduce so that we can reproduce the problem. -->
1. Click on the TextFormField.
2. Now click on any other area of screen.
**Expected results:** <!-- what did you want to see? -->
Expected behaviour is textfield must lose focus.
**Actual results:** <!-- what did you see? -->
The focus is still on Textformfield
**Note:**
This functionality previously worked upto version 1.19.0-4.3.pre (BETA channel)
**Demo Code:**
main.dart
`
import 'package:flutter/material.dart';
import 'dart:html' as html;
void main() {
runApp(MyApp());
}
class MyApp extends StatelessWidget {
@override
Widget build(BuildContext context) {
return MaterialApp(
title: 'TextField Issue',
theme: ThemeData(
primarySwatch: Colors.purple,
),
home: MyHomePage(title: 'TextField Issue'),
);
}
}
class MyHomePage extends StatefulWidget {
MyHomePage({Key key, this.title}) : super(key: key);
final String title;
@override
_MyHomePageState createState() => _MyHomePageState();
}
class _MyHomePageState extends State<MyHomePage> {
FocusNode _focusNode;
@override
void initState() {
_focusNode = FocusNode(
canRequestFocus: true,
);
_focusNode.addListener(() => print("FocusNode changed"));
super.initState();
}
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(
title: Text(widget.title),
),
body: Center(
child: Column(
crossAxisAlignment: CrossAxisAlignment.start,
children: <Widget>[
Container(
padding: EdgeInsets.all(24),
child: TextFormField(
focusNode: _focusNode,
),
),
Container(
padding: EdgeInsets.all(24),
child: Column(
crossAxisAlignment: CrossAxisAlignment.start,
children: [
Text("Demo displaying blur issue on web"),
Text("\n1) Click on the top-most TextFormField above"),
Text("2) Now click on any other area of screen. Basic expected behaviour is textfield must lose focus but it doesnt lose focus."),
Wrap(
children: [
Text("3) The blur was removed in recent PR "),
InkWell(
child: Text('https://github.com/flutter/engine/pull/18743'),
onTap: () {
html.window.open('https://github.com/flutter/engine/pull/18743', '_blank');
},
)
],
),
Text("4) This functionality previously worked upto version 1.19.0-4.3.pre (BETA channel)"),
],
),
),
],
),
),
);
}
}
` | 1.0 | [web] TextFormField does not get blurred when tapped outside of the widget. - <!-- Thank you for using Flutter!
If you are looking for support, please check out our documentation
or consider asking a question on Stack Overflow:
* https://flutter.dev/
* https://api.flutter.dev/
* https://stackoverflow.com/questions/tagged/flutter?sort=frequent
If you have found a bug or if our documentation doesn't have an answer
to what you're looking for, then fill out the template below. Please read
our guide to filing a bug first: https://flutter.dev/docs/resources/bug-reports
-->
## Steps to Reproduce
<!-- You must include full steps to reproduce so that we can reproduce the problem. -->
1. Click on the TextFormField.
2. Now click on any other area of screen.
**Expected results:** <!-- what did you want to see? -->
Expected behaviour is textfield must lose focus.
**Actual results:** <!-- what did you see? -->
The focus is still on Textformfield
**Note:**
This functionality previously worked upto version 1.19.0-4.3.pre (BETA channel)
**Demo Code:**
main.dart
`
import 'package:flutter/material.dart';
import 'dart:html' as html;
void main() {
runApp(MyApp());
}
class MyApp extends StatelessWidget {
@override
Widget build(BuildContext context) {
return MaterialApp(
title: 'TextField Issue',
theme: ThemeData(
primarySwatch: Colors.purple,
),
home: MyHomePage(title: 'TextField Issue'),
);
}
}
class MyHomePage extends StatefulWidget {
MyHomePage({Key key, this.title}) : super(key: key);
final String title;
@override
_MyHomePageState createState() => _MyHomePageState();
}
class _MyHomePageState extends State<MyHomePage> {
FocusNode _focusNode;
@override
void initState() {
_focusNode = FocusNode(
canRequestFocus: true,
);
_focusNode.addListener(() => print("FocusNode changed"));
super.initState();
}
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(
title: Text(widget.title),
),
body: Center(
child: Column(
crossAxisAlignment: CrossAxisAlignment.start,
children: <Widget>[
Container(
padding: EdgeInsets.all(24),
child: TextFormField(
focusNode: _focusNode,
),
),
Container(
padding: EdgeInsets.all(24),
child: Column(
crossAxisAlignment: CrossAxisAlignment.start,
children: [
Text("Demo displaying blur issue on web"),
Text("\n1) Click on the top-most TextFormField above"),
Text("2) Now click on any other area of screen. Basic expected behaviour is textfield must lose focus but it doesnt lose focus."),
Wrap(
children: [
Text("3) The blur was removed in recent PR "),
InkWell(
child: Text('https://github.com/flutter/engine/pull/18743'),
onTap: () {
html.window.open('https://github.com/flutter/engine/pull/18743', '_blank');
},
)
],
),
Text("4) This functionality previously worked upto version 1.19.0-4.3.pre (BETA channel)"),
],
),
),
],
),
),
);
}
}
` | non_priority | textformfield does not get blurred when tapped outside of the widget thank you for using flutter if you are looking for support please check out our documentation or consider asking a question on stack overflow if you have found a bug or if our documentation doesn t have an answer to what you re looking for then fill out the template below please read our guide to filing a bug first steps to reproduce click on the textformfield now click on any other area of screen expected results expected behaviour is textfield must lose focus actual results the focus is still on textformfield note this functionality previously worked upto version pre beta channel demo code main dart import package flutter material dart import dart html as html void main runapp myapp class myapp extends statelesswidget override widget build buildcontext context return materialapp title textfield issue theme themedata primaryswatch colors purple home myhomepage title textfield issue class myhomepage extends statefulwidget myhomepage key key this title super key key final string title override myhomepagestate createstate myhomepagestate class myhomepagestate extends state focusnode focusnode override void initstate focusnode focusnode canrequestfocus true focusnode addlistener print focusnode changed super initstate override widget build buildcontext context return scaffold appbar appbar title text widget title body center child column crossaxisalignment crossaxisalignment start children container padding edgeinsets all child textformfield focusnode focusnode container padding edgeinsets all child column crossaxisalignment crossaxisalignment start children text demo displaying blur issue on web text click on the top most textformfield above text now click on any other area of screen basic expected behaviour is textfield must lose focus but it doesnt lose focus wrap children text the blur was removed in recent pr inkwell child text ontap html window open blank text this functionality previously worked upto version pre beta channel | 0 |
41,754 | 12,839,710,558 | IssuesEvent | 2020-07-07 19:46:13 | doc-ai/tensorio-android | https://api.github.com/repos/doc-ai/tensorio-android | closed | CVE-2017-17485 (High) detected in jackson-databind-2.2.3.jar | security vulnerability | ## CVE-2017-17485 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.2.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to vulnerable library: /tmp/ws-scm/tensorio-android/KotlinExample/app/build.gradle,/tensorio-android/KotlinExample/app/build.gradle</p>
<p>
Dependency Hierarchy:
- json-schema-validator-2.2.10.jar (Root Library)
- json-schema-core-1.2.10.jar
- jackson-coreutils-1.9.jar
- :x: **jackson-databind-2.2.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/doc-ai/tensorio-android/commit/80bb78e2fbf7f62e76a4c6c5538cbfa1ec4da245">80bb78e2fbf7f62e76a4c6c5538cbfa1ec4da245</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind through 2.8.10 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 deserialization flaw. This is exploitable by sending maliciously crafted JSON input to the readValue method of the ObjectMapper, bypassing a blacklist that is ineffective if the Spring libraries are available in the classpath.
<p>Publish Date: 2018-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-17485>CVE-2017-17485</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-17485">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-17485</a></p>
<p>Release Date: 2018-01-10</p>
<p>Fix Resolution: 2.9.4</p>
</p>
</details>
<p></p>
| True | CVE-2017-17485 (High) detected in jackson-databind-2.2.3.jar - ## CVE-2017-17485 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.2.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to vulnerable library: /tmp/ws-scm/tensorio-android/KotlinExample/app/build.gradle,/tensorio-android/KotlinExample/app/build.gradle</p>
<p>
Dependency Hierarchy:
- json-schema-validator-2.2.10.jar (Root Library)
- json-schema-core-1.2.10.jar
- jackson-coreutils-1.9.jar
- :x: **jackson-databind-2.2.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/doc-ai/tensorio-android/commit/80bb78e2fbf7f62e76a4c6c5538cbfa1ec4da245">80bb78e2fbf7f62e76a4c6c5538cbfa1ec4da245</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind through 2.8.10 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 deserialization flaw. This is exploitable by sending maliciously crafted JSON input to the readValue method of the ObjectMapper, bypassing a blacklist that is ineffective if the Spring libraries are available in the classpath.
<p>Publish Date: 2018-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-17485>CVE-2017-17485</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-17485">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-17485</a></p>
<p>Release Date: 2018-01-10</p>
<p>Fix Resolution: 2.9.4</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to vulnerable library tmp ws scm tensorio android kotlinexample app build gradle tensorio android kotlinexample app build gradle dependency hierarchy json schema validator jar root library json schema core jar jackson coreutils jar x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind through and x through allows unauthenticated remote code execution because of an incomplete fix for the cve deserialization flaw this is exploitable by sending maliciously crafted json input to the readvalue method of the objectmapper bypassing a blacklist that is ineffective if the spring libraries are available in the classpath publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
57,543 | 8,197,304,509 | IssuesEvent | 2018-08-31 13:00:43 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | DOC: BUG: typo in ndimage LowLevelCallable tutorial example | Documentation | Is there a typo in the [ndimage LowLevelCallable tutorial example](https://github.com/scipy/scipy/blob/master/doc/source/tutorial/ndimage.rst#extending-modscipyndimage-in-c)?
AFAICT the C-extension exports the Python function `example.get_transform` which is imported in the next section, but then it's passed as `transform()` to `LowLevelCallable`, shouldn't it be passed as `get_transform()`?
I'm working on implementing `LowLevelCallable` for [`cython_optimize`](https://github.com/scipy/scipy/pull/8431), and I was looking at the tutorial examples for `ndimage`, but I'm having a hard time interpreting when the C methods are passed to the `LowLevelCallable`. If this is indeed a typo, I'm happy to send a PR to fix it, but just wanted to check that I wasn't misunderstanding the usage. Thanks!
### Reproducing code example:
```python
import ctypes
import numpy as np
from scipy import ndimage, LowLevelCallable
from example import get_transform # <- this line never used?
shift = 0.5
user_data = ctypes.c_double(shift)
ptr = ctypes.cast(ctypes.pointer(user_data), ctypes.c_void_p)
callback = LowLevelCallable(transform(), ptr) # <- shouldn't this be get_transform() ?
im = np.arange(12).reshape(4, 3).astype(np.float64)
print(ndimage.geometric_transform(im, callback))
```
### Error message:
```python
>>> callback = LowLevelCallable(transform(), ptr)
---------------------------------------------------------------------------
NameError Traceback (most recent call last)
<ipython-input-2-88e997e74745> in <module>()
----> 1 callback = LowLevelCallable(transform(), ptr)
NameError: name 'transform' is not defined
```
### Scipy/Numpy/Python version information:
```python
import sys, scipy, numpy; print(scipy.__version__, numpy.__version__, sys.version_info)
1.1.0 1.15.0 sys.version_info(major=3, minor=6, micro=6, releaselevel='final', serial=0)
```
| 1.0 | DOC: BUG: typo in ndimage LowLevelCallable tutorial example - Is there a typo in the [ndimage LowLevelCallable tutorial example](https://github.com/scipy/scipy/blob/master/doc/source/tutorial/ndimage.rst#extending-modscipyndimage-in-c)?
AFAICT the C-extension exports the Python function `example.get_transform` which is imported in the next section, but then it's passed as `transform()` to `LowLevelCallable`, shouldn't it be passed as `get_transform()`?
I'm working on implementing `LowLevelCallable` for [`cython_optimize`](https://github.com/scipy/scipy/pull/8431), and I was looking at the tutorial examples for `ndimage`, but I'm having a hard time interpreting when the C methods are passed to the `LowLevelCallable`. If this is indeed a typo, I'm happy to send a PR to fix it, but just wanted to check that I wasn't misunderstanding the usage. Thanks!
### Reproducing code example:
```python
import ctypes
import numpy as np
from scipy import ndimage, LowLevelCallable
from example import get_transform # <- this line never used?
shift = 0.5
user_data = ctypes.c_double(shift)
ptr = ctypes.cast(ctypes.pointer(user_data), ctypes.c_void_p)
callback = LowLevelCallable(transform(), ptr) # <- shouldn't this be get_transform() ?
im = np.arange(12).reshape(4, 3).astype(np.float64)
print(ndimage.geometric_transform(im, callback))
```
### Error message:
```python
>>> callback = LowLevelCallable(transform(), ptr)
---------------------------------------------------------------------------
NameError Traceback (most recent call last)
<ipython-input-2-88e997e74745> in <module>()
----> 1 callback = LowLevelCallable(transform(), ptr)
NameError: name 'transform' is not defined
```
### Scipy/Numpy/Python version information:
```python
import sys, scipy, numpy; print(scipy.__version__, numpy.__version__, sys.version_info)
1.1.0 1.15.0 sys.version_info(major=3, minor=6, micro=6, releaselevel='final', serial=0)
```
| non_priority | doc bug typo in ndimage lowlevelcallable tutorial example is there a typo in the afaict the c extension exports the python function example get transform which is imported in the next section but then it s passed as transform to lowlevelcallable shouldn t it be passed as get transform i m working on implementing lowlevelcallable for and i was looking at the tutorial examples for ndimage but i m having a hard time interpreting when the c methods are passed to the lowlevelcallable if this is indeed a typo i m happy to send a pr to fix it but just wanted to check that i wasn t misunderstanding the usage thanks reproducing code example python import ctypes import numpy as np from scipy import ndimage lowlevelcallable from example import get transform this line never used shift user data ctypes c double shift ptr ctypes cast ctypes pointer user data ctypes c void p callback lowlevelcallable transform ptr shouldn t this be get transform im np arange reshape astype np print ndimage geometric transform im callback error message python callback lowlevelcallable transform ptr nameerror traceback most recent call last in callback lowlevelcallable transform ptr nameerror name transform is not defined scipy numpy python version information python import sys scipy numpy print scipy version numpy version sys version info sys version info major minor micro releaselevel final serial | 0 |
46,292 | 11,811,403,996 | IssuesEvent | 2020-03-19 18:08:24 | ShaikASK/Testing | https://api.github.com/repos/ShaikASK/Testing | opened | Production: Review and Certify : HR Admin : When HR user tries to download the approved document error screen (docs/askstaffing/echosign) | Beta Release #5 Build#7 Defect HR User Module Production Issue Release#7 Build#4 Review & Certify | Steps to Replicate :
1. launch the HR user url
2. login as admin
3. Navigate to New Hire screen >> Navigate to 'Review and certify ' screen of any new hire and click on approved document
Experienced Behavior :
Observed that the below error screen is displayed instead of displaying the respective PDF document ( Attaching screen shot for more reference)

| 2.0 | Production: Review and Certify : HR Admin : When HR user tries to download the approved document error screen (docs/askstaffing/echosign) - Steps to Replicate :
1. launch the HR user url
2. login as admin
3. Navigate to New Hire screen >> Navigate to 'Review and certify ' screen of any new hire and click on approved document
Experienced Behavior :
Observed that the below error screen is displayed instead of displaying the respective PDF document ( Attaching screen shot for more reference)

| non_priority | production review and certify hr admin when hr user tries to download the approved document error screen docs askstaffing echosign steps to replicate launch the hr user url login as admin navigate to new hire screen navigate to review and certify screen of any new hire and click on approved document experienced behavior observed that the below error screen is displayed instead of displaying the respective pdf document attaching screen shot for more reference | 0 |
227,052 | 17,371,914,168 | IssuesEvent | 2021-07-30 15:02:53 | hoffstadt/DearPyGui | https://api.github.com/repos/hoffstadt/DearPyGui | closed | Not clear that add_file_extension('.*') is required for dearpygui.file_dialog(): | documentation enhancement | ---
We'd love help and you are welcome to improve the the documentation with the exception of the api reference because
it is currently generated manually from the stub file. It needs to be integrated within our CI process soon.
If you'd like to change the API reference, update the stub file and we will generate the api docs during releases:
https://github.com/hoffstadt/DearPyGui/blob/master/DearPyGui/stubs/dearpygui.pyi
What can be modified related to docs:
* Wiki
* https://github.com/hoffstadt/DearPyGui/tree/gh-pages except (api_docs.html)
* Readme
* https://github.com/hoffstadt/DearPyGui/blob/master/DearPyGui/include/mvMarvel.h (this shows up in _show_documentation()_)
* https://github.com/hoffstadt/DearPyGui/blob/master/DearPyGui/src/Core/StandardWindows/mvDocWindow.cpp (this shows up in _show_documentation()_)
* https://github.com/hoffstadt/DearPyGui/tree/master/Examples (adding examples)
Currently the CSS formatting is inherited from the autogenerated AIP docs done using pdoc3 to keep uniformity.
Delete the above comment between the lines and PLEASE FILL THE TEMPLATE BELOW before submitting your issue.
Thank you!
----
**My Improvement**
A clear and concise description of what the improvement is. Please provide as much context as possible.
**Necessary Assets**
- [ ] Pictures
- [x] Code Sample
- [ ] GIFs
| 1.0 | Not clear that add_file_extension('.*') is required for dearpygui.file_dialog(): - ---
We'd love help and you are welcome to improve the the documentation with the exception of the api reference because
it is currently generated manually from the stub file. It needs to be integrated within our CI process soon.
If you'd like to change the API reference, update the stub file and we will generate the api docs during releases:
https://github.com/hoffstadt/DearPyGui/blob/master/DearPyGui/stubs/dearpygui.pyi
What can be modified related to docs:
* Wiki
* https://github.com/hoffstadt/DearPyGui/tree/gh-pages except (api_docs.html)
* Readme
* https://github.com/hoffstadt/DearPyGui/blob/master/DearPyGui/include/mvMarvel.h (this shows up in _show_documentation()_)
* https://github.com/hoffstadt/DearPyGui/blob/master/DearPyGui/src/Core/StandardWindows/mvDocWindow.cpp (this shows up in _show_documentation()_)
* https://github.com/hoffstadt/DearPyGui/tree/master/Examples (adding examples)
Currently the CSS formatting is inherited from the autogenerated AIP docs done using pdoc3 to keep uniformity.
Delete the above comment between the lines and PLEASE FILL THE TEMPLATE BELOW before submitting your issue.
Thank you!
----
**My Improvement**
A clear and concise description of what the improvement is. Please provide as much context as possible.
**Necessary Assets**
- [ ] Pictures
- [x] Code Sample
- [ ] GIFs
| non_priority | not clear that add file extension is required for dearpygui file dialog we d love help and you are welcome to improve the the documentation with the exception of the api reference because it is currently generated manually from the stub file it needs to be integrated within our ci process soon if you d like to change the api reference update the stub file and we will generate the api docs during releases what can be modified related to docs wiki except api docs html readme this shows up in show documentation this shows up in show documentation adding examples currently the css formatting is inherited from the autogenerated aip docs done using to keep uniformity delete the above comment between the lines and please fill the template below before submitting your issue thank you my improvement a clear and concise description of what the improvement is please provide as much context as possible necessary assets pictures code sample gifs | 0 |
127,952 | 18,024,749,848 | IssuesEvent | 2021-09-17 01:58:56 | AlexRogalskiy/kotlin-patterns | https://api.github.com/repos/AlexRogalskiy/kotlin-patterns | opened | CVE-2021-23337 (High) detected in lodash-2.4.2.tgz | security vulnerability | ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, & extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p>
<p>Path to dependency file: kotlin-patterns/package.json</p>
<p>Path to vulnerable library: kotlin-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- dockerfile_lint-0.3.4.tgz (Root Library)
- :x: **lodash-2.4.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23337 (High) detected in lodash-2.4.2.tgz - ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, & extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p>
<p>Path to dependency file: kotlin-patterns/package.json</p>
<p>Path to vulnerable library: kotlin-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- dockerfile_lint-0.3.4.tgz (Root Library)
- :x: **lodash-2.4.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file kotlin patterns package json path to vulnerable library kotlin patterns node modules dockerfile lint node modules lodash package json dependency hierarchy dockerfile lint tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource | 0 |
110,824 | 11,710,371,957 | IssuesEvent | 2020-03-09 00:21:21 | bjverde/formDin | https://api.github.com/repos/bjverde/formDin | opened | REST: Alterar tabela JWT | documentation enhancement | Alterar a tabela de banco com tokens JWT e incluir clains reservador
* https://pt.slideshare.net/brunonm/aplicacoes-stateless-com-php-e-jwt slide 26
* https://tools.ietf.org/html/rfc7519#section-4.1
* https://jwt.io/introduction/
Usar base64 para grava id
https://pt.wikipedia.org/wiki/Base64 | 1.0 | REST: Alterar tabela JWT - Alterar a tabela de banco com tokens JWT e incluir clains reservador
* https://pt.slideshare.net/brunonm/aplicacoes-stateless-com-php-e-jwt slide 26
* https://tools.ietf.org/html/rfc7519#section-4.1
* https://jwt.io/introduction/
Usar base64 para grava id
https://pt.wikipedia.org/wiki/Base64 | non_priority | rest alterar tabela jwt alterar a tabela de banco com tokens jwt e incluir clains reservador slide usar para grava id | 0 |
280,730 | 21,315,141,222 | IssuesEvent | 2022-04-16 06:19:56 | brandonrhan/pe | https://api.github.com/repos/brandonrhan/pe | opened | Inconsistence in usage of 'other' | type.DocumentationBug severity.VeryLow | Others is used in faculty, while other is used in role. It should be consistent.

<!--session: 1650088334924-14f9646a-82c3-4f8e-a7ee-2eb0f6619db9-->
<!--Version: Desktop v3.4.2--> | 1.0 | Inconsistence in usage of 'other' - Others is used in faculty, while other is used in role. It should be consistent.

<!--session: 1650088334924-14f9646a-82c3-4f8e-a7ee-2eb0f6619db9-->
<!--Version: Desktop v3.4.2--> | non_priority | inconsistence in usage of other others is used in faculty while other is used in role it should be consistent | 0 |
127,170 | 18,010,316,416 | IssuesEvent | 2021-09-16 07:50:29 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | opened | CVE-2019-11884 (Low) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2019-11884 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The do_hidp_sock_ioctl function in net/bluetooth/hidp/sock.c in the Linux kernel before 5.0.15 allows a local user to obtain potentially sensitive information from kernel stack memory via a HIDPCONNADD command, because a name field may not end with a '\0' character.
<p>Publish Date: 2019-05-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11884>CVE-2019-11884</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11884">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11884</a></p>
<p>Release Date: 2019-05-10</p>
<p>Fix Resolution: 5.0.15</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-11884 (Low) detected in linux-stable-rtv4.1.33 - ## CVE-2019-11884 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The do_hidp_sock_ioctl function in net/bluetooth/hidp/sock.c in the Linux kernel before 5.0.15 allows a local user to obtain potentially sensitive information from kernel stack memory via a HIDPCONNADD command, because a name field may not end with a '\0' character.
<p>Publish Date: 2019-05-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11884>CVE-2019-11884</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11884">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11884</a></p>
<p>Release Date: 2019-05-10</p>
<p>Fix Resolution: 5.0.15</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve low detected in linux stable cve low severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href vulnerable source files vulnerability details the do hidp sock ioctl function in net bluetooth hidp sock c in the linux kernel before allows a local user to obtain potentially sensitive information from kernel stack memory via a hidpconnadd command because a name field may not end with a character publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
14,607 | 17,791,895,822 | IssuesEvent | 2021-08-31 17:09:58 | dosbox-staging/dosbox-staging | https://api.github.com/repos/dosbox-staging/dosbox-staging | closed | 4D Sports Boxing crashes on start when configured for Tandy output | bug external game compatibility | **Describe the bug**
When both dosbox and the game are configured for Tandy video and audio output, dosbox crashes on startup.
**To Reproduce**
Game: 4D Sports Boxing 1.0 (1990)
setup.exe:
- Video display "Tandy"
- Sound option "Tandy with DAC" or "Tandy without DAC/Tandy Other"
dosbox.conf:
``` text
[dosbox]
machine = tandy
[speaker]
tandy = on
pcspeaker = on
disney = off
[gus]
gus = off
[sblaster]
sbtype = none
oplmode = none
[midi]
mpu401 = none
mididevice = none
midiconfig = none
[dos]
xms = false
ems = false
umb = false
```
**Environment (please complete the following information):**
- Ubuntu 20.04
**Version of dosbox-staging:**
- current master branch
**Additional context**
The also affects upstream, but the crash is different. It's reported here: https://sourceforge.net/p/dosbox/bugs/535/
| True | 4D Sports Boxing crashes on start when configured for Tandy output - **Describe the bug**
When both dosbox and the game are configured for Tandy video and audio output, dosbox crashes on startup.
**To Reproduce**
Game: 4D Sports Boxing 1.0 (1990)
setup.exe:
- Video display "Tandy"
- Sound option "Tandy with DAC" or "Tandy without DAC/Tandy Other"
dosbox.conf:
``` text
[dosbox]
machine = tandy
[speaker]
tandy = on
pcspeaker = on
disney = off
[gus]
gus = off
[sblaster]
sbtype = none
oplmode = none
[midi]
mpu401 = none
mididevice = none
midiconfig = none
[dos]
xms = false
ems = false
umb = false
```
**Environment (please complete the following information):**
- Ubuntu 20.04
**Version of dosbox-staging:**
- current master branch
**Additional context**
The also affects upstream, but the crash is different. It's reported here: https://sourceforge.net/p/dosbox/bugs/535/
| non_priority | sports boxing crashes on start when configured for tandy output describe the bug when both dosbox and the game are configured for tandy video and audio output dosbox crashes on startup to reproduce game sports boxing setup exe video display tandy sound option tandy with dac or tandy without dac tandy other dosbox conf text machine tandy tandy on pcspeaker on disney off gus off sbtype none oplmode none none mididevice none midiconfig none xms false ems false umb false environment please complete the following information ubuntu version of dosbox staging current master branch additional context the also affects upstream but the crash is different it s reported here | 0 |
104,951 | 16,623,556,099 | IssuesEvent | 2021-06-03 06:39:11 | Thanraj/_OpenSSL_ | https://api.github.com/repos/Thanraj/_OpenSSL_ | opened | CVE-2015-1792 (Medium) detected in openssl87a37cbadd4b56fb4dc21008c5bcbe929f0a52b6 | security vulnerability | ## CVE-2015-1792 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>openssl87a37cbadd4b56fb4dc21008c5bcbe929f0a52b6</b></p></summary>
<p>
<p>TLS/SSL and crypto library</p>
<p>Library home page: <a href=https://github.com/openssl/openssl.git>https://github.com/openssl/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/_OpenSSL_/commit/f81a773989d97d3776092fa85d76f388e96817b9">f81a773989d97d3776092fa85d76f388e96817b9</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>_OpenSSL_/crypto/cms/cms_smime.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>_OpenSSL_/crypto/cms/cms_smime.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The do_free_upto function in crypto/cms/cms_smime.c in OpenSSL before 0.9.8zg, 1.0.0 before 1.0.0s, 1.0.1 before 1.0.1n, and 1.0.2 before 1.0.2b allows remote attackers to cause a denial of service (infinite loop) via vectors that trigger a NULL value of a BIO data structure, as demonstrated by an unrecognized X.660 OID for a hash function.
<p>Publish Date: 2015-06-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-1792>CVE-2015-1792</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-1792">https://nvd.nist.gov/vuln/detail/CVE-2015-1792</a></p>
<p>Release Date: 2015-06-12</p>
<p>Fix Resolution: 0.9.8zg,1.0.0s,1.0.1n,1.0.2b</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2015-1792 (Medium) detected in openssl87a37cbadd4b56fb4dc21008c5bcbe929f0a52b6 - ## CVE-2015-1792 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>openssl87a37cbadd4b56fb4dc21008c5bcbe929f0a52b6</b></p></summary>
<p>
<p>TLS/SSL and crypto library</p>
<p>Library home page: <a href=https://github.com/openssl/openssl.git>https://github.com/openssl/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/_OpenSSL_/commit/f81a773989d97d3776092fa85d76f388e96817b9">f81a773989d97d3776092fa85d76f388e96817b9</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>_OpenSSL_/crypto/cms/cms_smime.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>_OpenSSL_/crypto/cms/cms_smime.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The do_free_upto function in crypto/cms/cms_smime.c in OpenSSL before 0.9.8zg, 1.0.0 before 1.0.0s, 1.0.1 before 1.0.1n, and 1.0.2 before 1.0.2b allows remote attackers to cause a denial of service (infinite loop) via vectors that trigger a NULL value of a BIO data structure, as demonstrated by an unrecognized X.660 OID for a hash function.
<p>Publish Date: 2015-06-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-1792>CVE-2015-1792</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-1792">https://nvd.nist.gov/vuln/detail/CVE-2015-1792</a></p>
<p>Release Date: 2015-06-12</p>
<p>Fix Resolution: 0.9.8zg,1.0.0s,1.0.1n,1.0.2b</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in cve medium severity vulnerability vulnerable library tls ssl and crypto library library home page a href found in head commit a href found in base branch master vulnerable source files openssl crypto cms cms smime c openssl crypto cms cms smime c vulnerability details the do free upto function in crypto cms cms smime c in openssl before before before and before allows remote attackers to cause a denial of service infinite loop via vectors that trigger a null value of a bio data structure as demonstrated by an unrecognized x oid for a hash function publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
332,964 | 29,502,987,348 | IssuesEvent | 2023-06-03 01:35:37 | SynBioHub/synbiohub3 | https://api.github.com/repos/SynBioHub/synbiohub3 | opened | admin/registries error from SBH3 | backend test failure admin | [synbiohub test] test_admin_registries starting
http://localhost:7777/admin/registries
SBH1
{"registries":[{"uri":"https://synbiohub.org","url":"http://localhost:7777"},{"uri":"https://synbiohub.programmingbiology.org","url":"https://synbiohub.programmingbiology.org"},{"uri":"http://sevahub.es","url":"http://sevahub.es"}],"registered":false,"wor":"https://wor.synbiohub.org","errors":[]}
http://localhost:6789/admin/registries
Traceback (most recent call last):
File "/Users/anjala/Desktop/DLA/synbiohub/sbh3/synbiohub3/tests/test_functions.py", line 85, in get_request
response.raise_for_status()
File "/Users/anjala/.pyenv/versions/3.9.0/lib/python3.9/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 406 Client Error: for url: http://localhost:6789/admin/registries | 1.0 | admin/registries error from SBH3 - [synbiohub test] test_admin_registries starting
http://localhost:7777/admin/registries
SBH1
{"registries":[{"uri":"https://synbiohub.org","url":"http://localhost:7777"},{"uri":"https://synbiohub.programmingbiology.org","url":"https://synbiohub.programmingbiology.org"},{"uri":"http://sevahub.es","url":"http://sevahub.es"}],"registered":false,"wor":"https://wor.synbiohub.org","errors":[]}
http://localhost:6789/admin/registries
Traceback (most recent call last):
File "/Users/anjala/Desktop/DLA/synbiohub/sbh3/synbiohub3/tests/test_functions.py", line 85, in get_request
response.raise_for_status()
File "/Users/anjala/.pyenv/versions/3.9.0/lib/python3.9/site-packages/requests/models.py", line 1021, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 406 Client Error: for url: http://localhost:6789/admin/registries | non_priority | admin registries error from test admin registries starting registries registered false wor traceback most recent call last file users anjala desktop dla synbiohub tests test functions py line in get request response raise for status file users anjala pyenv versions lib site packages requests models py line in raise for status raise httperror http error msg response self requests exceptions httperror client error for url | 0 |
44,727 | 5,641,749,245 | IssuesEvent | 2017-04-06 19:27:05 | NIHOPA/NLPre | https://api.github.com/repos/NIHOPA/NLPre | closed | Add unittests to unidecoder | unit_tests | Add reasonable unittests to unidecoder, Greek letters are the most common, but it might be worth checking for a few other diacritics like
```
α-Helix β-sheet Αα Νν Ββ Ξξ Γγ Οο
Lëtzebuergesch
vóórkomen
perispōménē
```
| 1.0 | Add unittests to unidecoder - Add reasonable unittests to unidecoder, Greek letters are the most common, but it might be worth checking for a few other diacritics like
```
α-Helix β-sheet Αα Νν Ββ Ξξ Γγ Οο
Lëtzebuergesch
vóórkomen
perispōménē
```
| non_priority | add unittests to unidecoder add reasonable unittests to unidecoder greek letters are the most common but it might be worth checking for a few other diacritics like α helix β sheet αα νν ββ ξξ γγ οο lëtzebuergesch vóórkomen perispōménē | 0 |
112,837 | 24,334,776,865 | IssuesEvent | 2022-10-01 00:50:31 | marcosdosea/AgendeMe | https://api.github.com/repos/marcosdosea/AgendeMe | closed | Adicionar pacote para geração de codigo (versão 6.0.10) | 3.0 code | Acrescentar pacote Microsoft.VisualStudio.Web.CodeGeneration.Design versão 6.0.10 | 1.0 | Adicionar pacote para geração de codigo (versão 6.0.10) - Acrescentar pacote Microsoft.VisualStudio.Web.CodeGeneration.Design versão 6.0.10 | non_priority | adicionar pacote para geração de codigo versão acrescentar pacote microsoft visualstudio web codegeneration design versão | 0 |
8,767 | 3,005,833,909 | IssuesEvent | 2015-07-27 05:13:38 | SIGPLAN/online | https://api.github.com/repos/SIGPLAN/online | closed | `legacycopyright` option should function gracefully without a DOI | beta-test report LaTeX class legacycopyright option | I question whether the DOI warning is reasonable under the
legacycopyright option. And if the DOI is not provided, it shouldn't
get a line.
| 1.0 | `legacycopyright` option should function gracefully without a DOI - I question whether the DOI warning is reasonable under the
legacycopyright option. And if the DOI is not provided, it shouldn't
get a line.
| non_priority | legacycopyright option should function gracefully without a doi i question whether the doi warning is reasonable under the legacycopyright option and if the doi is not provided it shouldn t get a line | 0 |
196 | 2,575,579,994 | IssuesEvent | 2015-02-12 00:14:32 | globaleaks/GlobaLeaks | https://api.github.com/repos/globaleaks/GlobaLeaks | opened | Requirements for end-2-end encryption should be clearly defined | F: Security | With this ticket i wanna open the discussion on the requirements we want globaleaks to satisfy when implementing the end-2-end encryption. this in order to organize the work and performing a complete security analysis and planning the development ticket by ticket.
To this aim this initial message will act as a resume of the discussion with reference to all the different tickets we will open and address working on this topic.
Before starting opening tons of tickets let's us this one as bitchy ticket in order to discuss it all and then i will open the child ones.
Requirements not so tricky that i consider quite all useful and staightforward and needed as first step in order to improve the system without adding new capabilities; could be considered the first step into the integration: The only one really critical for usability is 8 and we should take care to discuss that in particular
* [ ] 1 The interface should enable the upload an ascii armored pgp keys
* [ ] 2 The interface should enable the upload an also binary pgp keys (load only, needed in order to support preexisting keys but globaleaks will always export ascii armored keys)
* [ ] 3 The interface should always provide/get ascii armored pgp keys to/from the server
* [ ] 4 The interface should be able to validate uploaded keys
* [ ] 5 The interface should enable to generate a key from the interface
* [ ] 6 Generated keys through the interface should be mapped on the receiver mail
* [ ] 7 Generated keys through the interface should have a 4096 private key (with no choice for the user; we should use automagical secure defaults at least in first implementations)
* [ ] 8 The private key should be protected with a password chosen by the user
* [] 9 Generated keys should be simply downloadable by the user with separated files for private key (private.asc) and public key (publickey.asc)
* [] 10 Downlodable keys by the user should be bundled into a common zip container including both the private and the public keys
Requirements needed in order to go forward the current work:
* [] 11 The interface should encrypt fields of the submission for each receiver of the submission (single encryption); this involve discussing if we should have a single encryption or independent encryptions for each receiver. in addition we should think if we need to have a single encryption of the submission or a per field encryption and the benefit of the two approach
* [] 12 The interface should encrypt files of the submission for each receiver of the submission (single encryption); this involve discussing if we should have a single encryption or independent encryptions for each receiver. quite for sure here we should go for the single encryption in order to minimize the network transfers (1GBx10 = 10GB! maths! arg)
* [] 13 The interface should generate automatically a key for the whistleblower and give the wistleblower a secret in order to recreate the key deterministically when he returns onto the plaform
* [] 14 Upon submission the public key of the whistleblower should be saved together with the submission
* [] 14 I know that 12 wont be easy replicable also for receiver due to the different validity of the keys but probably we should think to a similar solution also for receivers: how to avoid that the receiver should load his key on the platform any time he need to access the submissions?
this is simply the start; i've a plethora of ideas but this should be enough to start asking your comments:
@hellais @mmaker @vecna @fpietrosanti @origliante | True | Requirements for end-2-end encryption should be clearly defined - With this ticket i wanna open the discussion on the requirements we want globaleaks to satisfy when implementing the end-2-end encryption. this in order to organize the work and performing a complete security analysis and planning the development ticket by ticket.
To this aim this initial message will act as a resume of the discussion with reference to all the different tickets we will open and address working on this topic.
Before starting opening tons of tickets let's us this one as bitchy ticket in order to discuss it all and then i will open the child ones.
Requirements not so tricky that i consider quite all useful and staightforward and needed as first step in order to improve the system without adding new capabilities; could be considered the first step into the integration: The only one really critical for usability is 8 and we should take care to discuss that in particular
* [ ] 1 The interface should enable the upload an ascii armored pgp keys
* [ ] 2 The interface should enable the upload an also binary pgp keys (load only, needed in order to support preexisting keys but globaleaks will always export ascii armored keys)
* [ ] 3 The interface should always provide/get ascii armored pgp keys to/from the server
* [ ] 4 The interface should be able to validate uploaded keys
* [ ] 5 The interface should enable to generate a key from the interface
* [ ] 6 Generated keys through the interface should be mapped on the receiver mail
* [ ] 7 Generated keys through the interface should have a 4096 private key (with no choice for the user; we should use automagical secure defaults at least in first implementations)
* [ ] 8 The private key should be protected with a password chosen by the user
* [] 9 Generated keys should be simply downloadable by the user with separated files for private key (private.asc) and public key (publickey.asc)
* [] 10 Downlodable keys by the user should be bundled into a common zip container including both the private and the public keys
Requirements needed in order to go forward the current work:
* [] 11 The interface should encrypt fields of the submission for each receiver of the submission (single encryption); this involve discussing if we should have a single encryption or independent encryptions for each receiver. in addition we should think if we need to have a single encryption of the submission or a per field encryption and the benefit of the two approach
* [] 12 The interface should encrypt files of the submission for each receiver of the submission (single encryption); this involve discussing if we should have a single encryption or independent encryptions for each receiver. quite for sure here we should go for the single encryption in order to minimize the network transfers (1GBx10 = 10GB! maths! arg)
* [] 13 The interface should generate automatically a key for the whistleblower and give the wistleblower a secret in order to recreate the key deterministically when he returns onto the plaform
* [] 14 Upon submission the public key of the whistleblower should be saved together with the submission
* [] 14 I know that 12 wont be easy replicable also for receiver due to the different validity of the keys but probably we should think to a similar solution also for receivers: how to avoid that the receiver should load his key on the platform any time he need to access the submissions?
this is simply the start; i've a plethora of ideas but this should be enough to start asking your comments:
@hellais @mmaker @vecna @fpietrosanti @origliante | non_priority | requirements for end end encryption should be clearly defined with this ticket i wanna open the discussion on the requirements we want globaleaks to satisfy when implementing the end end encryption this in order to organize the work and performing a complete security analysis and planning the development ticket by ticket to this aim this initial message will act as a resume of the discussion with reference to all the different tickets we will open and address working on this topic before starting opening tons of tickets let s us this one as bitchy ticket in order to discuss it all and then i will open the child ones requirements not so tricky that i consider quite all useful and staightforward and needed as first step in order to improve the system without adding new capabilities could be considered the first step into the integration the only one really critical for usability is and we should take care to discuss that in particular the interface should enable the upload an ascii armored pgp keys the interface should enable the upload an also binary pgp keys load only needed in order to support preexisting keys but globaleaks will always export ascii armored keys the interface should always provide get ascii armored pgp keys to from the server the interface should be able to validate uploaded keys the interface should enable to generate a key from the interface generated keys through the interface should be mapped on the receiver mail generated keys through the interface should have a private key with no choice for the user we should use automagical secure defaults at least in first implementations the private key should be protected with a password chosen by the user generated keys should be simply downloadable by the user with separated files for private key private asc and public key publickey asc downlodable keys by the user should be bundled into a common zip container including both the private and the public keys requirements needed in order to go forward the current work the interface should encrypt fields of the submission for each receiver of the submission single encryption this involve discussing if we should have a single encryption or independent encryptions for each receiver in addition we should think if we need to have a single encryption of the submission or a per field encryption and the benefit of the two approach the interface should encrypt files of the submission for each receiver of the submission single encryption this involve discussing if we should have a single encryption or independent encryptions for each receiver quite for sure here we should go for the single encryption in order to minimize the network transfers maths arg the interface should generate automatically a key for the whistleblower and give the wistleblower a secret in order to recreate the key deterministically when he returns onto the plaform upon submission the public key of the whistleblower should be saved together with the submission i know that wont be easy replicable also for receiver due to the different validity of the keys but probably we should think to a similar solution also for receivers how to avoid that the receiver should load his key on the platform any time he need to access the submissions this is simply the start i ve a plethora of ideas but this should be enough to start asking your comments hellais mmaker vecna fpietrosanti origliante | 0 |
416,786 | 28,100,389,654 | IssuesEvent | 2023-03-30 18:59:55 | dfetch-org/dfetch | https://api.github.com/repos/dfetch-org/dfetch | closed | Create basic landing page | documentation | Create a basic landing page for https://dfetch.github.io/
Some basic bootstrap gave me:
https://htmlpreview.github.io/?https://github.com/dfetch-org/dfetch/blob/feature/landing-page/doc/landing.html | 1.0 | Create basic landing page - Create a basic landing page for https://dfetch.github.io/
Some basic bootstrap gave me:
https://htmlpreview.github.io/?https://github.com/dfetch-org/dfetch/blob/feature/landing-page/doc/landing.html | non_priority | create basic landing page create a basic landing page for some basic bootstrap gave me | 0 |
59,131 | 24,659,989,312 | IssuesEvent | 2022-10-18 05:30:58 | hashicorp/terraform-provider-azurerm | https://api.github.com/repos/hashicorp/terraform-provider-azurerm | closed | Importing Frontdoor route causes Panic | bug crash service/cdn | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.3.0
### AzureRM Provider Version
3.27.0
### Affected Resource(s)/Data Source(s)
azurerm_cdn_frontdoor_route
### Terraform Configuration Files
```hcl
terraform {
}
provider "azurerm" {
features {}
}
resource "azurerm_cdn_frontdoor_route" "shell" {
cache {}
}
```
### Debug Output/Panic Output
```shell
Error: Plugin did not respond
│
│ The plugin encountered an error, and failed to respond to the plugin.(*GRPCProvider).ReadResource call. The plugin logs may contain more details.
╵
Stack trace from the terraform-provider-azurerm_v3.27.0_x5.exe plugin:
panic: runtime error: invalid memory address or nil pointer dereference
[signal 0xc0000005 code=0x0 addr=0x0 pc=0x4632065]
goroutine 131 [running]:
github.com/hashicorp/terraform-provider-azurerm/internal/services/cdn.flattenRuleSetResourceArray(0xc0019870b0)
github.com/hashicorp/terraform-provider-azurerm/internal/services/cdn/cdn_frontdoor_route_resource.go:617 +0xc5
github.com/hashicorp/terraform-provider-azurerm/internal/services/cdn.resourceCdnFrontDoorRouteRead(0xc00014b900, {0x59022e0?, 0xc000598c00?})
github.com/hashicorp/terraform-provider-azurerm/internal/services/cdn/cdn_frontdoor_route_resource.go:375 +0xab4
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*Resource).read(0x6a78e48?, {0x6a78e48?, 0xc0012f31a0?}, 0xd?, {0x59022e0?, 0xc000598c00?})
github.com/hashicorp/terraform-plugin-sdk/v2@v2.18.0/helper/schema/resource.go:712 +0x178
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*Resource).RefreshWithoutUpgrade(0xc00035d960, {0x6a78e48, 0xc0012f31a0}, 0xc001961380, {0x59022e0, 0xc000598c00})
github.com/hashicorp/terraform-plugin-sdk/v2@v2.18.0/helper/schema/resource.go:1015 +0x585
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*GRPCProviderServer).ReadResource(0xc00011d7a0, {0x6a78e48?, 0xc0012f30e0?}, 0xc00196cd80)
github.com/hashicorp/terraform-plugin-sdk/v2@v2.18.0/helper/schema/grpc_provider.go:613 +0x497
github.com/hashicorp/terraform-plugin-go/tfprotov5/tf5server.(*server).ReadResource(0xc000dfb860, {0x6a78e48?, 0xc0012f29c0?}, 0xc001957e00)
github.com/hashicorp/terraform-plugin-go@v0.10.0/tfprotov5/tf5server/server.go:747 +0x41e
github.com/hashicorp/terraform-plugin-go/tfprotov5/internal/tfplugin5._Provider_ReadResource_Handler({0x5f00000?, 0xc000dfb860}, {0x6a78e48, 0xc0012f29c0}, 0xc001957da0, 0x0)
github.com/hashicorp/terraform-plugin-go@v0.10.0/tfprotov5/internal/tfplugin5/tfplugin5_grpc.pb.go:349 +0x170
google.golang.org/grpc.(*Server).processUnaryRPC(0xc0002b0a80, {0x6a8aaa0, 0xc000ae0000}, 0xc000d64900, 0xc000cdaf60, 0xab76830, 0x0)
google.golang.org/grpc@v1.47.0/server.go:1283 +0xcfd
google.golang.org/grpc.(*Server).handleStream(0xc0002b0a80, {0x6a8aaa0, 0xc000ae0000}, 0xc000d64900, 0x0)
google.golang.org/grpc@v1.47.0/server.go:1620 +0xa1b
google.golang.org/grpc.(*Server).serveStreams.func1.2()
google.golang.org/grpc@v1.47.0/server.go:922 +0x98
created by google.golang.org/grpc.(*Server).serveStreams.func1
google.golang.org/grpc@v1.47.0/server.go:920 +0x28a
Error: The terraform-provider-azurerm_v3.27.0_x5.exe plugin crashed!
This is always indicative of a bug within the plugin. It would be immensely
helpful if you could report the crash with the plugin's maintainers so that it
can be fixed. The output above should help diagnose the issue.
```
### Expected Behaviour
Resource should be imported
### Actual Behaviour
Terraform crashes
### Steps to Reproduce
terraform init
terraform import azurerm_cdn_frontdoor_route.shell "<resource_id>"
### Important Factoids
_No response_
### References
_No response_ | 1.0 | Importing Frontdoor route causes Panic - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.3.0
### AzureRM Provider Version
3.27.0
### Affected Resource(s)/Data Source(s)
azurerm_cdn_frontdoor_route
### Terraform Configuration Files
```hcl
terraform {
}
provider "azurerm" {
features {}
}
resource "azurerm_cdn_frontdoor_route" "shell" {
cache {}
}
```
### Debug Output/Panic Output
```shell
Error: Plugin did not respond
│
│ The plugin encountered an error, and failed to respond to the plugin.(*GRPCProvider).ReadResource call. The plugin logs may contain more details.
╵
Stack trace from the terraform-provider-azurerm_v3.27.0_x5.exe plugin:
panic: runtime error: invalid memory address or nil pointer dereference
[signal 0xc0000005 code=0x0 addr=0x0 pc=0x4632065]
goroutine 131 [running]:
github.com/hashicorp/terraform-provider-azurerm/internal/services/cdn.flattenRuleSetResourceArray(0xc0019870b0)
github.com/hashicorp/terraform-provider-azurerm/internal/services/cdn/cdn_frontdoor_route_resource.go:617 +0xc5
github.com/hashicorp/terraform-provider-azurerm/internal/services/cdn.resourceCdnFrontDoorRouteRead(0xc00014b900, {0x59022e0?, 0xc000598c00?})
github.com/hashicorp/terraform-provider-azurerm/internal/services/cdn/cdn_frontdoor_route_resource.go:375 +0xab4
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*Resource).read(0x6a78e48?, {0x6a78e48?, 0xc0012f31a0?}, 0xd?, {0x59022e0?, 0xc000598c00?})
github.com/hashicorp/terraform-plugin-sdk/v2@v2.18.0/helper/schema/resource.go:712 +0x178
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*Resource).RefreshWithoutUpgrade(0xc00035d960, {0x6a78e48, 0xc0012f31a0}, 0xc001961380, {0x59022e0, 0xc000598c00})
github.com/hashicorp/terraform-plugin-sdk/v2@v2.18.0/helper/schema/resource.go:1015 +0x585
github.com/hashicorp/terraform-plugin-sdk/v2/helper/schema.(*GRPCProviderServer).ReadResource(0xc00011d7a0, {0x6a78e48?, 0xc0012f30e0?}, 0xc00196cd80)
github.com/hashicorp/terraform-plugin-sdk/v2@v2.18.0/helper/schema/grpc_provider.go:613 +0x497
github.com/hashicorp/terraform-plugin-go/tfprotov5/tf5server.(*server).ReadResource(0xc000dfb860, {0x6a78e48?, 0xc0012f29c0?}, 0xc001957e00)
github.com/hashicorp/terraform-plugin-go@v0.10.0/tfprotov5/tf5server/server.go:747 +0x41e
github.com/hashicorp/terraform-plugin-go/tfprotov5/internal/tfplugin5._Provider_ReadResource_Handler({0x5f00000?, 0xc000dfb860}, {0x6a78e48, 0xc0012f29c0}, 0xc001957da0, 0x0)
github.com/hashicorp/terraform-plugin-go@v0.10.0/tfprotov5/internal/tfplugin5/tfplugin5_grpc.pb.go:349 +0x170
google.golang.org/grpc.(*Server).processUnaryRPC(0xc0002b0a80, {0x6a8aaa0, 0xc000ae0000}, 0xc000d64900, 0xc000cdaf60, 0xab76830, 0x0)
google.golang.org/grpc@v1.47.0/server.go:1283 +0xcfd
google.golang.org/grpc.(*Server).handleStream(0xc0002b0a80, {0x6a8aaa0, 0xc000ae0000}, 0xc000d64900, 0x0)
google.golang.org/grpc@v1.47.0/server.go:1620 +0xa1b
google.golang.org/grpc.(*Server).serveStreams.func1.2()
google.golang.org/grpc@v1.47.0/server.go:922 +0x98
created by google.golang.org/grpc.(*Server).serveStreams.func1
google.golang.org/grpc@v1.47.0/server.go:920 +0x28a
Error: The terraform-provider-azurerm_v3.27.0_x5.exe plugin crashed!
This is always indicative of a bug within the plugin. It would be immensely
helpful if you could report the crash with the plugin's maintainers so that it
can be fixed. The output above should help diagnose the issue.
```
### Expected Behaviour
Resource should be imported
### Actual Behaviour
Terraform crashes
### Steps to Reproduce
terraform init
terraform import azurerm_cdn_frontdoor_route.shell "<resource_id>"
### Important Factoids
_No response_
### References
_No response_ | non_priority | importing frontdoor route causes panic is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s azurerm cdn frontdoor route terraform configuration files hcl terraform provider azurerm features resource azurerm cdn frontdoor route shell cache debug output panic output shell error plugin did not respond │ │ the plugin encountered an error and failed to respond to the plugin grpcprovider readresource call the plugin logs may contain more details ╵ stack trace from the terraform provider azurerm exe plugin panic runtime error invalid memory address or nil pointer dereference goroutine github com hashicorp terraform provider azurerm internal services cdn flattenrulesetresourcearray github com hashicorp terraform provider azurerm internal services cdn cdn frontdoor route resource go github com hashicorp terraform provider azurerm internal services cdn resourcecdnfrontdoorrouteread github com hashicorp terraform provider azurerm internal services cdn cdn frontdoor route resource go github com hashicorp terraform plugin sdk helper schema resource read github com hashicorp terraform plugin sdk helper schema resource go github com hashicorp terraform plugin sdk helper schema resource refreshwithoutupgrade github com hashicorp terraform plugin sdk helper schema resource go github com hashicorp terraform plugin sdk helper schema grpcproviderserver readresource github com hashicorp terraform plugin sdk helper schema grpc provider go github com hashicorp terraform plugin go server readresource github com hashicorp terraform plugin go server go github com hashicorp terraform plugin go internal provider readresource handler github com hashicorp terraform plugin go internal grpc pb go google golang org grpc server processunaryrpc google golang org grpc server go google golang org grpc server handlestream google golang org grpc server go google golang org grpc server servestreams google golang org grpc server go created by google golang org grpc server servestreams google golang org grpc server go error the terraform provider azurerm exe plugin crashed this is always indicative of a bug within the plugin it would be immensely helpful if you could report the crash with the plugin s maintainers so that it can be fixed the output above should help diagnose the issue expected behaviour resource should be imported actual behaviour terraform crashes steps to reproduce terraform init terraform import azurerm cdn frontdoor route shell important factoids no response references no response | 0 |
91,834 | 15,856,656,938 | IssuesEvent | 2021-04-08 02:52:12 | benald/liferay-link-bookmark | https://api.github.com/repos/benald/liferay-link-bookmark | opened | WS-2019-0310 (High) detected in https-proxy-agent-2.2.1.tgz | security vulnerability | ## WS-2019-0310 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https-proxy-agent-2.2.1.tgz</b></p></summary>
<p>An HTTP(s) proxy `http.Agent` implementation for HTTPS</p>
<p>Library home page: <a href="https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-2.2.1.tgz">https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-2.2.1.tgz</a></p>
<p>Path to dependency file: /liferay-link-bookmark/package.json</p>
<p>Path to vulnerable library: liferay-link-bookmark/node_modules/https-proxy-agent/package.json</p>
<p>
Dependency Hierarchy:
- protractor-5.4.2.tgz (Root Library)
- browserstack-1.5.2.tgz
- :x: **https-proxy-agent-2.2.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
"in 'https-proxy-agent', before v2.2.3, there is a failure of TLS enforcement on the socket. Attacker may intercept unencrypted communications.
<p>Publish Date: 2019-10-07
<p>URL: <a href=https://github.com/TooTallNate/node-https-proxy-agent/commit/36d8cf509f877fa44f4404fce57ebaf9410fe51b>WS-2019-0310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1184">https://www.npmjs.com/advisories/1184</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: https-proxy-agent - 2.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0310 (High) detected in https-proxy-agent-2.2.1.tgz - ## WS-2019-0310 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https-proxy-agent-2.2.1.tgz</b></p></summary>
<p>An HTTP(s) proxy `http.Agent` implementation for HTTPS</p>
<p>Library home page: <a href="https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-2.2.1.tgz">https://registry.npmjs.org/https-proxy-agent/-/https-proxy-agent-2.2.1.tgz</a></p>
<p>Path to dependency file: /liferay-link-bookmark/package.json</p>
<p>Path to vulnerable library: liferay-link-bookmark/node_modules/https-proxy-agent/package.json</p>
<p>
Dependency Hierarchy:
- protractor-5.4.2.tgz (Root Library)
- browserstack-1.5.2.tgz
- :x: **https-proxy-agent-2.2.1.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
"in 'https-proxy-agent', before v2.2.3, there is a failure of TLS enforcement on the socket. Attacker may intercept unencrypted communications.
<p>Publish Date: 2019-10-07
<p>URL: <a href=https://github.com/TooTallNate/node-https-proxy-agent/commit/36d8cf509f877fa44f4404fce57ebaf9410fe51b>WS-2019-0310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1184">https://www.npmjs.com/advisories/1184</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: https-proxy-agent - 2.2.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws high detected in https proxy agent tgz ws high severity vulnerability vulnerable library https proxy agent tgz an http s proxy http agent implementation for https library home page a href path to dependency file liferay link bookmark package json path to vulnerable library liferay link bookmark node modules https proxy agent package json dependency hierarchy protractor tgz root library browserstack tgz x https proxy agent tgz vulnerable library vulnerability details in https proxy agent before there is a failure of tls enforcement on the socket attacker may intercept unencrypted communications publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution https proxy agent step up your open source security game with whitesource | 0 |
116,035 | 17,358,315,725 | IssuesEvent | 2021-07-29 16:56:24 | Tim-sandbox/webgoat-trng | https://api.github.com/repos/Tim-sandbox/webgoat-trng | opened | CVE-2019-10742 (High) detected in axios-0.17.1.tgz | security vulnerability | ## CVE-2019-10742 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>axios-0.17.1.tgz</b></p></summary>
<p>Promise based HTTP client for the browser and node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.17.1.tgz">https://registry.npmjs.org/axios/-/axios-0.17.1.tgz</a></p>
<p>Path to dependency file: webgoat-trng/docs/package.json</p>
<p>Path to vulnerable library: webgoat-trng/docs/node_modules/axios/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- localtunnel-1.9.1.tgz
- :x: **axios-0.17.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/webgoat-trng/commit/35c20cdfd5acad60536ca6b0402f619e18097690">35c20cdfd5acad60536ca6b0402f619e18097690</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Axios up to and including 0.18.0 allows attackers to cause a denial of service (application crash) by continuing to accepting content after maxContentLength is exceeded.
<p>Publish Date: 2019-05-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10742>CVE-2019-10742</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/axios/axios/issues/1098">https://github.com/axios/axios/issues/1098</a></p>
<p>Release Date: 2019-05-31</p>
<p>Fix Resolution: 0.19.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"axios","packageVersion":"0.17.1","packageFilePaths":["/docs/package.json"],"isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;localtunnel:1.9.1;axios:0.17.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.19.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-10742","vulnerabilityDetails":"Axios up to and including 0.18.0 allows attackers to cause a denial of service (application crash) by continuing to accepting content after maxContentLength is exceeded.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10742","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-10742 (High) detected in axios-0.17.1.tgz - ## CVE-2019-10742 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>axios-0.17.1.tgz</b></p></summary>
<p>Promise based HTTP client for the browser and node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/axios/-/axios-0.17.1.tgz">https://registry.npmjs.org/axios/-/axios-0.17.1.tgz</a></p>
<p>Path to dependency file: webgoat-trng/docs/package.json</p>
<p>Path to vulnerable library: webgoat-trng/docs/node_modules/axios/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.3.tgz (Root Library)
- localtunnel-1.9.1.tgz
- :x: **axios-0.17.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Tim-sandbox/webgoat-trng/commit/35c20cdfd5acad60536ca6b0402f619e18097690">35c20cdfd5acad60536ca6b0402f619e18097690</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Axios up to and including 0.18.0 allows attackers to cause a denial of service (application crash) by continuing to accepting content after maxContentLength is exceeded.
<p>Publish Date: 2019-05-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10742>CVE-2019-10742</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/axios/axios/issues/1098">https://github.com/axios/axios/issues/1098</a></p>
<p>Release Date: 2019-05-31</p>
<p>Fix Resolution: 0.19.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"axios","packageVersion":"0.17.1","packageFilePaths":["/docs/package.json"],"isTransitiveDependency":true,"dependencyTree":"browser-sync:2.26.3;localtunnel:1.9.1;axios:0.17.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.19.0"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-10742","vulnerabilityDetails":"Axios up to and including 0.18.0 allows attackers to cause a denial of service (application crash) by continuing to accepting content after maxContentLength is exceeded.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10742","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in axios tgz cve high severity vulnerability vulnerable library axios tgz promise based http client for the browser and node js library home page a href path to dependency file webgoat trng docs package json path to vulnerable library webgoat trng docs node modules axios package json dependency hierarchy browser sync tgz root library localtunnel tgz x axios tgz vulnerable library found in head commit a href found in base branch main vulnerability details axios up to and including allows attackers to cause a denial of service application crash by continuing to accepting content after maxcontentlength is exceeded publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree browser sync localtunnel axios isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails axios up to and including allows attackers to cause a denial of service application crash by continuing to accepting content after maxcontentlength is exceeded vulnerabilityurl | 0 |
8,735 | 12,266,635,453 | IssuesEvent | 2020-05-07 09:17:27 | AzureAD/microsoft-identity-web | https://api.github.com/repos/AzureAD/microsoft-identity-web | closed | [Feature Request] Leverage the ASP.NET Core IHttpClientFactory | Spec'd enhancement fixed performance requirement-aspnetcore-team | ### This issue is for a: (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x ] feature request
- [ ] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)
```
### Why?
Architectural guidance recommand to [Use IHttpClientFactory to implement resilient HTTP requests](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/implement-resilient-applications/use-httpclientfactory-to-implement-resilient-http-requests).
ASP.NET Core [IHttpClientFactory](https://docs.microsoft.com/en-us/aspnet/core/fundamentals/http-requests) manages the pooling and lifetime of underlying HttpClientMessageHandler instances, which avoids port exhaustion and common DNS (Domain Name System) problems that occur when manually managing HttpClient lifetimes.
It also allows for configurable logging experience (via ILogger) for all requests sent through clients created by the factory.
Today, when instanciating a IConfidentialClientApplication in https://github.com/AzureAD/microsoft-identity-web/blob/72939218c85f65d42e7d5989e5dffcf0cdee6000/src/Microsoft.Identity.Web/TokenAcquisition.cs#L348-L366
we leave it to MSAL.NET to manage the HttpClient
### Possible design
For the moment, we could have the following design
1. Add a new parameter `httpClientFactory` of type `IHttpClientFactory` in the constructor of `TokenAcquisition`: https://github.com/AzureAD/microsoft-identity-web/blob/72939218c85f65d42e7d5989e5dffcf0cdee6000/src/Microsoft.Identity.Web/TokenAcquisition.cs#L46-L51 so that an HttpClientFactor can be injected by dependency injection. Also add a corresponding property to `TokenAcquisition`
2. Add a new internal class (for instance `MsalAspNetCoreHttpClientFactory`) implementing [IMsalHttpClientFactory](https://github.com/AzureAD/microsoft-authentication-library-for-dotnet/wiki/HttpClient) and having a constructor with an ASP.NET Core `IHttpClientFactory`, and which `CreateClient()` method would just delegate to `IHttpClientFactory`
3. When building MSAL.NET confidential client application, use the `.WithHttpClientFactory(httpClientFactory)` modifyer and instanciate a `MsalAspNetCoreHttpClientFactory` from the `httpClientFactory` property
4. In `AddWebAppCallsWebApi` and `AddProtectedWebApiCallsWebApi`, add a service: `services.AddHttpClient();`
### Possible future improvement
I don't think we need to use [Named clients](https://docs.microsoft.com/en-us/aspnet/core/fundamentals/http-requests?view=aspnetcore-3.1#named-clients) ?
Copied from https://github.com/Azure-Samples/active-directory-aspnetcore-webapp-openidconnect-v2/issues/267 | 1.0 | [Feature Request] Leverage the ASP.NET Core IHttpClientFactory - ### This issue is for a: (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x ] feature request
- [ ] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)
```
### Why?
Architectural guidance recommand to [Use IHttpClientFactory to implement resilient HTTP requests](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/implement-resilient-applications/use-httpclientfactory-to-implement-resilient-http-requests).
ASP.NET Core [IHttpClientFactory](https://docs.microsoft.com/en-us/aspnet/core/fundamentals/http-requests) manages the pooling and lifetime of underlying HttpClientMessageHandler instances, which avoids port exhaustion and common DNS (Domain Name System) problems that occur when manually managing HttpClient lifetimes.
It also allows for configurable logging experience (via ILogger) for all requests sent through clients created by the factory.
Today, when instanciating a IConfidentialClientApplication in https://github.com/AzureAD/microsoft-identity-web/blob/72939218c85f65d42e7d5989e5dffcf0cdee6000/src/Microsoft.Identity.Web/TokenAcquisition.cs#L348-L366
we leave it to MSAL.NET to manage the HttpClient
### Possible design
For the moment, we could have the following design
1. Add a new parameter `httpClientFactory` of type `IHttpClientFactory` in the constructor of `TokenAcquisition`: https://github.com/AzureAD/microsoft-identity-web/blob/72939218c85f65d42e7d5989e5dffcf0cdee6000/src/Microsoft.Identity.Web/TokenAcquisition.cs#L46-L51 so that an HttpClientFactor can be injected by dependency injection. Also add a corresponding property to `TokenAcquisition`
2. Add a new internal class (for instance `MsalAspNetCoreHttpClientFactory`) implementing [IMsalHttpClientFactory](https://github.com/AzureAD/microsoft-authentication-library-for-dotnet/wiki/HttpClient) and having a constructor with an ASP.NET Core `IHttpClientFactory`, and which `CreateClient()` method would just delegate to `IHttpClientFactory`
3. When building MSAL.NET confidential client application, use the `.WithHttpClientFactory(httpClientFactory)` modifyer and instanciate a `MsalAspNetCoreHttpClientFactory` from the `httpClientFactory` property
4. In `AddWebAppCallsWebApi` and `AddProtectedWebApiCallsWebApi`, add a service: `services.AddHttpClient();`
### Possible future improvement
I don't think we need to use [Named clients](https://docs.microsoft.com/en-us/aspnet/core/fundamentals/http-requests?view=aspnetcore-3.1#named-clients) ?
Copied from https://github.com/Azure-Samples/active-directory-aspnetcore-webapp-openidconnect-v2/issues/267 | non_priority | leverage the asp net core ihttpclientfactory this issue is for a mark with an x bug report please search issues before submitting feature request documentation issue or request regression a behavior that used to work and stopped in a new release why architectural guidance recommand to asp net core manages the pooling and lifetime of underlying httpclientmessagehandler instances which avoids port exhaustion and common dns domain name system problems that occur when manually managing httpclient lifetimes it also allows for configurable logging experience via ilogger for all requests sent through clients created by the factory today when instanciating a iconfidentialclientapplication in we leave it to msal net to manage the httpclient possible design for the moment we could have the following design add a new parameter httpclientfactory of type ihttpclientfactory in the constructor of tokenacquisition so that an httpclientfactor can be injected by dependency injection also add a corresponding property to tokenacquisition add a new internal class for instance msalaspnetcorehttpclientfactory implementing and having a constructor with an asp net core ihttpclientfactory and which createclient method would just delegate to ihttpclientfactory when building msal net confidential client application use the withhttpclientfactory httpclientfactory modifyer and instanciate a msalaspnetcorehttpclientfactory from the httpclientfactory property in addwebappcallswebapi and addprotectedwebapicallswebapi add a service services addhttpclient possible future improvement i don t think we need to use copied from | 0 |
32,693 | 15,587,815,554 | IssuesEvent | 2021-03-18 05:03:59 | flutter/flutter | https://api.github.com/repos/flutter/flutter | opened | Regression in performance from Flutter 1.23.0-13.0.pre and above on older iOS devices | created via performance template | <!-- Title: Regression in performance from Flutter 1.23.0-13.0.pre and above on older iOS devices -->
## Details
<!--
1. Please tell us exactly how to reproduce the problem you are running into.
2. Please attach a small application (ideally just one main.dart file) that
reproduces the problem. You could use https://gist.github.com/ for this.
3. Switch flutter to master channel and run this app on a physical device
using profile mode with Skia tracing enabled, as follows:
flutter channel master
flutter run --profile --trace-skia
The bleeding edge master channel is encouraged here because Flutter is
constantly fixing bugs and improving its performance. Your problem in an
older Flutter version may have already been solved in the master channel.
4. Record a video of the performance issue using another phone so we
can have an intuitive understanding of what happened. Don’t use
"adb screenrecord", as that affects the performance of the profile run.
5. Open Observatory and save a timeline trace of the performance issue
so we know which functions might be causing it. See "How to Collect
and Read Timeline Traces" on this blog post:
https://medium.com/flutter/profiling-flutter-applications-using-the-timeline-a1a434964af3#a499
Make sure the performance overlay is turned OFF when recording the
trace as that may affect the performance of the profile run.
(Pressing ‘P’ on the command line toggles the overlay.)
-->
I'm very happy to see the desktop support now accessible from the stable release of Flutter! So I upgraded my application to the stable release. However, I'm getting a performance regression on older iOS devices, by that I mean that the application freeze during animation (popup animation or during a page navigation) and while playing videos.
At the beginning I was thinking this could be related to the jank problem everyone is talking about. However, the devices that I'm having problems with are older iOS devices (iPad 2 and iPad 4) which should use the openGL renderer instead of the Metal renderer.
By further investigation, I found that the problem started from Flutter 1.23.0-13.0.pre and above since everything is fine on Flutter 1.23.0-7.0.pre and below; by that I mean there's still some jank, but the application doesn't freeze and is a lot more responsible and usable.
If this is useful to help solve this problem, the commit messages of Flutter between 2020-10-01 and 2020-10-12, which is the timeframe between Flutter 1.23.0-7.0.pre and 1.23.0-13.0.pre, can be found here: https://github.com/flutter/flutter/commits/dev?since=2020-10-01&until=2020-10-12
I've tested with Flutter Stable 2.0.2 and the Master version and both have the problem and seems to be related to the drawing to the surface which takes a very long time.
I've tested by creating 2 small applications:
1. The default application created by Flutter (the simple counter create with `flutter create`)
2. A simple video player based on the video_player plugin. Gist: https://gist.github.com/guibarrette/2cf49c010be2ae231b259d8de7bf2e19
The first application feel more janky on newer Flutter version than 1.23.0-7.0.pre, but the later simply doesn't let the video play; it simply freeze, but is working fine on 1.23.0-7.0.pre.
### Here are 3 videos showing the different performance of those Flutter versions with the second test application:
I've tested with `video_player 1.0.1` to make it compatible with Flutter 1.23.0-7.0.pre and 2.0.2, but upgraded to `video_player 2.0.2` for the Flutter Master version. Sorry, I did the videos with the performance overlay since I didn't see at that time that it was preferable without and it helped myself to see the performance freeze, but you can see on the video of the Master version that the freeze is there on a `--release` version and without the performance overlay.
#### 1.23.0-7.0.pre:
Using video_player 1.0.1
Observatory timeline trace: [dart-timeline-2021-3-17-Flutter-1_23-Video.zip](https://github.com/flutter/flutter/files/6161321/dart-timeline-2021-3-17-Flutter-1_23-Video.zip)
Video: https://d.pr/v/tLbQj4
#### Stable 2.0.2:
Using video_player 1.0.1
Observatory timeline trace: [dart-timeline-2021-3-17-Flutter-2_0_2-Video.zip](https://github.com/flutter/flutter/files/6161336/dart-timeline-2021-3-17-Flutter-2_0_2-Video.zip)
Video: https://d.pr/v/TcW4Rx
#### Stable 2.0.2 (same build, but longer video):
Using video_player 1.0.1
Observatory timeline trace: [dart-timeline-2021-3-17-Flutter-2_0_2-Video-2.zip](https://github.com/flutter/flutter/files/6161342/dart-timeline-2021-3-17-Flutter-2_0_2-Video-2.zip)
Video: https://d.pr/v/t3oG8S
#### Master (2021-03-17):
Using video_player 2.0.2
Video: https://d.pr/v/KiVKL0
With the latest Master version I'm not able to `flutter run --profile --trace-skia` on my iOS device since I'm using macOS Mojave (10.14.6) with Xcode 11.3.1, but I changed the minimum Xcode version of Flutter so it let me compile and I uploaded manually with ios-deploy. So for this video I've built with `flutter build ios --release`. Therefore, I'm simply attaching the video here without the observatory timeline to show that it freezes the same as with Flutter 2.0.2 and even when the performance overlay is not active.
<!--
Please tell us which target platform(s) the problem occurs (Android / iOS / Web / macOS / Linux / Windows)
Which target OS version, for Web, browser, is the test system running?
Does the problem occur on emulator/simulator as well as on physical devices?
-->
**Target Platform:**
- iOS
**Target OS version/browser:**
Tested and problematic on:
- iOS 9.3.5
- iOS 10.3.3
**Devices:**
- iPad 2 (iOS 9.3.5)
- iPad 4 (iOS 10.3.3)
## Logs
<details>
<summary>Logs</summary>
<!--
Run `flutter analyze` and attach any output of that command below.
If there are any analysis errors, try resolving them before filing this issue.
-->
```
Running "flutter pub get" in flutter_ios_video_freeze_master... 1 652ms
Analyzing flutter_ios_video_freeze_master...
No issues found! (ran in 4.8s)
```
<!-- Finally, paste the output of running `flutter doctor -v` here, with your device plugged in. -->
### Flutter Doctor:
#### 1.23.0-7.0.pre:
```
[✓] Flutter (Channel dev, 1.23.0-7.0.pre, on Mac OS X 10.14.6 18G6032 x86_64, locale fr-CA)
• Flutter version 1.23.0-7.0.pre at /Users/GuillaumeBarrette/Documents/Programmation/Flutter/flutter
• Framework revision db6e2d8aa5 (il y a 6 mois), 2020-09-25 06:47:03 -0400
• Engine revision 3a73d073c8
• Dart version 2.11.0 (build 2.11.0-161.0.dev)
[✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3)
• Android SDK at /Users/GuillaumeBarrette/Library/Android/sdk
• Platform android-29, build-tools 28.0.3
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS (Xcode 11.3.1)
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 11.3.1, Build version 11C505
• CocoaPods version 1.10.1
[✓] Android Studio (version 3.5)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 43.0.1
• Dart plugin version 191.8593
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
[✓] VS Code (version 1.54.3)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.20.0
[✓] Connected device (2 available)
• IPad PSM-11 (mobile) • bdfb2550f9cb07f766bde4185e0ae060c833cb23 • ios • iOS 10.3.3
• macOS (desktop) • macos • darwin-x64 • Mac OS X 10.14.6 18G6032 x86_64
• No issues found!
```
#### Stable 2.0.2:
```
[✓] Flutter (Channel stable, 2.0.2, on Mac OS X 10.14.6 18G6032 darwin-x64, locale fr-CA)
• Flutter version 2.0.2 at /Users/GuillaumeBarrette/Documents/Programmation/Flutter/flutter
• Framework revision 8962f6dc68 (il y a 6 jours), 2021-03-11 13:22:20 -0800
• Engine revision 5d8bf811b3
• Dart version 2.12.1
[✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3)
• Android SDK at /Users/GuillaumeBarrette/Library/Android/sdk
• Platform android-29, build-tools 28.0.3
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
• All Android licenses accepted.
[!] Xcode - develop for iOS and macOS
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 11.3.1, Build version 11C505
! Xcode 11.3.1 out of date (12.0.1 is recommended).
Download the latest version or update via the Mac App Store.
• CocoaPods version 1.10.1
[✓] Chrome - develop for the web
• Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome
[✓] Android Studio (version 3.5)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 43.0.1
• Dart plugin version 191.8593
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
[✓] VS Code (version 1.54.3)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.20.0
[✓] Connected device (3 available)
• IPad PSM-11 (mobile) • bdfb2550f9cb07f766bde4185e0ae060c833cb23 • ios • iOS 10.3.3
• macOS (desktop) • macos • darwin-x64 • Mac OS X 10.14.6 18G6032 darwin-x64
• Chrome (web) • chrome • web-javascript • Google Chrome 89.0.4389.90
! Doctor found issues in 1 category.
```
#### Master (2021-03-17):
```
[✓] Flutter (Channel master, 2.1.0-13.0.pre.149, on Mac OS X 10.14.6 18G6032 darwin-x64, locale fr-CA)
• Flutter version 2.1.0-13.0.pre.149 at /Users/GuillaumeBarrette/Documents/Programmation/Flutter/flutter
• Framework revision 637fb5b9be (il y a 38 minutes), 2021-03-17 18:25:04 -0400
• Engine revision c0e6763e6b
• Dart version 2.13.0 (build 2.13.0-147.0.dev)
[✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3)
• Android SDK at /Users/GuillaumeBarrette/Library/Android/sdk
• Platform android-29, build-tools 28.0.3
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 11.3.1, Build version 11C505
• CocoaPods version 1.10.1
[✓] Chrome - develop for the web
• Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome
[✓] Android Studio (version 3.5)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 43.0.1
• Dart plugin version 191.8593
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
[✓] VS Code (version 1.54.3)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.20.0
[✓] Connected device (3 available)
• IPad PSM-11 (mobile) • bdfb2550f9cb07f766bde4185e0ae060c833cb23 • ios • iOS 10.3.3
• macOS (desktop) • macos • darwin-x64 • Mac OS X 10.14.6 18G6032 darwin-x64
• Chrome (web) • chrome • web-javascript • Google Chrome 89.0.4389.90
• No issues found!
```
</details> | True | Regression in performance from Flutter 1.23.0-13.0.pre and above on older iOS devices - <!-- Title: Regression in performance from Flutter 1.23.0-13.0.pre and above on older iOS devices -->
## Details
<!--
1. Please tell us exactly how to reproduce the problem you are running into.
2. Please attach a small application (ideally just one main.dart file) that
reproduces the problem. You could use https://gist.github.com/ for this.
3. Switch flutter to master channel and run this app on a physical device
using profile mode with Skia tracing enabled, as follows:
flutter channel master
flutter run --profile --trace-skia
The bleeding edge master channel is encouraged here because Flutter is
constantly fixing bugs and improving its performance. Your problem in an
older Flutter version may have already been solved in the master channel.
4. Record a video of the performance issue using another phone so we
can have an intuitive understanding of what happened. Don’t use
"adb screenrecord", as that affects the performance of the profile run.
5. Open Observatory and save a timeline trace of the performance issue
so we know which functions might be causing it. See "How to Collect
and Read Timeline Traces" on this blog post:
https://medium.com/flutter/profiling-flutter-applications-using-the-timeline-a1a434964af3#a499
Make sure the performance overlay is turned OFF when recording the
trace as that may affect the performance of the profile run.
(Pressing ‘P’ on the command line toggles the overlay.)
-->
I'm very happy to see the desktop support now accessible from the stable release of Flutter! So I upgraded my application to the stable release. However, I'm getting a performance regression on older iOS devices, by that I mean that the application freeze during animation (popup animation or during a page navigation) and while playing videos.
At the beginning I was thinking this could be related to the jank problem everyone is talking about. However, the devices that I'm having problems with are older iOS devices (iPad 2 and iPad 4) which should use the openGL renderer instead of the Metal renderer.
By further investigation, I found that the problem started from Flutter 1.23.0-13.0.pre and above since everything is fine on Flutter 1.23.0-7.0.pre and below; by that I mean there's still some jank, but the application doesn't freeze and is a lot more responsible and usable.
If this is useful to help solve this problem, the commit messages of Flutter between 2020-10-01 and 2020-10-12, which is the timeframe between Flutter 1.23.0-7.0.pre and 1.23.0-13.0.pre, can be found here: https://github.com/flutter/flutter/commits/dev?since=2020-10-01&until=2020-10-12
I've tested with Flutter Stable 2.0.2 and the Master version and both have the problem and seems to be related to the drawing to the surface which takes a very long time.
I've tested by creating 2 small applications:
1. The default application created by Flutter (the simple counter create with `flutter create`)
2. A simple video player based on the video_player plugin. Gist: https://gist.github.com/guibarrette/2cf49c010be2ae231b259d8de7bf2e19
The first application feel more janky on newer Flutter version than 1.23.0-7.0.pre, but the later simply doesn't let the video play; it simply freeze, but is working fine on 1.23.0-7.0.pre.
### Here are 3 videos showing the different performance of those Flutter versions with the second test application:
I've tested with `video_player 1.0.1` to make it compatible with Flutter 1.23.0-7.0.pre and 2.0.2, but upgraded to `video_player 2.0.2` for the Flutter Master version. Sorry, I did the videos with the performance overlay since I didn't see at that time that it was preferable without and it helped myself to see the performance freeze, but you can see on the video of the Master version that the freeze is there on a `--release` version and without the performance overlay.
#### 1.23.0-7.0.pre:
Using video_player 1.0.1
Observatory timeline trace: [dart-timeline-2021-3-17-Flutter-1_23-Video.zip](https://github.com/flutter/flutter/files/6161321/dart-timeline-2021-3-17-Flutter-1_23-Video.zip)
Video: https://d.pr/v/tLbQj4
#### Stable 2.0.2:
Using video_player 1.0.1
Observatory timeline trace: [dart-timeline-2021-3-17-Flutter-2_0_2-Video.zip](https://github.com/flutter/flutter/files/6161336/dart-timeline-2021-3-17-Flutter-2_0_2-Video.zip)
Video: https://d.pr/v/TcW4Rx
#### Stable 2.0.2 (same build, but longer video):
Using video_player 1.0.1
Observatory timeline trace: [dart-timeline-2021-3-17-Flutter-2_0_2-Video-2.zip](https://github.com/flutter/flutter/files/6161342/dart-timeline-2021-3-17-Flutter-2_0_2-Video-2.zip)
Video: https://d.pr/v/t3oG8S
#### Master (2021-03-17):
Using video_player 2.0.2
Video: https://d.pr/v/KiVKL0
With the latest Master version I'm not able to `flutter run --profile --trace-skia` on my iOS device since I'm using macOS Mojave (10.14.6) with Xcode 11.3.1, but I changed the minimum Xcode version of Flutter so it let me compile and I uploaded manually with ios-deploy. So for this video I've built with `flutter build ios --release`. Therefore, I'm simply attaching the video here without the observatory timeline to show that it freezes the same as with Flutter 2.0.2 and even when the performance overlay is not active.
<!--
Please tell us which target platform(s) the problem occurs (Android / iOS / Web / macOS / Linux / Windows)
Which target OS version, for Web, browser, is the test system running?
Does the problem occur on emulator/simulator as well as on physical devices?
-->
**Target Platform:**
- iOS
**Target OS version/browser:**
Tested and problematic on:
- iOS 9.3.5
- iOS 10.3.3
**Devices:**
- iPad 2 (iOS 9.3.5)
- iPad 4 (iOS 10.3.3)
## Logs
<details>
<summary>Logs</summary>
<!--
Run `flutter analyze` and attach any output of that command below.
If there are any analysis errors, try resolving them before filing this issue.
-->
```
Running "flutter pub get" in flutter_ios_video_freeze_master... 1 652ms
Analyzing flutter_ios_video_freeze_master...
No issues found! (ran in 4.8s)
```
<!-- Finally, paste the output of running `flutter doctor -v` here, with your device plugged in. -->
### Flutter Doctor:
#### 1.23.0-7.0.pre:
```
[✓] Flutter (Channel dev, 1.23.0-7.0.pre, on Mac OS X 10.14.6 18G6032 x86_64, locale fr-CA)
• Flutter version 1.23.0-7.0.pre at /Users/GuillaumeBarrette/Documents/Programmation/Flutter/flutter
• Framework revision db6e2d8aa5 (il y a 6 mois), 2020-09-25 06:47:03 -0400
• Engine revision 3a73d073c8
• Dart version 2.11.0 (build 2.11.0-161.0.dev)
[✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3)
• Android SDK at /Users/GuillaumeBarrette/Library/Android/sdk
• Platform android-29, build-tools 28.0.3
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS (Xcode 11.3.1)
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 11.3.1, Build version 11C505
• CocoaPods version 1.10.1
[✓] Android Studio (version 3.5)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 43.0.1
• Dart plugin version 191.8593
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
[✓] VS Code (version 1.54.3)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.20.0
[✓] Connected device (2 available)
• IPad PSM-11 (mobile) • bdfb2550f9cb07f766bde4185e0ae060c833cb23 • ios • iOS 10.3.3
• macOS (desktop) • macos • darwin-x64 • Mac OS X 10.14.6 18G6032 x86_64
• No issues found!
```
#### Stable 2.0.2:
```
[✓] Flutter (Channel stable, 2.0.2, on Mac OS X 10.14.6 18G6032 darwin-x64, locale fr-CA)
• Flutter version 2.0.2 at /Users/GuillaumeBarrette/Documents/Programmation/Flutter/flutter
• Framework revision 8962f6dc68 (il y a 6 jours), 2021-03-11 13:22:20 -0800
• Engine revision 5d8bf811b3
• Dart version 2.12.1
[✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3)
• Android SDK at /Users/GuillaumeBarrette/Library/Android/sdk
• Platform android-29, build-tools 28.0.3
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
• All Android licenses accepted.
[!] Xcode - develop for iOS and macOS
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 11.3.1, Build version 11C505
! Xcode 11.3.1 out of date (12.0.1 is recommended).
Download the latest version or update via the Mac App Store.
• CocoaPods version 1.10.1
[✓] Chrome - develop for the web
• Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome
[✓] Android Studio (version 3.5)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 43.0.1
• Dart plugin version 191.8593
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
[✓] VS Code (version 1.54.3)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.20.0
[✓] Connected device (3 available)
• IPad PSM-11 (mobile) • bdfb2550f9cb07f766bde4185e0ae060c833cb23 • ios • iOS 10.3.3
• macOS (desktop) • macos • darwin-x64 • Mac OS X 10.14.6 18G6032 darwin-x64
• Chrome (web) • chrome • web-javascript • Google Chrome 89.0.4389.90
! Doctor found issues in 1 category.
```
#### Master (2021-03-17):
```
[✓] Flutter (Channel master, 2.1.0-13.0.pre.149, on Mac OS X 10.14.6 18G6032 darwin-x64, locale fr-CA)
• Flutter version 2.1.0-13.0.pre.149 at /Users/GuillaumeBarrette/Documents/Programmation/Flutter/flutter
• Framework revision 637fb5b9be (il y a 38 minutes), 2021-03-17 18:25:04 -0400
• Engine revision c0e6763e6b
• Dart version 2.13.0 (build 2.13.0-147.0.dev)
[✓] Android toolchain - develop for Android devices (Android SDK version 28.0.3)
• Android SDK at /Users/GuillaumeBarrette/Library/Android/sdk
• Platform android-29, build-tools 28.0.3
• Java binary at: /Applications/Android Studio.app/Contents/jre/jdk/Contents/Home/bin/java
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
• All Android licenses accepted.
[✓] Xcode - develop for iOS and macOS
• Xcode at /Applications/Xcode.app/Contents/Developer
• Xcode 11.3.1, Build version 11C505
• CocoaPods version 1.10.1
[✓] Chrome - develop for the web
• Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome
[✓] Android Studio (version 3.5)
• Android Studio at /Applications/Android Studio.app/Contents
• Flutter plugin version 43.0.1
• Dart plugin version 191.8593
• Java version OpenJDK Runtime Environment (build 1.8.0_202-release-1483-b49-5587405)
[✓] VS Code (version 1.54.3)
• VS Code at /Applications/Visual Studio Code.app/Contents
• Flutter extension version 3.20.0
[✓] Connected device (3 available)
• IPad PSM-11 (mobile) • bdfb2550f9cb07f766bde4185e0ae060c833cb23 • ios • iOS 10.3.3
• macOS (desktop) • macos • darwin-x64 • Mac OS X 10.14.6 18G6032 darwin-x64
• Chrome (web) • chrome • web-javascript • Google Chrome 89.0.4389.90
• No issues found!
```
</details> | non_priority | regression in performance from flutter pre and above on older ios devices details please tell us exactly how to reproduce the problem you are running into please attach a small application ideally just one main dart file that reproduces the problem you could use for this switch flutter to master channel and run this app on a physical device using profile mode with skia tracing enabled as follows flutter channel master flutter run profile trace skia the bleeding edge master channel is encouraged here because flutter is constantly fixing bugs and improving its performance your problem in an older flutter version may have already been solved in the master channel record a video of the performance issue using another phone so we can have an intuitive understanding of what happened don’t use adb screenrecord as that affects the performance of the profile run open observatory and save a timeline trace of the performance issue so we know which functions might be causing it see how to collect and read timeline traces on this blog post make sure the performance overlay is turned off when recording the trace as that may affect the performance of the profile run pressing ‘p’ on the command line toggles the overlay i m very happy to see the desktop support now accessible from the stable release of flutter so i upgraded my application to the stable release however i m getting a performance regression on older ios devices by that i mean that the application freeze during animation popup animation or during a page navigation and while playing videos at the beginning i was thinking this could be related to the jank problem everyone is talking about however the devices that i m having problems with are older ios devices ipad and ipad which should use the opengl renderer instead of the metal renderer by further investigation i found that the problem started from flutter pre and above since everything is fine on flutter pre and below by that i mean there s still some jank but the application doesn t freeze and is a lot more responsible and usable if this is useful to help solve this problem the commit messages of flutter between and which is the timeframe between flutter pre and pre can be found here i ve tested with flutter stable and the master version and both have the problem and seems to be related to the drawing to the surface which takes a very long time i ve tested by creating small applications the default application created by flutter the simple counter create with flutter create a simple video player based on the video player plugin gist the first application feel more janky on newer flutter version than pre but the later simply doesn t let the video play it simply freeze but is working fine on pre here are videos showing the different performance of those flutter versions with the second test application i ve tested with video player to make it compatible with flutter pre and but upgraded to video player for the flutter master version sorry i did the videos with the performance overlay since i didn t see at that time that it was preferable without and it helped myself to see the performance freeze but you can see on the video of the master version that the freeze is there on a release version and without the performance overlay pre using video player observatory timeline trace video stable using video player observatory timeline trace video stable same build but longer video using video player observatory timeline trace video master using video player video with the latest master version i m not able to flutter run profile trace skia on my ios device since i m using macos mojave with xcode but i changed the minimum xcode version of flutter so it let me compile and i uploaded manually with ios deploy so for this video i ve built with flutter build ios release therefore i m simply attaching the video here without the observatory timeline to show that it freezes the same as with flutter and even when the performance overlay is not active please tell us which target platform s the problem occurs android ios web macos linux windows which target os version for web browser is the test system running does the problem occur on emulator simulator as well as on physical devices target platform ios target os version browser tested and problematic on ios ios devices ipad ios ipad ios logs logs run flutter analyze and attach any output of that command below if there are any analysis errors try resolving them before filing this issue running flutter pub get in flutter ios video freeze master analyzing flutter ios video freeze master no issues found ran in flutter doctor pre flutter channel dev pre on mac os x locale fr ca • flutter version pre at users guillaumebarrette documents programmation flutter flutter • framework revision il y a mois • engine revision • dart version build dev android toolchain develop for android devices android sdk version • android sdk at users guillaumebarrette library android sdk • platform android build tools • java binary at applications android studio app contents jre jdk contents home bin java • java version openjdk runtime environment build release • all android licenses accepted xcode develop for ios and macos xcode • xcode at applications xcode app contents developer • xcode build version • cocoapods version android studio version • android studio at applications android studio app contents • flutter plugin version • dart plugin version • java version openjdk runtime environment build release vs code version • vs code at applications visual studio code app contents • flutter extension version connected device available • ipad psm mobile • • ios • ios • macos desktop • macos • darwin • mac os x • no issues found stable flutter channel stable on mac os x darwin locale fr ca • flutter version at users guillaumebarrette documents programmation flutter flutter • framework revision il y a jours • engine revision • dart version android toolchain develop for android devices android sdk version • android sdk at users guillaumebarrette library android sdk • platform android build tools • java binary at applications android studio app contents jre jdk contents home bin java • java version openjdk runtime environment build release • all android licenses accepted xcode develop for ios and macos • xcode at applications xcode app contents developer • xcode build version xcode out of date is recommended download the latest version or update via the mac app store • cocoapods version chrome develop for the web • chrome at applications google chrome app contents macos google chrome android studio version • android studio at applications android studio app contents • flutter plugin version • dart plugin version • java version openjdk runtime environment build release vs code version • vs code at applications visual studio code app contents • flutter extension version connected device available • ipad psm mobile • • ios • ios • macos desktop • macos • darwin • mac os x darwin • chrome web • chrome • web javascript • google chrome doctor found issues in category master flutter channel master pre on mac os x darwin locale fr ca • flutter version pre at users guillaumebarrette documents programmation flutter flutter • framework revision il y a minutes • engine revision • dart version build dev android toolchain develop for android devices android sdk version • android sdk at users guillaumebarrette library android sdk • platform android build tools • java binary at applications android studio app contents jre jdk contents home bin java • java version openjdk runtime environment build release • all android licenses accepted xcode develop for ios and macos • xcode at applications xcode app contents developer • xcode build version • cocoapods version chrome develop for the web • chrome at applications google chrome app contents macos google chrome android studio version • android studio at applications android studio app contents • flutter plugin version • dart plugin version • java version openjdk runtime environment build release vs code version • vs code at applications visual studio code app contents • flutter extension version connected device available • ipad psm mobile • • ios • ios • macos desktop • macos • darwin • mac os x darwin • chrome web • chrome • web javascript • google chrome • no issues found | 0 |
46,091 | 13,055,850,864 | IssuesEvent | 2020-07-30 02:55:26 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | Mac linker is getting repeated boost libraries during build (Trac #570) | Incomplete Migration Migrated from Trac cmake defect | Migrated from https://code.icecube.wisc.edu/ticket/570
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "On Mac OS X targets (including teufel), building libraries you'll see:\n\n\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_python-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_system-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_signals-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_thread-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_date_time-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_serialization-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_filesystem-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_program_options-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_regex-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_iostreams-mt.dylib\nld warning: duplicate dylib /usr/lib/libpthread.dylib\n\n\nA VERBOSE link looks like:\nLinking CXX shared library ../lib/libI3Db.dylib\ncd /Users/blaufuss/icework/offline-software/trunk/build_release/I3Db && mkdir -p /Users/blaufuss/icework/offline-software/trunk/build_release/lib\ncd /Users/blaufuss/icework/offline-software/trunk/build_release/I3Db && /Users/blaufuss/icework/i3tools/bin/cmake -E cmake_link_script CMakeFiles/I3Db.dir/link.txt --verbose=1\n/usr/bin/env MACOSX_DEPLOYMENT_TARGET=10.4 /usr/bin/c++ -O3 -Wno-unused-variable -DNDEBUG -DI3_OPTIMIZE -dynamiclib -headerpad_max_install_names -single_module -undefined dynamic_lookup -flat_namespace -o ../lib/libI3Db.dylib -install_name /Users/blaufuss/icework/offline-software/trunk/build_release/lib/libI3Db.dylib CMakeFiles/I3Db.dir/private/I3Db/AmandaTWRGlobalConstants_t.cxx.o CMakeFiles/I3Db.dir/private/I3Db/AmandaTWRTrigger_t.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3Db.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbCalibrationService.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbCalibrationServiceFactory.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbDetectorStatusService.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbDetectorStatusServiceFactory.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbGeometryService.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbGeometryServiceFactory.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbOMKey2ChannelID.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbOMKey2ChannelIDFactory.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbOMKey2MBID.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbOMKey2MBIDFactory.cxx.o CMakeFiles/I3Db.dir/private/I3OmDb/I3OmDb.cxx.o CMakeFiles/I3Db.dir/private/I3OmDb/I3OmDbFactory.cxx.o CMakeFiles/I3Db.dir/private/I3OmDb/ZSql.cxx.o -lm -ldl -lstdc++ ../lib/libinterfaces.dylib ../lib/libphys-services.dylib ../lib/libicetray.dylib ../lib/libdataclasses.dylib ../lib/libdaq-decode.dylib /Users/blaufuss/icework/i3tools/lib/mysql-4.1.20/mysql/libmysqlclient.dylib /Users/blaufuss/icework/i3tools/lib/log4cplus-1.0.2/liblog4cplus.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_python-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_system-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_signals-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_thread-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_date_time-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_serialization-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_filesystem-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_program_options-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_regex-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_iostreams-mt.dylib /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libCore.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libCint.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libRIO.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libNet.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libHist.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGraf.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGraf3d.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGpad.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libTree.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libRint.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libPostscript.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMatrix.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libPhysics.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMathCore.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libThread.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMinuit.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGui.so /usr/lib/libpthread.dylib ../lib/libphys-services.dylib ../lib/libinterfaces.dylib /Users/blaufuss/icework/i3tools/lib/sprng-2.0a/libsprng.a /Users/blaufuss/icework/i3tools/lib/gsl-1.8/libgsl.dylib /Users/blaufuss/icework/i3tools/lib/gsl-1.8/libgslcblas.dylib ../lib/libpfclasses.dylib ../lib/libjebclasses.dylib ../lib/libdataclasses.dylib ../lib/libicetray.dylib -lm -ldl -lstdc++ -framework Python /Users/blaufuss/icework/i3tools/lib/log4cplus-1.0.2/liblog4cplus.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_python-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_system-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_signals-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_thread-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_date_time-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_serialization-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_filesystem-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_program_options-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_regex-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_iostreams-mt.dylib /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libCore.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libCint.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libRIO.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libNet.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libHist.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGraf.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGraf3d.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGpad.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libTree.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libRint.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libPostscript.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMatrix.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libPhysics.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMathCore.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libThread.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMinuit.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGui.so /usr/lib/libpthread.dylib \nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_python-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_system-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_signals-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_thread-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_date_time-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_serialization-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_filesystem-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_program_options-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_regex-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_iostreams-mt.dylib\nld warning: duplicate dylib /usr/lib/libpthread.dylib\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1416713876862109",
"component": "cmake",
"summary": "Mac linker is getting repeated boost libraries during build",
"priority": "normal",
"keywords": "",
"time": "2009-07-31T20:09:58",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| 1.0 | Mac linker is getting repeated boost libraries during build (Trac #570) - Migrated from https://code.icecube.wisc.edu/ticket/570
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "On Mac OS X targets (including teufel), building libraries you'll see:\n\n\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_python-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_system-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_signals-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_thread-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_date_time-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_serialization-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_filesystem-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_program_options-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_regex-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_iostreams-mt.dylib\nld warning: duplicate dylib /usr/lib/libpthread.dylib\n\n\nA VERBOSE link looks like:\nLinking CXX shared library ../lib/libI3Db.dylib\ncd /Users/blaufuss/icework/offline-software/trunk/build_release/I3Db && mkdir -p /Users/blaufuss/icework/offline-software/trunk/build_release/lib\ncd /Users/blaufuss/icework/offline-software/trunk/build_release/I3Db && /Users/blaufuss/icework/i3tools/bin/cmake -E cmake_link_script CMakeFiles/I3Db.dir/link.txt --verbose=1\n/usr/bin/env MACOSX_DEPLOYMENT_TARGET=10.4 /usr/bin/c++ -O3 -Wno-unused-variable -DNDEBUG -DI3_OPTIMIZE -dynamiclib -headerpad_max_install_names -single_module -undefined dynamic_lookup -flat_namespace -o ../lib/libI3Db.dylib -install_name /Users/blaufuss/icework/offline-software/trunk/build_release/lib/libI3Db.dylib CMakeFiles/I3Db.dir/private/I3Db/AmandaTWRGlobalConstants_t.cxx.o CMakeFiles/I3Db.dir/private/I3Db/AmandaTWRTrigger_t.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3Db.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbCalibrationService.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbCalibrationServiceFactory.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbDetectorStatusService.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbDetectorStatusServiceFactory.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbGeometryService.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbGeometryServiceFactory.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbOMKey2ChannelID.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbOMKey2ChannelIDFactory.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbOMKey2MBID.cxx.o CMakeFiles/I3Db.dir/private/I3Db/I3DbOMKey2MBIDFactory.cxx.o CMakeFiles/I3Db.dir/private/I3OmDb/I3OmDb.cxx.o CMakeFiles/I3Db.dir/private/I3OmDb/I3OmDbFactory.cxx.o CMakeFiles/I3Db.dir/private/I3OmDb/ZSql.cxx.o -lm -ldl -lstdc++ ../lib/libinterfaces.dylib ../lib/libphys-services.dylib ../lib/libicetray.dylib ../lib/libdataclasses.dylib ../lib/libdaq-decode.dylib /Users/blaufuss/icework/i3tools/lib/mysql-4.1.20/mysql/libmysqlclient.dylib /Users/blaufuss/icework/i3tools/lib/log4cplus-1.0.2/liblog4cplus.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_python-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_system-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_signals-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_thread-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_date_time-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_serialization-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_filesystem-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_program_options-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_regex-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_iostreams-mt.dylib /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libCore.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libCint.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libRIO.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libNet.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libHist.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGraf.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGraf3d.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGpad.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libTree.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libRint.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libPostscript.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMatrix.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libPhysics.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMathCore.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libThread.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMinuit.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGui.so /usr/lib/libpthread.dylib ../lib/libphys-services.dylib ../lib/libinterfaces.dylib /Users/blaufuss/icework/i3tools/lib/sprng-2.0a/libsprng.a /Users/blaufuss/icework/i3tools/lib/gsl-1.8/libgsl.dylib /Users/blaufuss/icework/i3tools/lib/gsl-1.8/libgslcblas.dylib ../lib/libpfclasses.dylib ../lib/libjebclasses.dylib ../lib/libdataclasses.dylib ../lib/libicetray.dylib -lm -ldl -lstdc++ -framework Python /Users/blaufuss/icework/i3tools/lib/log4cplus-1.0.2/liblog4cplus.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_python-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_system-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_signals-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_thread-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_date_time-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_serialization-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_filesystem-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_program_options-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_regex-mt.dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_iostreams-mt.dylib /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libCore.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libCint.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libRIO.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libNet.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libHist.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGraf.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGraf3d.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGpad.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libTree.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libRint.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libPostscript.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMatrix.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libPhysics.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMathCore.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libThread.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libMinuit.so /Users/blaufuss/icework/i3tools/root-v5.20.00/lib/libGui.so /usr/lib/libpthread.dylib \nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_python-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_system-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_signals-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_thread-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_date_time-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_serialization-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_filesystem-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_program_options-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_regex-mt.dylib\nld warning: duplicate dylib /Users/blaufuss/icework/i3tools/lib/boost-1.38.0/libboost_iostreams-mt.dylib\nld warning: duplicate dylib /usr/lib/libpthread.dylib\n",
"reporter": "blaufuss",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1416713876862109",
"component": "cmake",
"summary": "Mac linker is getting repeated boost libraries during build",
"priority": "normal",
"keywords": "",
"time": "2009-07-31T20:09:58",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
| non_priority | mac linker is getting repeated boost libraries during build trac migrated from json status closed changetime description on mac os x targets including teufel building libraries you ll see n n nld warning duplicate dylib users blaufuss icework lib boost libboost python mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost system mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost signals mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost thread mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost date time mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost serialization mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost filesystem mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost program options mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost regex mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost iostreams mt dylib nld warning duplicate dylib usr lib libpthread dylib n n na verbose link looks like nlinking cxx shared library lib dylib ncd users blaufuss icework offline software trunk build release mkdir p users blaufuss icework offline software trunk build release lib ncd users blaufuss icework offline software trunk build release users blaufuss icework bin cmake e cmake link script cmakefiles dir link txt verbose n usr bin env macosx deployment target usr bin c wno unused variable dndebug optimize dynamiclib headerpad max install names single module undefined dynamic lookup flat namespace o lib dylib install name users blaufuss icework offline software trunk build release lib dylib cmakefiles dir private amandatwrglobalconstants t cxx o cmakefiles dir private amandatwrtrigger t cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private cxx o cmakefiles dir private zsql cxx o lm ldl lstdc lib libinterfaces dylib lib libphys services dylib lib libicetray dylib lib libdataclasses dylib lib libdaq decode dylib users blaufuss icework lib mysql mysql libmysqlclient dylib users blaufuss icework lib dylib users blaufuss icework lib boost libboost python mt dylib users blaufuss icework lib boost libboost system mt dylib users blaufuss icework lib boost libboost signals mt dylib users blaufuss icework lib boost libboost thread mt dylib users blaufuss icework lib boost libboost date time mt dylib users blaufuss icework lib boost libboost serialization mt dylib users blaufuss icework lib boost libboost filesystem mt dylib users blaufuss icework lib boost libboost program options mt dylib users blaufuss icework lib boost libboost regex mt dylib users blaufuss icework lib boost libboost iostreams mt dylib users blaufuss icework root lib libcore so users blaufuss icework root lib libcint so users blaufuss icework root lib librio so users blaufuss icework root lib libnet so users blaufuss icework root lib libhist so users blaufuss icework root lib libgraf so users blaufuss icework root lib so users blaufuss icework root lib libgpad so users blaufuss icework root lib libtree so users blaufuss icework root lib librint so users blaufuss icework root lib libpostscript so users blaufuss icework root lib libmatrix so users blaufuss icework root lib libphysics so users blaufuss icework root lib libmathcore so users blaufuss icework root lib libthread so users blaufuss icework root lib libminuit so users blaufuss icework root lib libgui so usr lib libpthread dylib lib libphys services dylib lib libinterfaces dylib users blaufuss icework lib sprng libsprng a users blaufuss icework lib gsl libgsl dylib users blaufuss icework lib gsl libgslcblas dylib lib libpfclasses dylib lib libjebclasses dylib lib libdataclasses dylib lib libicetray dylib lm ldl lstdc framework python users blaufuss icework lib dylib users blaufuss icework lib boost libboost python mt dylib users blaufuss icework lib boost libboost system mt dylib users blaufuss icework lib boost libboost signals mt dylib users blaufuss icework lib boost libboost thread mt dylib users blaufuss icework lib boost libboost date time mt dylib users blaufuss icework lib boost libboost serialization mt dylib users blaufuss icework lib boost libboost filesystem mt dylib users blaufuss icework lib boost libboost program options mt dylib users blaufuss icework lib boost libboost regex mt dylib users blaufuss icework lib boost libboost iostreams mt dylib users blaufuss icework root lib libcore so users blaufuss icework root lib libcint so users blaufuss icework root lib librio so users blaufuss icework root lib libnet so users blaufuss icework root lib libhist so users blaufuss icework root lib libgraf so users blaufuss icework root lib so users blaufuss icework root lib libgpad so users blaufuss icework root lib libtree so users blaufuss icework root lib librint so users blaufuss icework root lib libpostscript so users blaufuss icework root lib libmatrix so users blaufuss icework root lib libphysics so users blaufuss icework root lib libmathcore so users blaufuss icework root lib libthread so users blaufuss icework root lib libminuit so users blaufuss icework root lib libgui so usr lib libpthread dylib nld warning duplicate dylib users blaufuss icework lib boost libboost python mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost system mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost signals mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost thread mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost date time mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost serialization mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost filesystem mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost program options mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost regex mt dylib nld warning duplicate dylib users blaufuss icework lib boost libboost iostreams mt dylib nld warning duplicate dylib usr lib libpthread dylib n reporter blaufuss cc resolution wont or cant fix ts component cmake summary mac linker is getting repeated boost libraries during build priority normal keywords time milestone owner troy type defect | 0 |
230,666 | 25,482,744,264 | IssuesEvent | 2022-11-26 01:23:17 | Satheesh575555/linux-4.1.15 | https://api.github.com/repos/Satheesh575555/linux-4.1.15 | reopened | CVE-2016-7425 (High) detected in linuxlinux-4.6 | security vulnerability | ## CVE-2016-7425 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15/commit/951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1">951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/arcmsr/arcmsr_hba.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/arcmsr/arcmsr_hba.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The arcmsr_iop_message_xfer function in drivers/scsi/arcmsr/arcmsr_hba.c in the Linux kernel through 4.8.2 does not restrict a certain length field, which allows local users to gain privileges or cause a denial of service (heap-based buffer overflow) via an ARCMSR_MESSAGE_WRITE_WQBUFFER control code.
<p>Publish Date: 2016-10-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-7425>CVE-2016-7425</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-7425">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-7425</a></p>
<p>Release Date: 2016-10-16</p>
<p>Fix Resolution: v4.9-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-7425 (High) detected in linuxlinux-4.6 - ## CVE-2016-7425 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15/commit/951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1">951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/arcmsr/arcmsr_hba.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/scsi/arcmsr/arcmsr_hba.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The arcmsr_iop_message_xfer function in drivers/scsi/arcmsr/arcmsr_hba.c in the Linux kernel through 4.8.2 does not restrict a certain length field, which allows local users to gain privileges or cause a denial of service (heap-based buffer overflow) via an ARCMSR_MESSAGE_WRITE_WQBUFFER control code.
<p>Publish Date: 2016-10-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-7425>CVE-2016-7425</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-7425">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2016-7425</a></p>
<p>Release Date: 2016-10-16</p>
<p>Fix Resolution: v4.9-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers scsi arcmsr arcmsr hba c drivers scsi arcmsr arcmsr hba c vulnerability details the arcmsr iop message xfer function in drivers scsi arcmsr arcmsr hba c in the linux kernel through does not restrict a certain length field which allows local users to gain privileges or cause a denial of service heap based buffer overflow via an arcmsr message write wqbuffer control code publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
203,456 | 15,882,269,622 | IssuesEvent | 2021-04-09 15:48:37 | gadicc/node-yahoo-finance2 | https://api.github.com/repos/gadicc/node-yahoo-finance2 | opened | Interface Docs | documentation | # Docs Improvment
```typescript
export interface SummaryDetail {
...
bid?: number;
ask?: number;
bidSize?: number;
askSize?: number;
...
}
```
## Suggested Improvement
```typescript
export interface SummaryDetail {
...
bid?: number;
/**
* Returns the current market price for one share of this stock.
*/
ask?: number;
bidSize?: number;
askSize?: number;
...
}
```
## Reason For Suggestion (if not related to a module)
Currently many interfaces does not have any documentation about the attributes they contain and what the attributes are meant for. This makes it hard to guess what the attribute could be used for, when you are not familiar with the yahoo api. Look at the [SummaryDetail Interface](https://github.com/gadicc/node-yahoo-finance2/blob/devel/src/modules/quoteSummary-iface.ts) in the quoteSummary functionality. When you read the attribute `ask` for example you may ask yourself if the attribute represents the current market price of this stock or the current volume activity? I think its not needed to document every single attribute since the meaning of some are obvious but some should be documented.
If you agree with me I would be happy to contribute to this improvement.
| 1.0 | Interface Docs - # Docs Improvment
```typescript
export interface SummaryDetail {
...
bid?: number;
ask?: number;
bidSize?: number;
askSize?: number;
...
}
```
## Suggested Improvement
```typescript
export interface SummaryDetail {
...
bid?: number;
/**
* Returns the current market price for one share of this stock.
*/
ask?: number;
bidSize?: number;
askSize?: number;
...
}
```
## Reason For Suggestion (if not related to a module)
Currently many interfaces does not have any documentation about the attributes they contain and what the attributes are meant for. This makes it hard to guess what the attribute could be used for, when you are not familiar with the yahoo api. Look at the [SummaryDetail Interface](https://github.com/gadicc/node-yahoo-finance2/blob/devel/src/modules/quoteSummary-iface.ts) in the quoteSummary functionality. When you read the attribute `ask` for example you may ask yourself if the attribute represents the current market price of this stock or the current volume activity? I think its not needed to document every single attribute since the meaning of some are obvious but some should be documented.
If you agree with me I would be happy to contribute to this improvement.
| non_priority | interface docs docs improvment typescript export interface summarydetail bid number ask number bidsize number asksize number suggested improvement typescript export interface summarydetail bid number returns the current market price for one share of this stock ask number bidsize number asksize number reason for suggestion if not related to a module currently many interfaces does not have any documentation about the attributes they contain and what the attributes are meant for this makes it hard to guess what the attribute could be used for when you are not familiar with the yahoo api look at the in the quotesummary functionality when you read the attribute ask for example you may ask yourself if the attribute represents the current market price of this stock or the current volume activity i think its not needed to document every single attribute since the meaning of some are obvious but some should be documented if you agree with me i would be happy to contribute to this improvement | 0 |
277,870 | 21,057,829,833 | IssuesEvent | 2022-04-01 06:23:14 | eugenechiaay/ped | https://api.github.com/repos/eugenechiaay/ped | opened | Tip format | type.DocumentationBug severity.VeryLow | 
Can be mor eclear with phrasing: typing view 2 and clicking on... produce the same results
<!--session: 1648793695787-d92a4f87-7a60-4be9-bb4c-c3f0db7bace1-->
<!--Version: Web v3.4.2--> | 1.0 | Tip format - 
Can be mor eclear with phrasing: typing view 2 and clicking on... produce the same results
<!--session: 1648793695787-d92a4f87-7a60-4be9-bb4c-c3f0db7bace1-->
<!--Version: Web v3.4.2--> | non_priority | tip format can be mor eclear with phrasing typing view and clicking on produce the same results | 0 |
272,926 | 29,795,134,884 | IssuesEvent | 2023-06-16 01:13:52 | billmcchesney1/pacbot | https://api.github.com/repos/billmcchesney1/pacbot | closed | CVE-2022-0686 (Critical) detected in url-parse-1.4.7.tgz - autoclosed | Mend: dependency security vulnerability | ## CVE-2022-0686 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /webapp/package.json</p>
<p>Path to vulnerable library: /webapp/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- cli-1.6.8.tgz (Root Library)
- webpack-dev-server-2.11.5.tgz
- sockjs-client-1.1.5.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.8.
<p>Publish Date: 2022-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0686>CVE-2022-0686</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686</a></p>
<p>Release Date: 2022-02-20</p>
<p>Fix Resolution: url-parse - 1.5.8</p>
</p>
</details>
<p></p>
| True | CVE-2022-0686 (Critical) detected in url-parse-1.4.7.tgz - autoclosed - ## CVE-2022-0686 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /webapp/package.json</p>
<p>Path to vulnerable library: /webapp/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- cli-1.6.8.tgz (Root Library)
- webpack-dev-server-2.11.5.tgz
- sockjs-client-1.1.5.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.8.
<p>Publish Date: 2022-02-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0686>CVE-2022-0686</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686</a></p>
<p>Release Date: 2022-02-20</p>
<p>Fix Resolution: url-parse - 1.5.8</p>
</p>
</details>
<p></p>
| non_priority | cve critical detected in url parse tgz autoclosed cve critical severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file webapp package json path to vulnerable library webapp node modules url parse package json dependency hierarchy cli tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse | 0 |
40,198 | 10,471,027,744 | IssuesEvent | 2019-09-23 06:35:38 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Failed to get convolution algorithm | TF 1.14 stat:awaiting response type:build/install | <em>Please make sure that this is a bug. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:bug_template</em>
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): yes / no
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): 16.04 LTS Ubuntu
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: n/a
- TensorFlow installed from (source or binary): conda
- TensorFlow version (use command below): 1.14
- Python version: 2.6
- Bazel version (if compiling from source): n/a
- GCC/Compiler version (if compiling from source): n/a
- CUDA/cuDNN version: 9.0
- GPU model and memory: Quadro GP100
You can collect some of this information using our environment capture
[script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh)
You can also obtain the TensorFlow version with: 1. TF 1.0: `python -c "import
tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"` 2. TF 2.0: `python -c
"import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"`
**Describe the current behavior**
I have a simple sequential `tf.keras` model with convolution that runs on
a tfrecords Dataset. Suddenly, convolutions stopped working.
No known changes made to system.
**Describe the expected behavior**
It runs as per usual
**Code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary to generate the problem.
Won't help reproduce what may be system's issue.
```
ds = tf.data.TFRecordsDataset(...).shuffle(...)
model = tf.keras.Sequential([
tf.keras.layers.Conv2D(...),
...
])
model.compile(...)
model.fit(...)
```
**Other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
name: Quadro GP100, pci bus id: 0000:af:00.0, compute capability: 6.0)
2019-09-14 08:13:43.313823: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x56244eb0db50 executing computations on platform CUDA. Devices:
2019-09-14 08:13:43.313866: I tensorflow/compiler/xla/service/service.cc:175] StreamExecutor device (0): Quadro GP100, Compute Capability 6.0
2019-09-14 08:14:35.273134: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:111] Filling up shuffle buffer (this may take a while): 4255 of 129446
....
2019-09-14 08:14:36.878088: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:162] Shuffle buffer filled.
2019-09-14 08:14:48.210416: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
2019-09-14 08:14:49.011180: E tensorflow/stream_executor/cuda/cuda_dnn.cc:329] Could not create cudnn handle: CUDNN_STATUS_INTERNAL_ERROR
2019-09-14 08:14:49.016488: E tensorflow/stream_executor/cuda/cuda_dnn.cc:329] Could not create cudnn handle: CUDNN_STATUS_INTERNAL_ERROR
2019-09-14 08:14:49.016616: W tensorflow/core/common_runtime/base_collective_executor.cc:216] BaseCollectiveExecutor::StartAbort Unknown: Failed to get convolution algorithm. This is probably because cuDNN failed to initialize, so try looking to see if a warning log message was printed above.
[[{{node conv1d/conv1d}}]]
```
no warning message was printed above. | 1.0 | Failed to get convolution algorithm - <em>Please make sure that this is a bug. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:bug_template</em>
**System information**
- Have I written custom code (as opposed to using a stock example script provided in TensorFlow): yes / no
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): 16.04 LTS Ubuntu
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device: n/a
- TensorFlow installed from (source or binary): conda
- TensorFlow version (use command below): 1.14
- Python version: 2.6
- Bazel version (if compiling from source): n/a
- GCC/Compiler version (if compiling from source): n/a
- CUDA/cuDNN version: 9.0
- GPU model and memory: Quadro GP100
You can collect some of this information using our environment capture
[script](https://github.com/tensorflow/tensorflow/tree/master/tools/tf_env_collect.sh)
You can also obtain the TensorFlow version with: 1. TF 1.0: `python -c "import
tensorflow as tf; print(tf.GIT_VERSION, tf.VERSION)"` 2. TF 2.0: `python -c
"import tensorflow as tf; print(tf.version.GIT_VERSION, tf.version.VERSION)"`
**Describe the current behavior**
I have a simple sequential `tf.keras` model with convolution that runs on
a tfrecords Dataset. Suddenly, convolutions stopped working.
No known changes made to system.
**Describe the expected behavior**
It runs as per usual
**Code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary to generate the problem.
Won't help reproduce what may be system's issue.
```
ds = tf.data.TFRecordsDataset(...).shuffle(...)
model = tf.keras.Sequential([
tf.keras.layers.Conv2D(...),
...
])
model.compile(...)
model.fit(...)
```
**Other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
name: Quadro GP100, pci bus id: 0000:af:00.0, compute capability: 6.0)
2019-09-14 08:13:43.313823: I tensorflow/compiler/xla/service/service.cc:168] XLA service 0x56244eb0db50 executing computations on platform CUDA. Devices:
2019-09-14 08:13:43.313866: I tensorflow/compiler/xla/service/service.cc:175] StreamExecutor device (0): Quadro GP100, Compute Capability 6.0
2019-09-14 08:14:35.273134: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:111] Filling up shuffle buffer (this may take a while): 4255 of 129446
....
2019-09-14 08:14:36.878088: I tensorflow/core/kernels/data/shuffle_dataset_op.cc:162] Shuffle buffer filled.
2019-09-14 08:14:48.210416: I tensorflow/stream_executor/platform/default/dso_loader.cc:42] Successfully opened dynamic library libcudnn.so.7
2019-09-14 08:14:49.011180: E tensorflow/stream_executor/cuda/cuda_dnn.cc:329] Could not create cudnn handle: CUDNN_STATUS_INTERNAL_ERROR
2019-09-14 08:14:49.016488: E tensorflow/stream_executor/cuda/cuda_dnn.cc:329] Could not create cudnn handle: CUDNN_STATUS_INTERNAL_ERROR
2019-09-14 08:14:49.016616: W tensorflow/core/common_runtime/base_collective_executor.cc:216] BaseCollectiveExecutor::StartAbort Unknown: Failed to get convolution algorithm. This is probably because cuDNN failed to initialize, so try looking to see if a warning log message was printed above.
[[{{node conv1d/conv1d}}]]
```
no warning message was printed above. | non_priority | failed to get convolution algorithm please make sure that this is a bug as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag bug template system information have i written custom code as opposed to using a stock example script provided in tensorflow yes no os platform and distribution e g linux ubuntu lts ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device n a tensorflow installed from source or binary conda tensorflow version use command below python version bazel version if compiling from source n a gcc compiler version if compiling from source n a cuda cudnn version gpu model and memory quadro you can collect some of this information using our environment capture you can also obtain the tensorflow version with tf python c import tensorflow as tf print tf git version tf version tf python c import tensorflow as tf print tf version git version tf version version describe the current behavior i have a simple sequential tf keras model with convolution that runs on a tfrecords dataset suddenly convolutions stopped working no known changes made to system describe the expected behavior it runs as per usual code to reproduce the issue provide a reproducible test case that is the bare minimum necessary to generate the problem won t help reproduce what may be system s issue ds tf data tfrecordsdataset shuffle model tf keras sequential tf keras layers model compile model fit other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached name quadro pci bus id af compute capability i tensorflow compiler xla service service cc xla service executing computations on platform cuda devices i tensorflow compiler xla service service cc streamexecutor device quadro compute capability i tensorflow core kernels data shuffle dataset op cc filling up shuffle buffer this may take a while of i tensorflow core kernels data shuffle dataset op cc shuffle buffer filled i tensorflow stream executor platform default dso loader cc successfully opened dynamic library libcudnn so e tensorflow stream executor cuda cuda dnn cc could not create cudnn handle cudnn status internal error e tensorflow stream executor cuda cuda dnn cc could not create cudnn handle cudnn status internal error w tensorflow core common runtime base collective executor cc basecollectiveexecutor startabort unknown failed to get convolution algorithm this is probably because cudnn failed to initialize so try looking to see if a warning log message was printed above no warning message was printed above | 0 |
54,995 | 6,886,141,838 | IssuesEvent | 2017-11-21 18:23:25 | pennsignals/gentry | https://api.github.com/repos/pennsignals/gentry | closed | Create Slack Attachment model | design | [Attachments](https://api.slack.com/docs/message-attachments) let you add more context to a message, making them more useful and effective.
Below is a sample `Attachment` structure:
```json
[
{
"author_icon": "http://flickr.com/icons/bobby.jpg",
"author_link": "http://flickr.com/bobby/",
"author_name": "Bobby Tables",
"color": "danger",
"fallback": "Required plain-text summary of the attachment.",
"fields": [
{
"short": false,
"title": "Priority",
"value": "High"
}
],
"footer": "Slack API",
"footer_icon": "https://platform.slack-edge.com/img/default_application_icon.png",
"image_url": "http://my-website.com/path/to/image.jpg",
"pretext": "Optional text that appears above the attachment block",
"text": "Optional text that appears within the attachment",
"thumb_url": "http://example.com/path/to/thumb.png",
"title": "Slack API Documentation",
"title_link": "https://api.slack.com/",
"ts": 123456789
}
]
``` | 1.0 | Create Slack Attachment model - [Attachments](https://api.slack.com/docs/message-attachments) let you add more context to a message, making them more useful and effective.
Below is a sample `Attachment` structure:
```json
[
{
"author_icon": "http://flickr.com/icons/bobby.jpg",
"author_link": "http://flickr.com/bobby/",
"author_name": "Bobby Tables",
"color": "danger",
"fallback": "Required plain-text summary of the attachment.",
"fields": [
{
"short": false,
"title": "Priority",
"value": "High"
}
],
"footer": "Slack API",
"footer_icon": "https://platform.slack-edge.com/img/default_application_icon.png",
"image_url": "http://my-website.com/path/to/image.jpg",
"pretext": "Optional text that appears above the attachment block",
"text": "Optional text that appears within the attachment",
"thumb_url": "http://example.com/path/to/thumb.png",
"title": "Slack API Documentation",
"title_link": "https://api.slack.com/",
"ts": 123456789
}
]
``` | non_priority | create slack attachment model let you add more context to a message making them more useful and effective below is a sample attachment structure json author icon author link author name bobby tables color danger fallback required plain text summary of the attachment fields short false title priority value high footer slack api footer icon image url pretext optional text that appears above the attachment block text optional text that appears within the attachment thumb url title slack api documentation title link ts | 0 |
272,106 | 20,733,325,222 | IssuesEvent | 2022-03-14 11:27:58 | opfab/operatorfabric-core | https://api.github.com/repos/opfab/operatorfabric-core | closed | Rework UI dev documentation | Documentation | src/docs/asciidoc/dev_env/ui.adoc
Rework documentation to point to the back dev documentation for the steps to get the back running instead of repeating them | 1.0 | Rework UI dev documentation - src/docs/asciidoc/dev_env/ui.adoc
Rework documentation to point to the back dev documentation for the steps to get the back running instead of repeating them | non_priority | rework ui dev documentation src docs asciidoc dev env ui adoc rework documentation to point to the back dev documentation for the steps to get the back running instead of repeating them | 0 |
263,238 | 28,029,770,605 | IssuesEvent | 2023-03-28 11:36:21 | RG4421/ampere-centos-kernel | https://api.github.com/repos/RG4421/ampere-centos-kernel | reopened | CVE-2019-18680 (High) detected in linuxv5.2 | Mend: dependency security vulnerability | ## CVE-2019-18680 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/tcp.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/tcp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel 4.4.x before 4.4.195. There is a NULL pointer dereference in rds_tcp_kill_sock() in net/rds/tcp.c that will cause denial of service, aka CID-91573ae4aed0.
<p>Publish Date: 2019-11-04
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-18680>CVE-2019-18680</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-18680">https://nvd.nist.gov/vuln/detail/CVE-2019-18680</a></p>
<p>Release Date: 2019-12-05</p>
<p>Fix Resolution: linux-libc-headers - 5.8;linux-yocto - 4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68,5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1</p>
</p>
</details>
<p></p>
| True | CVE-2019-18680 (High) detected in linuxv5.2 - ## CVE-2019-18680 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv5.2</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/torvalds/linux.git>https://github.com/torvalds/linux.git</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/tcp.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rds/tcp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel 4.4.x before 4.4.195. There is a NULL pointer dereference in rds_tcp_kill_sock() in net/rds/tcp.c that will cause denial of service, aka CID-91573ae4aed0.
<p>Publish Date: 2019-11-04
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-18680>CVE-2019-18680</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-18680">https://nvd.nist.gov/vuln/detail/CVE-2019-18680</a></p>
<p>Release Date: 2019-12-05</p>
<p>Fix Resolution: linux-libc-headers - 5.8;linux-yocto - 4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68,5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1</p>
</p>
</details>
<p></p>
| non_priority | cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in base branch amp centos kernel vulnerable source files net rds tcp c net rds tcp c vulnerability details an issue was discovered in the linux kernel x before there is a null pointer dereference in rds tcp kill sock in net rds tcp c that will cause denial of service aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux libc headers linux yocto gitautoinc gitautoinc | 0 |
449,555 | 31,850,391,790 | IssuesEvent | 2023-09-15 00:48:43 | bazelbuild/bazel | https://api.github.com/repos/bazelbuild/bazel | opened | typo is bazel module docs | type: documentation (cleanup) untriaged team-Documentation | ### Page link:
https://bazel.build/rules/lib/globals/module#module
### Problem description (include actual vs expected text, if applicable):
Text for `bazel_compatibility` includes ambiguous colons. E.g.
"Three constraints are supported: <=X.X.X: The Bazel version must be equal or older than X.X.X. "
which reads as: `<=X.X.X:` This occurs in several places.
Grammatical forms should be typographically distinguished to make sure they are set off from surrounding prose.
### Where do you see this issue? (include link to specific section of the page, if applicable)
_No response_
### Any other information you'd like to share?
_No response_ | 2.0 | typo is bazel module docs - ### Page link:
https://bazel.build/rules/lib/globals/module#module
### Problem description (include actual vs expected text, if applicable):
Text for `bazel_compatibility` includes ambiguous colons. E.g.
"Three constraints are supported: <=X.X.X: The Bazel version must be equal or older than X.X.X. "
which reads as: `<=X.X.X:` This occurs in several places.
Grammatical forms should be typographically distinguished to make sure they are set off from surrounding prose.
### Where do you see this issue? (include link to specific section of the page, if applicable)
_No response_
### Any other information you'd like to share?
_No response_ | non_priority | typo is bazel module docs page link problem description include actual vs expected text if applicable text for bazel compatibility includes ambiguous colons e g three constraints are supported x x x the bazel version must be equal or older than x x x which reads as x x x this occurs in several places grammatical forms should be typographically distinguished to make sure they are set off from surrounding prose where do you see this issue include link to specific section of the page if applicable no response any other information you d like to share no response | 0 |
146,361 | 11,734,489,439 | IssuesEvent | 2020-03-11 09:26:42 | luvit/luvit | https://api.github.com/repos/luvit/luvit | closed | make test: "tls-set-ciphers - set null cipher" fails with luvi 2.7.6 | tests | When running `make test` on the master branch of luvit the test "tls-set-ciphers - set null cipher" fails on my Arch Linux system:
```
# Starting Test: tls-set-ciphers - set null cipher
'CONNECTED(00000003)\n---\nno peer certificate available\n---\nNo client certificate CA names sent\n---\nSSL handshake has read 0 bytes and written 7 bytes\nVerification: OK\n---\nNew, (NONE), Cipher is (NONE)\nSecure Renegotiation IS NOT supported\nCompression: NONE\nExpansion: NONE\nNo ALPN negotiated\nSSL-Session:\n Protocol : TLSv1\n Cipher : 0000\n Session-ID: \n Session-ID-ctx: \n Master-Key: \n PSK identity: None\n PSK identity hint: None\n SRP username: None\n Start Time: 1560370549\n Timeout : 7200 (sec)\n Verify return code: 0 (ok)\n Extended master secret: no\n---\n'
/home/luvit/tests/libs/tap.lua:83: Missing 1 expected call
stack traceback:
[C]: in function 'error'
/home/luvit/tests/libs/tap.lua:83: in function </home/luvit/tests/libs/tap.lua:64>
[C]: in function 'xpcall'
/home/luvit/tests/libs/tap.lua:64: in function 'run'
/home/luvit/tests/libs/tap.lua:165: in function 'tap'
/home/luvit/tests/run.lua:42: in function 'fn'
[string "bundle:deps/require.lua"]:310: in function 'require'
/home/luvit/main.lua:118: in function 'main'
/home/luvit/init.lua:49: in function </home/luvit/init.lua:47>
[C]: in function 'xpcall'
/home/luvit/init.lua:47: in function 'fn'
[string "bundle:deps/require.lua"]:310: in function <[string "bundle:deps/require.lua"]:266>
not ok 198 tls-set-ciphers - set null cipher
```
EDIT: Noticed, that `make test` runs the test using a fetched luvi version 2.7.6:
```
luvi v2.7.6
rex: 8.37 2015-04-28
libuv: 1.9.1
ssl: OpenSSL 1.0.2h 3 May 2016, lua-openssl 0.5.1
```
EDIT: Running `luvi . -- tests/run.lua` manually with luvi from master branch, all test passes:
```
luvi v2.9.3-11-g17c1c9b
rex: 8.37 2015-04-28
libuv: 1.29.0
ssl: OpenSSL 1.1.1b 26 Feb 2019, lua-openssl 0.7.5-2
``` | 1.0 | make test: "tls-set-ciphers - set null cipher" fails with luvi 2.7.6 - When running `make test` on the master branch of luvit the test "tls-set-ciphers - set null cipher" fails on my Arch Linux system:
```
# Starting Test: tls-set-ciphers - set null cipher
'CONNECTED(00000003)\n---\nno peer certificate available\n---\nNo client certificate CA names sent\n---\nSSL handshake has read 0 bytes and written 7 bytes\nVerification: OK\n---\nNew, (NONE), Cipher is (NONE)\nSecure Renegotiation IS NOT supported\nCompression: NONE\nExpansion: NONE\nNo ALPN negotiated\nSSL-Session:\n Protocol : TLSv1\n Cipher : 0000\n Session-ID: \n Session-ID-ctx: \n Master-Key: \n PSK identity: None\n PSK identity hint: None\n SRP username: None\n Start Time: 1560370549\n Timeout : 7200 (sec)\n Verify return code: 0 (ok)\n Extended master secret: no\n---\n'
/home/luvit/tests/libs/tap.lua:83: Missing 1 expected call
stack traceback:
[C]: in function 'error'
/home/luvit/tests/libs/tap.lua:83: in function </home/luvit/tests/libs/tap.lua:64>
[C]: in function 'xpcall'
/home/luvit/tests/libs/tap.lua:64: in function 'run'
/home/luvit/tests/libs/tap.lua:165: in function 'tap'
/home/luvit/tests/run.lua:42: in function 'fn'
[string "bundle:deps/require.lua"]:310: in function 'require'
/home/luvit/main.lua:118: in function 'main'
/home/luvit/init.lua:49: in function </home/luvit/init.lua:47>
[C]: in function 'xpcall'
/home/luvit/init.lua:47: in function 'fn'
[string "bundle:deps/require.lua"]:310: in function <[string "bundle:deps/require.lua"]:266>
not ok 198 tls-set-ciphers - set null cipher
```
EDIT: Noticed, that `make test` runs the test using a fetched luvi version 2.7.6:
```
luvi v2.7.6
rex: 8.37 2015-04-28
libuv: 1.9.1
ssl: OpenSSL 1.0.2h 3 May 2016, lua-openssl 0.5.1
```
EDIT: Running `luvi . -- tests/run.lua` manually with luvi from master branch, all test passes:
```
luvi v2.9.3-11-g17c1c9b
rex: 8.37 2015-04-28
libuv: 1.29.0
ssl: OpenSSL 1.1.1b 26 Feb 2019, lua-openssl 0.7.5-2
``` | non_priority | make test tls set ciphers set null cipher fails with luvi when running make test on the master branch of luvit the test tls set ciphers set null cipher fails on my arch linux system starting test tls set ciphers set null cipher connected n nno peer certificate available n nno client certificate ca names sent n nssl handshake has read bytes and written bytes nverification ok n nnew none cipher is none nsecure renegotiation is not supported ncompression none nexpansion none nno alpn negotiated nssl session n protocol n cipher n session id n session id ctx n master key n psk identity none n psk identity hint none n srp username none n start time n timeout sec n verify return code ok n extended master secret no n n home luvit tests libs tap lua missing expected call stack traceback in function error home luvit tests libs tap lua in function in function xpcall home luvit tests libs tap lua in function run home luvit tests libs tap lua in function tap home luvit tests run lua in function fn in function require home luvit main lua in function main home luvit init lua in function in function xpcall home luvit init lua in function fn in function not ok tls set ciphers set null cipher edit noticed that make test runs the test using a fetched luvi version luvi rex libuv ssl openssl may lua openssl edit running luvi tests run lua manually with luvi from master branch all test passes luvi rex libuv ssl openssl feb lua openssl | 0 |
177,120 | 28,355,440,322 | IssuesEvent | 2023-04-12 06:57:36 | codeforboston/maple | https://api.github.com/repos/codeforboston/maple | closed | Homepage - What People are Saying | design beta-2-feedback | - not sure what this section was at first
- thumbs were a bit confusing
- “it’s not clear to me whether that is the most recent comments that the site has decided to show, like the last 5 or if they are the highest rated/liked.” Would like to see a value assigned to the comments, like if people thumbed up or support the comment.
| 1.0 | Homepage - What People are Saying - - not sure what this section was at first
- thumbs were a bit confusing
- “it’s not clear to me whether that is the most recent comments that the site has decided to show, like the last 5 or if they are the highest rated/liked.” Would like to see a value assigned to the comments, like if people thumbed up or support the comment.
| non_priority | homepage what people are saying not sure what this section was at first thumbs were a bit confusing “it’s not clear to me whether that is the most recent comments that the site has decided to show like the last or if they are the highest rated liked ” would like to see a value assigned to the comments like if people thumbed up or support the comment | 0 |
14,594 | 3,868,986,956 | IssuesEvent | 2016-04-10 10:04:07 | symfony/symfony-docs | https://api.github.com/repos/symfony/symfony-docs | closed | [PhpUnitBridge] Mock DNS functions | Actionable hasPR Missing Documentation PHPUnitBridge | The PHPUnit bridge got support for mocking PHP's DNS related functions (see symfony/symfony#18181). | 1.0 | [PhpUnitBridge] Mock DNS functions - The PHPUnit bridge got support for mocking PHP's DNS related functions (see symfony/symfony#18181). | non_priority | mock dns functions the phpunit bridge got support for mocking php s dns related functions see symfony symfony | 0 |
23,096 | 21,004,454,081 | IssuesEvent | 2022-03-29 20:54:54 | tl-its-umich-edu/canvas-course-manager-next | https://api.github.com/repos/tl-its-umich-edu/canvas-course-manager-next | opened | Change error messages with field input to use directive | enhancement front end usability | Text below from @janelilr:
Update error message when a user has to type in an input. For example:
**Adding a non-UM user’s (one at a time) first and last name.**
Current error message: “Value for the last name may not be blank.”
New error message: “Please enter last name”
Current error message: “Value for the first name may not be blank.”
New error message: “Please enter first name”
**Editing course name**
Current error message: “Value for the course name may not be blank.”
New error message: “Please enter the course name”
**For Add U-M users, Add non-UM users, Format Third-Party Gradebook, and Create Sections features.**
The current CSV file error messages can stay. | True | Change error messages with field input to use directive - Text below from @janelilr:
Update error message when a user has to type in an input. For example:
**Adding a non-UM user’s (one at a time) first and last name.**
Current error message: “Value for the last name may not be blank.”
New error message: “Please enter last name”
Current error message: “Value for the first name may not be blank.”
New error message: “Please enter first name”
**Editing course name**
Current error message: “Value for the course name may not be blank.”
New error message: “Please enter the course name”
**For Add U-M users, Add non-UM users, Format Third-Party Gradebook, and Create Sections features.**
The current CSV file error messages can stay. | non_priority | change error messages with field input to use directive text below from janelilr update error message when a user has to type in an input for example adding a non um user’s one at a time first and last name current error message “value for the last name may not be blank ” new error message “please enter last name” current error message “value for the first name may not be blank ” new error message “please enter first name” editing course name current error message “value for the course name may not be blank ” new error message “please enter the course name” for add u m users add non um users format third party gradebook and create sections features the current csv file error messages can stay | 0 |
196,772 | 15,609,760,428 | IssuesEvent | 2021-03-19 12:20:16 | kirbydesign/designsystem | https://api.github.com/repos/kirbydesign/designsystem | opened | [Documentation] Write "The good: pull-request"-section for contribution guidelines | documentation | **This is part of a larger issue - please read: https://github.com/kirbydesign/designsystem/issues/1393#issuecomment-802766408 before continuing!**
**Short description**
<!-- Replace the paragraph below with your own description -->
A "The good: pull-request" section should be added to the contribution guidelines. An outline already exists on the branch [`documentation/contribution-guidelines`](https://github.com/kirbydesign/designsystem/tree/documentation/contribution-guidelines) which should be expanded upon.
| 1.0 | [Documentation] Write "The good: pull-request"-section for contribution guidelines - **This is part of a larger issue - please read: https://github.com/kirbydesign/designsystem/issues/1393#issuecomment-802766408 before continuing!**
**Short description**
<!-- Replace the paragraph below with your own description -->
A "The good: pull-request" section should be added to the contribution guidelines. An outline already exists on the branch [`documentation/contribution-guidelines`](https://github.com/kirbydesign/designsystem/tree/documentation/contribution-guidelines) which should be expanded upon.
| non_priority | write the good pull request section for contribution guidelines this is part of a larger issue please read before continuing short description a the good pull request section should be added to the contribution guidelines an outline already exists on the branch which should be expanded upon | 0 |
47,641 | 19,685,968,632 | IssuesEvent | 2022-01-11 22:12:26 | microsoft/botframework-sdk | https://api.github.com/repos/microsoft/botframework-sdk | closed | Please include the current channel name in the Conversation for Microsoft Teams | customer-replied-to customer-reported Bot Services Area: Teams feature-request | ## Issue
When using Bot Framework to create a bot for Microsoft Teams, the `turnContext.Activtity.Conversation` property always has null for the `Name`.
Here's an example of a conversation I'm getting.
```
Conversation:
Id: 19:some-guid@thread.tacv2;messageid=1617657461612
Name:
IsGroup: True
Role:
TenantId: my-tenant-guid
ConversationType: channel
Properties: {}
```
I understand I can get a list of ALL channels via `TeamsInfo.GetTeamChannelsAsync(turnContext)` and then find the one that matches the current conversation, but that seems like a lot of effort just to return the name of the current channel and leaves a lot of room for implementors to get it wrong.
## Proposed change
Set the `Name` property of `turnContext.Conversation` to be the name of the current channel in Teams in cases where it makes sense (ConversationType: "channel" or "teams")
## Component Impact
???
## Customer Impact
Any customer currently trying to get the room name for the current conversation has to implement an API call to get every channel. This leaves a lot of room for user error.
| 1.0 | Please include the current channel name in the Conversation for Microsoft Teams - ## Issue
When using Bot Framework to create a bot for Microsoft Teams, the `turnContext.Activtity.Conversation` property always has null for the `Name`.
Here's an example of a conversation I'm getting.
```
Conversation:
Id: 19:some-guid@thread.tacv2;messageid=1617657461612
Name:
IsGroup: True
Role:
TenantId: my-tenant-guid
ConversationType: channel
Properties: {}
```
I understand I can get a list of ALL channels via `TeamsInfo.GetTeamChannelsAsync(turnContext)` and then find the one that matches the current conversation, but that seems like a lot of effort just to return the name of the current channel and leaves a lot of room for implementors to get it wrong.
## Proposed change
Set the `Name` property of `turnContext.Conversation` to be the name of the current channel in Teams in cases where it makes sense (ConversationType: "channel" or "teams")
## Component Impact
???
## Customer Impact
Any customer currently trying to get the room name for the current conversation has to implement an API call to get every channel. This leaves a lot of room for user error.
| non_priority | please include the current channel name in the conversation for microsoft teams issue when using bot framework to create a bot for microsoft teams the turncontext activtity conversation property always has null for the name here s an example of a conversation i m getting conversation id some guid thread messageid name isgroup true role tenantid my tenant guid conversationtype channel properties i understand i can get a list of all channels via teamsinfo getteamchannelsasync turncontext and then find the one that matches the current conversation but that seems like a lot of effort just to return the name of the current channel and leaves a lot of room for implementors to get it wrong proposed change set the name property of turncontext conversation to be the name of the current channel in teams in cases where it makes sense conversationtype channel or teams component impact customer impact any customer currently trying to get the room name for the current conversation has to implement an api call to get every channel this leaves a lot of room for user error | 0 |
279,383 | 24,221,103,780 | IssuesEvent | 2022-09-26 10:57:06 | rhinstaller/kickstart-tests | https://api.github.com/repos/rhinstaller/kickstart-tests | closed | network-prefixdevname flakes on rhel8 with *** Failed check: kstest3 configuration added to GUI | test flake | Since 04/25/2022 we are seeing flakes for network-prefixdevname on rhel8:
```
2022-04-26 00:58:13,990 DEBUG: Failed to run subprocess: '['/opt/kstest/kickstart-tests/scripts/launcher/lib/launcher_interface.sh', '-i', '/opt/kstest/data/images/boot.iso', '-k', '1', '-w', '/var/tmp/kstest-network-prefixdevname.2022_04_26-00_36_35.7ycybonw', '-t', '/opt/kstest/kickstart-tests/network-prefixdevname.sh', 'validate']'
stdout:
*** Failed check: kstest3 configuration added to GUI
*** Failed check: team0 has slave kstest3 yes
*** Failed check: team0 has option priority \(port:kstest3\) 100
2022-04-26 00:58:13,991 INFO: RESULT:network-prefixdevname:80dcd3d6276a:FAILED:Validation failed with return code 1
```
[kstest.log](https://github.com/rhinstaller/kickstart-tests/files/8618424/kstest.log)
[virt-install.log](https://github.com/rhinstaller/kickstart-tests/files/8618425/virt-install.log)
| 1.0 | network-prefixdevname flakes on rhel8 with *** Failed check: kstest3 configuration added to GUI - Since 04/25/2022 we are seeing flakes for network-prefixdevname on rhel8:
```
2022-04-26 00:58:13,990 DEBUG: Failed to run subprocess: '['/opt/kstest/kickstart-tests/scripts/launcher/lib/launcher_interface.sh', '-i', '/opt/kstest/data/images/boot.iso', '-k', '1', '-w', '/var/tmp/kstest-network-prefixdevname.2022_04_26-00_36_35.7ycybonw', '-t', '/opt/kstest/kickstart-tests/network-prefixdevname.sh', 'validate']'
stdout:
*** Failed check: kstest3 configuration added to GUI
*** Failed check: team0 has slave kstest3 yes
*** Failed check: team0 has option priority \(port:kstest3\) 100
2022-04-26 00:58:13,991 INFO: RESULT:network-prefixdevname:80dcd3d6276a:FAILED:Validation failed with return code 1
```
[kstest.log](https://github.com/rhinstaller/kickstart-tests/files/8618424/kstest.log)
[virt-install.log](https://github.com/rhinstaller/kickstart-tests/files/8618425/virt-install.log)
| non_priority | network prefixdevname flakes on with failed check configuration added to gui since we are seeing flakes for network prefixdevname on debug failed to run subprocess apos apos stdout failed check configuration added to gui failed check has slave yes failed check has option priority port info result network prefixdevname failed validation failed with return code | 0 |
14,031 | 16,843,938,520 | IssuesEvent | 2021-06-19 04:14:37 | cseelhoff/RimThreaded | https://api.github.com/repos/cseelhoff/RimThreaded | closed | "Rimefeller" Compatibility Issues | 1.3.1 - 1.3.2 Bug Confirmed Fixed In Preview Mod Incompatibility Reproducible |
**Describe the bug**
Rimefeller - Do Until X orders on refiners dont stop on X and produce lots of errors in the debug log
UPDATE: The error only seems to occur when the bill is set to look everywhere instead of in a specific stockpile zone
**Steps to reproduce the behavior (VERY IMPORTANT)**
1. Get Rimefeller and Rimthreaded
2. Build any refiner, connect it with pipes to a full chemfuel tank, build a resource console, set a do until X bill and connect all to power
4. get a colonist to man the resource console so the refiners turn on
5. check debug log
**Error Log**
[Player.log](https://github.com/cseelhoff/RimThreaded/files/6392049/Player.log)
**Save file**
https://drive.google.com/file/d/1qdCu8yVhs0pKJFgu_O6iR-nJKU8dt0NA/view?usp=sharing
**Screenshots**

**Mod list (Preferably a RimPy compatible list.)**
* Harmony
* Core
* Royalty (optional)
* Mod Manager (By Fluffy)
* Rimefeller
* RimThreaded
| True | "Rimefeller" Compatibility Issues -
**Describe the bug**
Rimefeller - Do Until X orders on refiners dont stop on X and produce lots of errors in the debug log
UPDATE: The error only seems to occur when the bill is set to look everywhere instead of in a specific stockpile zone
**Steps to reproduce the behavior (VERY IMPORTANT)**
1. Get Rimefeller and Rimthreaded
2. Build any refiner, connect it with pipes to a full chemfuel tank, build a resource console, set a do until X bill and connect all to power
4. get a colonist to man the resource console so the refiners turn on
5. check debug log
**Error Log**
[Player.log](https://github.com/cseelhoff/RimThreaded/files/6392049/Player.log)
**Save file**
https://drive.google.com/file/d/1qdCu8yVhs0pKJFgu_O6iR-nJKU8dt0NA/view?usp=sharing
**Screenshots**

**Mod list (Preferably a RimPy compatible list.)**
* Harmony
* Core
* Royalty (optional)
* Mod Manager (By Fluffy)
* Rimefeller
* RimThreaded
| non_priority | rimefeller compatibility issues describe the bug rimefeller do until x orders on refiners dont stop on x and produce lots of errors in the debug log update the error only seems to occur when the bill is set to look everywhere instead of in a specific stockpile zone steps to reproduce the behavior very important get rimefeller and rimthreaded build any refiner connect it with pipes to a full chemfuel tank build a resource console set a do until x bill and connect all to power get a colonist to man the resource console so the refiners turn on check debug log error log save file screenshots mod list preferably a rimpy compatible list harmony core royalty optional mod manager by fluffy rimefeller rimthreaded | 0 |
52,955 | 6,287,542,755 | IssuesEvent | 2017-07-19 15:09:37 | CLARIAH/wp5_mediasuite | https://api.github.com/repos/CLARIAH/wp5_mediasuite | closed | Connect user annotations (video) to user who added them | Effort: improvement Function: annotation (video) Importance: medium MS-Component-function ToDo: testing needed Work: functionality | This is for users to see which annotations they made (attached to their user, as logged in the surfConnex). | 1.0 | Connect user annotations (video) to user who added them - This is for users to see which annotations they made (attached to their user, as logged in the surfConnex). | non_priority | connect user annotations video to user who added them this is for users to see which annotations they made attached to their user as logged in the surfconnex | 0 |
63,222 | 15,502,247,011 | IssuesEvent | 2021-03-11 11:35:22 | ARM-software/armnn | https://api.github.com/repos/ARM-software/armnn | closed | A system's native Boost preferred over the one specifically compiled for ArmNN | Build issue Future work | Dear ArmNN developers,
It took me a few days to diagnose a weird family of issues where compilation of ArmNN and consumer programs happened without errors (with some warnings though), and either "Bus Error" or "Segmentation Fault" would appear during the runtime.
This happened on a `hikey960` "host" platform running Debian Linux 9, which also had a pre-installed version of Boost 1.58 or 1.62 (this detail proved to be essential).
According to the documentation ( https://developer.arm.com/solutions/machine-learning-on-arm/developer-material/how-to-guides/configuring-the-arm-nn-sdk-build-environment-for-tensorflow/single-page ) we need Boost 1.64 with -fPIC. So this is what I compiled and passed on to cmake building the program. When cmake runs, it reports that it found our Boost v.1.64 and PRETENDS to take it into account. However when I switch on all logging I can see that the path to this newly compiled Boost is nowhere to be seen in the command used to link the actual `armnn.so` !
The linking command looks like this (I'm skipping the object files in the middle):
```
/usr/bin/cmake -E cmake_link_script CMakeFiles/armnn.dir/link.txt --verbose=1
/usr/bin/clang++-6.0 -fPIC -std=c++14 -Werror -std=c++14 -Wall -Wold-style-cast -Wno-missing-braces -Wconversion -Wsign-con
version -DNDEBUG -O3 -shared -Wl,-soname,libarmnn.so -o libarmnn.so CMakeFiles/armnn.dir/src/armnn/layers/ActivationLayer.cpp.o CMakeFiles/armnn.dir/src/armnn/layers/AdditionLayer.cpp.o CMakeFiles/armnn.dir/src/armnn/layers/BatchNormalizationLayer.cpp.o CMakeFiles/armnn.dir/src/armnn/layers/BatchToSpaceNdLayer.cpp.o
...
src/backends/reference/workloads/CMakeFiles/armnnRefBackendWorkloads.dir/RefMeanFloat32Workload.cpp.o src/backends/reference/workloads/CMakeFiles/armnnRefBackendWorkloads.dir/RefMeanUint8Workload.cpp.o libarmnnUtils.a -ldl -Wl,-Bstatic -lboost_log -lboost_thread -lboost_system -Wl,-Bdynamic
```
Which is immediately followed by reaction from the linker:
```
/usr/bin/ld: /usr/bin/../lib/gcc/aarch64-linux-gnu/7.2.0/../../../aarch64-linux-gnu/libboost_log.a(core.o): relocation R_AARCH64_ADR_PREL_PG_HI21 against symbol `_ZTVN5boost16exception_detail19error_info_injectorISt13runtime_errorEE' which may bind externally can not be used when making a shared object; recompile with -fPIC
/usr/bin/ld: /usr/bin/../lib/gcc/aarch64-linux-gnu/7.2.0/../../../aarch64-linux-gnu/libboost_log.a(core.o): relocation R_AARCH64_ADR_PREL_PG_HI21 against symbol `_ZTVN5boost9exceptionE' which may bind externally can not be used when making a shared object; recompile with -fPIC
/usr/bin/ld: /usr/bin/../lib/gcc/aarch64-linux-gnu/7.2.0/../../../aarch64-linux-gnu/libboost_log.a(core.o): relocation R_AARCH64_ADR_PREL_PG_HI21 against symbol `_ZTVN5boost16exception_detail10clone_baseE' which may bind externally can not be used when making a shared object; recompile with -fPIC
...
```
There is either an error or inconsistency in the `cmake/GlobalConfig.cmake` included from the main `CMakeLists.txt` which contains the following bit related to Boost:
```
add_definitions("-DBOOST_ALL_NO_LIB") # Turn off auto-linking as we specify the libs manually
set(Boost_USE_STATIC_LIBS ON)
find_package(Boost 1.59 REQUIRED COMPONENTS unit_test_framework system filesystem log program_options)
include_directories(SYSTEM "${Boost_INCLUDE_DIR}")
link_directories(${Boost_LIBRARY_DIR})
```
It suddenly requires Boost v.1.59 (which does not match 1.64 required by the documentation), however when I patched this to 1.64 my system still insisted on linking with the system's Boost. As you can see from the log above, the `-L` flag necessary to enforce specific Boost's path was missing from the linking command.
However AS SOON AS I uninstalled the system's Boost, everything both compiled and ran without any problems. Please review your `cmake` code related to linking with Boost. | 1.0 | A system's native Boost preferred over the one specifically compiled for ArmNN - Dear ArmNN developers,
It took me a few days to diagnose a weird family of issues where compilation of ArmNN and consumer programs happened without errors (with some warnings though), and either "Bus Error" or "Segmentation Fault" would appear during the runtime.
This happened on a `hikey960` "host" platform running Debian Linux 9, which also had a pre-installed version of Boost 1.58 or 1.62 (this detail proved to be essential).
According to the documentation ( https://developer.arm.com/solutions/machine-learning-on-arm/developer-material/how-to-guides/configuring-the-arm-nn-sdk-build-environment-for-tensorflow/single-page ) we need Boost 1.64 with -fPIC. So this is what I compiled and passed on to cmake building the program. When cmake runs, it reports that it found our Boost v.1.64 and PRETENDS to take it into account. However when I switch on all logging I can see that the path to this newly compiled Boost is nowhere to be seen in the command used to link the actual `armnn.so` !
The linking command looks like this (I'm skipping the object files in the middle):
```
/usr/bin/cmake -E cmake_link_script CMakeFiles/armnn.dir/link.txt --verbose=1
/usr/bin/clang++-6.0 -fPIC -std=c++14 -Werror -std=c++14 -Wall -Wold-style-cast -Wno-missing-braces -Wconversion -Wsign-con
version -DNDEBUG -O3 -shared -Wl,-soname,libarmnn.so -o libarmnn.so CMakeFiles/armnn.dir/src/armnn/layers/ActivationLayer.cpp.o CMakeFiles/armnn.dir/src/armnn/layers/AdditionLayer.cpp.o CMakeFiles/armnn.dir/src/armnn/layers/BatchNormalizationLayer.cpp.o CMakeFiles/armnn.dir/src/armnn/layers/BatchToSpaceNdLayer.cpp.o
...
src/backends/reference/workloads/CMakeFiles/armnnRefBackendWorkloads.dir/RefMeanFloat32Workload.cpp.o src/backends/reference/workloads/CMakeFiles/armnnRefBackendWorkloads.dir/RefMeanUint8Workload.cpp.o libarmnnUtils.a -ldl -Wl,-Bstatic -lboost_log -lboost_thread -lboost_system -Wl,-Bdynamic
```
Which is immediately followed by reaction from the linker:
```
/usr/bin/ld: /usr/bin/../lib/gcc/aarch64-linux-gnu/7.2.0/../../../aarch64-linux-gnu/libboost_log.a(core.o): relocation R_AARCH64_ADR_PREL_PG_HI21 against symbol `_ZTVN5boost16exception_detail19error_info_injectorISt13runtime_errorEE' which may bind externally can not be used when making a shared object; recompile with -fPIC
/usr/bin/ld: /usr/bin/../lib/gcc/aarch64-linux-gnu/7.2.0/../../../aarch64-linux-gnu/libboost_log.a(core.o): relocation R_AARCH64_ADR_PREL_PG_HI21 against symbol `_ZTVN5boost9exceptionE' which may bind externally can not be used when making a shared object; recompile with -fPIC
/usr/bin/ld: /usr/bin/../lib/gcc/aarch64-linux-gnu/7.2.0/../../../aarch64-linux-gnu/libboost_log.a(core.o): relocation R_AARCH64_ADR_PREL_PG_HI21 against symbol `_ZTVN5boost16exception_detail10clone_baseE' which may bind externally can not be used when making a shared object; recompile with -fPIC
...
```
There is either an error or inconsistency in the `cmake/GlobalConfig.cmake` included from the main `CMakeLists.txt` which contains the following bit related to Boost:
```
add_definitions("-DBOOST_ALL_NO_LIB") # Turn off auto-linking as we specify the libs manually
set(Boost_USE_STATIC_LIBS ON)
find_package(Boost 1.59 REQUIRED COMPONENTS unit_test_framework system filesystem log program_options)
include_directories(SYSTEM "${Boost_INCLUDE_DIR}")
link_directories(${Boost_LIBRARY_DIR})
```
It suddenly requires Boost v.1.59 (which does not match 1.64 required by the documentation), however when I patched this to 1.64 my system still insisted on linking with the system's Boost. As you can see from the log above, the `-L` flag necessary to enforce specific Boost's path was missing from the linking command.
However AS SOON AS I uninstalled the system's Boost, everything both compiled and ran without any problems. Please review your `cmake` code related to linking with Boost. | non_priority | a system s native boost preferred over the one specifically compiled for armnn dear armnn developers it took me a few days to diagnose a weird family of issues where compilation of armnn and consumer programs happened without errors with some warnings though and either bus error or segmentation fault would appear during the runtime this happened on a host platform running debian linux which also had a pre installed version of boost or this detail proved to be essential according to the documentation we need boost with fpic so this is what i compiled and passed on to cmake building the program when cmake runs it reports that it found our boost v and pretends to take it into account however when i switch on all logging i can see that the path to this newly compiled boost is nowhere to be seen in the command used to link the actual armnn so the linking command looks like this i m skipping the object files in the middle usr bin cmake e cmake link script cmakefiles armnn dir link txt verbose usr bin clang fpic std c werror std c wall wold style cast wno missing braces wconversion wsign con version dndebug shared wl soname libarmnn so o libarmnn so cmakefiles armnn dir src armnn layers activationlayer cpp o cmakefiles armnn dir src armnn layers additionlayer cpp o cmakefiles armnn dir src armnn layers batchnormalizationlayer cpp o cmakefiles armnn dir src armnn layers batchtospacendlayer cpp o src backends reference workloads cmakefiles armnnrefbackendworkloads dir cpp o src backends reference workloads cmakefiles armnnrefbackendworkloads dir cpp o libarmnnutils a ldl wl bstatic lboost log lboost thread lboost system wl bdynamic which is immediately followed by reaction from the linker usr bin ld usr bin lib gcc linux gnu linux gnu libboost log a core o relocation r adr prel pg against symbol info erroree which may bind externally can not be used when making a shared object recompile with fpic usr bin ld usr bin lib gcc linux gnu linux gnu libboost log a core o relocation r adr prel pg against symbol which may bind externally can not be used when making a shared object recompile with fpic usr bin ld usr bin lib gcc linux gnu linux gnu libboost log a core o relocation r adr prel pg against symbol basee which may bind externally can not be used when making a shared object recompile with fpic there is either an error or inconsistency in the cmake globalconfig cmake included from the main cmakelists txt which contains the following bit related to boost add definitions dboost all no lib turn off auto linking as we specify the libs manually set boost use static libs on find package boost required components unit test framework system filesystem log program options include directories system boost include dir link directories boost library dir it suddenly requires boost v which does not match required by the documentation however when i patched this to my system still insisted on linking with the system s boost as you can see from the log above the l flag necessary to enforce specific boost s path was missing from the linking command however as soon as i uninstalled the system s boost everything both compiled and ran without any problems please review your cmake code related to linking with boost | 0 |
40,076 | 12,746,037,147 | IssuesEvent | 2020-06-26 15:14:34 | RG4421/developers | https://api.github.com/repos/RG4421/developers | opened | WS-2018-0625 (High) detected in xmlbuilder-4.2.1.tgz | security vulnerability | ## WS-2018-0625 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmlbuilder-4.2.1.tgz</b></p></summary>
<p>An XML builder for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmlbuilder/-/xmlbuilder-4.2.1.tgz">https://registry.npmjs.org/xmlbuilder/-/xmlbuilder-4.2.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/developers/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/developers/node_modules/xmlbuilder/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-plugin-react-svg-2.1.2.tgz (Root Library)
- svg-react-loader-0.4.6.tgz
- xml2js-0.4.17.tgz
- :x: **xmlbuilder-4.2.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RG4421/developers/commit/09bff0d3b38e28079c6c900ddef39d33d88ab428">09bff0d3b38e28079c6c900ddef39d33d88ab428</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package xmlbuilder-js before 9.0.5 is vulnerable to denial of service due to a regular expression issue.
<p>Publish Date: 2018-02-08
<p>URL: <a href=https://github.com/oozcitak/xmlbuilder-js/commit/bbf929a8a54f0d012bdc44cbe622fdeda2509230>WS-2018-0625</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/oozcitak/xmlbuilder-js/commit/bbf929a8a54f0d012bdc44cbe622fdeda2509230">https://github.com/oozcitak/xmlbuilder-js/commit/bbf929a8a54f0d012bdc44cbe622fdeda2509230</a></p>
<p>Release Date: 2020-03-23</p>
<p>Fix Resolution: 9.0.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"xmlbuilder","packageVersion":"4.2.1","isTransitiveDependency":true,"dependencyTree":"gatsby-plugin-react-svg:2.1.2;svg-react-loader:0.4.6;xml2js:0.4.17;xmlbuilder:4.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"9.0.5"}],"vulnerabilityIdentifier":"WS-2018-0625","vulnerabilityDetails":"The package xmlbuilder-js before 9.0.5 is vulnerable to denial of service due to a regular expression issue.","vulnerabilityUrl":"https://github.com/oozcitak/xmlbuilder-js/commit/bbf929a8a54f0d012bdc44cbe622fdeda2509230","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | WS-2018-0625 (High) detected in xmlbuilder-4.2.1.tgz - ## WS-2018-0625 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xmlbuilder-4.2.1.tgz</b></p></summary>
<p>An XML builder for node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmlbuilder/-/xmlbuilder-4.2.1.tgz">https://registry.npmjs.org/xmlbuilder/-/xmlbuilder-4.2.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/developers/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/developers/node_modules/xmlbuilder/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-plugin-react-svg-2.1.2.tgz (Root Library)
- svg-react-loader-0.4.6.tgz
- xml2js-0.4.17.tgz
- :x: **xmlbuilder-4.2.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RG4421/developers/commit/09bff0d3b38e28079c6c900ddef39d33d88ab428">09bff0d3b38e28079c6c900ddef39d33d88ab428</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package xmlbuilder-js before 9.0.5 is vulnerable to denial of service due to a regular expression issue.
<p>Publish Date: 2018-02-08
<p>URL: <a href=https://github.com/oozcitak/xmlbuilder-js/commit/bbf929a8a54f0d012bdc44cbe622fdeda2509230>WS-2018-0625</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/oozcitak/xmlbuilder-js/commit/bbf929a8a54f0d012bdc44cbe622fdeda2509230">https://github.com/oozcitak/xmlbuilder-js/commit/bbf929a8a54f0d012bdc44cbe622fdeda2509230</a></p>
<p>Release Date: 2020-03-23</p>
<p>Fix Resolution: 9.0.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"xmlbuilder","packageVersion":"4.2.1","isTransitiveDependency":true,"dependencyTree":"gatsby-plugin-react-svg:2.1.2;svg-react-loader:0.4.6;xml2js:0.4.17;xmlbuilder:4.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"9.0.5"}],"vulnerabilityIdentifier":"WS-2018-0625","vulnerabilityDetails":"The package xmlbuilder-js before 9.0.5 is vulnerable to denial of service due to a regular expression issue.","vulnerabilityUrl":"https://github.com/oozcitak/xmlbuilder-js/commit/bbf929a8a54f0d012bdc44cbe622fdeda2509230","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | ws high detected in xmlbuilder tgz ws high severity vulnerability vulnerable library xmlbuilder tgz an xml builder for node js library home page a href path to dependency file tmp ws scm developers package json path to vulnerable library tmp ws scm developers node modules xmlbuilder package json dependency hierarchy gatsby plugin react svg tgz root library svg react loader tgz tgz x xmlbuilder tgz vulnerable library found in head commit a href vulnerability details the package xmlbuilder js before is vulnerable to denial of service due to a regular expression issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails the package xmlbuilder js before is vulnerable to denial of service due to a regular expression issue vulnerabilityurl | 0 |
103,228 | 8,883,394,748 | IssuesEvent | 2019-01-14 15:35:47 | khartec/waltz | https://api.github.com/repos/khartec/waltz | closed | Taxonomy Management: DDL mistake. Params is Long, not String | DDL change bug fixed (test & close) | Currently mapping `params` into the unused `description` field. Will need to migrate any existing data in this field when this task is worked on. | 1.0 | Taxonomy Management: DDL mistake. Params is Long, not String - Currently mapping `params` into the unused `description` field. Will need to migrate any existing data in this field when this task is worked on. | non_priority | taxonomy management ddl mistake params is long not string currently mapping params into the unused description field will need to migrate any existing data in this field when this task is worked on | 0 |
71,313 | 8,643,675,337 | IssuesEvent | 2018-11-25 20:10:47 | cockpit-project/cockpit | https://api.github.com/repos/cockpit-project/cockpit | closed | persistent metrics vs. Store Performance data | needsdesign | It's called one thing before it's installed, and another once it's installed.
Maybe best to call it Persistent metrics in both places, since it's the shorter label.
@garrett thoughts? | 1.0 | persistent metrics vs. Store Performance data - It's called one thing before it's installed, and another once it's installed.
Maybe best to call it Persistent metrics in both places, since it's the shorter label.
@garrett thoughts? | non_priority | persistent metrics vs store performance data it s called one thing before it s installed and another once it s installed maybe best to call it persistent metrics in both places since it s the shorter label garrett thoughts | 0 |
163,787 | 13,927,318,912 | IssuesEvent | 2020-10-21 19:38:45 | phalcon/cphalcon | https://api.github.com/repos/phalcon/cphalcon | closed | [BUG] Custom class filtering/sanitizing returns null | 4.1.0 bug documentation | The result of sanitizing ipv4 with class:
https://docs.phalcon.io/4.0/en/filter#adding-sanitizers
returns null.
PHP 7.4.6
Phalcon 4.0.6 | 1.0 | [BUG] Custom class filtering/sanitizing returns null - The result of sanitizing ipv4 with class:
https://docs.phalcon.io/4.0/en/filter#adding-sanitizers
returns null.
PHP 7.4.6
Phalcon 4.0.6 | non_priority | custom class filtering sanitizing returns null the result of sanitizing with class returns null php phalcon | 0 |
246,411 | 26,611,696,734 | IssuesEvent | 2023-01-24 01:05:00 | MohamedElashri/snibox-1 | https://api.github.com/repos/MohamedElashri/snibox-1 | opened | CVE-2020-10663 (High) detected in json-2.2.0.gem | security vulnerability | ## CVE-2020-10663 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-2.2.0.gem</b></p></summary>
<p>This is a JSON implementation as a Ruby extension in C.</p>
<p>Library home page: <a href="https://rubygems.org/gems/json-2.2.0.gem">https://rubygems.org/gems/json-2.2.0.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/json-2.2.0.gem</p>
<p>
Dependency Hierarchy:
- simplecov-0.14.1.gem (Root Library)
- :x: **json-2.2.0.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The JSON gem through 2.2.0 for Ruby, as used in Ruby 2.4 through 2.4.9, 2.5 through 2.5.7, and 2.6 through 2.6.5, has an Unsafe Object Creation Vulnerability. This is quite similar to CVE-2013-0269, but does not rely on poor garbage-collection behavior within Ruby. Specifically, use of JSON parsing methods can lead to creation of a malicious object within the interpreter, with adverse effects that are application-dependent.
<p>Publish Date: 2020-04-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-10663>CVE-2020-10663</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.ruby-lang.org/en/news/2020/03/19/json-dos-cve-2020-10663/">https://www.ruby-lang.org/en/news/2020/03/19/json-dos-cve-2020-10663/</a></p>
<p>Release Date: 2020-04-28</p>
<p>Fix Resolution: 2.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-10663 (High) detected in json-2.2.0.gem - ## CVE-2020-10663 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>json-2.2.0.gem</b></p></summary>
<p>This is a JSON implementation as a Ruby extension in C.</p>
<p>Library home page: <a href="https://rubygems.org/gems/json-2.2.0.gem">https://rubygems.org/gems/json-2.2.0.gem</a></p>
<p>Path to dependency file: /Gemfile.lock</p>
<p>Path to vulnerable library: /home/wss-scanner/.gem/ruby/2.7.0/cache/json-2.2.0.gem</p>
<p>
Dependency Hierarchy:
- simplecov-0.14.1.gem (Root Library)
- :x: **json-2.2.0.gem** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The JSON gem through 2.2.0 for Ruby, as used in Ruby 2.4 through 2.4.9, 2.5 through 2.5.7, and 2.6 through 2.6.5, has an Unsafe Object Creation Vulnerability. This is quite similar to CVE-2013-0269, but does not rely on poor garbage-collection behavior within Ruby. Specifically, use of JSON parsing methods can lead to creation of a malicious object within the interpreter, with adverse effects that are application-dependent.
<p>Publish Date: 2020-04-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-10663>CVE-2020-10663</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.ruby-lang.org/en/news/2020/03/19/json-dos-cve-2020-10663/">https://www.ruby-lang.org/en/news/2020/03/19/json-dos-cve-2020-10663/</a></p>
<p>Release Date: 2020-04-28</p>
<p>Fix Resolution: 2.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in json gem cve high severity vulnerability vulnerable library json gem this is a json implementation as a ruby extension in c library home page a href path to dependency file gemfile lock path to vulnerable library home wss scanner gem ruby cache json gem dependency hierarchy simplecov gem root library x json gem vulnerable library found in base branch master vulnerability details the json gem through for ruby as used in ruby through through and through has an unsafe object creation vulnerability this is quite similar to cve but does not rely on poor garbage collection behavior within ruby specifically use of json parsing methods can lead to creation of a malicious object within the interpreter with adverse effects that are application dependent publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
5,724 | 13,050,819,580 | IssuesEvent | 2020-07-29 16:04:13 | zielu/GitToolBox | https://api.github.com/repos/zielu/GitToolBox | closed | Version 202 is compiled with Java 11 bugs on IntelliJ when run with jbrsdk8 | 202 T-shirt: S architecture | Jetbrains IntelliJ IDEA can be run with a JRE of version 8 (due to various reasons, enterprise constraints ...).
However when enabling the last version of the GitToolBox plugin after the 202.0.3 update, it breaks some of IntelliJ's gui : View -> Appearance menu doesn't open anymore, and most of the toolbars are missing. Disabling GitToolBox gets things back to normal.
`Caused by: java.lang.UnsupportedClassVersionError: zielu/gittoolbox/changes/ChangeListSubscriberChangeListListener has been compiled by a more recent version of the Java Runtime (class file version 55.0), this version of the Java Runtime only recognizes class file versions up to 52.0` | 1.0 | Version 202 is compiled with Java 11 bugs on IntelliJ when run with jbrsdk8 - Jetbrains IntelliJ IDEA can be run with a JRE of version 8 (due to various reasons, enterprise constraints ...).
However when enabling the last version of the GitToolBox plugin after the 202.0.3 update, it breaks some of IntelliJ's gui : View -> Appearance menu doesn't open anymore, and most of the toolbars are missing. Disabling GitToolBox gets things back to normal.
`Caused by: java.lang.UnsupportedClassVersionError: zielu/gittoolbox/changes/ChangeListSubscriberChangeListListener has been compiled by a more recent version of the Java Runtime (class file version 55.0), this version of the Java Runtime only recognizes class file versions up to 52.0` | non_priority | version is compiled with java bugs on intellij when run with jetbrains intellij idea can be run with a jre of version due to various reasons enterprise constraints however when enabling the last version of the gittoolbox plugin after the update it breaks some of intellij s gui view appearance menu doesn t open anymore and most of the toolbars are missing disabling gittoolbox gets things back to normal caused by java lang unsupportedclassversionerror zielu gittoolbox changes changelistsubscriberchangelistlistener has been compiled by a more recent version of the java runtime class file version this version of the java runtime only recognizes class file versions up to | 0 |
86,442 | 24,850,316,283 | IssuesEvent | 2022-10-26 19:26:55 | NVIDIA/spark-rapids | https://api.github.com/repos/NVIDIA/spark-rapids | closed | Spark-3.4 build failure in AggregateFunctions | build P0 Spark 3.4+ | **Is your feature request related to a problem? Please describe.**
Spark3.4 build is failing in AggregateFunctions with below errors. Need to fix these errors:
```
06:04:19 [[1;31mERROR[m] [Error] /home/jenkins/agent/workspace/spark340_nightly-github/sql-plugin/src/main/scala/org/apache/spark/sql/rapids/AggregateFunctions.scala:1084: value checkForNumericExpr is not a member of object org.apache.spark.sql.catalyst.util.TypeUtils
06:04:19 [[1;31mERROR[m] [Error] /home/jenkins/agent/workspace/spark340_nightly-github/sql-plugin/src/main/scala/org/apache/spark/sql/rapids/AggregateFunctions.scala:1575: value checkForNumericExpr is not a member of object org.apache.spark.sql.catalyst.util.TypeUtils
``` | 1.0 | Spark-3.4 build failure in AggregateFunctions - **Is your feature request related to a problem? Please describe.**
Spark3.4 build is failing in AggregateFunctions with below errors. Need to fix these errors:
```
06:04:19 [[1;31mERROR[m] [Error] /home/jenkins/agent/workspace/spark340_nightly-github/sql-plugin/src/main/scala/org/apache/spark/sql/rapids/AggregateFunctions.scala:1084: value checkForNumericExpr is not a member of object org.apache.spark.sql.catalyst.util.TypeUtils
06:04:19 [[1;31mERROR[m] [Error] /home/jenkins/agent/workspace/spark340_nightly-github/sql-plugin/src/main/scala/org/apache/spark/sql/rapids/AggregateFunctions.scala:1575: value checkForNumericExpr is not a member of object org.apache.spark.sql.catalyst.util.TypeUtils
``` | non_priority | spark build failure in aggregatefunctions is your feature request related to a problem please describe build is failing in aggregatefunctions with below errors need to fix these errors home jenkins agent workspace nightly github sql plugin src main scala org apache spark sql rapids aggregatefunctions scala value checkfornumericexpr is not a member of object org apache spark sql catalyst util typeutils home jenkins agent workspace nightly github sql plugin src main scala org apache spark sql rapids aggregatefunctions scala value checkfornumericexpr is not a member of object org apache spark sql catalyst util typeutils | 0 |
471 | 2,731,835,343 | IssuesEvent | 2015-04-16 22:53:26 | crewapp/crewapp | https://api.github.com/repos/crewapp/crewapp | closed | Passwords hashes are not being saved into the DB | bug MySQL production security server | passwords hash's are not being saved right now we new users are being created. | True | Passwords hashes are not being saved into the DB - passwords hash's are not being saved right now we new users are being created. | non_priority | passwords hashes are not being saved into the db passwords hash s are not being saved right now we new users are being created | 0 |
220,203 | 17,157,418,120 | IssuesEvent | 2021-07-14 08:45:53 | masatakashiwagi/paper | https://api.github.com/repos/masatakashiwagi/paper | closed | A/Bテストに必要な期間の予測方法とその決定要因 | A/B test article | # A/Bテストに必要な期間の予測方法とその決定要因
* Link: https://www.assion.co.jp/blog/abtest-required-term/
## どのようなもの?
- A/Bテストの結果が出るためのポイントは,Aバージョン・Bバージョンの目標達成数の間にある「差」
- 目標達成数とは,テストすることで改善させたい数値(特定ページへの遷移数とか...)
- 一定のUU数に対して,この目標の達成数に「差」が開くことでAかBどちらが優れたページかを判断する
- 差は有意性検定で調べる
- 有意差が出るためには,以下を考える
- サンプル数を増やす
- 差が大きく開いている
- UU数を確保するためには,テスト期間が長くなってしまうことがよくある
- テスト期間はだいたい1週間~2週間、長くても1ヶ月
- その期間の経過後もテストに結論がでていなければ別のテストに移行する | 1.0 | A/Bテストに必要な期間の予測方法とその決定要因 - # A/Bテストに必要な期間の予測方法とその決定要因
* Link: https://www.assion.co.jp/blog/abtest-required-term/
## どのようなもの?
- A/Bテストの結果が出るためのポイントは,Aバージョン・Bバージョンの目標達成数の間にある「差」
- 目標達成数とは,テストすることで改善させたい数値(特定ページへの遷移数とか...)
- 一定のUU数に対して,この目標の達成数に「差」が開くことでAかBどちらが優れたページかを判断する
- 差は有意性検定で調べる
- 有意差が出るためには,以下を考える
- サンプル数を増やす
- 差が大きく開いている
- UU数を確保するためには,テスト期間が長くなってしまうことがよくある
- テスト期間はだいたい1週間~2週間、長くても1ヶ月
- その期間の経過後もテストに結論がでていなければ別のテストに移行する | non_priority | a bテストに必要な期間の予測方法とその決定要因 a bテストに必要な期間の予測方法とその決定要因 link どのようなもの? a bテストの結果が出るためのポイントは,aバージョン・bバージョンの目標達成数の間にある「差」 目標達成数とは,テストすることで改善させたい数値(特定ページへの遷移数とか ) 一定のuu数に対して,この目標の達成数に「差」が開くことでaかbどちらが優れたページかを判断する 差は有意性検定で調べる 有意差が出るためには,以下を考える サンプル数を増やす 差が大きく開いている uu数を確保するためには,テスト期間が長くなってしまうことがよくある ~ 、 その期間の経過後もテストに結論がでていなければ別のテストに移行する | 0 |
146,079 | 11,725,307,831 | IssuesEvent | 2020-03-10 12:42:39 | dynasist/ALObjectParser | https://api.github.com/repos/dynasist/ALObjectParser | closed | Separate parser from objects | testing needed | Can I suggest that we make the ALObjectParser a separate, static and stateless class (ALObjectParser)? The only thing it should do, as far as I'm concerned, is something like this:
```csharp
static IEnumerable<ALObject> Read(string path)
```
I chose an IEnumerable as the return value type since a file could contain >1 object. For writing, I think we need the mirrored version, again in a separate, static and stateless class (ALObjectWriter?)
```csharp
static void Write(IEnumerable<ALObject> objects, string path)
``` | 1.0 | Separate parser from objects - Can I suggest that we make the ALObjectParser a separate, static and stateless class (ALObjectParser)? The only thing it should do, as far as I'm concerned, is something like this:
```csharp
static IEnumerable<ALObject> Read(string path)
```
I chose an IEnumerable as the return value type since a file could contain >1 object. For writing, I think we need the mirrored version, again in a separate, static and stateless class (ALObjectWriter?)
```csharp
static void Write(IEnumerable<ALObject> objects, string path)
``` | non_priority | separate parser from objects can i suggest that we make the alobjectparser a separate static and stateless class alobjectparser the only thing it should do as far as i m concerned is something like this csharp static ienumerable read string path i chose an ienumerable as the return value type since a file could contain object for writing i think we need the mirrored version again in a separate static and stateless class alobjectwriter csharp static void write ienumerable objects string path | 0 |
213,846 | 16,541,099,435 | IssuesEvent | 2021-05-27 16:53:53 | dotnet/efcore | https://api.github.com/repos/dotnet/efcore | opened | Add code coverage report to AzDo | area-test type-enhancement | https://docs.microsoft.com/azure/devops/pipelines/tasks/test/publish-code-coverage-results
https://docs.microsoft.com/dotnet/core/testing/unit-testing-code-coverage?tabs=windows
```dotnet test --collect:"XPlat Code Coverage"```
```dotnet test /p:CollectCoverage=true /p:CoverletOutputFormat=cobertura```
Related: https://github.com/dotnet/efcore/issues/22589 | 1.0 | Add code coverage report to AzDo - https://docs.microsoft.com/azure/devops/pipelines/tasks/test/publish-code-coverage-results
https://docs.microsoft.com/dotnet/core/testing/unit-testing-code-coverage?tabs=windows
```dotnet test --collect:"XPlat Code Coverage"```
```dotnet test /p:CollectCoverage=true /p:CoverletOutputFormat=cobertura```
Related: https://github.com/dotnet/efcore/issues/22589 | non_priority | add code coverage report to azdo dotnet test collect xplat code coverage dotnet test p collectcoverage true p coverletoutputformat cobertura related | 0 |
277,974 | 21,057,929,237 | IssuesEvent | 2022-04-01 06:30:48 | jessicajacelyn/ped | https://api.github.com/repos/jessicajacelyn/ped | closed | Command given in UG resulted in error | type.DocumentationBug severity.Low | 
I copied the command example from the UserGuide but it resulted in error.
<!--session: 1648793891700-2e7c0b6b-2767-41e2-ab2c-530cfe873aec-->
<!--Version: Web v3.4.2--> | 1.0 | Command given in UG resulted in error - 
I copied the command example from the UserGuide but it resulted in error.
<!--session: 1648793891700-2e7c0b6b-2767-41e2-ab2c-530cfe873aec-->
<!--Version: Web v3.4.2--> | non_priority | command given in ug resulted in error i copied the command example from the userguide but it resulted in error | 0 |
12,024 | 3,563,564,214 | IssuesEvent | 2016-01-25 04:55:23 | pullapprove/support | https://api.github.com/repos/pullapprove/support | closed | "At least N other collaborators" mode? | idea in progress needs documentation | When using PullApprove with a simple project consisting of maybe one or two main collaborators and the occasional side developer, I find myself wishing I could get rid of the notion of "these people are PR reviewers" and replace it with "as long as at least 1/2/N other people with access to the repo approve it, you're good". Then I don't have to configure who has approval or worry about self approval, or update the list of reviewers when a new collaborator joins. This also would make setting up new repos much simpler. | 1.0 | "At least N other collaborators" mode? - When using PullApprove with a simple project consisting of maybe one or two main collaborators and the occasional side developer, I find myself wishing I could get rid of the notion of "these people are PR reviewers" and replace it with "as long as at least 1/2/N other people with access to the repo approve it, you're good". Then I don't have to configure who has approval or worry about self approval, or update the list of reviewers when a new collaborator joins. This also would make setting up new repos much simpler. | non_priority | at least n other collaborators mode when using pullapprove with a simple project consisting of maybe one or two main collaborators and the occasional side developer i find myself wishing i could get rid of the notion of these people are pr reviewers and replace it with as long as at least n other people with access to the repo approve it you re good then i don t have to configure who has approval or worry about self approval or update the list of reviewers when a new collaborator joins this also would make setting up new repos much simpler | 0 |
120,668 | 10,131,404,625 | IssuesEvent | 2019-08-01 19:27:02 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Tracking Issue - Conformance Coverage for Graceful Termination | area/conformance kind/feature sig/testing | <!-- Please only use this template for submitting enhancement requests -->
**What would you like to be added**:
This is a place holder to track Conformance Coverage on the Topic.
This issue should remain open until all related work is accomplished in the k/k repo
This issue will contain analysis of the coverage. Existing tests and additional tests requested with links to those issues/PR's
**Why is this needed**:
Add or promote Graceful Termination test coverage for conformance
/area conformance
/assign @pontiyaraja
| 1.0 | Tracking Issue - Conformance Coverage for Graceful Termination - <!-- Please only use this template for submitting enhancement requests -->
**What would you like to be added**:
This is a place holder to track Conformance Coverage on the Topic.
This issue should remain open until all related work is accomplished in the k/k repo
This issue will contain analysis of the coverage. Existing tests and additional tests requested with links to those issues/PR's
**Why is this needed**:
Add or promote Graceful Termination test coverage for conformance
/area conformance
/assign @pontiyaraja
| non_priority | tracking issue conformance coverage for graceful termination what would you like to be added this is a place holder to track conformance coverage on the topic this issue should remain open until all related work is accomplished in the k k repo this issue will contain analysis of the coverage existing tests and additional tests requested with links to those issues pr s why is this needed add or promote graceful termination test coverage for conformance area conformance assign pontiyaraja | 0 |
219,192 | 24,457,155,384 | IssuesEvent | 2022-10-07 07:52:56 | adoptium/containers | https://api.github.com/repos/adoptium/containers | closed | [Security] Critical CVE-2022-40674 in expat | security | Hello,
There is this critical CVE [CVE-2022-40674](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-40674)
for expat which affect lots of temurin docker images.
For alpine based images which affects me directly fixed version of expat 2.4.9-r0 is already in alpine repository. If understand correctly, you need to "refresh" temurin docker image to avoid using existing image layer.


Regards | True | [Security] Critical CVE-2022-40674 in expat - Hello,
There is this critical CVE [CVE-2022-40674](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-40674)
for expat which affect lots of temurin docker images.
For alpine based images which affects me directly fixed version of expat 2.4.9-r0 is already in alpine repository. If understand correctly, you need to "refresh" temurin docker image to avoid using existing image layer.


Regards | non_priority | critical cve in expat hello there is this critical cve for expat which affect lots of temurin docker images for alpine based images which affects me directly fixed version of expat is already in alpine repository if understand correctly you need to refresh temurin docker image to avoid using existing image layer regards | 0 |
205,352 | 15,610,281,664 | IssuesEvent | 2021-03-19 13:00:12 | NationalSecurityAgency/skills-service | https://api.github.com/repos/NationalSecurityAgency/skills-service | closed | Client Display: Improve Point History chart | enhancement test | expand maximum number of days loaded; make the property configurable | 1.0 | Client Display: Improve Point History chart - expand maximum number of days loaded; make the property configurable | non_priority | client display improve point history chart expand maximum number of days loaded make the property configurable | 0 |
148,780 | 11,864,858,035 | IssuesEvent | 2020-03-25 22:40:11 | longhorn/longhorn | https://api.github.com/repos/longhorn/longhorn | opened | [BUG] Nightly Upgrade Test: test_restore_inc sometimes failed | area/test bug | **Describe the bug**
**To Reproduce**
Steps to reproduce the behavior:
1. Install Longhorn v0.7.0
2. Upgrade Longhorn to `master`
3. Run `test_restore_inc` test
**Expected behavior**
test should pass
**Log**
```
clients = {'longhorn-tests-01': <longhorn.Client object at 0x7fcf376a6fd0>, 'longhorn-tests-02': <longhorn.Client object at 0x7fcf37df1c10>, 'longhorn-tests-03': <longhorn.Client object at 0x7fcf37624350>}
core_api = <kubernetes.client.apis.core_v1_api.CoreV1Api object at 0x7fcf376ec190>
volume_name = 'longhorn-testvol-1xosau'
pod = {'apiVersion': 'v1', 'kind': 'Pod', 'metadata': {'name': 'pod-sb-2-longhorn-testvol-1xosau'}, 'spec': {'containers': [...ep', ...}], 'volumes': [{'name': 'pod-data', 'persistentVolumeClaim': {'claimName': 'sb-2-longhorn-testvol-1xosau'}}]}}
@pytest.mark.coretest # NOQA
def test_restore_inc(clients, core_api, volume_name, pod): # NOQA
for _, client in iter(clients.items()):
break
setting = client.by_id_setting(common.SETTING_BACKUP_TARGET)
# test backupTarget for multiple settings
backupstores = common.get_backupstore_url()
for backupstore in backupstores:
if common.is_backupTarget_s3(backupstore):
backupsettings = backupstore.split("$")
setting = client.update(setting, value=backupsettings[0])
assert setting.value == backupsettings[0]
credential = client.by_id_setting(
common.SETTING_BACKUP_TARGET_CREDENTIAL_SECRET)
credential = client.update(credential, value=backupsettings[1])
assert credential.value == backupsettings[1]
else:
setting = client.update(setting, value=backupstore)
assert setting.value == backupstore
credential = client.by_id_setting(
common.SETTING_BACKUP_TARGET_CREDENTIAL_SECRET)
credential = client.update(credential, value="")
assert credential.value == ""
> restore_inc_test(client, core_api, volume_name, pod)
test_basic.py:595:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
test_basic.py:731: in restore_inc_test
_, backup2, _, data2 = create_backup(client, volume_name, data2)
common.py:222: in create_backup
bv, b = find_backup(client, volname, snap.name)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
client = <longhorn.Client object at 0x7fcf376a6fd0>
vol_name = 'longhorn-testvol-1xosau'
snap_name = 'b0f5359c-58e3-4d3b-88bf-67d3ed1952d0'
def find_backup(client, vol_name, snap_name):
found = False
for i in range(100):
bvs = client.list_backupVolume()
for bv in bvs:
if bv.name == vol_name:
found = True
break
if found:
break
time.sleep(1)
assert found
found = False
for i in range(20):
backups = bv.backupList().data
for b in backups:
if b.snapshotName == snap_name:
found = True
break
if found:
break
time.sleep(1)
> assert found
E AssertionError
common.py:2169: AssertionError
```
**Environment:**
- Longhorn version: 0.8.0
- Kubernetes version: v1.17.2
- Node OS type and version: Ubuntu 18.04
**Additional context:**
longhorn-upgrade-tests/18 | 1.0 | [BUG] Nightly Upgrade Test: test_restore_inc sometimes failed - **Describe the bug**
**To Reproduce**
Steps to reproduce the behavior:
1. Install Longhorn v0.7.0
2. Upgrade Longhorn to `master`
3. Run `test_restore_inc` test
**Expected behavior**
test should pass
**Log**
```
clients = {'longhorn-tests-01': <longhorn.Client object at 0x7fcf376a6fd0>, 'longhorn-tests-02': <longhorn.Client object at 0x7fcf37df1c10>, 'longhorn-tests-03': <longhorn.Client object at 0x7fcf37624350>}
core_api = <kubernetes.client.apis.core_v1_api.CoreV1Api object at 0x7fcf376ec190>
volume_name = 'longhorn-testvol-1xosau'
pod = {'apiVersion': 'v1', 'kind': 'Pod', 'metadata': {'name': 'pod-sb-2-longhorn-testvol-1xosau'}, 'spec': {'containers': [...ep', ...}], 'volumes': [{'name': 'pod-data', 'persistentVolumeClaim': {'claimName': 'sb-2-longhorn-testvol-1xosau'}}]}}
@pytest.mark.coretest # NOQA
def test_restore_inc(clients, core_api, volume_name, pod): # NOQA
for _, client in iter(clients.items()):
break
setting = client.by_id_setting(common.SETTING_BACKUP_TARGET)
# test backupTarget for multiple settings
backupstores = common.get_backupstore_url()
for backupstore in backupstores:
if common.is_backupTarget_s3(backupstore):
backupsettings = backupstore.split("$")
setting = client.update(setting, value=backupsettings[0])
assert setting.value == backupsettings[0]
credential = client.by_id_setting(
common.SETTING_BACKUP_TARGET_CREDENTIAL_SECRET)
credential = client.update(credential, value=backupsettings[1])
assert credential.value == backupsettings[1]
else:
setting = client.update(setting, value=backupstore)
assert setting.value == backupstore
credential = client.by_id_setting(
common.SETTING_BACKUP_TARGET_CREDENTIAL_SECRET)
credential = client.update(credential, value="")
assert credential.value == ""
> restore_inc_test(client, core_api, volume_name, pod)
test_basic.py:595:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
test_basic.py:731: in restore_inc_test
_, backup2, _, data2 = create_backup(client, volume_name, data2)
common.py:222: in create_backup
bv, b = find_backup(client, volname, snap.name)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
client = <longhorn.Client object at 0x7fcf376a6fd0>
vol_name = 'longhorn-testvol-1xosau'
snap_name = 'b0f5359c-58e3-4d3b-88bf-67d3ed1952d0'
def find_backup(client, vol_name, snap_name):
found = False
for i in range(100):
bvs = client.list_backupVolume()
for bv in bvs:
if bv.name == vol_name:
found = True
break
if found:
break
time.sleep(1)
assert found
found = False
for i in range(20):
backups = bv.backupList().data
for b in backups:
if b.snapshotName == snap_name:
found = True
break
if found:
break
time.sleep(1)
> assert found
E AssertionError
common.py:2169: AssertionError
```
**Environment:**
- Longhorn version: 0.8.0
- Kubernetes version: v1.17.2
- Node OS type and version: Ubuntu 18.04
**Additional context:**
longhorn-upgrade-tests/18 | non_priority | nightly upgrade test test restore inc sometimes failed describe the bug to reproduce steps to reproduce the behavior install longhorn upgrade longhorn to master run test restore inc test expected behavior test should pass log clients longhorn tests longhorn tests longhorn tests core api volume name longhorn testvol pod apiversion kind pod metadata name pod sb longhorn testvol spec containers volumes pytest mark coretest noqa def test restore inc clients core api volume name pod noqa for client in iter clients items break setting client by id setting common setting backup target test backuptarget for multiple settings backupstores common get backupstore url for backupstore in backupstores if common is backuptarget backupstore backupsettings backupstore split setting client update setting value backupsettings assert setting value backupsettings credential client by id setting common setting backup target credential secret credential client update credential value backupsettings assert credential value backupsettings else setting client update setting value backupstore assert setting value backupstore credential client by id setting common setting backup target credential secret credential client update credential value assert credential value restore inc test client core api volume name pod test basic py test basic py in restore inc test create backup client volume name common py in create backup bv b find backup client volname snap name client vol name longhorn testvol snap name def find backup client vol name snap name found false for i in range bvs client list backupvolume for bv in bvs if bv name vol name found true break if found break time sleep assert found found false for i in range backups bv backuplist data for b in backups if b snapshotname snap name found true break if found break time sleep assert found e assertionerror common py assertionerror environment longhorn version kubernetes version node os type and version ubuntu additional context longhorn upgrade tests | 0 |
71,022 | 9,476,283,421 | IssuesEvent | 2019-04-19 14:35:08 | fga-eps-mds/2019.1-unbrake | https://api.github.com/repos/fga-eps-mds/2019.1-unbrake | closed | Configurar mkdocs | documentation | ## Descrição
Substituir pelican por mkdocs na geração do GitHub pages
.
## Tarefas
- [ ] Trocar configuração para mkdocs;
- [ ] Ajustar arquivos .md e imagens nas pastas corretas.
## Informação adicional
| 1.0 | Configurar mkdocs - ## Descrição
Substituir pelican por mkdocs na geração do GitHub pages
.
## Tarefas
- [ ] Trocar configuração para mkdocs;
- [ ] Ajustar arquivos .md e imagens nas pastas corretas.
## Informação adicional
| non_priority | configurar mkdocs descrição substituir pelican por mkdocs na geração do github pages tarefas trocar configuração para mkdocs ajustar arquivos md e imagens nas pastas corretas informação adicional | 0 |
113,106 | 14,368,054,833 | IssuesEvent | 2020-12-01 07:47:21 | teamforus/general | https://api.github.com/repos/teamforus/general | closed | Me app should show a euro symbol before each amount | Approval: Granted Phase 2: Design Scope: Small Status: Not Planned Type: Improvement Proposal | Learn more about change requests here: https://bit.ly/39CWeEE
### Requested by:
Jamal
### Change description
Right now we display numbers as:
5
it should be formatted nicer e.g.
€5,-
or
€5,00
or
€5,25
Please be aware. Here in the netherlands we **don't** use smaller numbers (superscript) to show decimals.
| 1.0 | Me app should show a euro symbol before each amount - Learn more about change requests here: https://bit.ly/39CWeEE
### Requested by:
Jamal
### Change description
Right now we display numbers as:
5
it should be formatted nicer e.g.
€5,-
or
€5,00
or
€5,25
Please be aware. Here in the netherlands we **don't** use smaller numbers (superscript) to show decimals.
| non_priority | me app should show a euro symbol before each amount learn more about change requests here requested by jamal change description right now we display numbers as it should be formatted nicer e g € or € or € please be aware here in the netherlands we don t use smaller numbers superscript to show decimals | 0 |
308,602 | 26,617,131,794 | IssuesEvent | 2023-01-24 08:18:41 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix creation_ops.test_torch_tensor | PyTorch Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3954630919/jobs/6772168731" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/3954630919/jobs/6772168731" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3954630919/jobs/6772168731" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3954630919/jobs/6772168731" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_creation_ops.py::test_torch_tensor[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-19T02:08:26.3351531Z E AssertionError: -128 != 127
2023-01-19T02:08:26.3351942Z E Falsifying example: test_torch_tensor(
2023-01-19T02:08:26.3352489Z E dtype_and_x=(['bfloat16'], [array(-129, dtype=bfloat16)]),
2023-01-19T02:08:26.3352972Z E dtype=['int8'],
2023-01-19T02:08:26.3353504Z E fn_tree='ivy.functional.frontends.torch.tensor',
2023-01-19T02:08:26.3354106Z E test_flags=num_positional_args=0. with_out=False. inplace=False. native_arrays=[False]. as_variable=[False]. ,
2023-01-19T02:08:26.3354640Z E on_device='cpu',
2023-01-19T02:08:26.3355041Z E frontend='torch',
2023-01-19T02:08:26.3355438Z E )
2023-01-19T02:08:26.3355726Z E
2023-01-19T02:08:26.3356462Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2AAAkYGCGiEMAADsACE') as a decorator on your test case
</details>
| 1.0 | Fix creation_ops.test_torch_tensor - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3954630919/jobs/6772168731" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/3954630919/jobs/6772168731" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3954630919/jobs/6772168731" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3954630919/jobs/6772168731" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_torch/test_creation_ops.py::test_torch_tensor[cpu-ivy.functional.backends.jax-False-False]</summary>
2023-01-19T02:08:26.3351531Z E AssertionError: -128 != 127
2023-01-19T02:08:26.3351942Z E Falsifying example: test_torch_tensor(
2023-01-19T02:08:26.3352489Z E dtype_and_x=(['bfloat16'], [array(-129, dtype=bfloat16)]),
2023-01-19T02:08:26.3352972Z E dtype=['int8'],
2023-01-19T02:08:26.3353504Z E fn_tree='ivy.functional.frontends.torch.tensor',
2023-01-19T02:08:26.3354106Z E test_flags=num_positional_args=0. with_out=False. inplace=False. native_arrays=[False]. as_variable=[False]. ,
2023-01-19T02:08:26.3354640Z E on_device='cpu',
2023-01-19T02:08:26.3355041Z E frontend='torch',
2023-01-19T02:08:26.3355438Z E )
2023-01-19T02:08:26.3355726Z E
2023-01-19T02:08:26.3356462Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2AAAkYGCGiEMAADsACE') as a decorator on your test case
</details>
| non_priority | fix creation ops test torch tensor tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test torch test creation ops py test torch tensor e assertionerror e falsifying example test torch tensor e dtype and x e dtype e fn tree ivy functional frontends torch tensor e test flags num positional args with out false inplace false native arrays as variable e on device cpu e frontend torch e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case | 0 |
200,234 | 15,094,923,121 | IssuesEvent | 2021-02-07 08:50:36 | Realm667/WolfenDoom | https://api.github.com/repos/Realm667/WolfenDoom | closed | Chapter 3 feedback | bug gameplay playtesting suggestion | Please enclose the screenshots you post into spoilers, so the issue does not become a monstrosity like #178.
---
The resolved issues have moved to 3rd post here!
---
General
---
- [x] No Cyrillic Chalkboard font usable for the Mission Summary's 'Phase' string. _[assigned to @Talon1024]_
> Added font fallback in 22fad8a2a9cf59cae4274160c7a1d291cb8980fa so that a different font gets used if the original font doesn't have all of the characters. Not a fix, but a workaround... Maybe good enough, unless @Talon1024 is going to be doing a lot more font work - AFA (16 Dec 20)
> The work with fonts is underway, and there is a separate issue, need not to have this checkbox anymore --N00b
C3INTRO
---
C3 INTERMAP
---
C3M1_A
---
- [x] SS guards just disappear after the explosion (which is also too small for the accompanying text), lowering the killcount. _[assigned to @Talon1024]_
C3M1_B
---
C3M0_A
---
C3M2
---
- [x] Due to nature of C3M2 (constantly spawning and de-spawning enemies) it may not be possible to achieve 100% during a normal playthrough. _[assigned to @AFADoomer @Talon1024]_
> I'm not sure how much of a problem this actually is... I haven't done a full playthrough, just checked specific spots that I thought could be problematic. But the enemies spawned by the never-ending spawners don't count against the total kill count, and the rest are relatively standard in their placement. - AFA (14 Jan 21)
> There is no such problem anymore. --N00b
- [x] In C3M2, sometimes the invasion script (with that "Kill X more Kraut(s)") stops at 1 remaining enemy (this is actually not a recent problem). I used iddt x2 but could not find the actor. But what's more entertaining is that the label kill count starts increasing later during the map _[assigned to @AFADoomer @Talon1024]_
> The only time I've seen this is when there was one enemy tucked away somewhere that I had missed... Though the count script should probably be stopped if you do continue playing without killing them all. - AFA (14 Jan 21)
<details><summary>Screenshot</summary>

</details>
> Likely irreproducible. --N00b
C3M3_A
---
More to come after next playthroughs!
C3M3_B
---
C3M0_B
---
- [x] The map does not end properly (see December videos). _[assigned to @Talon1024]_
> @Tormentor667 You said that you have fixed this problem (the player cannot return to INTERMAP, they have to reload a savegame), and also the ability to escape the vault before its door closes, but here are two videos from today's build in which they can be seen again: [c3m0_b.zip](https://github.com/Realm667/WolfenDoom/files/5782613/c3m0_b.zip). For the polyobject, I suggest using a Polyobj_OR_... special. --N00b, 7 Jan 2021
C3M4
---
???
C3M5_A
---
C3M5_B
---
C3M5_C
---
C3M6_A
---
C3M6_B
---
- [x] (Added by AFADoomer - 16 Jan) How is Ascher supposed to be handled in this map? Right now, he gets deactivated via script 46 when you cross the line at -5540, 644 (entry to the lab with Schabbs Jrs and Zombie Butcher), but the dialogue gives no indication as to why, so he just stands there stupidly and he never goes active again. I think he originally was deactivated via script 44 - the dialogue line there makes more sense - but I can't find anywhere in the map that calls that script. Rather than fully deactivating him, I can set him to properly stay in place (but still respond if shot or if he sees an enemy), but it would make more sense to do this later in the map, after the player opens the big all-keys door (and that wouldn't require changing the LANGUAGE strings). | 1.0 | Chapter 3 feedback - Please enclose the screenshots you post into spoilers, so the issue does not become a monstrosity like #178.
---
The resolved issues have moved to 3rd post here!
---
General
---
- [x] No Cyrillic Chalkboard font usable for the Mission Summary's 'Phase' string. _[assigned to @Talon1024]_
> Added font fallback in 22fad8a2a9cf59cae4274160c7a1d291cb8980fa so that a different font gets used if the original font doesn't have all of the characters. Not a fix, but a workaround... Maybe good enough, unless @Talon1024 is going to be doing a lot more font work - AFA (16 Dec 20)
> The work with fonts is underway, and there is a separate issue, need not to have this checkbox anymore --N00b
C3INTRO
---
C3 INTERMAP
---
C3M1_A
---
- [x] SS guards just disappear after the explosion (which is also too small for the accompanying text), lowering the killcount. _[assigned to @Talon1024]_
C3M1_B
---
C3M0_A
---
C3M2
---
- [x] Due to nature of C3M2 (constantly spawning and de-spawning enemies) it may not be possible to achieve 100% during a normal playthrough. _[assigned to @AFADoomer @Talon1024]_
> I'm not sure how much of a problem this actually is... I haven't done a full playthrough, just checked specific spots that I thought could be problematic. But the enemies spawned by the never-ending spawners don't count against the total kill count, and the rest are relatively standard in their placement. - AFA (14 Jan 21)
> There is no such problem anymore. --N00b
- [x] In C3M2, sometimes the invasion script (with that "Kill X more Kraut(s)") stops at 1 remaining enemy (this is actually not a recent problem). I used iddt x2 but could not find the actor. But what's more entertaining is that the label kill count starts increasing later during the map _[assigned to @AFADoomer @Talon1024]_
> The only time I've seen this is when there was one enemy tucked away somewhere that I had missed... Though the count script should probably be stopped if you do continue playing without killing them all. - AFA (14 Jan 21)
<details><summary>Screenshot</summary>

</details>
> Likely irreproducible. --N00b
C3M3_A
---
More to come after next playthroughs!
C3M3_B
---
C3M0_B
---
- [x] The map does not end properly (see December videos). _[assigned to @Talon1024]_
> @Tormentor667 You said that you have fixed this problem (the player cannot return to INTERMAP, they have to reload a savegame), and also the ability to escape the vault before its door closes, but here are two videos from today's build in which they can be seen again: [c3m0_b.zip](https://github.com/Realm667/WolfenDoom/files/5782613/c3m0_b.zip). For the polyobject, I suggest using a Polyobj_OR_... special. --N00b, 7 Jan 2021
C3M4
---
???
C3M5_A
---
C3M5_B
---
C3M5_C
---
C3M6_A
---
C3M6_B
---
- [x] (Added by AFADoomer - 16 Jan) How is Ascher supposed to be handled in this map? Right now, he gets deactivated via script 46 when you cross the line at -5540, 644 (entry to the lab with Schabbs Jrs and Zombie Butcher), but the dialogue gives no indication as to why, so he just stands there stupidly and he never goes active again. I think he originally was deactivated via script 44 - the dialogue line there makes more sense - but I can't find anywhere in the map that calls that script. Rather than fully deactivating him, I can set him to properly stay in place (but still respond if shot or if he sees an enemy), but it would make more sense to do this later in the map, after the player opens the big all-keys door (and that wouldn't require changing the LANGUAGE strings). | non_priority | chapter feedback please enclose the screenshots you post into spoilers so the issue does not become a monstrosity like the resolved issues have moved to post here general no cyrillic chalkboard font usable for the mission summary s phase string added font fallback in so that a different font gets used if the original font doesn t have all of the characters not a fix but a workaround maybe good enough unless is going to be doing a lot more font work afa dec the work with fonts is underway and there is a separate issue need not to have this checkbox anymore intermap a ss guards just disappear after the explosion which is also too small for the accompanying text lowering the killcount b a due to nature of constantly spawning and de spawning enemies it may not be possible to achieve during a normal playthrough i m not sure how much of a problem this actually is i haven t done a full playthrough just checked specific spots that i thought could be problematic but the enemies spawned by the never ending spawners don t count against the total kill count and the rest are relatively standard in their placement afa jan there is no such problem anymore in sometimes the invasion script with that kill x more kraut s stops at remaining enemy this is actually not a recent problem i used iddt but could not find the actor but what s more entertaining is that the label kill count starts increasing later during the map the only time i ve seen this is when there was one enemy tucked away somewhere that i had missed though the count script should probably be stopped if you do continue playing without killing them all afa jan screenshot likely irreproducible a more to come after next playthroughs b b the map does not end properly see december videos you said that you have fixed this problem the player cannot return to intermap they have to reload a savegame and also the ability to escape the vault before its door closes but here are two videos from today s build in which they can be seen again for the polyobject i suggest using a polyobj or special jan a b c a b added by afadoomer jan how is ascher supposed to be handled in this map right now he gets deactivated via script when you cross the line at entry to the lab with schabbs jrs and zombie butcher but the dialogue gives no indication as to why so he just stands there stupidly and he never goes active again i think he originally was deactivated via script the dialogue line there makes more sense but i can t find anywhere in the map that calls that script rather than fully deactivating him i can set him to properly stay in place but still respond if shot or if he sees an enemy but it would make more sense to do this later in the map after the player opens the big all keys door and that wouldn t require changing the language strings | 0 |
184,317 | 14,287,491,454 | IssuesEvent | 2020-11-23 16:25:49 | NationalSecurityAgency/skills-service | https://api.github.com/repos/NationalSecurityAgency/skills-service | closed | Do not allow enabling badges without any skills | bug good first issue test | Badges should not be allowed to be enabled if they do not contain any skills. Likewise, once a badge is enabled, you should not be allowed to remove the last skill from the badge. | 1.0 | Do not allow enabling badges without any skills - Badges should not be allowed to be enabled if they do not contain any skills. Likewise, once a badge is enabled, you should not be allowed to remove the last skill from the badge. | non_priority | do not allow enabling badges without any skills badges should not be allowed to be enabled if they do not contain any skills likewise once a badge is enabled you should not be allowed to remove the last skill from the badge | 0 |
158,139 | 20,007,805,910 | IssuesEvent | 2022-02-01 00:22:25 | RG4421/openedr | https://api.github.com/repos/RG4421/openedr | reopened | CVE-2019-1549 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2019-1549 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
OpenSSL 1.1.1 introduced a rewritten random number generator (RNG). This was intended to include protection in the event of a fork() system call in order to ensure that the parent and child processes did not share the same RNG state. However this protection was not being used in the default case. A partial mitigation for this issue is that the output from a high precision timer is mixed into the RNG state so the likelihood of a parent and child process sharing state is significantly reduced. If an application already calls OPENSSL_init_crypto() explicitly using OPENSSL_INIT_ATFORK then this problem does not occur at all. Fixed in OpenSSL 1.1.1d (Affected 1.1.1-1.1.1c).
<p>Publish Date: 2019-09-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1549>CVE-2019-1549</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openssl.org/news/secadv/20190910.txt">https://www.openssl.org/news/secadv/20190910.txt</a></p>
<p>Release Date: 2019-09-10</p>
<p>Fix Resolution: 1.1.1d</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-1549","vulnerabilityDetails":"OpenSSL 1.1.1 introduced a rewritten random number generator (RNG). This was intended to include protection in the event of a fork() system call in order to ensure that the parent and child processes did not share the same RNG state. However this protection was not being used in the default case. A partial mitigation for this issue is that the output from a high precision timer is mixed into the RNG state so the likelihood of a parent and child process sharing state is significantly reduced. If an application already calls OPENSSL_init_crypto() explicitly using OPENSSL_INIT_ATFORK then this problem does not occur at all. Fixed in OpenSSL 1.1.1d (Affected 1.1.1-1.1.1c).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1549","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-1549 (Medium) detected in multiple libraries - ## CVE-2019-1549 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b>, <b>opensslOpenSSL_1_1_1b</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
OpenSSL 1.1.1 introduced a rewritten random number generator (RNG). This was intended to include protection in the event of a fork() system call in order to ensure that the parent and child processes did not share the same RNG state. However this protection was not being used in the default case. A partial mitigation for this issue is that the output from a high precision timer is mixed into the RNG state so the likelihood of a parent and child process sharing state is significantly reduced. If an application already calls OPENSSL_init_crypto() explicitly using OPENSSL_INIT_ATFORK then this problem does not occur at all. Fixed in OpenSSL 1.1.1d (Affected 1.1.1-1.1.1c).
<p>Publish Date: 2019-09-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1549>CVE-2019-1549</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openssl.org/news/secadv/20190910.txt">https://www.openssl.org/news/secadv/20190910.txt</a></p>
<p>Release Date: 2019-09-10</p>
<p>Fix Resolution: 1.1.1d</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-1549","vulnerabilityDetails":"OpenSSL 1.1.1 introduced a rewritten random number generator (RNG). This was intended to include protection in the event of a fork() system call in order to ensure that the parent and child processes did not share the same RNG state. However this protection was not being used in the default case. A partial mitigation for this issue is that the output from a high precision timer is mixed into the RNG state so the likelihood of a parent and child process sharing state is significantly reduced. If an application already calls OPENSSL_init_crypto() explicitly using OPENSSL_INIT_ATFORK then this problem does not occur at all. Fixed in OpenSSL 1.1.1d (Affected 1.1.1-1.1.1c).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1549","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl vulnerability details openssl introduced a rewritten random number generator rng this was intended to include protection in the event of a fork system call in order to ensure that the parent and child processes did not share the same rng state however this protection was not being used in the default case a partial mitigation for this issue is that the output from a high precision timer is mixed into the rng state so the likelihood of a parent and child process sharing state is significantly reduced if an application already calls openssl init crypto explicitly using openssl init atfork then this problem does not occur at all fixed in openssl affected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages basebranches vulnerabilityidentifier cve vulnerabilitydetails openssl introduced a rewritten random number generator rng this was intended to include protection in the event of a fork system call in order to ensure that the parent and child processes did not share the same rng state however this protection was not being used in the default case a partial mitigation for this issue is that the output from a high precision timer is mixed into the rng state so the likelihood of a parent and child process sharing state is significantly reduced if an application already calls openssl init crypto explicitly using openssl init atfork then this problem does not occur at all fixed in openssl affected vulnerabilityurl | 0 |
10,533 | 8,616,091,449 | IssuesEvent | 2018-11-19 22:42:10 | regal/regal | https://api.github.com/repos/regal/regal | closed | Research bundling solution | infrastructure research | ## Overview
Bundling JavaScript packages is an important problem for which there doesn't seem to be a single clear answer. Regarding the Regal Framework, there are two immediate and distinct ways bundling will be used:
1. To distribute the Regal Game Library itself so it can be used by game developers to make Regal games.
2. To package Regal games via a bundling service into a self-contained artifact that can be run on multiple platforms via a common API.
These two functions will be implemented separately, but the concepts behind implementing them should be largely the same. The plan for implementing function one (bundling the Regal Game Library) is described at the bottom of this issue, and function two (the game-bundling script) is included in issue #41.
## Background
A non-trivial amount of research went into deciding on a bundling solution. To preserve that effort, the notes I took are listed below. I don't have a ton of experience in bundling JS packages, so I'm basing much of my decision-making off of what the successful libraries are doing.
### Bundling Options
* [Webpack](https://webpack.js.org/)
* [Rollup](https://rollupjs.org/guide/en)
* [Parcel](https://parceljs.org/)
*Helpful link for comparing the three can be found [here](https://medium.com/js-imaginea/comparing-bundlers-webpack-rollup-parcel-f8f5dc609cfd)
### Distribution Structures used by Popular Libraries
* Angular
* Appears to use Webpack
* Has multiple packages that makeup components of the app
* Appears that there is a *test* and a normal (production) variant of the source
* Each package has separate directories for each platform
* `bundles`
* Contains bundled source files, including all comments
* These are the `umd` bundles
* For source types `<pkgname>` and `<pkgname>-testing`, there are the following files:
* `<srctype>.umd.js`
* `<srctype>.umd.js.map`
* `<srctype>.umd.min.js`
* `<srctype>.umd.min.js.map`
* `esm5`
* Has the following structure:
* `src`
* `testing`
* `core.js`
* `index.js`
* `public_api.js`
* Each file has a sourcemap string at the end of it
* It looks like this source maintains its original directory structure
* `esm2015`
* Similar to `esm5`, with an extra file `core.externs.js` that doesn't seem too relevant
* `fesm5`
* Bundled, but not minified. Contains the following files for `<core>` and `<testing>`:
* `<srctype>.js`
* `<srctype>.js.map`
* `fesm2015`
* Same structure as `fesm5`
* `src`
* Types for non-test source, seems to be following original directory structure
* `testing`
* Same as `src`, but for test environment
* Top level of the package directory also contains `core.d.ts`, `index.d.ts`, `testing.d.ts`, and a few other files that are all pretty intuitive given the structure seen already
* Each package has fields in the `package.json` that correspond to each build platform (`es2015`, `esm2015`, `esm5`, etc.)
* See explanation [here](https://github.com/angular/angular/issues/21260).
* Apparently these are used by webpack and the cli
* Notably, `main` points to `./bundles/core.umd.js` and `module` points to `./fesm5/core.js`.
* Redux
* Uses rollup
* Has the following chunk in `package.json` (annotated by a [blog post](https://babeljs.io/blog/2018/06/26/on-consuming-and-publishing-es2015+-packages) on babel):
```
// redux package.json
{
...
"main": "lib/redux.js", // ES5 + Common JS
"module": "es/redux.js", // ES5 + JS Modules
}
```
* Looking at the [rollup file](https://github.com/reduxjs/redux/blob/master/rollup.config.js), they also export:
* A compressed `es`-format file to `es/redux.mjs` with the comment "ES for Browsers"
* A bundled `umd` file (with `process.env.NODE_ENV` replaced with `development`) to `dist/redux.js` and the comment "UMD Development"
* A bundled `umd` file like above except for `production` exported to `dist/redux.min.js`
* The `package.json` file also includes `"unpkg": "dist/redux.js"`. Unpkg is a CDN.
* Has an `index.d.ts` file, but it looks like this was made manually
* React
* Uses some form of rollup
* Build directory structure (all of the following files are bundled, and some minified):
* `cjs`
* `react.development.js`
* `react.production.min.js`
* `umd`
* `react.development.js`
* `react.production.min.js`
* `react.profiling.min.js`
* `package.json` only has `main`, which points to `index.js` (which itself points to the `cjs` directory)
* Rollup
* Has `dist` with the following files:
* `rollup.browser.js` - looks like minified UMD
* `rollup.d.ts`
* `rollup.es.js` - bundled but not minified
* `rollup.js`- CJS bundled but not minified
* `package.json` has the following properties:
* `"main": "dist/rollup.js"`
* `"module": "dist/rollup.es.js"`
* `"jsnext:main": "dist/rollup.es.js"`
* `rollup.config.js` also has a command for generating `dist/rollup.browser.mjs` in `es` instead of `umd`.
* Vue
* Uses Rollup
* Exports to `umd`, `cjs`, and `esm`. [See explanation here.](https://github.com/vuejs/vue/tree/dev/dist)
### Output Formats
* `umd`
* `cjs`
* `esm`/`mjs`
* Rollup's justification for using `esm` in conjunction with `pkg.module` can be found [here](https://github.com/rollup/rollup/wiki/pkg.module).
* `amd`
### Useful Links
https://webpack.js.org/guides/typescript/
https://webpack.js.org/guides/author-libraries/
https://github.com/rollup/rollup-plugin-typescript
https://rollupjs.org/guide/en#faqs
## Description
**Requirement 1. Distribution Structure**
Based on the research above, it seems the optimal way to distribute the Regal Game Library is in the following structure (note that all of these files will be inside one `dist` folder):
* `regal.umd.min.js`: Bundled and minified `umd` format to run in the browser.<sup>1</sup>
* `regal.cjs.js`: Bundled CommonJS format to use with older module loaders.
* `regal.esm.js`: Bundled ES format to use with newer module loaders (especially desired by Rollup).<sup>2</sup>
* `types`: Directory containing all `*.d.ts` files, in their original project structure.<sup>3</sup>
<sup>1</sup>*Note that `regal-bundler` (#41) will not be able to bundle projects that use Regal as a `umd` dependency, but this format is included to allow maximum portability of the library.*
<sup>2</sup>*While generating an experimental `.mjs` file was considered, that doesn't seem to be a priority at this point. Once the format becomes more standard, it will be easy to add-on in the future.*
<sup>3</sup>*At the moment, there doesn't seem to be a way to bundle declaration files intuitively.*
**Requirement 2. Automatic Builds**
The distribution directory should be generated simply by running `npm run build`.
**Requirement 3. Comment Management**
This isn't a strict requirement, but ideally the file-level comments from each of the source files will be removed (as they have a lot of repeated information), and a single comment with a description of the library, copyright information, and the license will be placed on the top of the bundle.
### Bundler Choice
Based on everything above, I'm deciding to go with Rollup as the bundling tool for the Regal Game Library. If circumstances change in the future, other options may be reevaluated. | 1.0 | Research bundling solution - ## Overview
Bundling JavaScript packages is an important problem for which there doesn't seem to be a single clear answer. Regarding the Regal Framework, there are two immediate and distinct ways bundling will be used:
1. To distribute the Regal Game Library itself so it can be used by game developers to make Regal games.
2. To package Regal games via a bundling service into a self-contained artifact that can be run on multiple platforms via a common API.
These two functions will be implemented separately, but the concepts behind implementing them should be largely the same. The plan for implementing function one (bundling the Regal Game Library) is described at the bottom of this issue, and function two (the game-bundling script) is included in issue #41.
## Background
A non-trivial amount of research went into deciding on a bundling solution. To preserve that effort, the notes I took are listed below. I don't have a ton of experience in bundling JS packages, so I'm basing much of my decision-making off of what the successful libraries are doing.
### Bundling Options
* [Webpack](https://webpack.js.org/)
* [Rollup](https://rollupjs.org/guide/en)
* [Parcel](https://parceljs.org/)
*Helpful link for comparing the three can be found [here](https://medium.com/js-imaginea/comparing-bundlers-webpack-rollup-parcel-f8f5dc609cfd)
### Distribution Structures used by Popular Libraries
* Angular
* Appears to use Webpack
* Has multiple packages that makeup components of the app
* Appears that there is a *test* and a normal (production) variant of the source
* Each package has separate directories for each platform
* `bundles`
* Contains bundled source files, including all comments
* These are the `umd` bundles
* For source types `<pkgname>` and `<pkgname>-testing`, there are the following files:
* `<srctype>.umd.js`
* `<srctype>.umd.js.map`
* `<srctype>.umd.min.js`
* `<srctype>.umd.min.js.map`
* `esm5`
* Has the following structure:
* `src`
* `testing`
* `core.js`
* `index.js`
* `public_api.js`
* Each file has a sourcemap string at the end of it
* It looks like this source maintains its original directory structure
* `esm2015`
* Similar to `esm5`, with an extra file `core.externs.js` that doesn't seem too relevant
* `fesm5`
* Bundled, but not minified. Contains the following files for `<core>` and `<testing>`:
* `<srctype>.js`
* `<srctype>.js.map`
* `fesm2015`
* Same structure as `fesm5`
* `src`
* Types for non-test source, seems to be following original directory structure
* `testing`
* Same as `src`, but for test environment
* Top level of the package directory also contains `core.d.ts`, `index.d.ts`, `testing.d.ts`, and a few other files that are all pretty intuitive given the structure seen already
* Each package has fields in the `package.json` that correspond to each build platform (`es2015`, `esm2015`, `esm5`, etc.)
* See explanation [here](https://github.com/angular/angular/issues/21260).
* Apparently these are used by webpack and the cli
* Notably, `main` points to `./bundles/core.umd.js` and `module` points to `./fesm5/core.js`.
* Redux
* Uses rollup
* Has the following chunk in `package.json` (annotated by a [blog post](https://babeljs.io/blog/2018/06/26/on-consuming-and-publishing-es2015+-packages) on babel):
```
// redux package.json
{
...
"main": "lib/redux.js", // ES5 + Common JS
"module": "es/redux.js", // ES5 + JS Modules
}
```
* Looking at the [rollup file](https://github.com/reduxjs/redux/blob/master/rollup.config.js), they also export:
* A compressed `es`-format file to `es/redux.mjs` with the comment "ES for Browsers"
* A bundled `umd` file (with `process.env.NODE_ENV` replaced with `development`) to `dist/redux.js` and the comment "UMD Development"
* A bundled `umd` file like above except for `production` exported to `dist/redux.min.js`
* The `package.json` file also includes `"unpkg": "dist/redux.js"`. Unpkg is a CDN.
* Has an `index.d.ts` file, but it looks like this was made manually
* React
* Uses some form of rollup
* Build directory structure (all of the following files are bundled, and some minified):
* `cjs`
* `react.development.js`
* `react.production.min.js`
* `umd`
* `react.development.js`
* `react.production.min.js`
* `react.profiling.min.js`
* `package.json` only has `main`, which points to `index.js` (which itself points to the `cjs` directory)
* Rollup
* Has `dist` with the following files:
* `rollup.browser.js` - looks like minified UMD
* `rollup.d.ts`
* `rollup.es.js` - bundled but not minified
* `rollup.js`- CJS bundled but not minified
* `package.json` has the following properties:
* `"main": "dist/rollup.js"`
* `"module": "dist/rollup.es.js"`
* `"jsnext:main": "dist/rollup.es.js"`
* `rollup.config.js` also has a command for generating `dist/rollup.browser.mjs` in `es` instead of `umd`.
* Vue
* Uses Rollup
* Exports to `umd`, `cjs`, and `esm`. [See explanation here.](https://github.com/vuejs/vue/tree/dev/dist)
### Output Formats
* `umd`
* `cjs`
* `esm`/`mjs`
* Rollup's justification for using `esm` in conjunction with `pkg.module` can be found [here](https://github.com/rollup/rollup/wiki/pkg.module).
* `amd`
### Useful Links
https://webpack.js.org/guides/typescript/
https://webpack.js.org/guides/author-libraries/
https://github.com/rollup/rollup-plugin-typescript
https://rollupjs.org/guide/en#faqs
## Description
**Requirement 1. Distribution Structure**
Based on the research above, it seems the optimal way to distribute the Regal Game Library is in the following structure (note that all of these files will be inside one `dist` folder):
* `regal.umd.min.js`: Bundled and minified `umd` format to run in the browser.<sup>1</sup>
* `regal.cjs.js`: Bundled CommonJS format to use with older module loaders.
* `regal.esm.js`: Bundled ES format to use with newer module loaders (especially desired by Rollup).<sup>2</sup>
* `types`: Directory containing all `*.d.ts` files, in their original project structure.<sup>3</sup>
<sup>1</sup>*Note that `regal-bundler` (#41) will not be able to bundle projects that use Regal as a `umd` dependency, but this format is included to allow maximum portability of the library.*
<sup>2</sup>*While generating an experimental `.mjs` file was considered, that doesn't seem to be a priority at this point. Once the format becomes more standard, it will be easy to add-on in the future.*
<sup>3</sup>*At the moment, there doesn't seem to be a way to bundle declaration files intuitively.*
**Requirement 2. Automatic Builds**
The distribution directory should be generated simply by running `npm run build`.
**Requirement 3. Comment Management**
This isn't a strict requirement, but ideally the file-level comments from each of the source files will be removed (as they have a lot of repeated information), and a single comment with a description of the library, copyright information, and the license will be placed on the top of the bundle.
### Bundler Choice
Based on everything above, I'm deciding to go with Rollup as the bundling tool for the Regal Game Library. If circumstances change in the future, other options may be reevaluated. | non_priority | research bundling solution overview bundling javascript packages is an important problem for which there doesn t seem to be a single clear answer regarding the regal framework there are two immediate and distinct ways bundling will be used to distribute the regal game library itself so it can be used by game developers to make regal games to package regal games via a bundling service into a self contained artifact that can be run on multiple platforms via a common api these two functions will be implemented separately but the concepts behind implementing them should be largely the same the plan for implementing function one bundling the regal game library is described at the bottom of this issue and function two the game bundling script is included in issue background a non trivial amount of research went into deciding on a bundling solution to preserve that effort the notes i took are listed below i don t have a ton of experience in bundling js packages so i m basing much of my decision making off of what the successful libraries are doing bundling options helpful link for comparing the three can be found distribution structures used by popular libraries angular appears to use webpack has multiple packages that makeup components of the app appears that there is a test and a normal production variant of the source each package has separate directories for each platform bundles contains bundled source files including all comments these are the umd bundles for source types and testing there are the following files umd js umd js map umd min js umd min js map has the following structure src testing core js index js public api js each file has a sourcemap string at the end of it it looks like this source maintains its original directory structure similar to with an extra file core externs js that doesn t seem too relevant bundled but not minified contains the following files for and js js map same structure as src types for non test source seems to be following original directory structure testing same as src but for test environment top level of the package directory also contains core d ts index d ts testing d ts and a few other files that are all pretty intuitive given the structure seen already each package has fields in the package json that correspond to each build platform etc see explanation apparently these are used by webpack and the cli notably main points to bundles core umd js and module points to core js redux uses rollup has the following chunk in package json annotated by a on babel redux package json main lib redux js common js module es redux js js modules looking at the they also export a compressed es format file to es redux mjs with the comment es for browsers a bundled umd file with process env node env replaced with development to dist redux js and the comment umd development a bundled umd file like above except for production exported to dist redux min js the package json file also includes unpkg dist redux js unpkg is a cdn has an index d ts file but it looks like this was made manually react uses some form of rollup build directory structure all of the following files are bundled and some minified cjs react development js react production min js umd react development js react production min js react profiling min js package json only has main which points to index js which itself points to the cjs directory rollup has dist with the following files rollup browser js looks like minified umd rollup d ts rollup es js bundled but not minified rollup js cjs bundled but not minified package json has the following properties main dist rollup js module dist rollup es js jsnext main dist rollup es js rollup config js also has a command for generating dist rollup browser mjs in es instead of umd vue uses rollup exports to umd cjs and esm output formats umd cjs esm mjs rollup s justification for using esm in conjunction with pkg module can be found amd useful links description requirement distribution structure based on the research above it seems the optimal way to distribute the regal game library is in the following structure note that all of these files will be inside one dist folder regal umd min js bundled and minified umd format to run in the browser regal cjs js bundled commonjs format to use with older module loaders regal esm js bundled es format to use with newer module loaders especially desired by rollup types directory containing all d ts files in their original project structure note that regal bundler will not be able to bundle projects that use regal as a umd dependency but this format is included to allow maximum portability of the library while generating an experimental mjs file was considered that doesn t seem to be a priority at this point once the format becomes more standard it will be easy to add on in the future at the moment there doesn t seem to be a way to bundle declaration files intuitively requirement automatic builds the distribution directory should be generated simply by running npm run build requirement comment management this isn t a strict requirement but ideally the file level comments from each of the source files will be removed as they have a lot of repeated information and a single comment with a description of the library copyright information and the license will be placed on the top of the bundle bundler choice based on everything above i m deciding to go with rollup as the bundling tool for the regal game library if circumstances change in the future other options may be reevaluated | 0 |
58,441 | 24,447,911,595 | IssuesEvent | 2022-10-06 19:42:55 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | Potential user stories for the Platform Console | console-services Console-API | - [ ] As a user, I’d like to be able to authenticate myself so that I can view my apps. (Dependent on AMT team, probably not for us)
- [ ] As a user, I’d like to be able to unauthenticate myself (logout).( Dependent on AMT team, probably not for us)
- [ ] As a user, I’d like to create and deploy a new app using a Dockerfile. (need to convert to issue)
- [ ] As a user, I’d like to see what version/commit is deployed. (need to convert to issue)
- [ ] As a user, I’d like to trigger a manual deployment. (need to convert to issue)
- [ ] As a user, I’d like to choose which branch to deploy. (need to convert to issue)
- [ ] As a user, I’d like to view the environments my app is deployed in. (need to convert to issue)
- [ ] As a user, I’d like to link my app to a GitHub repository for various reasons (e.g. view CI status).(need to convert to issue)
- [ ] As a user, I’d like to see my app’s CI status from GitHub Actions. (need to convert to issue)
- [ ] As a user, I’d like to be warned if I attempt to deploy with a red CI status. (need to convert to issue)
- [ ] As a user, I’d like to see an audit log of Console events that have occurred on my application (e.g. deploys, env var changes). (need to convert to issue) | 1.0 | Potential user stories for the Platform Console - - [ ] As a user, I’d like to be able to authenticate myself so that I can view my apps. (Dependent on AMT team, probably not for us)
- [ ] As a user, I’d like to be able to unauthenticate myself (logout).( Dependent on AMT team, probably not for us)
- [ ] As a user, I’d like to create and deploy a new app using a Dockerfile. (need to convert to issue)
- [ ] As a user, I’d like to see what version/commit is deployed. (need to convert to issue)
- [ ] As a user, I’d like to trigger a manual deployment. (need to convert to issue)
- [ ] As a user, I’d like to choose which branch to deploy. (need to convert to issue)
- [ ] As a user, I’d like to view the environments my app is deployed in. (need to convert to issue)
- [ ] As a user, I’d like to link my app to a GitHub repository for various reasons (e.g. view CI status).(need to convert to issue)
- [ ] As a user, I’d like to see my app’s CI status from GitHub Actions. (need to convert to issue)
- [ ] As a user, I’d like to be warned if I attempt to deploy with a red CI status. (need to convert to issue)
- [ ] As a user, I’d like to see an audit log of Console events that have occurred on my application (e.g. deploys, env var changes). (need to convert to issue) | non_priority | potential user stories for the platform console as a user i’d like to be able to authenticate myself so that i can view my apps dependent on amt team probably not for us as a user i’d like to be able to unauthenticate myself logout dependent on amt team probably not for us as a user i’d like to create and deploy a new app using a dockerfile need to convert to issue as a user i’d like to see what version commit is deployed need to convert to issue as a user i’d like to trigger a manual deployment need to convert to issue as a user i’d like to choose which branch to deploy need to convert to issue as a user i’d like to view the environments my app is deployed in need to convert to issue as a user i’d like to link my app to a github repository for various reasons e g view ci status need to convert to issue as a user i’d like to see my app’s ci status from github actions need to convert to issue as a user i’d like to be warned if i attempt to deploy with a red ci status need to convert to issue as a user i’d like to see an audit log of console events that have occurred on my application e g deploys env var changes need to convert to issue | 0 |
5,979 | 21,782,621,182 | IssuesEvent | 2022-05-13 20:44:22 | bcgov/api-services-portal | https://api.github.com/repos/bcgov/api-services-portal | closed | API Test - Authorization Profiles end points | automation | 1. Get the user session token
1.1 authenticates Janis (api owner) to get the user session token
2. API Tests for Authorization Profiles
2.1 Prepare the Request Specification for the API
2.2 Put the resource and verify the success code in the response
2.3 Get the resource and verify the success code in the response
2.4 Compare the values in response (in Get call) against the values passed in the request (in Put call)
2.5 Delete the authorization profile
2.6 Verify that the authorization is deleted | 1.0 | API Test - Authorization Profiles end points - 1. Get the user session token
1.1 authenticates Janis (api owner) to get the user session token
2. API Tests for Authorization Profiles
2.1 Prepare the Request Specification for the API
2.2 Put the resource and verify the success code in the response
2.3 Get the resource and verify the success code in the response
2.4 Compare the values in response (in Get call) against the values passed in the request (in Put call)
2.5 Delete the authorization profile
2.6 Verify that the authorization is deleted | non_priority | api test authorization profiles end points get the user session token authenticates janis api owner to get the user session token api tests for authorization profiles prepare the request specification for the api put the resource and verify the success code in the response get the resource and verify the success code in the response compare the values in response in get call against the values passed in the request in put call delete the authorization profile verify that the authorization is deleted | 0 |
402,073 | 27,350,283,724 | IssuesEvent | 2023-02-27 09:05:40 | microsoft/cloud-hubs | https://api.github.com/repos/microsoft/cloud-hubs | closed | Update docs based on revised plans | Area: Documentation Status: 🔭 Spec review | <!--
⚠️⚠️⚠️ BEFORE YOU SUBMIT ⚠️⚠️⚠️
1. Confirm there isn't an issue already. If so, vote it up (👍) and add comments.
2. Complete all TODO items below and remove the TODO lines after.
3. Internal: Add applicable labels: Type, Micro PR, Area
-->
## ⚠️ Problem
In an effort to accelerate timelines, we are merging with https://github.com/MSBrett/ccm_lite for the v0.1 release.
## 🛠️ Solution
Update docs based on revised plans for the 0.1 release.
## 🙋♀️ Ask for the community
<!-- TODO: Customize this section to let the community know how they can help. -->
We could use your help:
1. Please vote this issue up (👍) to prioritize it.
2. Leave comments to help us solidify the vision.
| 1.0 | Update docs based on revised plans - <!--
⚠️⚠️⚠️ BEFORE YOU SUBMIT ⚠️⚠️⚠️
1. Confirm there isn't an issue already. If so, vote it up (👍) and add comments.
2. Complete all TODO items below and remove the TODO lines after.
3. Internal: Add applicable labels: Type, Micro PR, Area
-->
## ⚠️ Problem
In an effort to accelerate timelines, we are merging with https://github.com/MSBrett/ccm_lite for the v0.1 release.
## 🛠️ Solution
Update docs based on revised plans for the 0.1 release.
## 🙋♀️ Ask for the community
<!-- TODO: Customize this section to let the community know how they can help. -->
We could use your help:
1. Please vote this issue up (👍) to prioritize it.
2. Leave comments to help us solidify the vision.
| non_priority | update docs based on revised plans ⚠️⚠️⚠️ before you submit ⚠️⚠️⚠️ confirm there isn t an issue already if so vote it up 👍 and add comments complete all todo items below and remove the todo lines after internal add applicable labels type micro pr area ⚠️ problem in an effort to accelerate timelines we are merging with for the release 🛠️ solution update docs based on revised plans for the release 🙋♀️ ask for the community we could use your help please vote this issue up 👍 to prioritize it leave comments to help us solidify the vision | 0 |
238,509 | 26,128,347,206 | IssuesEvent | 2022-12-28 22:34:15 | pulumi/pulumi-aws | https://api.github.com/repos/pulumi/pulumi-aws | closed | Warn when a secret is captured by a aws.lambda.CallbackFunction | kind/enhancement impact/security | In https://github.com/pulumi/pulumi-aws/pull/1282 we added the ability for CallbackFunctions which captured a secret value to turn the entire captured function body into a Pulumi secret so that the sensitive data would be encrypted in Pulumi state.
However, the sensitive values in this case would still be captured into the text of the Lambda function visibile inside AWS. Although this is standard and expected behaviour for Pulumi secrets generally, it is also an easy way to "accidentally" leak a secret into a context that may not be trusted.
We should consider warning when this happens, and using the opportunity to suggest one of the alternatives, ideally SecretsManager, but even environment variables would be "better". | True | Warn when a secret is captured by a aws.lambda.CallbackFunction - In https://github.com/pulumi/pulumi-aws/pull/1282 we added the ability for CallbackFunctions which captured a secret value to turn the entire captured function body into a Pulumi secret so that the sensitive data would be encrypted in Pulumi state.
However, the sensitive values in this case would still be captured into the text of the Lambda function visibile inside AWS. Although this is standard and expected behaviour for Pulumi secrets generally, it is also an easy way to "accidentally" leak a secret into a context that may not be trusted.
We should consider warning when this happens, and using the opportunity to suggest one of the alternatives, ideally SecretsManager, but even environment variables would be "better". | non_priority | warn when a secret is captured by a aws lambda callbackfunction in we added the ability for callbackfunctions which captured a secret value to turn the entire captured function body into a pulumi secret so that the sensitive data would be encrypted in pulumi state however the sensitive values in this case would still be captured into the text of the lambda function visibile inside aws although this is standard and expected behaviour for pulumi secrets generally it is also an easy way to accidentally leak a secret into a context that may not be trusted we should consider warning when this happens and using the opportunity to suggest one of the alternatives ideally secretsmanager but even environment variables would be better | 0 |
359,739 | 25,253,827,813 | IssuesEvent | 2022-11-15 16:29:22 | hyter99/AI_22-23_L1 | https://api.github.com/repos/hyter99/AI_22-23_L1 | closed | [OT] Github Wiki | documentation | So far I've created 2 pages in wiki:
- how to set up local environment
- workflow.
What else do you think is needed? | 1.0 | [OT] Github Wiki - So far I've created 2 pages in wiki:
- how to set up local environment
- workflow.
What else do you think is needed? | non_priority | github wiki so far i ve created pages in wiki how to set up local environment workflow what else do you think is needed | 0 |
223,712 | 17,618,339,003 | IssuesEvent | 2021-08-18 12:37:59 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: schemachange/during/tpcc failed | C-test-failure O-robot O-roachtest branch-master release-blocker | roachtest.schemachange/during/tpcc [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3317393&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3317393&tab=artifacts#/schemachange/during/tpcc) on master @ [472008d077a4282b7a5b0557ee6b6b5c12f75586](https://github.com/cockroachdb/cockroach/commits/472008d077a4282b7a5b0557ee6b6b5c12f75586):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/schemachange/during/tpcc/run_1
monitor.go:128,tpcc.go:279,schemachange.go:429,test_runner.go:777: monitor failure: unexpected node event: 2: dead (exit status 137)
(1) attached stack trace
-- stack trace:
| main.(*monitorImpl).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/monitor.go:116
| main.(*monitorImpl).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/monitor.go:124
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runTPCC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:279
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.makeSchemaChangeDuringTPCC.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/schemachange.go:429
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:777
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1371
Wraps: (2) monitor failure
Wraps: (3) unexpected node event: 2: dead (exit status 137)
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString
cluster.go:1249,context.go:89,cluster.go:1237,test_runner.go:866: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-3317393-1629267491-66-n5cpu16 --oneshot --ignore-empty-nodes: exit status 1 5: skipped
2: dead (exit status 137)
4: 11681
1: 12120
3: 11716
Error: UNCLASSIFIED_PROBLEM: 2: dead (exit status 137)
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
-- stack trace:
| main.glob..func14
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1173
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:281
| github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:856
| github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:960
| github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:897
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:2107
| runtime.main
| /usr/local/go/src/runtime/proc.go:225
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1371
Wraps: (3) 2: dead (exit status 137)
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/roachtest)
See: [CI job to stress roachtests](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestStress)
<p>For the CI stress job, click the ellipsis (...) next to the Run button and fill in:
* Changes / Build branch: master
* Parameters / `env.TESTS`: `^schemachange/during/tpcc$`
* Parameters / `env.COUNT`: <number of runs>
</p>
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #54316 roachtest: schemachange/during/tpcc failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-20.2]
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/during/tpcc.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: schemachange/during/tpcc failed - roachtest.schemachange/during/tpcc [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3317393&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3317393&tab=artifacts#/schemachange/during/tpcc) on master @ [472008d077a4282b7a5b0557ee6b6b5c12f75586](https://github.com/cockroachdb/cockroach/commits/472008d077a4282b7a5b0557ee6b6b5c12f75586):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/schemachange/during/tpcc/run_1
monitor.go:128,tpcc.go:279,schemachange.go:429,test_runner.go:777: monitor failure: unexpected node event: 2: dead (exit status 137)
(1) attached stack trace
-- stack trace:
| main.(*monitorImpl).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/monitor.go:116
| main.(*monitorImpl).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/monitor.go:124
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runTPCC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/tpcc.go:279
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.makeSchemaChangeDuringTPCC.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/schemachange.go:429
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:777
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1371
Wraps: (2) monitor failure
Wraps: (3) unexpected node event: 2: dead (exit status 137)
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString
cluster.go:1249,context.go:89,cluster.go:1237,test_runner.go:866: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-3317393-1629267491-66-n5cpu16 --oneshot --ignore-empty-nodes: exit status 1 5: skipped
2: dead (exit status 137)
4: 11681
1: 12120
3: 11716
Error: UNCLASSIFIED_PROBLEM: 2: dead (exit status 137)
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
-- stack trace:
| main.glob..func14
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1173
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:281
| github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:856
| github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:960
| github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:897
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:2107
| runtime.main
| /usr/local/go/src/runtime/proc.go:225
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1371
Wraps: (3) 2: dead (exit status 137)
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/roachtest)
See: [CI job to stress roachtests](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestStress)
<p>For the CI stress job, click the ellipsis (...) next to the Run button and fill in:
* Changes / Build branch: master
* Parameters / `env.TESTS`: `^schemachange/during/tpcc$`
* Parameters / `env.COUNT`: <number of runs>
</p>
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #54316 roachtest: schemachange/during/tpcc failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-20.2]
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/during/tpcc.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_priority | roachtest schemachange during tpcc failed roachtest schemachange during tpcc with on master the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts schemachange during tpcc run monitor go tpcc go schemachange go test runner go monitor failure unexpected node event dead exit status attached stack trace stack trace main monitorimpl waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest monitor go main monitorimpl wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest monitor go github com cockroachdb cockroach pkg cmd roachtest tests runtpcc home agent work go src github com cockroachdb cockroach pkg cmd roachtest tests tpcc go github com cockroachdb cockroach pkg cmd roachtest tests makeschemachangeduringtpcc home agent work go src github com cockroachdb cockroach pkg cmd roachtest tests schemachange go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go runtime goexit usr local go src runtime asm s wraps monitor failure wraps unexpected node event dead exit status error types withstack withstack errutil withprefix errors errorstring cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status skipped dead exit status error unclassified problem dead exit status unclassified problem wraps attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead exit status error types errors unclassified withstack withstack errutil leaferror reproduce see see for the ci stress job click the ellipsis next to the run button and fill in changes build branch master parameters env tests schemachange during tpcc parameters env count lt number of runs gt same failure on other branches roachtest schemachange during tpcc failed cc cockroachdb sql schema | 0 |
116,190 | 14,929,580,772 | IssuesEvent | 2021-01-25 00:00:04 | microsoft/fluentui | https://api.github.com/repos/microsoft/fluentui | closed | modifying 'height' of dropdown. | Component: Dropdown Resolution: By Design Type: Question ❔ |
### Describe the issue:
On changing 'height' of the dropdown control, the selected item/option is not center-aligned.
#### Actual behavior:

#### Expected behavior:
Selected item should be center aligned as it happens in searchbox control.

| 1.0 | modifying 'height' of dropdown. -
### Describe the issue:
On changing 'height' of the dropdown control, the selected item/option is not center-aligned.
#### Actual behavior:

#### Expected behavior:
Selected item should be center aligned as it happens in searchbox control.

| non_priority | modifying height of dropdown describe the issue on changing height of the dropdown control the selected item option is not center aligned actual behavior expected behavior selected item should be center aligned as it happens in searchbox control | 0 |
194,077 | 14,668,796,976 | IssuesEvent | 2020-12-29 22:19:28 | github-vet/rangeloop-pointer-findings | https://api.github.com/repos/github-vet/rangeloop-pointer-findings | closed | Setheck/oba: utils_test.go; 3 LoC | fresh test tiny |
Found a possible issue in [Setheck/oba](https://www.github.com/Setheck/oba) at [utils_test.go](https://github.com/Setheck/oba/blob/7e1cf15d04397a3842153c16a7e4b770ec47487e/utils_test.go#L182-L184)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> function call at line 183 passes reference to r to third-party code
[Click here to see the code in its original context.](https://github.com/Setheck/oba/blob/7e1cf15d04397a3842153c16a7e4b770ec47487e/utils_test.go#L182-L184)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, r := range s.Routes {
VerifyRoute(t, &r)
}
```
</details>
<details>
<summary>Click here to show extra information the analyzer produced.</summary>
```
No path was found through the callgraph that could lead to a function which writes a pointer argument.
The following graphviz dot graph describes paths through the callgraph that could lead to a function which passes a pointer to third-party code:
digraph G {
"(VerifyRoute, 2)" -> {}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 7e1cf15d04397a3842153c16a7e4b770ec47487e
| 1.0 | Setheck/oba: utils_test.go; 3 LoC -
Found a possible issue in [Setheck/oba](https://www.github.com/Setheck/oba) at [utils_test.go](https://github.com/Setheck/oba/blob/7e1cf15d04397a3842153c16a7e4b770ec47487e/utils_test.go#L182-L184)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message.
> function call at line 183 passes reference to r to third-party code
[Click here to see the code in its original context.](https://github.com/Setheck/oba/blob/7e1cf15d04397a3842153c16a7e4b770ec47487e/utils_test.go#L182-L184)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, r := range s.Routes {
VerifyRoute(t, &r)
}
```
</details>
<details>
<summary>Click here to show extra information the analyzer produced.</summary>
```
No path was found through the callgraph that could lead to a function which writes a pointer argument.
The following graphviz dot graph describes paths through the callgraph that could lead to a function which passes a pointer to third-party code:
digraph G {
"(VerifyRoute, 2)" -> {}
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 7e1cf15d04397a3842153c16a7e4b770ec47487e
| non_priority | setheck oba utils test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call at line passes reference to r to third party code click here to show the line s of go which triggered the analyzer go for r range s routes verifyroute t r click here to show extra information the analyzer produced no path was found through the callgraph that could lead to a function which writes a pointer argument the following graphviz dot graph describes paths through the callgraph that could lead to a function which passes a pointer to third party code digraph g verifyroute leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id | 0 |
52,612 | 27,663,328,619 | IssuesEvent | 2023-03-12 19:22:29 | MoleMan1024/audiowagon | https://api.github.com/repos/MoleMan1024/audiowagon | opened | Indexing takes long time with very many directories and no changes in filesystem | performance | Indexing many directories without any changes in filesystem is quite slow, see if this can be made faster.
See https://github.com/MoleMan1024/audiowagon/issues/108 | True | Indexing takes long time with very many directories and no changes in filesystem - Indexing many directories without any changes in filesystem is quite slow, see if this can be made faster.
See https://github.com/MoleMan1024/audiowagon/issues/108 | non_priority | indexing takes long time with very many directories and no changes in filesystem indexing many directories without any changes in filesystem is quite slow see if this can be made faster see | 0 |
18,510 | 25,717,944,344 | IssuesEvent | 2022-12-07 11:41:16 | grafana/xk6-browser | https://api.github.com/repos/grafana/xk6-browser | closed | xk6-browser should be importable as a k6 module | k6-core-compatibility | There is currently no way to import the xk6-browser and assign it a custom module name without also setting `k6/x/browser`.
To do that the [RootModule](https://github.com/grafana/xk6-browser/blob/main/main.go#L52) will need to be in package different from the "main" one and the main one to continue importing it.
You can see a very simplified example in https://github.com/grafana/xk6-timers repo for example. | True | xk6-browser should be importable as a k6 module - There is currently no way to import the xk6-browser and assign it a custom module name without also setting `k6/x/browser`.
To do that the [RootModule](https://github.com/grafana/xk6-browser/blob/main/main.go#L52) will need to be in package different from the "main" one and the main one to continue importing it.
You can see a very simplified example in https://github.com/grafana/xk6-timers repo for example. | non_priority | browser should be importable as a module there is currently no way to import the browser and assign it a custom module name without also setting x browser to do that the will need to be in package different from the main one and the main one to continue importing it you can see a very simplified example in repo for example | 0 |
375,389 | 26,160,860,896 | IssuesEvent | 2022-12-31 13:58:33 | BalticAmadeus/ProBro | https://api.github.com/repos/BalticAmadeus/ProBro | opened | Analyze possibilities to integrate with vscode-abl | documentation | Some ideas:
1. Clicking on db table buffer could open default query tab for that table.
2. Clicking on db table buffer could open table fields window.
3. Clikcing on oe query could open query tab with results.
... | 1.0 | Analyze possibilities to integrate with vscode-abl - Some ideas:
1. Clicking on db table buffer could open default query tab for that table.
2. Clicking on db table buffer could open table fields window.
3. Clikcing on oe query could open query tab with results.
... | non_priority | analyze possibilities to integrate with vscode abl some ideas clicking on db table buffer could open default query tab for that table clicking on db table buffer could open table fields window clikcing on oe query could open query tab with results | 0 |
133,418 | 12,540,865,751 | IssuesEvent | 2020-06-05 11:11:57 | ktorio/ktor | https://api.github.com/repos/ktorio/ktor | closed | TCP Socket File Descriptor Leak | documentation | ### Ktor Version
1.1.5
### Ktor Engine Used(client or server and name)
raw sockets
### JVM Version, Operating System and Relevant Context
macOS, linux
openjdk version "1.8.0_212"
OpenJDK Runtime Environment (AdoptOpenJDK)(build 1.8.0_212-b03)
OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.212-b03, mixed mode)
### Feedback
It seems like client tcp socket connections are not properly closed and leaking file descriptors. When testing the server with telnet no leak occurs. The issue is within the client code. The server part is only included for a complete, reproducable example.
Do the read and write channel need separate closing or is something else not handled correctly in the example?
The output is:
```
Server said: 'hello 0'
open file descriptors: 64
Server said: 'hello 1'
open file descriptors: 67
Server said: 'hello 2'
open file descriptors: 71
```
Example:
```kotlin
fun main() = runBlocking {
val osMxBean = ManagementFactory.getOperatingSystemMXBean() as UnixOperatingSystemMXBean
val address = InetSocketAddress("127.0.0.1", 2323)
val connectJob = launch {
// repeatedly connect to the server
repeat(3) {
delay(1_000)
aSocket(ActorSelectorManager(Dispatchers.IO)).tcp().connect(address).use { socket ->
val input = socket.openReadChannel()
val output = socket.openWriteChannel(autoFlush = true)
output.writeStringUtf8("hello $it\r\n")
val response = input.readUTF8Line()
println("Server said: '$response'")
}
println("open file descriptors: ${osMxBean.openFileDescriptorCount}")
}
}
// server - irrelevant for the issue
val serverJob = launch {
aSocket(ActorSelectorManager(Dispatchers.IO)).tcp().bind(address).use { server ->
while (true) {
val socket = server.accept()
launch {
socket.use {
val input = socket.openReadChannel()
val output = socket.openWriteChannel(autoFlush = true)
while (true) {
val line = input.readUTF8Line()
output.writeStringUtf8("$line\r\n")
}
}
}
}
}
}
connectJob.join()
serverJob.cancel()
}
``` | 1.0 | TCP Socket File Descriptor Leak - ### Ktor Version
1.1.5
### Ktor Engine Used(client or server and name)
raw sockets
### JVM Version, Operating System and Relevant Context
macOS, linux
openjdk version "1.8.0_212"
OpenJDK Runtime Environment (AdoptOpenJDK)(build 1.8.0_212-b03)
OpenJDK 64-Bit Server VM (AdoptOpenJDK)(build 25.212-b03, mixed mode)
### Feedback
It seems like client tcp socket connections are not properly closed and leaking file descriptors. When testing the server with telnet no leak occurs. The issue is within the client code. The server part is only included for a complete, reproducable example.
Do the read and write channel need separate closing or is something else not handled correctly in the example?
The output is:
```
Server said: 'hello 0'
open file descriptors: 64
Server said: 'hello 1'
open file descriptors: 67
Server said: 'hello 2'
open file descriptors: 71
```
Example:
```kotlin
fun main() = runBlocking {
val osMxBean = ManagementFactory.getOperatingSystemMXBean() as UnixOperatingSystemMXBean
val address = InetSocketAddress("127.0.0.1", 2323)
val connectJob = launch {
// repeatedly connect to the server
repeat(3) {
delay(1_000)
aSocket(ActorSelectorManager(Dispatchers.IO)).tcp().connect(address).use { socket ->
val input = socket.openReadChannel()
val output = socket.openWriteChannel(autoFlush = true)
output.writeStringUtf8("hello $it\r\n")
val response = input.readUTF8Line()
println("Server said: '$response'")
}
println("open file descriptors: ${osMxBean.openFileDescriptorCount}")
}
}
// server - irrelevant for the issue
val serverJob = launch {
aSocket(ActorSelectorManager(Dispatchers.IO)).tcp().bind(address).use { server ->
while (true) {
val socket = server.accept()
launch {
socket.use {
val input = socket.openReadChannel()
val output = socket.openWriteChannel(autoFlush = true)
while (true) {
val line = input.readUTF8Line()
output.writeStringUtf8("$line\r\n")
}
}
}
}
}
}
connectJob.join()
serverJob.cancel()
}
``` | non_priority | tcp socket file descriptor leak ktor version ktor engine used client or server and name raw sockets jvm version operating system and relevant context macos linux openjdk version openjdk runtime environment adoptopenjdk build openjdk bit server vm adoptopenjdk build mixed mode feedback it seems like client tcp socket connections are not properly closed and leaking file descriptors when testing the server with telnet no leak occurs the issue is within the client code the server part is only included for a complete reproducable example do the read and write channel need separate closing or is something else not handled correctly in the example the output is server said hello open file descriptors server said hello open file descriptors server said hello open file descriptors example kotlin fun main runblocking val osmxbean managementfactory getoperatingsystemmxbean as unixoperatingsystemmxbean val address inetsocketaddress val connectjob launch repeatedly connect to the server repeat delay asocket actorselectormanager dispatchers io tcp connect address use socket val input socket openreadchannel val output socket openwritechannel autoflush true output hello it r n val response input println server said response println open file descriptors osmxbean openfiledescriptorcount server irrelevant for the issue val serverjob launch asocket actorselectormanager dispatchers io tcp bind address use server while true val socket server accept launch socket use val input socket openreadchannel val output socket openwritechannel autoflush true while true val line input output line r n connectjob join serverjob cancel | 0 |
170,185 | 26,912,558,782 | IssuesEvent | 2023-02-07 01:52:11 | patternfly/patternfly-org | https://api.github.com/repos/patternfly/patternfly-org | closed | Add guidelines to forms and modals about focusing on the first field | PF4 design Guidelines | We don't currently have any guidance on when we should (or if we should) focus on the first field within a form or form within a modal. There are also circumstances where maybe we should be focusing on the cancel button in a destructive action (as [the guidelines in cloud.redhat.com say to do](https://docs.google.com/document/d/1Q58potYqoKfHSjzKY4yW-oUv_ESckI-vsbFnKp0cz3s/edit#)). I _think_ this type of documentation belongs at the PF level rather than leaving it up to products.
Add this guidance to the **modal guidelines**, about which buttons to focus on, and when. | 1.0 | Add guidelines to forms and modals about focusing on the first field - We don't currently have any guidance on when we should (or if we should) focus on the first field within a form or form within a modal. There are also circumstances where maybe we should be focusing on the cancel button in a destructive action (as [the guidelines in cloud.redhat.com say to do](https://docs.google.com/document/d/1Q58potYqoKfHSjzKY4yW-oUv_ESckI-vsbFnKp0cz3s/edit#)). I _think_ this type of documentation belongs at the PF level rather than leaving it up to products.
Add this guidance to the **modal guidelines**, about which buttons to focus on, and when. | non_priority | add guidelines to forms and modals about focusing on the first field we don t currently have any guidance on when we should or if we should focus on the first field within a form or form within a modal there are also circumstances where maybe we should be focusing on the cancel button in a destructive action as i think this type of documentation belongs at the pf level rather than leaving it up to products add this guidance to the modal guidelines about which buttons to focus on and when | 0 |
31,168 | 7,326,884,305 | IssuesEvent | 2018-03-04 01:58:44 | OneiricArts/CustomizeNewTab | https://api.github.com/repos/OneiricArts/CustomizeNewTab | opened | ugrade material design lite (mdl) library? | code enhancement | not currently in use, but:
`Dependencies defined in source/libs/mdl/package.json`
security vulnerability due to `jQuery`
this has been updated since the release I've included (and kept, even though not in use)
either delete, or upgrade | 1.0 | ugrade material design lite (mdl) library? - not currently in use, but:
`Dependencies defined in source/libs/mdl/package.json`
security vulnerability due to `jQuery`
this has been updated since the release I've included (and kept, even though not in use)
either delete, or upgrade | non_priority | ugrade material design lite mdl library not currently in use but dependencies defined in source libs mdl package json security vulnerability due to jquery this has been updated since the release i ve included and kept even though not in use either delete or upgrade | 0 |
129,846 | 27,577,509,842 | IssuesEvent | 2023-03-08 14:04:59 | Regalis11/Barotrauma | https://api.github.com/repos/Regalis11/Barotrauma | closed | [1.0] Pets trigger "Hostiles in x!" command | Bug Need more info Code | ### Disclaimers
- [X] I have searched the issue tracker to check if the issue has already been reported.
- [ ] My issue happened while using mods.
### What happened?
The only creature in the room with me here is the Psilotoad. Despite this, the intruders quick command (accessible via the headset) is prompting me to use it. This implies that there is a hostile creature in the same hull as me, which is impossible as the only creature with me is the Psilotoad (a non-hostile creature unless the owner is attacked, which they were not).

### Reproduction steps
1. Hatch a pet
2. Observe how the "intruders" command is now flashing
### Bug prevalence
Happens every time I play
### Version
Other
### -
1.0.2.0
### Which operating system did you encounter this bug on?
Windows
### Relevant error messages and crash reports
_No response_ | 1.0 | [1.0] Pets trigger "Hostiles in x!" command - ### Disclaimers
- [X] I have searched the issue tracker to check if the issue has already been reported.
- [ ] My issue happened while using mods.
### What happened?
The only creature in the room with me here is the Psilotoad. Despite this, the intruders quick command (accessible via the headset) is prompting me to use it. This implies that there is a hostile creature in the same hull as me, which is impossible as the only creature with me is the Psilotoad (a non-hostile creature unless the owner is attacked, which they were not).

### Reproduction steps
1. Hatch a pet
2. Observe how the "intruders" command is now flashing
### Bug prevalence
Happens every time I play
### Version
Other
### -
1.0.2.0
### Which operating system did you encounter this bug on?
Windows
### Relevant error messages and crash reports
_No response_ | non_priority | pets trigger hostiles in x command disclaimers i have searched the issue tracker to check if the issue has already been reported my issue happened while using mods what happened the only creature in the room with me here is the psilotoad despite this the intruders quick command accessible via the headset is prompting me to use it this implies that there is a hostile creature in the same hull as me which is impossible as the only creature with me is the psilotoad a non hostile creature unless the owner is attacked which they were not reproduction steps hatch a pet observe how the intruders command is now flashing bug prevalence happens every time i play version other which operating system did you encounter this bug on windows relevant error messages and crash reports no response | 0 |
278,837 | 30,702,410,455 | IssuesEvent | 2023-07-27 01:27:49 | panasalap/linux-4.19.72_test | https://api.github.com/repos/panasalap/linux-4.19.72_test | closed | CVE-2020-24490 (Medium) detected in linux-yoctov5.4.51 - autoclosed | Mend: dependency security vulnerability | ## CVE-2020-24490 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/fd7d4a1ec69b1b0468bf77f825beb235061a108e">fd7d4a1ec69b1b0468bf77f825beb235061a108e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_event.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Improper buffer restrictions in BlueZ may allow an unauthenticated user to potentially enable denial of service via adjacent access. This affects all Linux kernel versions that support BlueZ.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-24490>CVE-2020-24490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-24490">https://www.linuxkernelcves.com/cves/CVE-2020-24490</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution: v4.4.240, v4.9.240, v4.14.202, v4.19.137, v5.4.56, v5.7.13, v5.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-24490 (Medium) detected in linux-yoctov5.4.51 - autoclosed - ## CVE-2020-24490 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/fd7d4a1ec69b1b0468bf77f825beb235061a108e">fd7d4a1ec69b1b0468bf77f825beb235061a108e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_event.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Improper buffer restrictions in BlueZ may allow an unauthenticated user to potentially enable denial of service via adjacent access. This affects all Linux kernel versions that support BlueZ.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-24490>CVE-2020-24490</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-24490">https://www.linuxkernelcves.com/cves/CVE-2020-24490</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution: v4.4.240, v4.9.240, v4.14.202, v4.19.137, v5.4.56, v5.7.13, v5.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux autoclosed cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files net bluetooth hci event c vulnerability details improper buffer restrictions in bluez may allow an unauthenticated user to potentially enable denial of service via adjacent access this affects all linux kernel versions that support bluez publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
201,701 | 15,218,729,935 | IssuesEvent | 2021-02-17 18:13:22 | WeiXian042901/fyp_repository | https://api.github.com/repos/WeiXian042901/fyp_repository | opened | FA_033_Create Quiz(Adding A Question Input) | Acceptance Test Admin Quiz | **Test Scenario**
- Administrator wishes to create the quiz after filling in the information for the 2nd question of the quiz
**Test Case**
- Administrator enters all the available field in the Quiz Information and the Quest Information Panel except FIQ awarded and Time
**Pre-Conditions**
- Administrator must have successfully logged into the application
- Administrator must have selected the “Quizzes” option
- Administrator must have clicked on the “Add Question” button
- Administrator must have clicked on the “Create New” Button
**Test-Steps**
1. Type “Testing Title(2nd question)” into the Quiz Title box
2. Select “Finance” as Category
3. Select “25” in the FIQ per question box
4. Select “15” in the Time per question box
5. Type “Test Description” in the Quiz Description box
6. Type in “Testing Question” in the Question box
7. Type in “Correct Answer” in Option 1 box
8. Type in “Wrong Answer” in Option 2 box
9. Type in “Wrong Answer” in Option 3 box
10. Type in “Wrong Answer” in Option 4 box
11. Select the Option 1 as the Correct answer
12. Type “Testing Explanation” in the Answer Explanation box
13. Leave all the input fields in the Question Information Page empty
14. Click on “Add Question” button
15. Type in “Testing Question 2” in the Question box
16. Type in “Correct Answer” in Option 1 box
17. Type in “Wrong Answer” in Option 2 box
18. Type in “Wrong Answer” in Option 3 box
19. Type in “Wrong Answer” in Option 4 box
20. Select the Option 1 as the Correct answer
21. Click on “Create Quiz” button
**Test Data**
- Quiz Title: “Testing title(2nd question)”
- Category: “Finance”
- Time per question: “15”
- FIQ per question: “25”
- Question: ”Testing Question”
- Option 1 box: Correct Answer
- Option 2 box: Wrong Answer
- Option 3 box: Wrong Answer
- Option 4 box: Wrong Answer
- Checkbox: Option 1
- Answer Explanation: “Testing Explanation”
- Question: ”Testing Question 2”
- Option 1 box: Correct Answer
- Option 2 box: Wrong Answer
- Option 3 box: Wrong Answer
- Option 4 box: Wrong Answer
- Checkbox: Option 1
- Answer Explanation: “Testing Explanation 2
**Expected Results**
The administrator should be redirected to the Quiz page displaying the newly created quiz and a noty notifying that the quiz with the correct quiz title is created
**Actual Results**
-The administrator is redirected to the Quiz page displaying the newly created quiz and a noty notifying that the quiz with the correct quiz title is created
**Pass/Fail**
- Pass
**Date Tested**
- 8th February 2021
**Tested By**
- Zachary Tan | 1.0 | FA_033_Create Quiz(Adding A Question Input) - **Test Scenario**
- Administrator wishes to create the quiz after filling in the information for the 2nd question of the quiz
**Test Case**
- Administrator enters all the available field in the Quiz Information and the Quest Information Panel except FIQ awarded and Time
**Pre-Conditions**
- Administrator must have successfully logged into the application
- Administrator must have selected the “Quizzes” option
- Administrator must have clicked on the “Add Question” button
- Administrator must have clicked on the “Create New” Button
**Test-Steps**
1. Type “Testing Title(2nd question)” into the Quiz Title box
2. Select “Finance” as Category
3. Select “25” in the FIQ per question box
4. Select “15” in the Time per question box
5. Type “Test Description” in the Quiz Description box
6. Type in “Testing Question” in the Question box
7. Type in “Correct Answer” in Option 1 box
8. Type in “Wrong Answer” in Option 2 box
9. Type in “Wrong Answer” in Option 3 box
10. Type in “Wrong Answer” in Option 4 box
11. Select the Option 1 as the Correct answer
12. Type “Testing Explanation” in the Answer Explanation box
13. Leave all the input fields in the Question Information Page empty
14. Click on “Add Question” button
15. Type in “Testing Question 2” in the Question box
16. Type in “Correct Answer” in Option 1 box
17. Type in “Wrong Answer” in Option 2 box
18. Type in “Wrong Answer” in Option 3 box
19. Type in “Wrong Answer” in Option 4 box
20. Select the Option 1 as the Correct answer
21. Click on “Create Quiz” button
**Test Data**
- Quiz Title: “Testing title(2nd question)”
- Category: “Finance”
- Time per question: “15”
- FIQ per question: “25”
- Question: ”Testing Question”
- Option 1 box: Correct Answer
- Option 2 box: Wrong Answer
- Option 3 box: Wrong Answer
- Option 4 box: Wrong Answer
- Checkbox: Option 1
- Answer Explanation: “Testing Explanation”
- Question: ”Testing Question 2”
- Option 1 box: Correct Answer
- Option 2 box: Wrong Answer
- Option 3 box: Wrong Answer
- Option 4 box: Wrong Answer
- Checkbox: Option 1
- Answer Explanation: “Testing Explanation 2
**Expected Results**
The administrator should be redirected to the Quiz page displaying the newly created quiz and a noty notifying that the quiz with the correct quiz title is created
**Actual Results**
-The administrator is redirected to the Quiz page displaying the newly created quiz and a noty notifying that the quiz with the correct quiz title is created
**Pass/Fail**
- Pass
**Date Tested**
- 8th February 2021
**Tested By**
- Zachary Tan | non_priority | fa create quiz adding a question input test scenario administrator wishes to create the quiz after filling in the information for the question of the quiz test case administrator enters all the available field in the quiz information and the quest information panel except fiq awarded and time pre conditions administrator must have successfully logged into the application administrator must have selected the “quizzes” option administrator must have clicked on the “add question” button administrator must have clicked on the “create new” button test steps type “testing title question ” into the quiz title box select “finance” as category select “ ” in the fiq per question box select “ ” in the time per question box type “test description” in the quiz description box type in “testing question” in the question box type in “correct answer” in option box type in “wrong answer” in option box type in “wrong answer” in option box type in “wrong answer” in option box select the option as the correct answer type “testing explanation” in the answer explanation box leave all the input fields in the question information page empty click on “add question” button type in “testing question ” in the question box type in “correct answer” in option box type in “wrong answer” in option box type in “wrong answer” in option box type in “wrong answer” in option box select the option as the correct answer click on “create quiz” button test data quiz title “testing title question ” category “finance” time per question “ ” fiq per question “ ” question ”testing question” option box correct answer option box wrong answer option box wrong answer option box wrong answer checkbox option answer explanation “testing explanation” question ”testing question ” option box correct answer option box wrong answer option box wrong answer option box wrong answer checkbox option answer explanation “testing explanation expected results the administrator should be redirected to the quiz page displaying the newly created quiz and a noty notifying that the quiz with the correct quiz title is created actual results the administrator is redirected to the quiz page displaying the newly created quiz and a noty notifying that the quiz with the correct quiz title is created pass fail pass date tested february tested by zachary tan | 0 |
334,292 | 24,412,493,596 | IssuesEvent | 2022-10-05 13:30:27 | ViktoriaHillerud/dummy-project | https://api.github.com/repos/ViktoriaHillerud/dummy-project | opened | User story persona Kim Karlsson User story 2: | documentation | Som en IT-kunnig person vill jag kunna dela med mig kunskap och
knyta kontakter med andra programmerare.
Konversation:
Kunna se vilka programmeringsspråk de andra kan,
kunna skicka privata meddelanden till folk såväl som ett större forum. (@ för att tagga folk)
Konfirmation:
*Personliga profiler där man kan se vilka språk de kan
*funktionalitet för att tagga folk i kommentarer i forum
*skicka privata meddelanden
Estimerad komplexitet: 5 | 1.0 | User story persona Kim Karlsson User story 2: - Som en IT-kunnig person vill jag kunna dela med mig kunskap och
knyta kontakter med andra programmerare.
Konversation:
Kunna se vilka programmeringsspråk de andra kan,
kunna skicka privata meddelanden till folk såväl som ett större forum. (@ för att tagga folk)
Konfirmation:
*Personliga profiler där man kan se vilka språk de kan
*funktionalitet för att tagga folk i kommentarer i forum
*skicka privata meddelanden
Estimerad komplexitet: 5 | non_priority | user story persona kim karlsson user story som en it kunnig person vill jag kunna dela med mig kunskap och knyta kontakter med andra programmerare konversation kunna se vilka programmeringsspråk de andra kan kunna skicka privata meddelanden till folk såväl som ett större forum för att tagga folk konfirmation personliga profiler där man kan se vilka språk de kan funktionalitet för att tagga folk i kommentarer i forum skicka privata meddelanden estimerad komplexitet | 0 |
59,521 | 8,367,930,109 | IssuesEvent | 2018-10-04 13:35:43 | aspnetboilerplate/aspnetboilerplate | https://api.github.com/repos/aspnetboilerplate/aspnetboilerplate | closed | Documentation: Enrich the documentation with a sample custom module | documentation | Hello,
It would be great if you guys have time to enrich the documentation with a sample custom module just to show an example how would a new custom module look like, what code it can contain, to which DDD layer it belongs, etc.
Something that is different from AppService, Code, Web.Core, etc.
Thank you
Bilal H. | 1.0 | Documentation: Enrich the documentation with a sample custom module - Hello,
It would be great if you guys have time to enrich the documentation with a sample custom module just to show an example how would a new custom module look like, what code it can contain, to which DDD layer it belongs, etc.
Something that is different from AppService, Code, Web.Core, etc.
Thank you
Bilal H. | non_priority | documentation enrich the documentation with a sample custom module hello it would be great if you guys have time to enrich the documentation with a sample custom module just to show an example how would a new custom module look like what code it can contain to which ddd layer it belongs etc something that is different from appservice code web core etc thank you bilal h | 0 |
24,534 | 11,046,384,045 | IssuesEvent | 2019-12-09 16:45:50 | terraform-google-modules/terraform-google-lb | https://api.github.com/repos/terraform-google-modules/terraform-google-lb | closed | Add support for service account-based firewall targets | P2 enhancement security triaged | Firewall rules can only be configured by network tag. Better practice is to use Service Accounts, and this module should support that. | True | Add support for service account-based firewall targets - Firewall rules can only be configured by network tag. Better practice is to use Service Accounts, and this module should support that. | non_priority | add support for service account based firewall targets firewall rules can only be configured by network tag better practice is to use service accounts and this module should support that | 0 |
54,354 | 6,812,087,281 | IssuesEvent | 2017-11-06 00:16:04 | learn-anything/maps | https://api.github.com/repos/learn-anything/maps | closed | best path for learning design languages | design study plan | Take a look [here](https://my.mindnode.com/grUmLWNxBRYqRUbRVXqWc8uygE3SLPy7V9tp63nq#70.9,-171.3,2).
If you think there is a better way one can learn design languages or you think the way the nodes are structured is wrong, please say it here.
Also if you think there are some really amazing resources on design languages that are missing or you wish something was added, you can say it here. | 1.0 | best path for learning design languages - Take a look [here](https://my.mindnode.com/grUmLWNxBRYqRUbRVXqWc8uygE3SLPy7V9tp63nq#70.9,-171.3,2).
If you think there is a better way one can learn design languages or you think the way the nodes are structured is wrong, please say it here.
Also if you think there are some really amazing resources on design languages that are missing or you wish something was added, you can say it here. | non_priority | best path for learning design languages take a look if you think there is a better way one can learn design languages or you think the way the nodes are structured is wrong please say it here also if you think there are some really amazing resources on design languages that are missing or you wish something was added you can say it here | 0 |
269,074 | 20,370,728,801 | IssuesEvent | 2022-02-21 10:57:23 | gbif-norway/helpdesk | https://api.github.com/repos/gbif-norway/helpdesk | closed | AgentActions paper | documentation identifiers wikidata metadata | David has asked for feedback on a paper for the AgentActions extension:
https://docs.google.com/document/d/1VoqpGfyN_8FOILlrY8lZOZB3pZDol0S6-mom_kHGEHU/edit?ts=609d2bff#heading=h.eugmjzvvj9fv | 1.0 | AgentActions paper - David has asked for feedback on a paper for the AgentActions extension:
https://docs.google.com/document/d/1VoqpGfyN_8FOILlrY8lZOZB3pZDol0S6-mom_kHGEHU/edit?ts=609d2bff#heading=h.eugmjzvvj9fv | non_priority | agentactions paper david has asked for feedback on a paper for the agentactions extension | 0 |
86,561 | 15,755,680,458 | IssuesEvent | 2021-03-31 02:12:28 | Thezone1975/Wallie_Bot | https://api.github.com/repos/Thezone1975/Wallie_Bot | opened | CVE-2017-1000048 (High) detected in qs-5.1.0.tgz | security vulnerability | ## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-5.1.0.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-5.1.0.tgz">https://registry.npmjs.org/qs/-/qs-5.1.0.tgz</a></p>
<p>Path to dependency file: /watsonSocketBot/package.json</p>
<p>Path to vulnerable library: /tmp/git/watsonSocketBot/node_modules/solr-client/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- watson-developer-cloud-2.27.0.tgz (Root Library)
- solr-client-0.6.0.tgz
- request-2.63.0.tgz
- :x: **qs-5.1.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-1000048 (High) detected in qs-5.1.0.tgz - ## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-5.1.0.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-5.1.0.tgz">https://registry.npmjs.org/qs/-/qs-5.1.0.tgz</a></p>
<p>Path to dependency file: /watsonSocketBot/package.json</p>
<p>Path to vulnerable library: /tmp/git/watsonSocketBot/node_modules/solr-client/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- watson-developer-cloud-2.27.0.tgz (Root Library)
- solr-client-0.6.0.tgz
- request-2.63.0.tgz
- :x: **qs-5.1.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in qs tgz cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file watsonsocketbot package json path to vulnerable library tmp git watsonsocketbot node modules solr client node modules qs package json dependency hierarchy watson developer cloud tgz root library solr client tgz request tgz x qs tgz vulnerable library vulnerability details the web framework using ljharb s qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs step up your open source security game with whitesource | 0 |
13,082 | 9,856,595,089 | IssuesEvent | 2019-06-19 22:45:17 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | "az functionapp create" should create app insights by default | Feature Request Functions Functions-cli H2-2019 Service Attention | **Is your feature request related to a problem? Please describe.**
Currently all function apps created by any of the tools - except Az CLI - has app insights created and configured
**Describe the solution you'd like**
"az functionapp create" should create app insights by default
**Describe alternatives you've considered**
Create the app insights and configure the instrumentation key after the fact
**Additional context**
None
| 1.0 | "az functionapp create" should create app insights by default - **Is your feature request related to a problem? Please describe.**
Currently all function apps created by any of the tools - except Az CLI - has app insights created and configured
**Describe the solution you'd like**
"az functionapp create" should create app insights by default
**Describe alternatives you've considered**
Create the app insights and configure the instrumentation key after the fact
**Additional context**
None
| non_priority | az functionapp create should create app insights by default is your feature request related to a problem please describe currently all function apps created by any of the tools except az cli has app insights created and configured describe the solution you d like az functionapp create should create app insights by default describe alternatives you ve considered create the app insights and configure the instrumentation key after the fact additional context none | 0 |
52,632 | 13,020,351,839 | IssuesEvent | 2020-07-27 02:47:54 | proftpd/proftpd | https://api.github.com/repos/proftpd/proftpd | closed | Define OpenSSL API version compatibility macro | build enhancement | The scope of this ticket is to use the [`OPENSSL_API_COMPAT`](https://wiki.openssl.org/index.php/OpenSSL_1.1.0_Changes#OPENSSL_API_COMPAT) macro to define the OpenSSL API version with which ProFTPD is compatible; symbols from versions of OpenSSL older that the specified API version will cause build failures. See, for example, [this pyca PR](https://github.com/pyca/cryptography/pull/4313/files) which does the same thing for that project.
We will start by declaring that ProFTPD is compatible with OpenSSL 1.0.0. *Note* that for platforms where support for versions of OpenSSL older than 1.0.0 is still required, there will be a configure option to disable use of this macro. | 1.0 | Define OpenSSL API version compatibility macro - The scope of this ticket is to use the [`OPENSSL_API_COMPAT`](https://wiki.openssl.org/index.php/OpenSSL_1.1.0_Changes#OPENSSL_API_COMPAT) macro to define the OpenSSL API version with which ProFTPD is compatible; symbols from versions of OpenSSL older that the specified API version will cause build failures. See, for example, [this pyca PR](https://github.com/pyca/cryptography/pull/4313/files) which does the same thing for that project.
We will start by declaring that ProFTPD is compatible with OpenSSL 1.0.0. *Note* that for platforms where support for versions of OpenSSL older than 1.0.0 is still required, there will be a configure option to disable use of this macro. | non_priority | define openssl api version compatibility macro the scope of this ticket is to use the macro to define the openssl api version with which proftpd is compatible symbols from versions of openssl older that the specified api version will cause build failures see for example which does the same thing for that project we will start by declaring that proftpd is compatible with openssl note that for platforms where support for versions of openssl older than is still required there will be a configure option to disable use of this macro | 0 |
27,630 | 22,046,740,263 | IssuesEvent | 2022-05-30 03:13:19 | jguadagno/jjgnet-broadcast | https://api.github.com/repos/jguadagno/jjgnet-broadcast | closed | Update deployment for Azure Functions | Infrastructure | Update the Azure Functions pipeline to execute/deploy on commit to main only. | 1.0 | Update deployment for Azure Functions - Update the Azure Functions pipeline to execute/deploy on commit to main only. | non_priority | update deployment for azure functions update the azure functions pipeline to execute deploy on commit to main only | 0 |
194,599 | 22,262,042,068 | IssuesEvent | 2022-06-10 02:01:37 | ShaikUsaf/linux-4.19.72_CVE-2020-10757 | https://api.github.com/repos/ShaikUsaf/linux-4.19.72_CVE-2020-10757 | reopened | CVE-2020-27786 (High) detected in linuxlinux-4.19.236, linuxlinux-4.19.236 | security vulnerability | ## CVE-2020-27786 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.236</b>, <b>linuxlinux-4.19.236</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Linux kernel’s implementation of MIDI, where an attacker with a local account and the permissions to issue ioctl commands to midi devices could trigger a use-after-free issue. A write to this specific memory while freed and before use causes the flow of execution to change and possibly allow for memory corruption or privilege escalation. The highest threat from this vulnerability is to confidentiality, integrity, as well as system availability.
WhiteSource Note: After conducting further research, WhiteSource has determined that versions v2.6.12-rc2 through v4.4.223, ;v4.5-rc1 through v4.9.223, v4.10-rc1 through v4.14.180, v4.15-rc1 through v4.19.123, v5.0-rc1 through v5.4.41, v5.5-rc1 through v5.6.13 and v5.7-rc1 through v5.7-rc5 of Linux Kernel are vulnerable to CVE-2020-27786.
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27786>CVE-2020-27786</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-27786">https://www.linuxkernelcves.com/cves/CVE-2020-27786</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: v4.4.224, v4.9.224, v4.14.181, v4.19.124, v5.4.42, v5.6.14, v5.7-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-27786 (High) detected in linuxlinux-4.19.236, linuxlinux-4.19.236 - ## CVE-2020-27786 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linuxlinux-4.19.236</b>, <b>linuxlinux-4.19.236</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in the Linux kernel’s implementation of MIDI, where an attacker with a local account and the permissions to issue ioctl commands to midi devices could trigger a use-after-free issue. A write to this specific memory while freed and before use causes the flow of execution to change and possibly allow for memory corruption or privilege escalation. The highest threat from this vulnerability is to confidentiality, integrity, as well as system availability.
WhiteSource Note: After conducting further research, WhiteSource has determined that versions v2.6.12-rc2 through v4.4.223, ;v4.5-rc1 through v4.9.223, v4.10-rc1 through v4.14.180, v4.15-rc1 through v4.19.123, v5.0-rc1 through v5.4.41, v5.5-rc1 through v5.6.13 and v5.7-rc1 through v5.7-rc5 of Linux Kernel are vulnerable to CVE-2020-27786.
<p>Publish Date: 2020-12-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27786>CVE-2020-27786</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-27786">https://www.linuxkernelcves.com/cves/CVE-2020-27786</a></p>
<p>Release Date: 2020-12-11</p>
<p>Fix Resolution: v4.4.224, v4.9.224, v4.14.181, v4.19.124, v5.4.42, v5.6.14, v5.7-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in linuxlinux linuxlinux cve high severity vulnerability vulnerable libraries linuxlinux linuxlinux vulnerability details a flaw was found in the linux kernel’s implementation of midi where an attacker with a local account and the permissions to issue ioctl commands to midi devices could trigger a use after free issue a write to this specific memory while freed and before use causes the flow of execution to change and possibly allow for memory corruption or privilege escalation the highest threat from this vulnerability is to confidentiality integrity as well as system availability whitesource note after conducting further research whitesource has determined that versions through through through through through through and through of linux kernel are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
55,960 | 8,037,293,572 | IssuesEvent | 2018-07-30 12:09:25 | DevExpress/testcafe | https://api.github.com/repos/DevExpress/testcafe | closed | Describe new features | !IMPORTANT! AREA: docs DOCUMENTATION: required | - [x] Run TestCafe over https protocol (https://github.com/DevExpress/testcafe/issues/1985#issuecomment-389488983)
I've created a topic draft for this feature. See - https://github.com/miherlosev/tmp/issues/9
- [x] Screenshot file pattern (https://github.com/DevExpress/testcafe/pull/2086#issuecomment-365604782)
- [ ] No reload page url between tests (https://github.com/DevExpress/testcafe/issues/1770)
- [x] Debug mode with non-modified client scripts (https://github.com/DevExpress/testcafe/issues/2545) | 1.0 | Describe new features - - [x] Run TestCafe over https protocol (https://github.com/DevExpress/testcafe/issues/1985#issuecomment-389488983)
I've created a topic draft for this feature. See - https://github.com/miherlosev/tmp/issues/9
- [x] Screenshot file pattern (https://github.com/DevExpress/testcafe/pull/2086#issuecomment-365604782)
- [ ] No reload page url between tests (https://github.com/DevExpress/testcafe/issues/1770)
- [x] Debug mode with non-modified client scripts (https://github.com/DevExpress/testcafe/issues/2545) | non_priority | describe new features run testcafe over https protocol i ve created a topic draft for this feature see screenshot file pattern no reload page url between tests debug mode with non modified client scripts | 0 |
2,103 | 2,585,748,564 | IssuesEvent | 2015-02-17 03:32:34 | AbhinavMadahar/Amish-Tech-Support | https://api.github.com/repos/AbhinavMadahar/Amish-Tech-Support | closed | Test move.c | test | Make sure that the robot can still move well, even after converting move.c from percent speed to cm/s | 1.0 | Test move.c - Make sure that the robot can still move well, even after converting move.c from percent speed to cm/s | non_priority | test move c make sure that the robot can still move well even after converting move c from percent speed to cm s | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.