Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
3,624 | 14,659,927,445 | IssuesEvent | 2020-12-28 21:54:07 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | opened | Application to join: Pack & Upload | Maintainer application | Hello and welcome to the contrib application process! We're happy to have you :)
## Please note these 3 requirements for new contrib projects:
- [ ] Include a README.md file containing license and maintainer information.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md
- [ ] Include a LICENSE.txt file.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt.
- [ ] If porting a Drupal 7 project, Maintain the Git history from Drupal.
## Please provide the following information:
**The name of your module, theme, or layout**
Pack & Upload
**(Optional) Post a link here to an issue in the drupal.org queue notifying the Drupal 7 maintainers that you are working on a Backdrop port of their project**
https://www.drupal.org/project/pack_upload/issues/3189907
**Post a link to your new Backdrop project under your own GitHub account (option #1)**
https://github.com/djzwerg/pack_upload
**If you have chosen option #2 or #1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)**
YES
<!-- Once we have a chance to review your project, we will check for the 3 requirements at the top of this issue. If those requirements are met, you will be invited to the @backdrop-contrib group. At that point you will be able to transfer the project. -->
<!-- Please note that we may also include additional feedback in the code review, but anything else is only intended to be helpful, and is NOT a requirement for joining the contrib group. -->
| True | Application to join: Pack & Upload - Hello and welcome to the contrib application process! We're happy to have you :)
## Please note these 3 requirements for new contrib projects:
- [ ] Include a README.md file containing license and maintainer information.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md
- [ ] Include a LICENSE.txt file.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt.
- [ ] If porting a Drupal 7 project, Maintain the Git history from Drupal.
## Please provide the following information:
**The name of your module, theme, or layout**
Pack & Upload
**(Optional) Post a link here to an issue in the drupal.org queue notifying the Drupal 7 maintainers that you are working on a Backdrop port of their project**
https://www.drupal.org/project/pack_upload/issues/3189907
**Post a link to your new Backdrop project under your own GitHub account (option #1)**
https://github.com/djzwerg/pack_upload
**If you have chosen option #2 or #1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)**
YES
<!-- Once we have a chance to review your project, we will check for the 3 requirements at the top of this issue. If those requirements are met, you will be invited to the @backdrop-contrib group. At that point you will be able to transfer the project. -->
<!-- Please note that we may also include additional feedback in the code review, but anything else is only intended to be helpful, and is NOT a requirement for joining the contrib group. -->
| main | application to join pack upload hello and welcome to the contrib application process we re happy to have you please note these requirements for new contrib projects include a readme md file containing license and maintainer information you can use this example include a license txt file you can use this example if porting a drupal project maintain the git history from drupal please provide the following information the name of your module theme or layout pack upload optional post a link here to an issue in the drupal org queue notifying the drupal maintainers that you are working on a backdrop port of their project post a link to your new backdrop project under your own github account option if you have chosen option or above do you agree to the yes | 1 |
401,174 | 11,787,019,359 | IssuesEvent | 2020-03-17 13:21:36 | richelbilderbeek/pirouette | https://api.github.com/repos/richelbilderbeek/pirouette | closed | Add: create_dd_tree | low priority | Similar to `create_yule_tree`, `create_bd_tree` add a function called `create_dd_tree`. It should create a nice exemplary tree.
Also use similar tests as those in `test-create_bd_tree.R`. | 1.0 | Add: create_dd_tree - Similar to `create_yule_tree`, `create_bd_tree` add a function called `create_dd_tree`. It should create a nice exemplary tree.
Also use similar tests as those in `test-create_bd_tree.R`. | non_main | add create dd tree similar to create yule tree create bd tree add a function called create dd tree it should create a nice exemplary tree also use similar tests as those in test create bd tree r | 0 |
30,392 | 13,238,173,119 | IssuesEvent | 2020-08-18 23:32:40 | projectcontour/contour | https://api.github.com/repos/projectcontour/contour | opened | Decision: How do we implement *Class objects (IngressClass and GatewayClass) | area/ingress area/service-apis priority/important-soon | IngressClass (from Ingress v1) and GatewayClass (from service-apis) are both replacements-with-added-functionality for the `kubernetes.io/ingress.class` annotation. They both have fields that allow for Class-specific configuration to be applied.
Currently, Contour is configured from a "file" - that is, a ConfigMap, and those config parameters can only be overridden by command-line flags (when present). We're explicitly trying to move away from command line flags for Contour, however.
This issue is to discuss and answer the question: What do we do to implementt he two types of *Class object?
Options off the top of my head to get things started:
- One Contour deployment === One Class, either as currently configured, IngressClass, or GatewayClass. This would mean that you would need to run three (!) Contours to cover annotation, IngressClass, and GatewayClass. This does not seem ideal.
- One Contour deployment has multiple Class configuration. This is more flexible, but means that there would need to be a configuration hierarchy of some sort, where some parameters could be overridden by *Class params. How would that work?
- Something else I can't think of right now.
Ideas welcomed! | 1.0 | Decision: How do we implement *Class objects (IngressClass and GatewayClass) - IngressClass (from Ingress v1) and GatewayClass (from service-apis) are both replacements-with-added-functionality for the `kubernetes.io/ingress.class` annotation. They both have fields that allow for Class-specific configuration to be applied.
Currently, Contour is configured from a "file" - that is, a ConfigMap, and those config parameters can only be overridden by command-line flags (when present). We're explicitly trying to move away from command line flags for Contour, however.
This issue is to discuss and answer the question: What do we do to implementt he two types of *Class object?
Options off the top of my head to get things started:
- One Contour deployment === One Class, either as currently configured, IngressClass, or GatewayClass. This would mean that you would need to run three (!) Contours to cover annotation, IngressClass, and GatewayClass. This does not seem ideal.
- One Contour deployment has multiple Class configuration. This is more flexible, but means that there would need to be a configuration hierarchy of some sort, where some parameters could be overridden by *Class params. How would that work?
- Something else I can't think of right now.
Ideas welcomed! | non_main | decision how do we implement class objects ingressclass and gatewayclass ingressclass from ingress and gatewayclass from service apis are both replacements with added functionality for the kubernetes io ingress class annotation they both have fields that allow for class specific configuration to be applied currently contour is configured from a file that is a configmap and those config parameters can only be overridden by command line flags when present we re explicitly trying to move away from command line flags for contour however this issue is to discuss and answer the question what do we do to implementt he two types of class object options off the top of my head to get things started one contour deployment one class either as currently configured ingressclass or gatewayclass this would mean that you would need to run three contours to cover annotation ingressclass and gatewayclass this does not seem ideal one contour deployment has multiple class configuration this is more flexible but means that there would need to be a configuration hierarchy of some sort where some parameters could be overridden by class params how would that work something else i can t think of right now ideas welcomed | 0 |
3,507 | 13,689,186,653 | IssuesEvent | 2020-09-30 12:50:26 | frej/fast-export | https://api.github.com/repos/frej/fast-export | closed | git-fast-import died of signal 11 - [Errno 32] Broken pipe | not-available-to-maintainer user-support wintendo | Hi,
I'm running hg-fast-export to convert Mercurial to git. I'm using cygwin on Windows 10 to run on bash as I was having python issues trying to run the py files. I'm getting the following error during the numerous "Exporting simple delta revision" steps. Any thoughts on a solution?
```
error: git-fast-import died of signal 11
Traceback (most recent call last):
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 719, in <module>
plugins=plugins_dict))
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 571, in hg2git
plugins)
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 366, in export_commit
export_file_contents(ctx,man,added,hgtags,fn_encoding,plugins)
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 238, in export_file_contents
wr(d)
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 54, in wr
wr_no_nl(msg)
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 51, in wr_no_nl
stdout_buffer.write(msg)
IOError: [Errno 32] Broken pipe
``` | True | git-fast-import died of signal 11 - [Errno 32] Broken pipe - Hi,
I'm running hg-fast-export to convert Mercurial to git. I'm using cygwin on Windows 10 to run on bash as I was having python issues trying to run the py files. I'm getting the following error during the numerous "Exporting simple delta revision" steps. Any thoughts on a solution?
```
error: git-fast-import died of signal 11
Traceback (most recent call last):
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 719, in <module>
plugins=plugins_dict))
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 571, in hg2git
plugins)
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 366, in export_commit
export_file_contents(ctx,man,added,hgtags,fn_encoding,plugins)
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 238, in export_file_contents
wr(d)
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 54, in wr
wr_no_nl(msg)
File "/cygdrive/c/bitbucket/FastExport/fast-export/hg-fast-export.py", line 51, in wr_no_nl
stdout_buffer.write(msg)
IOError: [Errno 32] Broken pipe
``` | main | git fast import died of signal broken pipe hi i m running hg fast export to convert mercurial to git i m using cygwin on windows to run on bash as i was having python issues trying to run the py files i m getting the following error during the numerous exporting simple delta revision steps any thoughts on a solution error git fast import died of signal traceback most recent call last file cygdrive c bitbucket fastexport fast export hg fast export py line in plugins plugins dict file cygdrive c bitbucket fastexport fast export hg fast export py line in plugins file cygdrive c bitbucket fastexport fast export hg fast export py line in export commit export file contents ctx man added hgtags fn encoding plugins file cygdrive c bitbucket fastexport fast export hg fast export py line in export file contents wr d file cygdrive c bitbucket fastexport fast export hg fast export py line in wr wr no nl msg file cygdrive c bitbucket fastexport fast export hg fast export py line in wr no nl stdout buffer write msg ioerror broken pipe | 1 |
8,852 | 6,000,401,096 | IssuesEvent | 2017-06-05 04:43:40 | streaka/plantguard-issues | https://api.github.com/repos/streaka/plantguard-issues | closed | Add new checklist field types: Single Select, Multi Select and Switch | Type: Usability | _Please fill out this template thoroughly, for the sake of our sanity._
## **What did you expect to happen?**
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
## **What actually happened?**
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## **Steps to Reproduce Issue**
(preferably video in Chrome with inspector open with console visible)
## **What platform was this on?**
(Browser/iOS/Android)
| True | Add new checklist field types: Single Select, Multi Select and Switch - _Please fill out this template thoroughly, for the sake of our sanity._
## **What did you expect to happen?**
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
## **What actually happened?**
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## **Steps to Reproduce Issue**
(preferably video in Chrome with inspector open with console visible)
## **What platform was this on?**
(Browser/iOS/Android)
| non_main | add new checklist field types single select multi select and switch please fill out this template thoroughly for the sake of our sanity what did you expect to happen what actually happened steps to reproduce issue preferably video in chrome with inspector open with console visible what platform was this on browser ios android | 0 |
1,489 | 6,425,237,168 | IssuesEvent | 2017-08-09 15:02:41 | OpenRefine/OpenRefine | https://api.github.com/repos/OpenRefine/OpenRefine | closed | Travis: generate code coverage statistics | maintainability | We could use coveralls to report test coverage. It has a good Github integration. | True | Travis: generate code coverage statistics - We could use coveralls to report test coverage. It has a good Github integration. | main | travis generate code coverage statistics we could use coveralls to report test coverage it has a good github integration | 1 |
385 | 3,420,426,386 | IssuesEvent | 2015-12-08 14:48:03 | mesosphere/kubernetes-mesos | https://api.github.com/repos/mesosphere/kubernetes-mesos | closed | k8sm boostrap should sort env vars dumped to stdout | area/dcos-packaging class/enhancement integration/DCOS newbie pqm/maintainability WIP | ... to improve readability. current state:
```
* Environment:
SCHEDULER_PORT=25504
ETCD_MESOS_FRAMEWORK_NAME=etcd
PORT_25500=25500
PORT_25501=25505
S6_VERSION=v1.13.0.0
APISERVER_PORT=25503
DEBUG=false
PORT_25502=25502
HOSTNAME=ip-10-0-0-82.us-west-2.compute.internal
SHLVL=1
PORT_25503=25501
SCHEDULER_ACCOUNT_FOR_POD_RESOURCES=true
HOME=/root
FRAMEWORK_NAME=kubernetes
``` | True | k8sm boostrap should sort env vars dumped to stdout - ... to improve readability. current state:
```
* Environment:
SCHEDULER_PORT=25504
ETCD_MESOS_FRAMEWORK_NAME=etcd
PORT_25500=25500
PORT_25501=25505
S6_VERSION=v1.13.0.0
APISERVER_PORT=25503
DEBUG=false
PORT_25502=25502
HOSTNAME=ip-10-0-0-82.us-west-2.compute.internal
SHLVL=1
PORT_25503=25501
SCHEDULER_ACCOUNT_FOR_POD_RESOURCES=true
HOME=/root
FRAMEWORK_NAME=kubernetes
``` | main | boostrap should sort env vars dumped to stdout to improve readability current state environment scheduler port etcd mesos framework name etcd port port version apiserver port debug false port hostname ip us west compute internal shlvl port scheduler account for pod resources true home root framework name kubernetes | 1 |
75,955 | 26,175,561,488 | IssuesEvent | 2023-01-02 09:17:10 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | opened | Vulnerability in snakeyaml 1.32 used by Elasticsearch connectors | Type: Defect Source: Internal security severity:high Team: Platform | Elasticsearch 6 and 7 connector from HZ Jet is using snakeyaml in version 1.32 which includes following vulnerability:
- CVE-2022-1471 - https://nvd.nist.gov/vuln/detail/CVE-2022-1471
It affects Hazelcast streams `5.0.z`, `5.1.z`, `5.2.z` and `master`. | 1.0 | Vulnerability in snakeyaml 1.32 used by Elasticsearch connectors - Elasticsearch 6 and 7 connector from HZ Jet is using snakeyaml in version 1.32 which includes following vulnerability:
- CVE-2022-1471 - https://nvd.nist.gov/vuln/detail/CVE-2022-1471
It affects Hazelcast streams `5.0.z`, `5.1.z`, `5.2.z` and `master`. | non_main | vulnerability in snakeyaml used by elasticsearch connectors elasticsearch and connector from hz jet is using snakeyaml in version which includes following vulnerability cve it affects hazelcast streams z z z and master | 0 |
109,881 | 4,414,982,773 | IssuesEvent | 2016-08-13 19:47:27 | Sistema-Integrado-Gestao-Academica/SiGA | https://api.github.com/repos/Sistema-Integrado-Gestao-Academica/SiGA | closed | [Portal] Selecionar Secretários Distintos | bug [Medium Priority] | Na página do Portal, apresentar somente uma vez o nome do secretário. | 1.0 | [Portal] Selecionar Secretários Distintos - Na página do Portal, apresentar somente uma vez o nome do secretário. | non_main | selecionar secretários distintos na página do portal apresentar somente uma vez o nome do secretário | 0 |
115,663 | 9,808,428,138 | IssuesEvent | 2019-06-12 15:36:05 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] LangMustacheClientYamlTestSuiteIT.test {yaml=lang_mustache/30_search_template/Test with new response format} failed | :Core/Infra/Scripting >test-failure | `LangMustacheClientYamlTestSuiteIT.test {yaml=lang_mustache/30_search_template/Test with new response format}` failed on my PR build and seems unrelated:
https://scans.gradle.com/s/axrx54rul6s3c/tests/wvxejxkegxw3s-v3qw2yd63iq4w
```
java.lang.AssertionError:
Failure at [lang_mustache/30_search_template:148]: hits.total.value didn't match expected value:
hits.total.value: expected [0] but was [1]
```
Did not reproduce locally with the reproduction line. | 1.0 | [CI] LangMustacheClientYamlTestSuiteIT.test {yaml=lang_mustache/30_search_template/Test with new response format} failed - `LangMustacheClientYamlTestSuiteIT.test {yaml=lang_mustache/30_search_template/Test with new response format}` failed on my PR build and seems unrelated:
https://scans.gradle.com/s/axrx54rul6s3c/tests/wvxejxkegxw3s-v3qw2yd63iq4w
```
java.lang.AssertionError:
Failure at [lang_mustache/30_search_template:148]: hits.total.value didn't match expected value:
hits.total.value: expected [0] but was [1]
```
Did not reproduce locally with the reproduction line. | non_main | langmustacheclientyamltestsuiteit test yaml lang mustache search template test with new response format failed langmustacheclientyamltestsuiteit test yaml lang mustache search template test with new response format failed on my pr build and seems unrelated java lang assertionerror failure at hits total value didn t match expected value hits total value expected but was did not reproduce locally with the reproduction line | 0 |
1,667 | 6,574,070,847 | IssuesEvent | 2017-09-11 11:21:09 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | ec2_eip does not assign an ip to an eni | affects_2.3 aws bug_report cloud waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ec2_eip
##### ANSIBLE VERSION
ansible 2.3.0 (devel a83b00bbc0) last updated 2016/11/29 22:28:22 (GMT +000)
##### CONFIGURATION
standard
##### OS / ENVIRONMENT
centos
##### SUMMARY
If trying to assign an eip to an eni, the task will fail with a boto "MissingParameter" error.
##### STEPS TO REPRODUCE
Run the following task.
```
ec2_eip:
region: "{{ aws_region }}"
device_id: "{{ eni_a.interface.id }}"
reuse_existing_ip_allowed: yes
state: present
```
##### EXPECTED RESULTS
The ENI is assigned either a spare EIP or a new EIP
##### ACTUAL RESULTS
Task fails with the following message:
```
"EC2ResponseError: 400 Bad Request\n<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n<Response><Errors><Error><Code>MissingParameter</Code><Message>Either public IP or allocation id must be specified</Message></Error></Errors><RequestID>ab0b561f-47fa-4f43-80e4-8701b6666602</RequestID></Response>"
``` | True | ec2_eip does not assign an ip to an eni - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ec2_eip
##### ANSIBLE VERSION
ansible 2.3.0 (devel a83b00bbc0) last updated 2016/11/29 22:28:22 (GMT +000)
##### CONFIGURATION
standard
##### OS / ENVIRONMENT
centos
##### SUMMARY
If trying to assign an eip to an eni, the task will fail with a boto "MissingParameter" error.
##### STEPS TO REPRODUCE
Run the following task.
```
ec2_eip:
region: "{{ aws_region }}"
device_id: "{{ eni_a.interface.id }}"
reuse_existing_ip_allowed: yes
state: present
```
##### EXPECTED RESULTS
The ENI is assigned either a spare EIP or a new EIP
##### ACTUAL RESULTS
Task fails with the following message:
```
"EC2ResponseError: 400 Bad Request\n<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n<Response><Errors><Error><Code>MissingParameter</Code><Message>Either public IP or allocation id must be specified</Message></Error></Errors><RequestID>ab0b561f-47fa-4f43-80e4-8701b6666602</RequestID></Response>"
``` | main | eip does not assign an ip to an eni issue type bug report component name eip ansible version ansible devel last updated gmt configuration standard os environment centos summary if trying to assign an eip to an eni the task will fail with a boto missingparameter error steps to reproduce run the following task eip region aws region device id eni a interface id reuse existing ip allowed yes state present expected results the eni is assigned either a spare eip or a new eip actual results task fails with the following message bad request n n missingparameter either public ip or allocation id must be specified | 1 |
12,292 | 19,590,529,856 | IssuesEvent | 2022-01-05 12:26:55 | TycheSoftwares/Print-Invoice-Delivery-Notes-for-WooCommerce | https://api.github.com/repos/TycheSoftwares/Print-Invoice-Delivery-Notes-for-WooCommerce | closed | Print Invoice should support RTL | type: enhancement client's requirement | The client wants that our print invoice should support RTL.
Suppose, if the client is using the Persian language and he writes from right to left, so this format is not supported.
Ticket link: https://tychesoftwares.freshdesk.com/a/tickets/21559 | 1.0 | Print Invoice should support RTL - The client wants that our print invoice should support RTL.
Suppose, if the client is using the Persian language and he writes from right to left, so this format is not supported.
Ticket link: https://tychesoftwares.freshdesk.com/a/tickets/21559 | non_main | print invoice should support rtl the client wants that our print invoice should support rtl suppose if the client is using the persian language and he writes from right to left so this format is not supported ticket link | 0 |
427,846 | 12,399,725,996 | IssuesEvent | 2020-05-21 06:06:46 | grpc/grpc | https://api.github.com/repos/grpc/grpc | closed | Building PECL grpc-1.29.0 fails on Windows | kind/bug lang/php priority/P2 | The Windows builds of the freshly released PECL grpc-1.29.0 [fail](https://windows.php.net/downloads/pecl/releases/grpc/1.29.0/logs/), because the include path to third_party/boringssl-with-bazel/src/include is not properly set. I've fixed this directly in the shipped config.w32 to test the build locally, but probably the proper fix would be something like:
````.diff
templates/config.w32.template | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/templates/config.w32.template b/templates/config.w32.template
index f69cc08df65..84589544b7a 100644
--- a/templates/config.w32.template
+++ b/templates/config.w32.template
@@ -33,7 +33,7 @@
"/I"+configure_module_dirname+"\\src\\php\\ext\\grpc "+
"/I"+configure_module_dirname+"\\third_party\\abseil-cpp "+
"/I"+configure_module_dirname+"\\third_party\\address_sorting\\include "+
- "/I"+configure_module_dirname+"\\third_party\\boringssl\\include "+
+ "/I"+configure_module_dirname+"\\third_party\\boringssl-with-bazel\\src\\include "+
"/I"+configure_module_dirname+"\\third_party\\upb "+
"/I"+configure_module_dirname+"\\third_party\\zlib ");
````
| 1.0 | Building PECL grpc-1.29.0 fails on Windows - The Windows builds of the freshly released PECL grpc-1.29.0 [fail](https://windows.php.net/downloads/pecl/releases/grpc/1.29.0/logs/), because the include path to third_party/boringssl-with-bazel/src/include is not properly set. I've fixed this directly in the shipped config.w32 to test the build locally, but probably the proper fix would be something like:
````.diff
templates/config.w32.template | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/templates/config.w32.template b/templates/config.w32.template
index f69cc08df65..84589544b7a 100644
--- a/templates/config.w32.template
+++ b/templates/config.w32.template
@@ -33,7 +33,7 @@
"/I"+configure_module_dirname+"\\src\\php\\ext\\grpc "+
"/I"+configure_module_dirname+"\\third_party\\abseil-cpp "+
"/I"+configure_module_dirname+"\\third_party\\address_sorting\\include "+
- "/I"+configure_module_dirname+"\\third_party\\boringssl\\include "+
+ "/I"+configure_module_dirname+"\\third_party\\boringssl-with-bazel\\src\\include "+
"/I"+configure_module_dirname+"\\third_party\\upb "+
"/I"+configure_module_dirname+"\\third_party\\zlib ");
````
| non_main | building pecl grpc fails on windows the windows builds of the freshly released pecl grpc because the include path to third party boringssl with bazel src include is not properly set i ve fixed this directly in the shipped config to test the build locally but probably the proper fix would be something like diff templates config template file changed insertion deletion diff git a templates config template b templates config template index a templates config template b templates config template i configure module dirname src php ext grpc i configure module dirname third party abseil cpp i configure module dirname third party address sorting include i configure module dirname third party boringssl include i configure module dirname third party boringssl with bazel src include i configure module dirname third party upb i configure module dirname third party zlib | 0 |
959 | 4,703,613,210 | IssuesEvent | 2016-10-13 08:38:34 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | firewalld module should provide an option to add and remove interface to zone | affects_2.1 feature_idea waiting_on_maintainer | firewalld module should provide an option to add and remove interface to zone | True | firewalld module should provide an option to add and remove interface to zone - firewalld module should provide an option to add and remove interface to zone | main | firewalld module should provide an option to add and remove interface to zone firewalld module should provide an option to add and remove interface to zone | 1 |
446,321 | 12,855,002,456 | IssuesEvent | 2020-07-09 03:45:52 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.skechers.com.tr - "Browse" menu is empty | action-needssitepatch browser-fenix engine-gecko priority-normal severity-critical type-uaoverride | <!-- @browser: Firefox Mobile 75.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/51959 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.skechers.com.tr/p/go-walk-evolution-ultra-legac-15763-bbk
**Browser / Version**: Firefox Mobile 75.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
top left 3 line doesn't work.
some site features doesn't work
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.skechers.com.tr - "Browse" menu is empty - <!-- @browser: Firefox Mobile 75.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:75.0) Gecko/75.0 Firefox/75.0 -->
<!-- @reported_with: -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/51959 -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.skechers.com.tr/p/go-walk-evolution-ultra-legac-15763-bbk
**Browser / Version**: Firefox Mobile 75.0
**Operating System**: Android
**Tested Another Browser**: Yes Chrome
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
top left 3 line doesn't work.
some site features doesn't work
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | browse menu is empty url browser version firefox mobile operating system android tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce top left line doesn t work some site features doesn t work browser configuration none from with ❤️ | 0 |
5,275 | 26,671,419,479 | IssuesEvent | 2023-01-26 10:36:33 | beyarkay/eskom-calendar | https://api.github.com/repos/beyarkay/eskom-calendar | closed | Macassar - Western Cape area missing | bug waiting-on-maintainer missing-area-schedule | Hi @beyarkay , as previously discussed I think your open source work is great and lots of work has been put into the project. Could I ask that you add the Macassar area please as I stay in Somerset West but our schedule falls under this area. If you need help with adding areas or anything else in the future I don't mind helping with the right guidance. Keep up the good work. | True | Macassar - Western Cape area missing - Hi @beyarkay , as previously discussed I think your open source work is great and lots of work has been put into the project. Could I ask that you add the Macassar area please as I stay in Somerset West but our schedule falls under this area. If you need help with adding areas or anything else in the future I don't mind helping with the right guidance. Keep up the good work. | main | macassar western cape area missing hi beyarkay as previously discussed i think your open source work is great and lots of work has been put into the project could i ask that you add the macassar area please as i stay in somerset west but our schedule falls under this area if you need help with adding areas or anything else in the future i don t mind helping with the right guidance keep up the good work | 1 |
313,795 | 23,492,138,433 | IssuesEvent | 2022-08-17 19:53:36 | typescript-eslint/typescript-eslint | https://api.github.com/repos/typescript-eslint/typescript-eslint | closed | Docs: Automate the Options heading in rule docs pages | documentation accepting prs | ### Before You File a Documentation Request Please Confirm You Have Done The Following...
- [X] I have looked for existing [open or closed documentation requests](https://github.com/typescript-eslint/typescript-eslint/issues?q=is%3Aissue+label%3Adocumentation) that match my proposal.
- [X] I have [read the FAQ](https://typescript-eslint.io/docs/linting/troubleshooting) and my problem is not listed.
### Suggested Changes
Right now, rules pages manually include their options according to a light spattering of tests (#4367). But the part of the _Options_ section that includes the opening sentence(s) & code block description of the options can be generated using the rule's schema and/or code source.
Some rules, such as https://typescript-eslint.io/rules/adjacent-overload-signatures, have no options. They're the easier case.
Other rules, such as https://typescript-eslint.io/rules/array-type, have an options type, some ancillary type alias(es), and a `const defaultOptions: Options = { ... };`.
There are a few reasonable starting possibilities for how we could automate the creation of those options docs:
* Base them purely on the rule's `meta.schema[0]`
* Upside: simplest, and most portable (we could make a utility library for it!)
* Downside: some values might get duplicated (would we have to add logic to dedup?)
* Base them on the types & interfaces used in the rule's source file
* Upside: matches the internal implementation closer
* Downside: do we want it to match the internal implementation? and, is much more complicated
* Some combination of both
My proposal would be to base them purely on the rule's meta and see how that looks.
### Affected URL(s)
https://typescript-eslint.io/rules/* | 1.0 | Docs: Automate the Options heading in rule docs pages - ### Before You File a Documentation Request Please Confirm You Have Done The Following...
- [X] I have looked for existing [open or closed documentation requests](https://github.com/typescript-eslint/typescript-eslint/issues?q=is%3Aissue+label%3Adocumentation) that match my proposal.
- [X] I have [read the FAQ](https://typescript-eslint.io/docs/linting/troubleshooting) and my problem is not listed.
### Suggested Changes
Right now, rules pages manually include their options according to a light spattering of tests (#4367). But the part of the _Options_ section that includes the opening sentence(s) & code block description of the options can be generated using the rule's schema and/or code source.
Some rules, such as https://typescript-eslint.io/rules/adjacent-overload-signatures, have no options. They're the easier case.
Other rules, such as https://typescript-eslint.io/rules/array-type, have an options type, some ancillary type alias(es), and a `const defaultOptions: Options = { ... };`.
There are a few reasonable starting possibilities for how we could automate the creation of those options docs:
* Base them purely on the rule's `meta.schema[0]`
* Upside: simplest, and most portable (we could make a utility library for it!)
* Downside: some values might get duplicated (would we have to add logic to dedup?)
* Base them on the types & interfaces used in the rule's source file
* Upside: matches the internal implementation closer
* Downside: do we want it to match the internal implementation? and, is much more complicated
* Some combination of both
My proposal would be to base them purely on the rule's meta and see how that looks.
### Affected URL(s)
https://typescript-eslint.io/rules/* | non_main | docs automate the options heading in rule docs pages before you file a documentation request please confirm you have done the following i have looked for existing that match my proposal i have and my problem is not listed suggested changes right now rules pages manually include their options according to a light spattering of tests but the part of the options section that includes the opening sentence s code block description of the options can be generated using the rule s schema and or code source some rules such as have no options they re the easier case other rules such as have an options type some ancillary type alias es and a const defaultoptions options there are a few reasonable starting possibilities for how we could automate the creation of those options docs base them purely on the rule s meta schema upside simplest and most portable we could make a utility library for it downside some values might get duplicated would we have to add logic to dedup base them on the types interfaces used in the rule s source file upside matches the internal implementation closer downside do we want it to match the internal implementation and is much more complicated some combination of both my proposal would be to base them purely on the rule s meta and see how that looks affected url s | 0 |
63,751 | 14,656,775,551 | IssuesEvent | 2020-12-28 14:10:18 | fu1771695yongxie/quasar | https://api.github.com/repos/fu1771695yongxie/quasar | opened | CVE-2019-8331 (Medium) detected in bootstrap-3.3.5.min.js | security vulnerability | ## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p>
<p>Path to dependency file: quasar/docs/node_modules/autocomplete.js/test/playground_jquery.html</p>
<p>Path to vulnerable library: quasar/docs/node_modules/autocomplete.js/test/playground_jquery.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.5.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/quasar/commit/32f2b832b1bc2a748bc0c32b33f7cc0a6e03b95d">32f2b832b1bc2a748bc0c32b33f7cc0a6e03b95d</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-8331 (Medium) detected in bootstrap-3.3.5.min.js - ## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.5.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.5/js/bootstrap.min.js</a></p>
<p>Path to dependency file: quasar/docs/node_modules/autocomplete.js/test/playground_jquery.html</p>
<p>Path to vulnerable library: quasar/docs/node_modules/autocomplete.js/test/playground_jquery.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.5.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/quasar/commit/32f2b832b1bc2a748bc0c32b33f7cc0a6e03b95d">32f2b832b1bc2a748bc0c32b33f7cc0a6e03b95d</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file quasar docs node modules autocomplete js test playground jquery html path to vulnerable library quasar docs node modules autocomplete js test playground jquery html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch dev vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap bootstrap sass step up your open source security game with whitesource | 0 |
10,308 | 8,482,980,555 | IssuesEvent | 2018-10-25 20:10:11 | hiebj/stately | https://api.github.com/repos/hiebj/stately | closed | precommit.sh: reject cross-module commits | infrastructure | Automatically generated changelogs can only be kept sane if commits are isolated to a single module. For example, changing `stately-react` may require upstream changes in `stately-async`, but the commit message for `stately-async` needs to be independently clear; that is, it should only reference API changes or additions that are relevant to its own code.
To achieve this, changes that span multiple modules need to be split into separate commits - one per module. Commits that span multiple modules should be rejected, forcing the developer to create a separate commit for each module change. | 1.0 | precommit.sh: reject cross-module commits - Automatically generated changelogs can only be kept sane if commits are isolated to a single module. For example, changing `stately-react` may require upstream changes in `stately-async`, but the commit message for `stately-async` needs to be independently clear; that is, it should only reference API changes or additions that are relevant to its own code.
To achieve this, changes that span multiple modules need to be split into separate commits - one per module. Commits that span multiple modules should be rejected, forcing the developer to create a separate commit for each module change. | non_main | precommit sh reject cross module commits automatically generated changelogs can only be kept sane if commits are isolated to a single module for example changing stately react may require upstream changes in stately async but the commit message for stately async needs to be independently clear that is it should only reference api changes or additions that are relevant to its own code to achieve this changes that span multiple modules need to be split into separate commits one per module commits that span multiple modules should be rejected forcing the developer to create a separate commit for each module change | 0 |
500,379 | 14,497,612,957 | IssuesEvent | 2020-12-11 14:27:51 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | closed | Unable to databind Gantt with taskId and parentId string fields | Bug C: Gantt FP: Completed Kendo2 Next LIB Priority 5 S: Wrappers (ASP.NET Core) S: Wrappers (ASP.NET MVC) SEV: High | ### Bug report
In a Razor Pages project, the Gantt's tasks are not binding if the taskID and parentID fields are strings.
This is a regression introduced in version 2020.3.915.
### Reproduction of the problem
1. Open and run [this example](https://github.com/telerik/ui-for-aspnet-core-examples/tree/master/Telerik.Examples.RazorPages/Telerik.Examples.RazorPages/Pages/Gantt)
2. Switch to Kendo version after 2020.2.617
### Current behavior
There are no tasks displayed in the Gantt.
### Expected/desired behavior
The tasks should be displayed.
### Environment
* **Kendo UI version:** 2020.3.1118
* **Browser:** [all]
| 1.0 | Unable to databind Gantt with taskId and parentId string fields - ### Bug report
In a Razor Pages project, the Gantt's tasks are not binding if the taskID and parentID fields are strings.
This is a regression introduced in version 2020.3.915.
### Reproduction of the problem
1. Open and run [this example](https://github.com/telerik/ui-for-aspnet-core-examples/tree/master/Telerik.Examples.RazorPages/Telerik.Examples.RazorPages/Pages/Gantt)
2. Switch to Kendo version after 2020.2.617
### Current behavior
There are no tasks displayed in the Gantt.
### Expected/desired behavior
The tasks should be displayed.
### Environment
* **Kendo UI version:** 2020.3.1118
* **Browser:** [all]
| non_main | unable to databind gantt with taskid and parentid string fields bug report in a razor pages project the gantt s tasks are not binding if the taskid and parentid fields are strings this is a regression introduced in version reproduction of the problem open and run switch to kendo version after current behavior there are no tasks displayed in the gantt expected desired behavior the tasks should be displayed environment kendo ui version browser | 0 |
351,017 | 10,511,923,976 | IssuesEvent | 2019-09-27 16:34:17 | ehimennlab/ProjectE-2019 | https://api.github.com/repos/ehimennlab/ProjectE-2019 | opened | makeからsudoコマンドを除去 | Priority: low | ## WHY
- `sudo`権限を要求するのはあまり良くない
## WHAT
- `make`コマンド内から`sudo`を使用している箇所を修正 | 1.0 | makeからsudoコマンドを除去 - ## WHY
- `sudo`権限を要求するのはあまり良くない
## WHAT
- `make`コマンド内から`sudo`を使用している箇所を修正 | non_main | makeからsudoコマンドを除去 why sudo 権限を要求するのはあまり良くない what make コマンド内から sudo を使用している箇所を修正 | 0 |
794,882 | 28,053,267,294 | IssuesEvent | 2023-03-29 07:40:53 | AY2223S2-CS2103T-T09-4/tp | https://api.github.com/repos/AY2223S2-CS2103T-T09-4/tp | closed | Edit, Meetup Functions v1.3 | priority.High type.Task | Implement the functionality for:
* `editPolicy`
Implement appointment field for clients.
Integrate into storage. | 1.0 | Edit, Meetup Functions v1.3 - Implement the functionality for:
* `editPolicy`
Implement appointment field for clients.
Integrate into storage. | non_main | edit meetup functions implement the functionality for editpolicy implement appointment field for clients integrate into storage | 0 |
3,155 | 12,196,636,103 | IssuesEvent | 2020-04-29 19:22:22 | ipfs-shipyard/ipfs-desktop | https://api.github.com/repos/ipfs-shipyard/ipfs-desktop | closed | Unable to add folders via Files screen | P1 area/linux area/macos area/webui area/windows kind/bug need/analysis need/maintainer-input | macos : 10.13.6
Adding folders has no effect。
Using the command line ipfs add-r folders is fine | True | Unable to add folders via Files screen - macos : 10.13.6
Adding folders has no effect。
Using the command line ipfs add-r folders is fine | main | unable to add folders via files screen macos adding folders has no effect。 using the command line ipfs add r folders is fine | 1 |
4,758 | 24,524,425,770 | IssuesEvent | 2022-10-11 12:06:19 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Deterministic row ordering | type: enhancement work: backend work: frontend status: draft restricted: maintainers | ## Current behavior
- The records API returns records in an arbitrary, nondeterministic order.
- I find it somewhat disorienting to refresh a page and see different records, even when no data has changed.
## Desired behavior
- Ordering by PK (asc) is appended to the sorting criteria automatically.
- We could do this either on the back end or on the front end.
- We could even choose to display the auto-added sorting condition to the user, though this would require some additional UX design work.
| True | Deterministic row ordering - ## Current behavior
- The records API returns records in an arbitrary, nondeterministic order.
- I find it somewhat disorienting to refresh a page and see different records, even when no data has changed.
## Desired behavior
- Ordering by PK (asc) is appended to the sorting criteria automatically.
- We could do this either on the back end or on the front end.
- We could even choose to display the auto-added sorting condition to the user, though this would require some additional UX design work.
| main | deterministic row ordering current behavior the records api returns records in an arbitrary nondeterministic order i find it somewhat disorienting to refresh a page and see different records even when no data has changed desired behavior ordering by pk asc is appended to the sorting criteria automatically we could do this either on the back end or on the front end we could even choose to display the auto added sorting condition to the user though this would require some additional ux design work | 1 |
2,976 | 10,712,280,443 | IssuesEvent | 2019-10-25 08:38:51 | vostpt/mobile-app | https://api.github.com/repos/vostpt/mobile-app | closed | About Screen | Needs Maintainers Help good first issue hacktoberfest | **Description**
About screen for VOST, including contacts, links to the contributors and tutorial screen
**File Location**
```
- presentation
|__ ui
```
**Requirements**
- Include button to access *Contributors* screen
- Include button to access *Tutorial* screen
- All contacts should be clickable, either launch phone/email apps or use https://pub.dev/packages/url_launcher to open URLs
**UI**
<img width="361" alt="imagem" src="https://user-images.githubusercontent.com/10728633/63055869-b7f7d380-bede-11e9-8a71-3fd19c6bdfe6.png">
**NOTES**
Assume the following test:
```
A VOST Portugal - Associação de Voluntários Em Situações de Emergência - é uma equipa de voluntários digitais que recolhe, valida, e informa as populações em casos de emergência. Esta aplicação, feita integralmente por voluntários e em código aberto, tem como objectivo colocar numa só aplicação mensagens oficiais no âmbito da autoprotecção, desastres naturais, e fenómenos meteorológicos adversos.
```
Assume that in the grid you will receive a `List<Contacts>` in which a Contact can have the following structure:
```
String name;
String imageUrl;
String info;
ContactType type; // this is an enum with the values ["PHONE", "URL", "EMAIL"]
```
Depending on the `type`, a click on this button can either open the phone app with the number ,open an url or send an email.
The buttons at the bottom of the screen should be clickable, so a function should be exposed to place there the callback
| True | About Screen - **Description**
About screen for VOST, including contacts, links to the contributors and tutorial screen
**File Location**
```
- presentation
|__ ui
```
**Requirements**
- Include button to access *Contributors* screen
- Include button to access *Tutorial* screen
- All contacts should be clickable, either launch phone/email apps or use https://pub.dev/packages/url_launcher to open URLs
**UI**
<img width="361" alt="imagem" src="https://user-images.githubusercontent.com/10728633/63055869-b7f7d380-bede-11e9-8a71-3fd19c6bdfe6.png">
**NOTES**
Assume the following test:
```
A VOST Portugal - Associação de Voluntários Em Situações de Emergência - é uma equipa de voluntários digitais que recolhe, valida, e informa as populações em casos de emergência. Esta aplicação, feita integralmente por voluntários e em código aberto, tem como objectivo colocar numa só aplicação mensagens oficiais no âmbito da autoprotecção, desastres naturais, e fenómenos meteorológicos adversos.
```
Assume that in the grid you will receive a `List<Contacts>` in which a Contact can have the following structure:
```
String name;
String imageUrl;
String info;
ContactType type; // this is an enum with the values ["PHONE", "URL", "EMAIL"]
```
Depending on the `type`, a click on this button can either open the phone app with the number ,open an url or send an email.
The buttons at the bottom of the screen should be clickable, so a function should be exposed to place there the callback
| main | about screen description about screen for vost including contacts links to the contributors and tutorial screen file location presentation ui requirements include button to access contributors screen include button to access tutorial screen all contacts should be clickable either launch phone email apps or use to open urls ui img width alt imagem src notes assume the following test a vost portugal associação de voluntários em situações de emergência é uma equipa de voluntários digitais que recolhe valida e informa as populações em casos de emergência esta aplicação feita integralmente por voluntários e em código aberto tem como objectivo colocar numa só aplicação mensagens oficiais no âmbito da autoprotecção desastres naturais e fenómenos meteorológicos adversos assume that in the grid you will receive a list in which a contact can have the following structure string name string imageurl string info contacttype type this is an enum with the values depending on the type a click on this button can either open the phone app with the number open an url or send an email the buttons at the bottom of the screen should be clickable so a function should be exposed to place there the callback | 1 |
494 | 3,788,154,690 | IssuesEvent | 2016-03-21 13:49:49 | christoff-buerger/racr | https://api.github.com/repos/christoff-buerger/racr | closed | Refactoring of Questionnaires example for RACR-NET | low maintainability | The _C#_ implementation of the Questionnaires example (`examples-net/questionnaires`) has to be refactored regarding the following issues:
* the names of files `Attributes.cs` and `Questionnaires.cs` are missleading
* `Attributes.cs` consists of widget construction and rendering attributes only; it is about the GUI and should be renamed respectively
* `Questionnaires.cs` contains the actual logic implementation, i.e., AST scheme and attribution; it should be renamed to `Language.cs` according to the original Questionnaires implementation
* `Widgets.cs` is part of the GUI
* the source code should be properly formatted | True | Refactoring of Questionnaires example for RACR-NET - The _C#_ implementation of the Questionnaires example (`examples-net/questionnaires`) has to be refactored regarding the following issues:
* the names of files `Attributes.cs` and `Questionnaires.cs` are missleading
* `Attributes.cs` consists of widget construction and rendering attributes only; it is about the GUI and should be renamed respectively
* `Questionnaires.cs` contains the actual logic implementation, i.e., AST scheme and attribution; it should be renamed to `Language.cs` according to the original Questionnaires implementation
* `Widgets.cs` is part of the GUI
* the source code should be properly formatted | main | refactoring of questionnaires example for racr net the c implementation of the questionnaires example examples net questionnaires has to be refactored regarding the following issues the names of files attributes cs and questionnaires cs are missleading attributes cs consists of widget construction and rendering attributes only it is about the gui and should be renamed respectively questionnaires cs contains the actual logic implementation i e ast scheme and attribution it should be renamed to language cs according to the original questionnaires implementation widgets cs is part of the gui the source code should be properly formatted | 1 |
5,869 | 31,842,992,264 | IssuesEvent | 2023-09-14 17:41:43 | microsoft/mu_basecore | https://api.github.com/repos/microsoft/mu_basecore | closed | [Bug]: RuntimeLibExitBootServicesEvent function redefine in MU_BASECORE/PolicyServicePkg/Library/DxePolicyLib | state:needs-triage state:needs-owner type:bug state:needs-maintainer-feedback urgency:high | ### Is there an existing issue for this?
- [X] I have searched existing issues
### Current Behavior
RuntimeLibExitBootServicesEvent function is already implemented in MU_BASECORE/MdePkg/Library/UefiRuntimeLib and get called in constructor.
The same RuntimeLibExitBootServicesEvent function is also implemented in MU_BASECORE/PolicyServicePkg/Library/DxePolicyLib and get called in constructor too.
In that case, on one driver's INF add both UefiRuntimeLib and DxePolicyLib, then it will have functio redefine compiler error as below.

### Expected Behavior
DxePolicyLib should rename the RuntimeLibExitBootServicesEvent to prevent conflict with existing function in UefiRuntimeLib
### Steps To Reproduce
add both UefiRuntimeLib and DxePolicyLib to the same INF file and build this driver should be able to reproduce the build error.
### Build Environment
```markdown
- OS(s):Windows 11
- Tool Chain(s): VS2022
- Targets Impacted: RELEASE, DEBUG
```
### Version Information
```text
4a53c6a187d2179924136a18bf015ecc0f2f06da
```
### Urgency
High
### Are you going to fix this?
Someone else needs to fix it
### Do you need maintainer feedback?
Maintainer feedback requested
### Anything else?
_No response_ | True | [Bug]: RuntimeLibExitBootServicesEvent function redefine in MU_BASECORE/PolicyServicePkg/Library/DxePolicyLib - ### Is there an existing issue for this?
- [X] I have searched existing issues
### Current Behavior
RuntimeLibExitBootServicesEvent function is already implemented in MU_BASECORE/MdePkg/Library/UefiRuntimeLib and get called in constructor.
The same RuntimeLibExitBootServicesEvent function is also implemented in MU_BASECORE/PolicyServicePkg/Library/DxePolicyLib and get called in constructor too.
In that case, on one driver's INF add both UefiRuntimeLib and DxePolicyLib, then it will have functio redefine compiler error as below.

### Expected Behavior
DxePolicyLib should rename the RuntimeLibExitBootServicesEvent to prevent conflict with existing function in UefiRuntimeLib
### Steps To Reproduce
add both UefiRuntimeLib and DxePolicyLib to the same INF file and build this driver should be able to reproduce the build error.
### Build Environment
```markdown
- OS(s):Windows 11
- Tool Chain(s): VS2022
- Targets Impacted: RELEASE, DEBUG
```
### Version Information
```text
4a53c6a187d2179924136a18bf015ecc0f2f06da
```
### Urgency
High
### Are you going to fix this?
Someone else needs to fix it
### Do you need maintainer feedback?
Maintainer feedback requested
### Anything else?
_No response_ | main | runtimelibexitbootservicesevent function redefine in mu basecore policyservicepkg library dxepolicylib is there an existing issue for this i have searched existing issues current behavior runtimelibexitbootservicesevent function is already implemented in mu basecore mdepkg library uefiruntimelib and get called in constructor the same runtimelibexitbootservicesevent function is also implemented in mu basecore policyservicepkg library dxepolicylib and get called in constructor too in that case on one driver s inf add both uefiruntimelib and dxepolicylib then it will have functio redefine compiler error as below expected behavior dxepolicylib should rename the runtimelibexitbootservicesevent to prevent conflict with existing function in uefiruntimelib steps to reproduce add both uefiruntimelib and dxepolicylib to the same inf file and build this driver should be able to reproduce the build error build environment markdown os s windows tool chain s targets impacted release debug version information text urgency high are you going to fix this someone else needs to fix it do you need maintainer feedback maintainer feedback requested anything else no response | 1 |
1,216 | 5,197,285,328 | IssuesEvent | 2017-01-23 15:16:33 | Particular/PBot | https://api.github.com/repos/Particular/PBot | closed | Identify user that moved an issue with pbot | Impact: M Size: S Tag: Maintainer Prio Type: Feature Withdrawn: Won't Fix | Currently there's no way to see who executed `pbot move` commands. This makes it hard to audit who made the move and discuss any conflicts, especially if the user who issued the command did not participate in the original issue.
In addition to adding the name of who executed the command, it would also be useful for pbot to automatically CC the user who moved the issue on the original issue, so that if there's still discussion happening in the original issue, that user will be made aware of it.
Example issue that covers the 2 use cases above: https://github.com/Particular/Collaboration/issues/260
**Proposed format:**
- In closed issue, change pbot comment to `Moved to REPO/#ISSUE_NO by @USER`
- in new issue, change pbot comment to `Issue moved by @USER. Originally opened in ORIGINAL_REPO#ISSUE_NO`
CC // @adamralph
| True | Identify user that moved an issue with pbot - Currently there's no way to see who executed `pbot move` commands. This makes it hard to audit who made the move and discuss any conflicts, especially if the user who issued the command did not participate in the original issue.
In addition to adding the name of who executed the command, it would also be useful for pbot to automatically CC the user who moved the issue on the original issue, so that if there's still discussion happening in the original issue, that user will be made aware of it.
Example issue that covers the 2 use cases above: https://github.com/Particular/Collaboration/issues/260
**Proposed format:**
- In closed issue, change pbot comment to `Moved to REPO/#ISSUE_NO by @USER`
- in new issue, change pbot comment to `Issue moved by @USER. Originally opened in ORIGINAL_REPO#ISSUE_NO`
CC // @adamralph
| main | identify user that moved an issue with pbot currently there s no way to see who executed pbot move commands this makes it hard to audit who made the move and discuss any conflicts especially if the user who issued the command did not participate in the original issue in addition to adding the name of who executed the command it would also be useful for pbot to automatically cc the user who moved the issue on the original issue so that if there s still discussion happening in the original issue that user will be made aware of it example issue that covers the use cases above proposed format in closed issue change pbot comment to moved to repo issue no by user in new issue change pbot comment to issue moved by user originally opened in original repo issue no cc adamralph | 1 |
110,039 | 11,685,730,498 | IssuesEvent | 2020-03-05 09:37:40 | bcryptoregulatory/skzi-requirements | https://api.github.com/repos/bcryptoregulatory/skzi-requirements | closed | Избыточные требования к алгоритмам криптографического токена | documentation | В столбце с требованиями к криптографическим алгоритмам средства указывают только алгоритмы, которые предоставляются пользователю в качестве сервиса: в случае криптографического токена алгоритмы АШ и АИ1 пользователю не предоставляются и носят внутренний служебный характер | 1.0 | Избыточные требования к алгоритмам криптографического токена - В столбце с требованиями к криптографическим алгоритмам средства указывают только алгоритмы, которые предоставляются пользователю в качестве сервиса: в случае криптографического токена алгоритмы АШ и АИ1 пользователю не предоставляются и носят внутренний служебный характер | non_main | избыточные требования к алгоритмам криптографического токена в столбце с требованиями к криптографическим алгоритмам средства указывают только алгоритмы которые предоставляются пользователю в качестве сервиса в случае криптографического токена алгоритмы аш и пользователю не предоставляются и носят внутренний служебный характер | 0 |
256,752 | 19,457,267,455 | IssuesEvent | 2021-12-23 01:31:00 | seanpm2001/WacOS | https://api.github.com/repos/seanpm2001/WacOS | opened | [Maintenance] the project Wiki needs to be updated | documentation enhancement Wiki Issue |
***
### [Maintenance] the project Wiki needs to be updated
The project Wiki is a key part of the project and hasn't been updated in over a month. There are several things that need to be written about, including:
### New articles
- [ ] Apple Typography
- - [ ] Font:Chicago
- - [ ] Font:Monaco
- - [ ] Font:Garamond
- - [ ] Font:Charcoal
- - [ ] Font:Geneva
- - [ ] Font:Shaston
- - [ ] Font:Espy Sans
- - [ ] Font:Lucida Grande
- - [ ] Font:Podium Sans
- - [ ] Font:Helvetica
- - [ ] Font:San Francisco
- - [ ] Font:New York
- - [ ] Font:Motter Tektura
- - [ ] Font:Gill Sans
- - [ ] Font:Myriad
- [ ] Apple Hardware
- - [ ] Pippin
- - [ ] Pippin Bandai @world
- - [ ] Apple I
- - [ ] Apple II
- - [ ] Various other pieces of Apple hardware
Also, articles need to be updated. There are too many to list here.
I am also looking for public collaborators, more info can be found [here](https://github.com/seanpm2001/WacOS/issues/7/).
***
| 1.0 | [Maintenance] the project Wiki needs to be updated -
***
### [Maintenance] the project Wiki needs to be updated
The project Wiki is a key part of the project and hasn't been updated in over a month. There are several things that need to be written about, including:
### New articles
- [ ] Apple Typography
- - [ ] Font:Chicago
- - [ ] Font:Monaco
- - [ ] Font:Garamond
- - [ ] Font:Charcoal
- - [ ] Font:Geneva
- - [ ] Font:Shaston
- - [ ] Font:Espy Sans
- - [ ] Font:Lucida Grande
- - [ ] Font:Podium Sans
- - [ ] Font:Helvetica
- - [ ] Font:San Francisco
- - [ ] Font:New York
- - [ ] Font:Motter Tektura
- - [ ] Font:Gill Sans
- - [ ] Font:Myriad
- [ ] Apple Hardware
- - [ ] Pippin
- - [ ] Pippin Bandai @world
- - [ ] Apple I
- - [ ] Apple II
- - [ ] Various other pieces of Apple hardware
Also, articles need to be updated. There are too many to list here.
I am also looking for public collaborators, more info can be found [here](https://github.com/seanpm2001/WacOS/issues/7/).
***
| non_main | the project wiki needs to be updated the project wiki needs to be updated the project wiki is a key part of the project and hasn t been updated in over a month there are several things that need to be written about including new articles apple typography font chicago font monaco font garamond font charcoal font geneva font shaston font espy sans font lucida grande font podium sans font helvetica font san francisco font new york font motter tektura font gill sans font myriad apple hardware pippin pippin bandai world apple i apple ii various other pieces of apple hardware also articles need to be updated there are too many to list here i am also looking for public collaborators more info can be found | 0 |
2,797 | 10,021,126,362 | IssuesEvent | 2019-07-16 14:03:16 | ipfs/package-managers | https://api.github.com/repos/ipfs/package-managers | closed | Adding IPFS support to Homebrew | Audience: Package manager maintainers Package Manager: Homebrew Type: Experiment | As mentioned in https://docs.google.com/document/d/1-HtUiRpMzYq9to56ShCGyCr-NCZ6TR49Zl-b5HHJdm0/edit#heading=h.m17gf4vdw67, adding IPFS support to "portable registry" package managers is fairly easy due to the fact that you can fork the whole registry and add hashes to the metadata.
As an experiment, I thought I'd give it a try with Homebrew.
I picked a simple ["Formula"](https://github.com/Homebrew/homebrew-core/blob/master/Formula/sl.rb) and added the tarball to ipfs:
$ ipfs urlstore add https://github.com/mtoyoda/sl/archive/5.02.tar.gz
Which gives you the hash: `zb2rhYgkf18LAi1FjyZPZfU43UrxwBZgV59XrpKR3mCXfNAyX`
I then added that hash to the formula in my fork of homebrew-core: https://github.com/andrew/homebrew-core/commit/50b776ce8f2841457610a232a815267910a45f5c

This extra field won't just work with the brew command, so I forked and patched homebrew to add support for the `ipfs` method: https://github.com/andrew/brew/commit/17492b84b5da49181eb9a6c601747adc5ebafc96
This quick hack makes a few of assumptions:
- you have the `ipfs` command installed in `/usr/local/bin/ipfs`, which is the default on mac when installed using homebrew
- you have `ipfs daemon` running, it doesn't try to start it for you
- you want to build from source, it doesn't work with prebuilt binaries, which homebrew calls `bottles`
Luckily with homebrew, you can install and run it from any directory, so it's easy to download my fork:
$ git clone https://github.com/andrew/brew && cd brew
Then install my forked `sl` formula via IPFS:
$ ./bin/brew reinstall andrew/core/sl --build-from-source

Also of note, homebrew checks the file that was downloaded from IPFS against the exist sha256 integrity hash and it matches 🎉
Then you can run it:
$ ./bin/sl

IPFS hashes could easily be added to most of the other ~4,700 formula homebrew core using a script to find the url of each formulas tarball, add it to ipfs, update the formula with the hash and commit, and then have a script to watch the homebrew repo for new formula updates.
I also briefly mentioned it to @mikemcquaid over SMS, overall he thought it was very neat, adding:
> FYI to avoid disappointment: we’re unlikely to add a new download strategy to Homebrew/brew until it’s required by a core formula as the main download method.
Some other homebrew notes:
Homebrew is very heavily used in the mac community for installing databases, languages and other system level packages, they also have public install analytics: https://formulae.brew.sh/analytics/install/30d/

They're also moving towards offering more precompiled binaries (bottles), currently hosted on https://bintray.com/homebrew
Homebrew is also preinstalled on all Travis CI and Circle CI MacOS projects | True | Adding IPFS support to Homebrew - As mentioned in https://docs.google.com/document/d/1-HtUiRpMzYq9to56ShCGyCr-NCZ6TR49Zl-b5HHJdm0/edit#heading=h.m17gf4vdw67, adding IPFS support to "portable registry" package managers is fairly easy due to the fact that you can fork the whole registry and add hashes to the metadata.
As an experiment, I thought I'd give it a try with Homebrew.
I picked a simple ["Formula"](https://github.com/Homebrew/homebrew-core/blob/master/Formula/sl.rb) and added the tarball to ipfs:
$ ipfs urlstore add https://github.com/mtoyoda/sl/archive/5.02.tar.gz
Which gives you the hash: `zb2rhYgkf18LAi1FjyZPZfU43UrxwBZgV59XrpKR3mCXfNAyX`
I then added that hash to the formula in my fork of homebrew-core: https://github.com/andrew/homebrew-core/commit/50b776ce8f2841457610a232a815267910a45f5c

This extra field won't just work with the brew command, so I forked and patched homebrew to add support for the `ipfs` method: https://github.com/andrew/brew/commit/17492b84b5da49181eb9a6c601747adc5ebafc96
This quick hack makes a few of assumptions:
- you have the `ipfs` command installed in `/usr/local/bin/ipfs`, which is the default on mac when installed using homebrew
- you have `ipfs daemon` running, it doesn't try to start it for you
- you want to build from source, it doesn't work with prebuilt binaries, which homebrew calls `bottles`
Luckily with homebrew, you can install and run it from any directory, so it's easy to download my fork:
$ git clone https://github.com/andrew/brew && cd brew
Then install my forked `sl` formula via IPFS:
$ ./bin/brew reinstall andrew/core/sl --build-from-source

Also of note, homebrew checks the file that was downloaded from IPFS against the exist sha256 integrity hash and it matches 🎉
Then you can run it:
$ ./bin/sl

IPFS hashes could easily be added to most of the other ~4,700 formula homebrew core using a script to find the url of each formulas tarball, add it to ipfs, update the formula with the hash and commit, and then have a script to watch the homebrew repo for new formula updates.
I also briefly mentioned it to @mikemcquaid over SMS, overall he thought it was very neat, adding:
> FYI to avoid disappointment: we’re unlikely to add a new download strategy to Homebrew/brew until it’s required by a core formula as the main download method.
Some other homebrew notes:
Homebrew is very heavily used in the mac community for installing databases, languages and other system level packages, they also have public install analytics: https://formulae.brew.sh/analytics/install/30d/

They're also moving towards offering more precompiled binaries (bottles), currently hosted on https://bintray.com/homebrew
Homebrew is also preinstalled on all Travis CI and Circle CI MacOS projects | main | adding ipfs support to homebrew as mentioned in adding ipfs support to portable registry package managers is fairly easy due to the fact that you can fork the whole registry and add hashes to the metadata as an experiment i thought i d give it a try with homebrew i picked a simple and added the tarball to ipfs ipfs urlstore add which gives you the hash i then added that hash to the formula in my fork of homebrew core this extra field won t just work with the brew command so i forked and patched homebrew to add support for the ipfs method this quick hack makes a few of assumptions you have the ipfs command installed in usr local bin ipfs which is the default on mac when installed using homebrew you have ipfs daemon running it doesn t try to start it for you you want to build from source it doesn t work with prebuilt binaries which homebrew calls bottles luckily with homebrew you can install and run it from any directory so it s easy to download my fork git clone cd brew then install my forked sl formula via ipfs bin brew reinstall andrew core sl build from source also of note homebrew checks the file that was downloaded from ipfs against the exist integrity hash and it matches 🎉 then you can run it bin sl ipfs hashes could easily be added to most of the other formula homebrew core using a script to find the url of each formulas tarball add it to ipfs update the formula with the hash and commit and then have a script to watch the homebrew repo for new formula updates i also briefly mentioned it to mikemcquaid over sms overall he thought it was very neat adding fyi to avoid disappointment we’re unlikely to add a new download strategy to homebrew brew until it’s required by a core formula as the main download method some other homebrew notes homebrew is very heavily used in the mac community for installing databases languages and other system level packages they also have public install analytics they re also moving towards offering more precompiled binaries bottles currently hosted on homebrew is also preinstalled on all travis ci and circle ci macos projects | 1 |
26,372 | 4,682,654,841 | IssuesEvent | 2016-10-09 11:19:58 | phingofficial/phing | https://api.github.com/repos/phingofficial/phing | opened | PHPUnitReportTask fails with XSLTProcessor::importStylesheet() unable to read phar:/usr/local/bin/phing/etc/str.replace.function.xsl (Trac #1240) | defect Incomplete Migration Migrated from Trac phing-tasks-phpunit | Migrated from http://localhost:3530/trac/ticket/1240
```json
{
"status": "new",
"changetime": "2016-09-29T08:11:42",
"description": "Hi,\n\nmy build fails with:\n\nXSLTProcessor::importStylesheet(): I/O warning : failed to load external entity \"phar:/usr/local/bin/phing/etc/str.replace.function.xsl\", file: phar:///usr/local/bin/phing/classes/phing/tasks/ext/phpunit/PHPUnitReportTask.php\n\nSee buildfile below: It successfully executed the tests and creates a valid logfile.xml as expected. Converting test results with <<phpunitreport> then fails.\n\nUsing both Phing and PHPUnit PHAR archives.\n\nAfter extracing the Phing PHAR, I see that is DOES contain phar:/usr/local/bin/phing/etc/str.replace.function.xsl.\n\nEnvironment:\nPhing 2.13.0\nPHP 5.6.14-0+deb8u1 (cli) (built: Oct 4 2015 16:13:10)\nlibxslt Version 1.1.28 with XSL enabled\nRegistered PHP Streams: https, ftps, compress.zlib, compress.bzip2, php, file, glob, data, http, ftp, phar, zip\nPHPUnit 4.8.7 by Sebastian Bergmann and contributors.\nLinux version 3.16.0-4-amd64 (debian-kernel@lists.debian.org) (gcc version 4.8.4 (Debian 4.8.4-1) ) #1 SMP Debian 3.16.7-ckt11-1+deb8u3 (2015-08-04)\n\nBuildfile (paths changed, please adjust as needed):\n\n<?xml version=\"1.0\"?>\n\n<project name=\"MyProject\" basedir=\".\" default=\"tests_with_report\">\n <target name=\"prepare\">\n <mkdir dir=\"testreport\"/>\n </target>\n\n <target name=\"clean\">\n <delete dir=\"testreport\"/>\n </target>\n\n <target name=\"tests_with_report\" depends=\"prepare\">\n <phpunit bootstrap=\"bootstrap.php\" pharlocation=\"/usr/local/bin/phpunit\">\n <batchtest>\n <fileset dir=\"tests\">\n <include name=\"config/*Test.php\"/>\n </fileset>\n </batchtest>\n\n <formatter type=\"xml\" todir=\"testreport\" outfile=\"logfile.xml\" />\n </phpunit>\n\n <phpunitreport infile=\"logfile.xml\"\n format=\"frames\"\n todir=\"testreport\"/>\n\n\n </target>\n</project>\n\n-----\n\n<phpunitreport> is successfully executing and creating html base report on a different system with PHP 5.4.45-0+deb7u2.",
"reporter": "Ulrich",
"cc": "",
"resolution": "",
"_ts": "1475136702149451",
"component": "phing-tasks-phpunit",
"summary": "PHPUnitReportTask fails with XSLTProcessor::importStylesheet() unable to read phar:/usr/local/bin/phing/etc/str.replace.function.xsl",
"priority": "minor",
"keywords": "phar XSLTProcessor PHPUnit phpunitreport I/O warning",
"version": "2.13.0",
"time": "2016-01-15T11:45:41",
"milestone": "2.16.0",
"owner": "",
"type": "defect"
}
```
| 1.0 | PHPUnitReportTask fails with XSLTProcessor::importStylesheet() unable to read phar:/usr/local/bin/phing/etc/str.replace.function.xsl (Trac #1240) - Migrated from http://localhost:3530/trac/ticket/1240
```json
{
"status": "new",
"changetime": "2016-09-29T08:11:42",
"description": "Hi,\n\nmy build fails with:\n\nXSLTProcessor::importStylesheet(): I/O warning : failed to load external entity \"phar:/usr/local/bin/phing/etc/str.replace.function.xsl\", file: phar:///usr/local/bin/phing/classes/phing/tasks/ext/phpunit/PHPUnitReportTask.php\n\nSee buildfile below: It successfully executed the tests and creates a valid logfile.xml as expected. Converting test results with <<phpunitreport> then fails.\n\nUsing both Phing and PHPUnit PHAR archives.\n\nAfter extracing the Phing PHAR, I see that is DOES contain phar:/usr/local/bin/phing/etc/str.replace.function.xsl.\n\nEnvironment:\nPhing 2.13.0\nPHP 5.6.14-0+deb8u1 (cli) (built: Oct 4 2015 16:13:10)\nlibxslt Version 1.1.28 with XSL enabled\nRegistered PHP Streams: https, ftps, compress.zlib, compress.bzip2, php, file, glob, data, http, ftp, phar, zip\nPHPUnit 4.8.7 by Sebastian Bergmann and contributors.\nLinux version 3.16.0-4-amd64 (debian-kernel@lists.debian.org) (gcc version 4.8.4 (Debian 4.8.4-1) ) #1 SMP Debian 3.16.7-ckt11-1+deb8u3 (2015-08-04)\n\nBuildfile (paths changed, please adjust as needed):\n\n<?xml version=\"1.0\"?>\n\n<project name=\"MyProject\" basedir=\".\" default=\"tests_with_report\">\n <target name=\"prepare\">\n <mkdir dir=\"testreport\"/>\n </target>\n\n <target name=\"clean\">\n <delete dir=\"testreport\"/>\n </target>\n\n <target name=\"tests_with_report\" depends=\"prepare\">\n <phpunit bootstrap=\"bootstrap.php\" pharlocation=\"/usr/local/bin/phpunit\">\n <batchtest>\n <fileset dir=\"tests\">\n <include name=\"config/*Test.php\"/>\n </fileset>\n </batchtest>\n\n <formatter type=\"xml\" todir=\"testreport\" outfile=\"logfile.xml\" />\n </phpunit>\n\n <phpunitreport infile=\"logfile.xml\"\n format=\"frames\"\n todir=\"testreport\"/>\n\n\n </target>\n</project>\n\n-----\n\n<phpunitreport> is successfully executing and creating html base report on a different system with PHP 5.4.45-0+deb7u2.",
"reporter": "Ulrich",
"cc": "",
"resolution": "",
"_ts": "1475136702149451",
"component": "phing-tasks-phpunit",
"summary": "PHPUnitReportTask fails with XSLTProcessor::importStylesheet() unable to read phar:/usr/local/bin/phing/etc/str.replace.function.xsl",
"priority": "minor",
"keywords": "phar XSLTProcessor PHPUnit phpunitreport I/O warning",
"version": "2.13.0",
"time": "2016-01-15T11:45:41",
"milestone": "2.16.0",
"owner": "",
"type": "defect"
}
```
| non_main | phpunitreporttask fails with xsltprocessor importstylesheet unable to read phar usr local bin phing etc str replace function xsl trac migrated from json status new changetime description hi n nmy build fails with n nxsltprocessor importstylesheet i o warning failed to load external entity phar usr local bin phing etc str replace function xsl file phar usr local bin phing classes phing tasks ext phpunit phpunitreporttask php n nsee buildfile below it successfully executed the tests and creates a valid logfile xml as expected converting test results with then fails n nusing both phing and phpunit phar archives n nafter extracing the phing phar i see that is does contain phar usr local bin phing etc str replace function xsl n nenvironment nphing nphp cli built oct nlibxslt version with xsl enabled nregistered php streams https ftps compress zlib compress php file glob data http ftp phar zip nphpunit by sebastian bergmann and contributors nlinux version debian kernel lists debian org gcc version debian smp debian n nbuildfile paths changed please adjust as needed n n n n n n n n n n n n n n n n n n n n n n n n n n n n n n n n is successfully executing and creating html base report on a different system with php reporter ulrich cc resolution ts component phing tasks phpunit summary phpunitreporttask fails with xsltprocessor importstylesheet unable to read phar usr local bin phing etc str replace function xsl priority minor keywords phar xsltprocessor phpunit phpunitreport i o warning version time milestone owner type defect | 0 |
532,360 | 15,554,722,293 | IssuesEvent | 2021-03-16 04:32:56 | worldanvil/worldanvil-bug-tracker | https://api.github.com/repos/worldanvil/worldanvil-bug-tracker | closed | Myth template: placeholder texts cut off | Feature: Articles & Templates Priority: Low Severity: Trivial Status: Will Not Fix Type: Typo | On the Myth article template, many of the placeholder texts are cut off, "Make y" instead of "Make your choice".

| 1.0 | Myth template: placeholder texts cut off - On the Myth article template, many of the placeholder texts are cut off, "Make y" instead of "Make your choice".

| non_main | myth template placeholder texts cut off on the myth article template many of the placeholder texts are cut off make y instead of make your choice | 0 |
1,043 | 4,847,233,341 | IssuesEvent | 2016-11-10 14:25:37 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | cloudformation module on Ansible 2.2.0 throws "PhysicalResourceId" error intermittently. | affects_2.2 aws bug_report cloud in progress waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
cloudformation
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ubuntu 14.04
##### SUMMARY
<!--- Explain the problem briefly -->
cloudformation module on Ansible 2.2.0 throws "PhysicalResourceId" error intermittently.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Push a stack with Ansible cloudformation module on Ansible 2.2.0
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: KeyError: 'PhysicalResourceId'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_Kx3c0c/ansible_module_cloudformation.py\", line 483, in <module>\n main()\n File \"/tmp/ansible_Kx3c0c/ansible_module_cloudformation.py\", line 450, in main\n \"physical_resource_id\": res['PhysicalResourceId'],\nKeyError: 'PhysicalResourceId'\n", "module_stdout": "", "msg": "MODULE FAILURE"}
to retry, use: --limit
```
| True | cloudformation module on Ansible 2.2.0 throws "PhysicalResourceId" error intermittently. - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
cloudformation
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ubuntu 14.04
##### SUMMARY
<!--- Explain the problem briefly -->
cloudformation module on Ansible 2.2.0 throws "PhysicalResourceId" error intermittently.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Push a stack with Ansible cloudformation module on Ansible 2.2.0
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: KeyError: 'PhysicalResourceId'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_Kx3c0c/ansible_module_cloudformation.py\", line 483, in <module>\n main()\n File \"/tmp/ansible_Kx3c0c/ansible_module_cloudformation.py\", line 450, in main\n \"physical_resource_id\": res['PhysicalResourceId'],\nKeyError: 'PhysicalResourceId'\n", "module_stdout": "", "msg": "MODULE FAILURE"}
to retry, use: --limit
```
| main | cloudformation module on ansible throws physicalresourceid error intermittently issue type bug report component name cloudformation ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific ubuntu summary cloudformation module on ansible throws physicalresourceid error intermittently steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used push a stack with ansible cloudformation module on ansible expected results actual results an exception occurred during task execution to see the full traceback use vvv the error was keyerror physicalresourceid fatal failed changed false failed true module stderr traceback most recent call last n file tmp ansible ansible module cloudformation py line in n main n file tmp ansible ansible module cloudformation py line in main n physical resource id res nkeyerror physicalresourceid n module stdout msg module failure to retry use limit | 1 |
145,479 | 5,576,231,268 | IssuesEvent | 2017-03-28 05:51:53 | Citadel-Station-13/Citadel-Station-13 | https://api.github.com/repos/Citadel-Station-13/Citadel-Station-13 | closed | Sprite gets broken after augmentation | Bug Icon/Art Priority: High | Not sure how to reproduce it myself, but another player reported this. After being augmented their sprite has pretty much disappeared.
Picture for reference: https://gyazo.com/bc5581138b24ed0bb1a71c8ada2df104 | 1.0 | Sprite gets broken after augmentation - Not sure how to reproduce it myself, but another player reported this. After being augmented their sprite has pretty much disappeared.
Picture for reference: https://gyazo.com/bc5581138b24ed0bb1a71c8ada2df104 | non_main | sprite gets broken after augmentation not sure how to reproduce it myself but another player reported this after being augmented their sprite has pretty much disappeared picture for reference | 0 |
210,842 | 16,130,709,078 | IssuesEvent | 2021-04-29 04:00:13 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: sqlsmith/setup=empty/setting=no-mutations failed | C-test-failure O-roachtest O-robot branch-master | roachtest.sqlsmith/setup=empty/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2917180&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=2917180&tab=artifacts#/sqlsmith/setup=empty/setting=no-mutations) on master @ [9e821aa456bfbe8b5ec3a8ef79ae3954d52cb675](https://github.com/cockroachdb/cockroach/commits/9e821aa456bfbe8b5ec3a8ef79ae3954d52cb675):
```
('10dcf1c7-bc0e-4240-8dd4-1cded245becc':::UUID), ('a015aa85-d559-466f-861b-5fb6fbae2c28':::UUID)
)
AS tab_10765 (col_18487)
UNION ALL SELECT NULL AS col_18488 FROM with_3274 AS cte_ref_935 WHERE true
),
with_3276 (col_18489, col_18490)
AS (
SELECT
*
FROM
(
VALUES
(e'\x0fIb\x06b\x1a':::STRING, '30 years 8 mons 329 days 00:43:36.121574':::INTERVAL),
(e'\x00':::STRING, '55 years 10 mons 424 days 00:43:07.936553':::INTERVAL)
)
AS tab_10766 (col_18489, col_18490)
)
SELECT
'3d3bb807-e74f-47e1-b504-f0fe2a4ad91f':::UUID AS col_18491
FROM
with_3275 AS cte_ref_936
ORDER BY
cte_ref_936.col_18487 DESC, cte_ref_936.col_18487 ASC, cte_ref_936.col_18487 DESC, cte_ref_936.col_18487;
cluster.go:1716,context.go:140,cluster.go:1705,test_runner.go:858: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2917180-1619205254-05-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 1: dead
2: 11268
4: 10200
3: 11386
Error: UNCLASSIFIED_PROBLEM: 1: dead
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
-- stack trace:
| main.glob..func14
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1162
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:284
| github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:856
| github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:960
| github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:897
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:2057
| runtime.main
| /usr/local/go/src/runtime/proc.go:204
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (3) 1: dead
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError
```
<details><summary>Reproduce</summary>
<p>
<p>To reproduce, try:
```bash
# From https://go.crdb.dev/p/roachstress, perhaps edited lightly.
caffeinate ./roachstress.sh sqlsmith/setup=empty/setting=no-mutations
```
</p>
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=empty/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: sqlsmith/setup=empty/setting=no-mutations failed - roachtest.sqlsmith/setup=empty/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2917180&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=2917180&tab=artifacts#/sqlsmith/setup=empty/setting=no-mutations) on master @ [9e821aa456bfbe8b5ec3a8ef79ae3954d52cb675](https://github.com/cockroachdb/cockroach/commits/9e821aa456bfbe8b5ec3a8ef79ae3954d52cb675):
```
('10dcf1c7-bc0e-4240-8dd4-1cded245becc':::UUID), ('a015aa85-d559-466f-861b-5fb6fbae2c28':::UUID)
)
AS tab_10765 (col_18487)
UNION ALL SELECT NULL AS col_18488 FROM with_3274 AS cte_ref_935 WHERE true
),
with_3276 (col_18489, col_18490)
AS (
SELECT
*
FROM
(
VALUES
(e'\x0fIb\x06b\x1a':::STRING, '30 years 8 mons 329 days 00:43:36.121574':::INTERVAL),
(e'\x00':::STRING, '55 years 10 mons 424 days 00:43:07.936553':::INTERVAL)
)
AS tab_10766 (col_18489, col_18490)
)
SELECT
'3d3bb807-e74f-47e1-b504-f0fe2a4ad91f':::UUID AS col_18491
FROM
with_3275 AS cte_ref_936
ORDER BY
cte_ref_936.col_18487 DESC, cte_ref_936.col_18487 ASC, cte_ref_936.col_18487 DESC, cte_ref_936.col_18487;
cluster.go:1716,context.go:140,cluster.go:1705,test_runner.go:858: dead node detection: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod monitor teamcity-2917180-1619205254-05-n4cpu4 --oneshot --ignore-empty-nodes: exit status 1 1: dead
2: 11268
4: 10200
3: 11386
Error: UNCLASSIFIED_PROBLEM: 1: dead
(1) UNCLASSIFIED_PROBLEM
Wraps: (2) attached stack trace
-- stack trace:
| main.glob..func14
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1162
| main.wrap.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:284
| github.com/spf13/cobra.(*Command).execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:856
| github.com/spf13/cobra.(*Command).ExecuteC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:960
| github.com/spf13/cobra.(*Command).Execute
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:897
| main.main
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:2057
| runtime.main
| /usr/local/go/src/runtime/proc.go:204
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (3) 1: dead
Error types: (1) errors.Unclassified (2) *withstack.withStack (3) *errutil.leafError
```
<details><summary>Reproduce</summary>
<p>
<p>To reproduce, try:
```bash
# From https://go.crdb.dev/p/roachstress, perhaps edited lightly.
caffeinate ./roachstress.sh sqlsmith/setup=empty/setting=no-mutations
```
</p>
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=empty/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_main | roachtest sqlsmith setup empty setting no mutations failed roachtest sqlsmith setup empty setting no mutations with on master uuid uuid as tab col union all select null as col from with as cte ref where true with col col as select from values e string years mons days interval e string years mons days interval as tab col col select uuid as col from with as cte ref order by cte ref col desc cte ref col asc cte ref col desc cte ref col cluster go context go cluster go test runner go dead node detection home agent work go src github com cockroachdb cockroach bin roachprod monitor teamcity oneshot ignore empty nodes exit status dead error unclassified problem dead unclassified problem wraps attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types errors unclassified withstack withstack errutil leaferror reproduce to reproduce try bash from perhaps edited lightly caffeinate roachstress sh sqlsmith setup empty setting no mutations cc cockroachdb sql queries | 0 |
361 | 3,315,869,788 | IssuesEvent | 2015-11-06 14:30:31 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | closed | java version "1.8.0_66" in brew cask | awaiting maintainer feedback | After I run **brew cask install java** to install the 1.8.0_66 version, I found that **Java update** in **Java control panel** didn't work.
Is there a bug about this java version or brew cask of installing java? | True | java version "1.8.0_66" in brew cask - After I run **brew cask install java** to install the 1.8.0_66 version, I found that **Java update** in **Java control panel** didn't work.
Is there a bug about this java version or brew cask of installing java? | main | java version in brew cask after i run brew cask install java to install the version i found that java update in java control panel didn t work is there a bug about this java version or brew cask of installing java | 1 |
4,072 | 19,213,006,431 | IssuesEvent | 2021-12-07 05:40:33 | adda-team/adda | https://api.github.com/repos/adda-team/adda | opened | Remove deprecated specification of beam center as argument to -beam | comp-UI maintainability | Such specification has been marked as deprecated by #304 (and #285). At some time it should be removed completely. | True | Remove deprecated specification of beam center as argument to -beam - Such specification has been marked as deprecated by #304 (and #285). At some time it should be removed completely. | main | remove deprecated specification of beam center as argument to beam such specification has been marked as deprecated by and at some time it should be removed completely | 1 |
4,195 | 20,591,963,152 | IssuesEvent | 2022-03-05 00:39:27 | HPCL/code-analysis | https://api.github.com/repos/HPCL/code-analysis | closed | CWE-484 Omitted Break Statement in Switch | CLAIMED ISO/IEC 5055:2021 SwitchStatement WEAKNESS CATEGORY: MAINTAINABILITY | **Reference**
[https://cwe.mitre.org/data/definitions/484](https://cwe.mitre.org/data/definitions/484)
**Roles**
- the *SwitchStatement*
**Detection Patterns**
- 8.2.64 ASCQM Use Break in Switch Statement | True | CWE-484 Omitted Break Statement in Switch - **Reference**
[https://cwe.mitre.org/data/definitions/484](https://cwe.mitre.org/data/definitions/484)
**Roles**
- the *SwitchStatement*
**Detection Patterns**
- 8.2.64 ASCQM Use Break in Switch Statement | main | cwe omitted break statement in switch reference roles the switchstatement detection patterns ascqm use break in switch statement | 1 |
52,741 | 10,919,607,074 | IssuesEvent | 2019-11-21 19:23:42 | microsoft/AdaptiveCards | https://api.github.com/repos/microsoft/AdaptiveCards | closed | [JS][ShowCards] Show card inputs are cleared if closed and reopened | Bug Status-In Code Review Triage-Approved for Fix | # Platform
What platform is your issue or question related to? (Delete other platforms).
* JavaScript
# Details
If you show a show card, enter an input in that show card, and then close and re-open the show card, the inputted value is cleared:
1. In the below card, click the "Click here to personalize your item" ShowCard button.
2. Type something into the input
3. Close the show card and reopen it
Result: Typed input has been lost.
``` json
{
"type": "AdaptiveCard",
"version": "1.0",
"body": [
{
"type": "TextBlock",
"text": "Which color would you like?"
},
{
"type": "Input.ChoiceSet",
"placeholder": "Placeholder text",
"choices": [
{
"title": "Red",
"value": "Red"
},
{
"title": "Blue",
"value": "Blue"
}
],
"style": "expanded",
"id": "color"
},
{
"type": "ActionSet",
"actions": [
{
"type": "Action.ShowCard",
"title": "Click here to personalize your item",
"card": {
"type": "AdaptiveCard",
"$schema": "http://adaptivecards.io/schemas/adaptive-card.json",
"body": [
{
"type": "TextBlock",
"text": "Name (20 characters max):"
},
{
"type": "Input.Text",
"placeholder": "Name",
"id": "personalizationName"
}
]
}
}
]
},
{
"type": "TextBlock",
"text": "Choose your shipping type:"
},
{
"type": "Input.ChoiceSet",
"placeholder": "Placeholder text",
"choices": [
{
"title": "Standard (5-7 business days)",
"value": "Standard"
},
{
"title": "Express (2-3 business days)",
"value": "Express"
},
{
"title": "Overnight (next business day)",
"value": "Overnight"
}
],
"style": "expanded",
"id": "shipping"
}
],
"$schema": "http://adaptivecards.io/schemas/adaptive-card.json",
"actions": [
{
"type": "Action.Submit",
"title": "Submit Order"
}
]
}
```
| 1.0 | [JS][ShowCards] Show card inputs are cleared if closed and reopened - # Platform
What platform is your issue or question related to? (Delete other platforms).
* JavaScript
# Details
If you show a show card, enter an input in that show card, and then close and re-open the show card, the inputted value is cleared:
1. In the below card, click the "Click here to personalize your item" ShowCard button.
2. Type something into the input
3. Close the show card and reopen it
Result: Typed input has been lost.
``` json
{
"type": "AdaptiveCard",
"version": "1.0",
"body": [
{
"type": "TextBlock",
"text": "Which color would you like?"
},
{
"type": "Input.ChoiceSet",
"placeholder": "Placeholder text",
"choices": [
{
"title": "Red",
"value": "Red"
},
{
"title": "Blue",
"value": "Blue"
}
],
"style": "expanded",
"id": "color"
},
{
"type": "ActionSet",
"actions": [
{
"type": "Action.ShowCard",
"title": "Click here to personalize your item",
"card": {
"type": "AdaptiveCard",
"$schema": "http://adaptivecards.io/schemas/adaptive-card.json",
"body": [
{
"type": "TextBlock",
"text": "Name (20 characters max):"
},
{
"type": "Input.Text",
"placeholder": "Name",
"id": "personalizationName"
}
]
}
}
]
},
{
"type": "TextBlock",
"text": "Choose your shipping type:"
},
{
"type": "Input.ChoiceSet",
"placeholder": "Placeholder text",
"choices": [
{
"title": "Standard (5-7 business days)",
"value": "Standard"
},
{
"title": "Express (2-3 business days)",
"value": "Express"
},
{
"title": "Overnight (next business day)",
"value": "Overnight"
}
],
"style": "expanded",
"id": "shipping"
}
],
"$schema": "http://adaptivecards.io/schemas/adaptive-card.json",
"actions": [
{
"type": "Action.Submit",
"title": "Submit Order"
}
]
}
```
| non_main | show card inputs are cleared if closed and reopened platform what platform is your issue or question related to delete other platforms javascript details if you show a show card enter an input in that show card and then close and re open the show card the inputted value is cleared in the below card click the click here to personalize your item showcard button type something into the input close the show card and reopen it result typed input has been lost json type adaptivecard version body type textblock text which color would you like type input choiceset placeholder placeholder text choices title red value red title blue value blue style expanded id color type actionset actions type action showcard title click here to personalize your item card type adaptivecard schema body type textblock text name characters max type input text placeholder name id personalizationname type textblock text choose your shipping type type input choiceset placeholder placeholder text choices title standard business days value standard title express business days value express title overnight next business day value overnight style expanded id shipping schema actions type action submit title submit order | 0 |
4,473 | 23,335,801,265 | IssuesEvent | 2022-08-09 09:49:25 | precice/precice | https://api.github.com/repos/precice/precice | opened | Simplification of EventTimings | enhancement maintainability dependencies | _I open this issue in this repo to preserve the information and make it easier to find from people running into issues regarding this. It also may impact the preCICE lib in the future._
**Please describe the problem you are trying to solve.**
The EventTimings provide a system to:
* measure named sections in the code
* attach data to these sections (used in PETSc RBF mappings)
* synchronize the communicator prior to the recording using a barrier on requested aka `syncmode`
* aggregate and normalize these measurements across ranks
* write a summary of the results to a file as a text table
* write the aggregate results to a file as json
The additional `events2trace` script formats and merges multiple of these outputs into a single eventstracing json file that can be visualized with various tools.
Concerns of this approach:
* The EventsTimings need a way to synchronize all ranks. This currently requires passing a custom MPI comm. Not using MPI only supports a single rank.
* The data aggregation happens during the finalization of preCICE, which is a collective operation on the communicator. If any issue occurs with this communicator during the lifetime of preCICE, then the collective will fail, resulting in a crash/error.
* If preCICE runs into any error, then the events won't be aggregated nor written to a file.
* preCICE requires an additional dependency for the sole purpose of writing the aggregated data to disk. We are currently using a checked-in version of the json library, which will at some point collide with other versions on the system leading to strange problems such as #527 or https://github.com/precice/openfoam-adapter/issues/238 .
**Describe the solution you propose.**
1. Simplify the Events in preCICE as much as possible.
* The synchronization can be handled fully by preCICE, as IntraComm provides a barrier method. Its implementation also works if preCICE is compiled without MPI.
* Use independent rank files, essentially removing the aggregation from the preCICE core. This allows to output events on error.
* Optionally write these files continuously during the lifetime of the program, allowing to inspect the events on a crash. We could even implement a block-wise write to reduce the memory overhead.
* This serialization is so simple that it doesn't require a special library. ( Similar to the clang time-tracing implementation. )
* The above results in additional IO. So, a configuration option to disable the tracing could be beneficial.
2. Move the functionality to normalize, aggregate and format to a separate script (or `precice-tools`)
* Use python pandas or similar for normalizing and aggregating the data.
* Ship this as an extra tool in `/usr/share/precice` or similar.
* This allows us to remove the json dependency from the project.
* Alternatively, move the existing C++ implementation from the preCICE code into a separate executable, or into `precice-tools`.
* The python version would allow everyone to easily add custom functionality such as
* plotting given timings over time
* analyse the comm establishment to detect filesystem issues on some nodes
* find an imbalance of mapping cost over ranks of a participant
**Describe alternatives you've considered**
* Move the events2trace script into the preCICE library, essentially completely integrating the external project.
* Reimplement the events2trace as a subcommand of `precice-tools`.
**Additional context**
* https://github.com/precice/EventTimings/issues/17
* #419 | True | Simplification of EventTimings - _I open this issue in this repo to preserve the information and make it easier to find from people running into issues regarding this. It also may impact the preCICE lib in the future._
**Please describe the problem you are trying to solve.**
The EventTimings provide a system to:
* measure named sections in the code
* attach data to these sections (used in PETSc RBF mappings)
* synchronize the communicator prior to the recording using a barrier on requested aka `syncmode`
* aggregate and normalize these measurements across ranks
* write a summary of the results to a file as a text table
* write the aggregate results to a file as json
The additional `events2trace` script formats and merges multiple of these outputs into a single eventstracing json file that can be visualized with various tools.
Concerns of this approach:
* The EventsTimings need a way to synchronize all ranks. This currently requires passing a custom MPI comm. Not using MPI only supports a single rank.
* The data aggregation happens during the finalization of preCICE, which is a collective operation on the communicator. If any issue occurs with this communicator during the lifetime of preCICE, then the collective will fail, resulting in a crash/error.
* If preCICE runs into any error, then the events won't be aggregated nor written to a file.
* preCICE requires an additional dependency for the sole purpose of writing the aggregated data to disk. We are currently using a checked-in version of the json library, which will at some point collide with other versions on the system leading to strange problems such as #527 or https://github.com/precice/openfoam-adapter/issues/238 .
**Describe the solution you propose.**
1. Simplify the Events in preCICE as much as possible.
* The synchronization can be handled fully by preCICE, as IntraComm provides a barrier method. Its implementation also works if preCICE is compiled without MPI.
* Use independent rank files, essentially removing the aggregation from the preCICE core. This allows to output events on error.
* Optionally write these files continuously during the lifetime of the program, allowing to inspect the events on a crash. We could even implement a block-wise write to reduce the memory overhead.
* This serialization is so simple that it doesn't require a special library. ( Similar to the clang time-tracing implementation. )
* The above results in additional IO. So, a configuration option to disable the tracing could be beneficial.
2. Move the functionality to normalize, aggregate and format to a separate script (or `precice-tools`)
* Use python pandas or similar for normalizing and aggregating the data.
* Ship this as an extra tool in `/usr/share/precice` or similar.
* This allows us to remove the json dependency from the project.
* Alternatively, move the existing C++ implementation from the preCICE code into a separate executable, or into `precice-tools`.
* The python version would allow everyone to easily add custom functionality such as
* plotting given timings over time
* analyse the comm establishment to detect filesystem issues on some nodes
* find an imbalance of mapping cost over ranks of a participant
**Describe alternatives you've considered**
* Move the events2trace script into the preCICE library, essentially completely integrating the external project.
* Reimplement the events2trace as a subcommand of `precice-tools`.
**Additional context**
* https://github.com/precice/EventTimings/issues/17
* #419 | main | simplification of eventtimings i open this issue in this repo to preserve the information and make it easier to find from people running into issues regarding this it also may impact the precice lib in the future please describe the problem you are trying to solve the eventtimings provide a system to measure named sections in the code attach data to these sections used in petsc rbf mappings synchronize the communicator prior to the recording using a barrier on requested aka syncmode aggregate and normalize these measurements across ranks write a summary of the results to a file as a text table write the aggregate results to a file as json the additional script formats and merges multiple of these outputs into a single eventstracing json file that can be visualized with various tools concerns of this approach the eventstimings need a way to synchronize all ranks this currently requires passing a custom mpi comm not using mpi only supports a single rank the data aggregation happens during the finalization of precice which is a collective operation on the communicator if any issue occurs with this communicator during the lifetime of precice then the collective will fail resulting in a crash error if precice runs into any error then the events won t be aggregated nor written to a file precice requires an additional dependency for the sole purpose of writing the aggregated data to disk we are currently using a checked in version of the json library which will at some point collide with other versions on the system leading to strange problems such as or describe the solution you propose simplify the events in precice as much as possible the synchronization can be handled fully by precice as intracomm provides a barrier method its implementation also works if precice is compiled without mpi use independent rank files essentially removing the aggregation from the precice core this allows to output events on error optionally write these files continuously during the lifetime of the program allowing to inspect the events on a crash we could even implement a block wise write to reduce the memory overhead this serialization is so simple that it doesn t require a special library similar to the clang time tracing implementation the above results in additional io so a configuration option to disable the tracing could be beneficial move the functionality to normalize aggregate and format to a separate script or precice tools use python pandas or similar for normalizing and aggregating the data ship this as an extra tool in usr share precice or similar this allows us to remove the json dependency from the project alternatively move the existing c implementation from the precice code into a separate executable or into precice tools the python version would allow everyone to easily add custom functionality such as plotting given timings over time analyse the comm establishment to detect filesystem issues on some nodes find an imbalance of mapping cost over ranks of a participant describe alternatives you ve considered move the script into the precice library essentially completely integrating the external project reimplement the as a subcommand of precice tools additional context | 1 |
2,899 | 10,322,656,395 | IssuesEvent | 2019-08-31 14:22:28 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | closed | Request to join backdrop contrib team | Maintainer application | I wrote a Backdrop module to export iCal feeds using views: https://github.com/indigoxela/viewsicalexport
It's not actually a port, but is a complete rewrite of Drupal [Date iCal](https://www.drupal.org/project/date_ical) with less features and no external library dependencies.
There is already a [full port](https://github.com/backdrop-contrib/date_ical) for that module, but unfortunately it has showstopper issues, which are hard to fix.
My module tries to be a lightweight alternative for people not needing the iCal import/parse feature.
I've already contributed to Backdrop core (comments, testing, code, issues).
| True | Request to join backdrop contrib team - I wrote a Backdrop module to export iCal feeds using views: https://github.com/indigoxela/viewsicalexport
It's not actually a port, but is a complete rewrite of Drupal [Date iCal](https://www.drupal.org/project/date_ical) with less features and no external library dependencies.
There is already a [full port](https://github.com/backdrop-contrib/date_ical) for that module, but unfortunately it has showstopper issues, which are hard to fix.
My module tries to be a lightweight alternative for people not needing the iCal import/parse feature.
I've already contributed to Backdrop core (comments, testing, code, issues).
| main | request to join backdrop contrib team i wrote a backdrop module to export ical feeds using views it s not actually a port but is a complete rewrite of drupal with less features and no external library dependencies there is already a for that module but unfortunately it has showstopper issues which are hard to fix my module tries to be a lightweight alternative for people not needing the ical import parse feature i ve already contributed to backdrop core comments testing code issues | 1 |
525,299 | 15,243,138,800 | IssuesEvent | 2021-02-19 10:50:01 | googleapis/java-aiplatform | https://api.github.com/repos/googleapis/java-aiplatform | closed | The build failed | flakybot: issue priority: p1 type: bug | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: c49ac8463d342c141cf59af5c19f1ef7f28fd8a5
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/29beee2d-e69f-4629-94da-9fe3822ae7af), [Sponge](http://sponge2/29beee2d-e69f-4629-94da-9fe3822ae7af)
status: failed
<details><summary>Test output</summary><br><pre>junit.framework.AssertionFailedError: Environment variable 'VIDEO_ACTION_EVALUATION_ID' is required to perform these tests.
at junit.framework.Assert.fail(Assert.java:57)
at junit.framework.Assert.assertTrue(Assert.java:22)
at junit.framework.Assert.assertNotNull(Assert.java:256)
at junit.framework.TestCase.assertNotNull(TestCase.java:399)
at aiplatform.GetModelEvaluationVideoActionRecognitionSampleTest.requireEnvVar(GetModelEvaluationVideoActionRecognitionSampleTest.java:41)
at aiplatform.GetModelEvaluationVideoActionRecognitionSampleTest.checkRequirements(GetModelEvaluationVideoActionRecognitionSampleTest.java:49)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
</pre></details> | 1.0 | The build failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: c49ac8463d342c141cf59af5c19f1ef7f28fd8a5
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/29beee2d-e69f-4629-94da-9fe3822ae7af), [Sponge](http://sponge2/29beee2d-e69f-4629-94da-9fe3822ae7af)
status: failed
<details><summary>Test output</summary><br><pre>junit.framework.AssertionFailedError: Environment variable 'VIDEO_ACTION_EVALUATION_ID' is required to perform these tests.
at junit.framework.Assert.fail(Assert.java:57)
at junit.framework.Assert.assertTrue(Assert.java:22)
at junit.framework.Assert.assertNotNull(Assert.java:256)
at junit.framework.TestCase.assertNotNull(TestCase.java:399)
at aiplatform.GetModelEvaluationVideoActionRecognitionSampleTest.requireEnvVar(GetModelEvaluationVideoActionRecognitionSampleTest.java:41)
at aiplatform.GetModelEvaluationVideoActionRecognitionSampleTest.checkRequirements(GetModelEvaluationVideoActionRecognitionSampleTest.java:49)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272)
at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237)
at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
</pre></details> | non_main | the build failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output junit framework assertionfailederror environment variable video action evaluation id is required to perform these tests at junit framework assert fail assert java at junit framework assert asserttrue assert java at junit framework assert assertnotnull assert java at junit framework testcase assertnotnull testcase java at aiplatform getmodelevaluationvideoactionrecognitionsampletest requireenvvar getmodelevaluationvideoactionrecognitionsampletest java at aiplatform getmodelevaluationvideoactionrecognitionsampletest checkrequirements getmodelevaluationvideoactionrecognitionsampletest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements runbefores invokemethod runbefores java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java | 0 |
382 | 3,413,464,766 | IssuesEvent | 2015-12-06 18:23:53 | spyder-ide/spyder | https://api.github.com/repos/spyder-ide/spyder | closed | Enhancement: Add a 'dev' menu with common actions useful in development | Enhancement Maintainability | I know this might not be useful to all contributors/developers out there, but in the specific case of people using Spyder to develop Spyder I suggest we include a new `Development` Menu that includes actions useful when developing Spyder.
Some things that could be included:
- An action to open a signal tracker panel, as discussed in #1421
- Tools for using the translation scripts plus opening PoEditor when needed
- Actions to automate the packaging and release of Spyder (leveraging the scripts that are floating in the repo... by the way these scripts should be moved to the scripts folder)
- ...
| True | Enhancement: Add a 'dev' menu with common actions useful in development - I know this might not be useful to all contributors/developers out there, but in the specific case of people using Spyder to develop Spyder I suggest we include a new `Development` Menu that includes actions useful when developing Spyder.
Some things that could be included:
- An action to open a signal tracker panel, as discussed in #1421
- Tools for using the translation scripts plus opening PoEditor when needed
- Actions to automate the packaging and release of Spyder (leveraging the scripts that are floating in the repo... by the way these scripts should be moved to the scripts folder)
- ...
| main | enhancement add a dev menu with common actions useful in development i know this might not be useful to all contributors developers out there but in the specific case of people using spyder to develop spyder i suggest we include a new development menu that includes actions useful when developing spyder some things that could be included an action to open a signal tracker panel as discussed in tools for using the translation scripts plus opening poeditor when needed actions to automate the packaging and release of spyder leveraging the scripts that are floating in the repo by the way these scripts should be moved to the scripts folder | 1 |
147,944 | 19,526,184,462 | IssuesEvent | 2021-12-30 08:16:05 | panasalap/linux-4.1.15 | https://api.github.com/repos/panasalap/linux-4.1.15 | opened | CVE-2020-27171 (Medium) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2020-27171 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69">9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.11.8. kernel/bpf/verifier.c has an off-by-one error (with a resultant integer underflow) affecting out-of-bounds speculation on pointer arithmetic, leading to side-channel attacks that defeat Spectre mitigations and obtain sensitive information from kernel memory, aka CID-10d2bb2e6b1d.
<p>Publish Date: 2021-03-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27171>CVE-2020-27171</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/torvalds/linux/commit/10d2bb2e6b1d8c4576c56a748f697dbeb8388899">https://github.com/torvalds/linux/commit/10d2bb2e6b1d8c4576c56a748f697dbeb8388899</a></p>
<p>Release Date: 2021-03-17</p>
<p>Fix Resolution: Replace or update the following file: verifier.c</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-27171 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2020-27171 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69">9c15ec31637ff4ee4a4c14fb9b3264a31f75aa69</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/kernel/bpf/verifier.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.11.8. kernel/bpf/verifier.c has an off-by-one error (with a resultant integer underflow) affecting out-of-bounds speculation on pointer arithmetic, leading to side-channel attacks that defeat Spectre mitigations and obtain sensitive information from kernel memory, aka CID-10d2bb2e6b1d.
<p>Publish Date: 2021-03-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27171>CVE-2020-27171</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Change files</p>
<p>Origin: <a href="https://github.com/torvalds/linux/commit/10d2bb2e6b1d8c4576c56a748f697dbeb8388899">https://github.com/torvalds/linux/commit/10d2bb2e6b1d8c4576c56a748f697dbeb8388899</a></p>
<p>Release Date: 2021-03-17</p>
<p>Fix Resolution: Replace or update the following file: verifier.c</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files kernel bpf verifier c kernel bpf verifier c vulnerability details an issue was discovered in the linux kernel before kernel bpf verifier c has an off by one error with a resultant integer underflow affecting out of bounds speculation on pointer arithmetic leading to side channel attacks that defeat spectre mitigations and obtain sensitive information from kernel memory aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file verifier c step up your open source security game with whitesource | 0 |
3,463 | 13,258,519,187 | IssuesEvent | 2020-08-20 15:29:24 | pace/bricks | https://api.github.com/repos/pace/bricks | opened | Introduce *decimal.Decimal to make omitempty work for numbers | S::Pending T::Maintainance | # Issue
Golang json marshal default behaviour is to omit fields with "zero" value, this is only defined for primitive/integrated types, i.e., ints/floats/pointers etc.
`decimal.Decimal` by default does not have a "zero" value loosely speaking, because it always has some values which means that even `decimal.Zero` is not recognized as "zero" value, i.e., `json.Marshal(...)` will happily render it as 0.
This is fine unless you want to make a value optional and not include it in a http json response.
# Solution
Adjust `http/jsonapi/generator/generate_helper.go` to add `stmt.Op("*")` for decimals to allow omitempty to work. | True | Introduce *decimal.Decimal to make omitempty work for numbers - # Issue
Golang json marshal default behaviour is to omit fields with "zero" value, this is only defined for primitive/integrated types, i.e., ints/floats/pointers etc.
`decimal.Decimal` by default does not have a "zero" value loosely speaking, because it always has some values which means that even `decimal.Zero` is not recognized as "zero" value, i.e., `json.Marshal(...)` will happily render it as 0.
This is fine unless you want to make a value optional and not include it in a http json response.
# Solution
Adjust `http/jsonapi/generator/generate_helper.go` to add `stmt.Op("*")` for decimals to allow omitempty to work. | main | introduce decimal decimal to make omitempty work for numbers issue golang json marshal default behaviour is to omit fields with zero value this is only defined for primitive integrated types i e ints floats pointers etc decimal decimal by default does not have a zero value loosely speaking because it always has some values which means that even decimal zero is not recognized as zero value i e json marshal will happily render it as this is fine unless you want to make a value optional and not include it in a http json response solution adjust http jsonapi generator generate helper go to add stmt op for decimals to allow omitempty to work | 1 |
210 | 2,862,358,366 | IssuesEvent | 2015-06-04 03:37:17 | daemonraco/toobasic | https://api.github.com/repos/daemonraco/toobasic | closed | SQLite Database Structure Maintainer | Database Structure Maintainer next version | ## What to do
Create an database structure adapter for SQLlite. | True | SQLite Database Structure Maintainer - ## What to do
Create an database structure adapter for SQLlite. | main | sqlite database structure maintainer what to do create an database structure adapter for sqllite | 1 |
33,505 | 15,982,648,273 | IssuesEvent | 2021-04-18 05:19:19 | crystal-lang/crystal | https://api.github.com/repos/crystal-lang/crystal | closed | Crystal and GMP | kind:question performance topic:stdlib | The previous discussion was in 2015 but tested with pi digits using BigInteger and how do we implement GMP?
https://groups.google.com/forum/#!topic/crystal-lang/cY8bImdHWV4
Tested PHP 7.1 on macOS Mojave completed around 2 seconds for the same number of iterations:
https://benchmarksgame-team.pages.debian.net/benchmarksgame/program/pidigits-php-5.html
C, Ruby and Crystal Source code:
https://github.com/kostya/crystal-benchmarks-game#pidigits
https://benchmarksgame-team.pages.debian.net/benchmarksgame/program/pidigits-yarv-1.html | True | Crystal and GMP - The previous discussion was in 2015 but tested with pi digits using BigInteger and how do we implement GMP?
https://groups.google.com/forum/#!topic/crystal-lang/cY8bImdHWV4
Tested PHP 7.1 on macOS Mojave completed around 2 seconds for the same number of iterations:
https://benchmarksgame-team.pages.debian.net/benchmarksgame/program/pidigits-php-5.html
C, Ruby and Crystal Source code:
https://github.com/kostya/crystal-benchmarks-game#pidigits
https://benchmarksgame-team.pages.debian.net/benchmarksgame/program/pidigits-yarv-1.html | non_main | crystal and gmp the previous discussion was in but tested with pi digits using biginteger and how do we implement gmp tested php on macos mojave completed around seconds for the same number of iterations c ruby and crystal source code | 0 |
161,081 | 25,284,078,380 | IssuesEvent | 2022-11-16 17:49:00 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | closed | The secondary blob container of one leased blob container doesn't show as leased | 🧪 testing :gear: blobs :heavy_check_mark: by design :beetle: regression :gear: adls gen2 | **Storage Explorer Version**: 1.27.0-dev
**Build Number**: 20221114.4
**Branch**: main
**Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Ventura 13.0 (Apple M1 Pro)
**Architecture**: ia32/x64
**How Found**: Ad-hoc testing
**Regression From**: Previous release (1.23.1)
## Bug Descriptions ##
The leased blob container's secondary blob container doesn't show as leased while a blob container which just break lease, its secondary blob container shows as leased
## Steps to Reproduce ##
1. Expand one storage account -> Blob Containers.
2. Create a blob container -> Acquire lease for the blob container -> Click the blob container to make below properties show correct.

3. Right click the leased blob container -> Click 'Open Secondary Location'.
4. Check whether the secondary blob container shows as leased.
## Expected Experience ##
The secondary blob container shows as leased.
## Actual Experience ##
The secondary blob container doesn't show as leased.

## Additional Context ##
Here is the record:

| 1.0 | The secondary blob container of one leased blob container doesn't show as leased - **Storage Explorer Version**: 1.27.0-dev
**Build Number**: 20221114.4
**Branch**: main
**Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Ventura 13.0 (Apple M1 Pro)
**Architecture**: ia32/x64
**How Found**: Ad-hoc testing
**Regression From**: Previous release (1.23.1)
## Bug Descriptions ##
The leased blob container's secondary blob container doesn't show as leased while a blob container which just break lease, its secondary blob container shows as leased
## Steps to Reproduce ##
1. Expand one storage account -> Blob Containers.
2. Create a blob container -> Acquire lease for the blob container -> Click the blob container to make below properties show correct.

3. Right click the leased blob container -> Click 'Open Secondary Location'.
4. Check whether the secondary blob container shows as leased.
## Expected Experience ##
The secondary blob container shows as leased.
## Actual Experience ##
The secondary blob container doesn't show as leased.

## Additional Context ##
Here is the record:

| non_main | the secondary blob container of one leased blob container doesn t show as leased storage explorer version dev build number branch main platform os windows linux ubuntu macos ventura apple pro architecture how found ad hoc testing regression from previous release bug descriptions the leased blob container s secondary blob container doesn t show as leased while a blob container which just break lease its secondary blob container shows as leased steps to reproduce expand one storage account blob containers create a blob container acquire lease for the blob container click the blob container to make below properties show correct right click the leased blob container click open secondary location check whether the secondary blob container shows as leased expected experience the secondary blob container shows as leased actual experience the secondary blob container doesn t show as leased additional context here is the record | 0 |
3,346 | 12,972,351,145 | IssuesEvent | 2020-07-21 12:27:34 | ipfs-shipyard/ipld-explorer-components | https://api.github.com/repos/ipfs-shipyard/ipld-explorer-components | opened | Missing support for Filecoin codecs | P0 dif/expert effort/weeks kind/bug kind/enhancement need/analysis need/community-input need/maintainer-input | ## Problem
Even when we update to latest `cids` library, when user enters a CID with Filecoin-related codec, they get error because `ipld-filecoin` decoder does not exist:
> 
Test CIDs from https://github.com/multiformats/multihash/issues/129#issuecomment-661040091:
```
baga6ea4seaqggjjfh7whhdoxvhrix6jbcgobmdyhajcimfn33iedcp3kr23gruq
baga6ea4seaqidbk23bub2dmg2hur4aawpe44wzuu2lccflgsbcqaokjzjb7wtgi
bagboea4b5abcax5zbow3g7cyeg3nsvjguqnjkbdibnhhzc3whinr2sousvoijbrb
bagboea4b5abcb245dcsepbaelwd7hrt46itun2mvv5nckntzkg5kf73m2ry4ja7r
```
## Solution
IPLD Explorer already supports Bitcoin and Ethereum:
```
"ipld-bitcoin": "^0.3.0",
"ipld-ethereum": "^4.0.0",
```
I believe IPLD Explorer should support Filecoin CIDs.
@ribasushi @vmx @rvagg – were there any prior/ongoing discussions regarding creating `ipld-filecoin` ? | True | Missing support for Filecoin codecs - ## Problem
Even when we update to latest `cids` library, when user enters a CID with Filecoin-related codec, they get error because `ipld-filecoin` decoder does not exist:
> 
Test CIDs from https://github.com/multiformats/multihash/issues/129#issuecomment-661040091:
```
baga6ea4seaqggjjfh7whhdoxvhrix6jbcgobmdyhajcimfn33iedcp3kr23gruq
baga6ea4seaqidbk23bub2dmg2hur4aawpe44wzuu2lccflgsbcqaokjzjb7wtgi
bagboea4b5abcax5zbow3g7cyeg3nsvjguqnjkbdibnhhzc3whinr2sousvoijbrb
bagboea4b5abcb245dcsepbaelwd7hrt46itun2mvv5nckntzkg5kf73m2ry4ja7r
```
## Solution
IPLD Explorer already supports Bitcoin and Ethereum:
```
"ipld-bitcoin": "^0.3.0",
"ipld-ethereum": "^4.0.0",
```
I believe IPLD Explorer should support Filecoin CIDs.
@ribasushi @vmx @rvagg – were there any prior/ongoing discussions regarding creating `ipld-filecoin` ? | main | missing support for filecoin codecs problem even when we update to latest cids library when user enters a cid with filecoin related codec they get error because ipld filecoin decoder does not exist test cids from solution ipld explorer already supports bitcoin and ethereum ipld bitcoin ipld ethereum i believe ipld explorer should support filecoin cids ribasushi vmx rvagg – were there any prior ongoing discussions regarding creating ipld filecoin | 1 |
1,080 | 4,913,681,124 | IssuesEvent | 2016-11-23 13:23:13 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | lvol module isnt idempotent when using %FREE | affects_2.1 bug_report waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
lvol
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Debian Jessie
##### SUMMARY
<!--- Explain the problem briefly -->
`lvol` module fails at subsequent runs when the lv has been created at the first run.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: create srv lvol
lvol: vg=vsrv lv=srv size=100%FREE state=present
```
The lvol is created fine at the first run.
```
# vgs vsrv
VG #PV #LV #SN Attr VSize VFree
vsrv 1 1 0 wz--n- 907.71g 0
# lvs vsrv
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
srv vsrv -wi-ao---- 907.71g
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
`lvol` task should show [ok] for subsequent runs
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"force": false, "lv": "srv", "opts
": null, "size": "100%FREE", "snapshot": null, "state": "present", "vg": "vsrv"}, "module_name": "lvol"}, "msg": "Sorry, no shrinking of srv to 0 permitted."}
```
| True | lvol module isnt idempotent when using %FREE - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
lvol
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Debian Jessie
##### SUMMARY
<!--- Explain the problem briefly -->
`lvol` module fails at subsequent runs when the lv has been created at the first run.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: create srv lvol
lvol: vg=vsrv lv=srv size=100%FREE state=present
```
The lvol is created fine at the first run.
```
# vgs vsrv
VG #PV #LV #SN Attr VSize VFree
vsrv 1 1 0 wz--n- 907.71g 0
# lvs vsrv
LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
srv vsrv -wi-ao---- 907.71g
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
`lvol` task should show [ok] for subsequent runs
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"force": false, "lv": "srv", "opts
": null, "size": "100%FREE", "snapshot": null, "state": "present", "vg": "vsrv"}, "module_name": "lvol"}, "msg": "Sorry, no shrinking of srv to 0 permitted."}
```
| main | lvol module isnt idempotent when using free issue type bug report component name lvol ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific debian jessie summary lvol module fails at subsequent runs when the lv has been created at the first run steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name create srv lvol lvol vg vsrv lv srv size free state present the lvol is created fine at the first run vgs vsrv vg pv lv sn attr vsize vfree vsrv wz n lvs vsrv lv vg attr lsize pool origin data meta move log cpy sync convert srv vsrv wi ao expected results lvol task should show for subsequent runs actual results failed changed false failed true invocation module args force false lv srv opts null size free snapshot null state present vg vsrv module name lvol msg sorry no shrinking of srv to permitted | 1 |
230,174 | 17,604,607,570 | IssuesEvent | 2021-08-17 15:32:28 | jbetancur/react-data-table-component | https://api.github.com/repos/jbetancur/react-data-table-component | closed | Column sortFunction unclear in Readme | documentation | ## Issue Check list
- [x] Agree to the [Code of Conduct](https://github.com/jbetancur/react-data-table-component/blob/master/CODE-OF-CONDUCT.md)
- [x] Read the README
- [x] You are using React 16.8.0+
- [x] You installed `styled-components`
- [x] Include relevant code or preferably a [code sandbox](https://codesandbox.io/embed/react-data-table-sandbox-ccyuu
)
## Describe the bug
The documentation for the sortFunction property of a column is unclear. Firstly, the link goes to #-custom-sort-function which does not work as the id of the section is #CustomSortFunction. Secondly, the custom sort function section (12) shows an example of a custom sortFunction for the table itself and not for an individual column. The sortFunction for an individual column accepts two arguments, rowA and rowB (as defined on line 6 of src/DataTable/types.ts), which is different from the table sortFunction. This might be obvious to many users, but it isn't explicitly stated in that section of the documentation and I had to discover this for myself.
## To Reproduce
Steps to reproduce the behavior:
1. Visit this GitHub repo
2. Read the Readme
## Expected behavior
I would have expected separate descriptions and examples for how the overall sortFunction and the column sortFunction work, as they are different. | 1.0 | Column sortFunction unclear in Readme - ## Issue Check list
- [x] Agree to the [Code of Conduct](https://github.com/jbetancur/react-data-table-component/blob/master/CODE-OF-CONDUCT.md)
- [x] Read the README
- [x] You are using React 16.8.0+
- [x] You installed `styled-components`
- [x] Include relevant code or preferably a [code sandbox](https://codesandbox.io/embed/react-data-table-sandbox-ccyuu
)
## Describe the bug
The documentation for the sortFunction property of a column is unclear. Firstly, the link goes to #-custom-sort-function which does not work as the id of the section is #CustomSortFunction. Secondly, the custom sort function section (12) shows an example of a custom sortFunction for the table itself and not for an individual column. The sortFunction for an individual column accepts two arguments, rowA and rowB (as defined on line 6 of src/DataTable/types.ts), which is different from the table sortFunction. This might be obvious to many users, but it isn't explicitly stated in that section of the documentation and I had to discover this for myself.
## To Reproduce
Steps to reproduce the behavior:
1. Visit this GitHub repo
2. Read the Readme
## Expected behavior
I would have expected separate descriptions and examples for how the overall sortFunction and the column sortFunction work, as they are different. | non_main | column sortfunction unclear in readme issue check list agree to the read the readme you are using react you installed styled components include relevant code or preferably a describe the bug the documentation for the sortfunction property of a column is unclear firstly the link goes to custom sort function which does not work as the id of the section is customsortfunction secondly the custom sort function section shows an example of a custom sortfunction for the table itself and not for an individual column the sortfunction for an individual column accepts two arguments rowa and rowb as defined on line of src datatable types ts which is different from the table sortfunction this might be obvious to many users but it isn t explicitly stated in that section of the documentation and i had to discover this for myself to reproduce steps to reproduce the behavior visit this github repo read the readme expected behavior i would have expected separate descriptions and examples for how the overall sortfunction and the column sortfunction work as they are different | 0 |
1,143 | 5,000,406,985 | IssuesEvent | 2016-12-10 09:26:40 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | git always pulls from master instead from tracking branch | affects_1.9 bug_report waiting_on_maintainer | Issue Type: Bug report
Ansible Version: 1.9.2
Ansible Configuration: ?
Environment: Gentoo Linux 64bit
Summary: remote checkout tracks branch develop but git module always updates from master
Steps To Reproduce:
- Clone remote repo onto branch != master, with master/HEAD != branch/HEAD
- .... -m git -a "repo=git://..../foo.git dest=.../foo update=yes"
- check repo status
Expected Results: pull from tracking branch
Actual Results:
On remote:
```
git branch -av
* develop dc66049 Basic Celery support
remotes/origin/HEAD -> origin/master
remotes/origin/develop dc66049 Basic Celery support
remotes/origin/master 049a7c5 Merge branch 'release/0.4.2'
```
After running
`ansible foo-staging -m git -a "repo=git://.../foo.git dest=../foo update=yes"`
```
# git branch -av
* develop 049a7c5 [behind 4] Merge branch 'release/0.4.2'
remotes/origin/HEAD -> origin/master
remotes/origin/develop dc66049 Basic Celery support
remotes/origin/master 049a7c5 Merge branch 'release/0.4.2'
```
| True | git always pulls from master instead from tracking branch - Issue Type: Bug report
Ansible Version: 1.9.2
Ansible Configuration: ?
Environment: Gentoo Linux 64bit
Summary: remote checkout tracks branch develop but git module always updates from master
Steps To Reproduce:
- Clone remote repo onto branch != master, with master/HEAD != branch/HEAD
- .... -m git -a "repo=git://..../foo.git dest=.../foo update=yes"
- check repo status
Expected Results: pull from tracking branch
Actual Results:
On remote:
```
git branch -av
* develop dc66049 Basic Celery support
remotes/origin/HEAD -> origin/master
remotes/origin/develop dc66049 Basic Celery support
remotes/origin/master 049a7c5 Merge branch 'release/0.4.2'
```
After running
`ansible foo-staging -m git -a "repo=git://.../foo.git dest=../foo update=yes"`
```
# git branch -av
* develop 049a7c5 [behind 4] Merge branch 'release/0.4.2'
remotes/origin/HEAD -> origin/master
remotes/origin/develop dc66049 Basic Celery support
remotes/origin/master 049a7c5 Merge branch 'release/0.4.2'
```
| main | git always pulls from master instead from tracking branch issue type bug report ansible version ansible configuration environment gentoo linux summary remote checkout tracks branch develop but git module always updates from master steps to reproduce clone remote repo onto branch master with master head branch head m git a repo git foo git dest foo update yes check repo status expected results pull from tracking branch actual results on remote git branch av develop basic celery support remotes origin head origin master remotes origin develop basic celery support remotes origin master merge branch release after running ansible foo staging m git a repo git foo git dest foo update yes git branch av develop merge branch release remotes origin head origin master remotes origin develop basic celery support remotes origin master merge branch release | 1 |
4,922 | 25,306,280,566 | IssuesEvent | 2022-11-17 14:23:59 | precice/precice | https://api.github.com/repos/precice/precice | closed | Replace utils::ptr_vector<T> with smart pointers | maintainability good first issue | `utils::ptr_vector<T>` serves a similar purpose like smart pointers from `std`. We should remove this implementation in favor of smart pointers from `std`.
I already took care of replacing `utils::ptr_vector<T>` with `std::unique_ptr` in https://github.com/precice/precice/pull/1159, but there are many more places, where `utils::ptr_vector<T>` is used. | True | Replace utils::ptr_vector<T> with smart pointers - `utils::ptr_vector<T>` serves a similar purpose like smart pointers from `std`. We should remove this implementation in favor of smart pointers from `std`.
I already took care of replacing `utils::ptr_vector<T>` with `std::unique_ptr` in https://github.com/precice/precice/pull/1159, but there are many more places, where `utils::ptr_vector<T>` is used. | main | replace utils ptr vector with smart pointers utils ptr vector serves a similar purpose like smart pointers from std we should remove this implementation in favor of smart pointers from std i already took care of replacing utils ptr vector with std unique ptr in but there are many more places where utils ptr vector is used | 1 |
14,797 | 9,531,008,056 | IssuesEvent | 2019-04-29 15:04:14 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Freeze the debugger variable inspector scrolling position while debugging | bug junior job topic:editor usability | **Godot version:**
3.1 rc2
**OS/device including version:**
Windows 10 x64
**Issue description:**
While debugging, I want to watch one variable that is located below the first visible variables in the inspector.
So I scroll the inspector down until it shows me the desired variable.
But if I click in the "continue" button, the variable inspector scroll back to the first variables again.
So, on each "continue", I have to scroll to the desired variable again.
**Steps to reproduce:**
- Have a code with multiple variables
- Set a breakpoint in any line
- Start debugging
- The debugger will show the first variables:

- But I want to follow the 'x' variable, so I scroll down:

- Now I want to continue the script, so I click on the "continue" button, and then the variable list is again on the beginning:

| True | Freeze the debugger variable inspector scrolling position while debugging - **Godot version:**
3.1 rc2
**OS/device including version:**
Windows 10 x64
**Issue description:**
While debugging, I want to watch one variable that is located below the first visible variables in the inspector.
So I scroll the inspector down until it shows me the desired variable.
But if I click in the "continue" button, the variable inspector scroll back to the first variables again.
So, on each "continue", I have to scroll to the desired variable again.
**Steps to reproduce:**
- Have a code with multiple variables
- Set a breakpoint in any line
- Start debugging
- The debugger will show the first variables:

- But I want to follow the 'x' variable, so I scroll down:

- Now I want to continue the script, so I click on the "continue" button, and then the variable list is again on the beginning:

| non_main | freeze the debugger variable inspector scrolling position while debugging godot version os device including version windows issue description while debugging i want to watch one variable that is located below the first visible variables in the inspector so i scroll the inspector down until it shows me the desired variable but if i click in the continue button the variable inspector scroll back to the first variables again so on each continue i have to scroll to the desired variable again steps to reproduce have a code with multiple variables set a breakpoint in any line start debugging the debugger will show the first variables but i want to follow the x variable so i scroll down now i want to continue the script so i click on the continue button and then the variable list is again on the beginning | 0 |
9,388 | 2,615,147,268 | IssuesEvent | 2015-03-01 06:23:29 | chrsmith/html5rocks | https://api.github.com/repos/chrsmith/html5rocks | closed | Error in IE-8 | auto-migrated Priority-Medium Type-Defect | ```
Please describe the issue:
Webpage error details
User Agent: Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 5.1; Trident/4.0;
GTB6.5; .NET CLR 2.0.50727; .NET CLR 3.0.04506.30; .NET CLR 3.0.04506.648; .NET
CLR 3.0.4506.2152; InfoPath.2; .NET CLR 3.5.21022; yie8)
Timestamp: Thu, 23 Sep 2010 20:04:12 UTC
Message: Object doesn't support this property or method
Line: 629
Char: 3
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 693
Char: 13
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 833
Char: 9
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1022
Char: 11
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1114
Char: 15
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1245
Char: 11
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1561
Char: 13
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1619
Char: 15
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1705
Char: 13
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1765
Char: 13
Code: 0
URI: http://slides.html5rocks.com/
Message: Invalid property value.
Line: 2090
Char: 17
Code: 0
URI: http://slides.html5rocks.com/
Message: Invalid property value.
Line: 2136
Char: 17
Code: 0
URI: http://slides.html5rocks.com/
Message: JScript object expected
Line: 2663
Char: 11
Code: 0
URI: http://slides.html5rocks.com/
Please provide any additional information below.
```
Original issue reported on code.google.com by `maran001...@gmail.com` on 23 Sep 2010 at 8:05 | 1.0 | Error in IE-8 - ```
Please describe the issue:
Webpage error details
User Agent: Mozilla/4.0 (compatible; MSIE 8.0; Windows NT 5.1; Trident/4.0;
GTB6.5; .NET CLR 2.0.50727; .NET CLR 3.0.04506.30; .NET CLR 3.0.04506.648; .NET
CLR 3.0.4506.2152; InfoPath.2; .NET CLR 3.5.21022; yie8)
Timestamp: Thu, 23 Sep 2010 20:04:12 UTC
Message: Object doesn't support this property or method
Line: 629
Char: 3
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 693
Char: 13
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 833
Char: 9
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1022
Char: 11
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1114
Char: 15
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1245
Char: 11
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1561
Char: 13
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1619
Char: 15
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1705
Char: 13
Code: 0
URI: http://slides.html5rocks.com/
Message: Object doesn't support this property or method
Line: 1765
Char: 13
Code: 0
URI: http://slides.html5rocks.com/
Message: Invalid property value.
Line: 2090
Char: 17
Code: 0
URI: http://slides.html5rocks.com/
Message: Invalid property value.
Line: 2136
Char: 17
Code: 0
URI: http://slides.html5rocks.com/
Message: JScript object expected
Line: 2663
Char: 11
Code: 0
URI: http://slides.html5rocks.com/
Please provide any additional information below.
```
Original issue reported on code.google.com by `maran001...@gmail.com` on 23 Sep 2010 at 8:05 | non_main | error in ie please describe the issue webpage error details user agent mozilla compatible msie windows nt trident net clr net clr net clr net clr infopath net clr timestamp thu sep utc message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message object doesn t support this property or method line char code uri message invalid property value line char code uri message invalid property value line char code uri message jscript object expected line char code uri please provide any additional information below original issue reported on code google com by gmail com on sep at | 0 |
4,446 | 23,114,533,183 | IssuesEvent | 2022-07-27 15:31:51 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Eliminate superfluous API request when loading the Link Table dialog | type: enhancement work: frontend status: ready restricted: maintainers | ## Current behavior
- When the "Link Table" modal loads, there is an unnecessary API request to fetch the columns of the first listed table. This has to do with some logic within the Select component issuing unwanted 'change' events.
## Desired behavior
- The list of columns is only fetched when the user selects a target table.
| True | Eliminate superfluous API request when loading the Link Table dialog - ## Current behavior
- When the "Link Table" modal loads, there is an unnecessary API request to fetch the columns of the first listed table. This has to do with some logic within the Select component issuing unwanted 'change' events.
## Desired behavior
- The list of columns is only fetched when the user selects a target table.
| main | eliminate superfluous api request when loading the link table dialog current behavior when the link table modal loads there is an unnecessary api request to fetch the columns of the first listed table this has to do with some logic within the select component issuing unwanted change events desired behavior the list of columns is only fetched when the user selects a target table | 1 |
2,499 | 8,655,458,118 | IssuesEvent | 2018-11-27 16:00:19 | codestation/qcma | https://api.github.com/repos/codestation/qcma | closed | Backup Manager "Online ID/Username" doens't show multiple accounts | bug unmaintained | Backup Manager "Online ID/Username" only shows the last connected PS vita account.
I have 2 vitas with 2 different accounts (different regions too). I can only see the account of the last vita I transfered data from/to.
| True | Backup Manager "Online ID/Username" doens't show multiple accounts - Backup Manager "Online ID/Username" only shows the last connected PS vita account.
I have 2 vitas with 2 different accounts (different regions too). I can only see the account of the last vita I transfered data from/to.
| main | backup manager online id username doens t show multiple accounts backup manager online id username only shows the last connected ps vita account i have vitas with different accounts different regions too i can only see the account of the last vita i transfered data from to | 1 |
3,510 | 13,723,968,483 | IssuesEvent | 2020-10-03 12:05:44 | NaluKit/nalu | https://api.github.com/repos/NaluKit/nalu | closed | remove 'remove'-method in AbstractCompositeComponent | maintainance | This method is called just before the `onDetach`-method. So everything that should be done inside the `remove`-method can be done inside the `onDetach`-method. | True | remove 'remove'-method in AbstractCompositeComponent - This method is called just before the `onDetach`-method. So everything that should be done inside the `remove`-method can be done inside the `onDetach`-method. | main | remove remove method in abstractcompositecomponent this method is called just before the ondetach method so everything that should be done inside the remove method can be done inside the ondetach method | 1 |
2,285 | 8,135,486,230 | IssuesEvent | 2018-08-20 03:14:50 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | Enable configuration of VM metric account | affects_2.3 azure cloud deprecated feature needs_maintainer support:core | From @bearrito on 2016-10-13T14:52:26Z
ISSUE TYPE
Feature Idea
COMPONENT NAME
http://docs.ansible.com/ansible/azure_module.html
ANSIBLE VERSION
N/A
SUMMARY
The ask is to be able to enable setting metrics storage account via the `azure_rm_virtualmachine` task.
Copied from original issue: ansible/ansible-modules-core#5253
| True | Enable configuration of VM metric account - From @bearrito on 2016-10-13T14:52:26Z
ISSUE TYPE
Feature Idea
COMPONENT NAME
http://docs.ansible.com/ansible/azure_module.html
ANSIBLE VERSION
N/A
SUMMARY
The ask is to be able to enable setting metrics storage account via the `azure_rm_virtualmachine` task.
Copied from original issue: ansible/ansible-modules-core#5253
| main | enable configuration of vm metric account from bearrito on issue type feature idea component name ansible version n a summary the ask is to be able to enable setting metrics storage account via the azure rm virtualmachine task copied from original issue ansible ansible modules core | 1 |
319,570 | 9,746,526,894 | IssuesEvent | 2019-06-03 12:29:09 | opencollective/opencollective | https://api.github.com/repos/opencollective/opencollective | closed | Can't make a one time donation with decimals | bug complexity → simple frontend priority | Step to reproduce:
Go to opencollective.com/europe-collective/donate
Pick a profile, then enter "143.20". You can play with the plus/minus controls.
Select "Wire transfer", then click on "contribute"
500 error
Error in the API logs:
```
2019-05-31T09:31:28.881161+00:00 app[web.1]: error: GraphQL v1 error: Variable "$order" got invalid value {"paymentMethod":{"type":"manual"},"totalAmount":14319.999999999998,"taxAmount":0,"quantity":1,"currency":"EUR","interval":null,"fromCollective":{"id":37416,"type":"COLLECTIVE","name":"G4G"},"collective":{"id":868},"description":""};
Expected type Int at value.totalAmount; Int cannot represent non-integer value: 14319.999999999998
```
The process is still very broken, I just want to be able to do /:collective/donate/:amount/:description
It's still not working for amounts with decimals | 1.0 | Can't make a one time donation with decimals - Step to reproduce:
Go to opencollective.com/europe-collective/donate
Pick a profile, then enter "143.20". You can play with the plus/minus controls.
Select "Wire transfer", then click on "contribute"
500 error
Error in the API logs:
```
2019-05-31T09:31:28.881161+00:00 app[web.1]: error: GraphQL v1 error: Variable "$order" got invalid value {"paymentMethod":{"type":"manual"},"totalAmount":14319.999999999998,"taxAmount":0,"quantity":1,"currency":"EUR","interval":null,"fromCollective":{"id":37416,"type":"COLLECTIVE","name":"G4G"},"collective":{"id":868},"description":""};
Expected type Int at value.totalAmount; Int cannot represent non-integer value: 14319.999999999998
```
The process is still very broken, I just want to be able to do /:collective/donate/:amount/:description
It's still not working for amounts with decimals | non_main | can t make a one time donation with decimals step to reproduce go to opencollective com europe collective donate pick a profile then enter you can play with the plus minus controls select wire transfer then click on contribute error error in the api logs app error graphql error variable order got invalid value paymentmethod type manual totalamount taxamount quantity currency eur interval null fromcollective id type collective name collective id description expected type int at value totalamount int cannot represent non integer value the process is still very broken i just want to be able to do collective donate amount description it s still not working for amounts with decimals | 0 |
3,839 | 16,745,377,061 | IssuesEvent | 2021-06-11 14:54:10 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | closed | Local variable declarations and assignments should be preceded by a blank line | Area: analyzer Area: maintainability feature | A declaration (and assignment) to a local variable should be precededby a blank line if the preceding line does not contain another variable declaration/assignment. The reason is ease of reading (spotting variable declarations with ease).
Following _should report_ a violation:
```c#
DoSomething();
var x = 42;
var y = "something";
DoSomethingElse(x, y);
```
While following _should **NOT** report_ a violation:
```c#
DoSomething();
var x = 42;
var y = "something";
DoSomethingElse(x, y);
```
In addition, following _should report_ a violation:
```c#
var x = 42;
DoSomething();
x = 0815;
DoSomethingElse(x, y);
```
While following _should **NOT** report_ a violation:
```c#
var x = 42;
DoSomething();
x = 0815;
DoSomethingElse(x, y);
```
| True | Local variable declarations and assignments should be preceded by a blank line - A declaration (and assignment) to a local variable should be precededby a blank line if the preceding line does not contain another variable declaration/assignment. The reason is ease of reading (spotting variable declarations with ease).
Following _should report_ a violation:
```c#
DoSomething();
var x = 42;
var y = "something";
DoSomethingElse(x, y);
```
While following _should **NOT** report_ a violation:
```c#
DoSomething();
var x = 42;
var y = "something";
DoSomethingElse(x, y);
```
In addition, following _should report_ a violation:
```c#
var x = 42;
DoSomething();
x = 0815;
DoSomethingElse(x, y);
```
While following _should **NOT** report_ a violation:
```c#
var x = 42;
DoSomething();
x = 0815;
DoSomethingElse(x, y);
```
| main | local variable declarations and assignments should be preceded by a blank line a declaration and assignment to a local variable should be precededby a blank line if the preceding line does not contain another variable declaration assignment the reason is ease of reading spotting variable declarations with ease following should report a violation c dosomething var x var y something dosomethingelse x y while following should not report a violation c dosomething var x var y something dosomethingelse x y in addition following should report a violation c var x dosomething x dosomethingelse x y while following should not report a violation c var x dosomething x dosomethingelse x y | 1 |
352,455 | 32,070,070,814 | IssuesEvent | 2023-09-25 07:23:39 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | opened | The string 'Cold' is not localized on the 'Update Access Tier' dialog | 🧪 testing :gear: blobs :beetle: regression | **Storage Explorer Version:** 1.32.0-dev (93)
**Build Number:** 20230923.2
**Branch:** main
**Platform/OS:** Windows 10/Linux Ubuntu 20.04/MacOS Ventura 13.5.2 (Apple M1 Pro)
**Language:** German
**Architecture:** x64/x64/x64
**How Found:** From running test cases
**Regression From:** Previous release (1.29.2)
## Steps to Reproduce ##
1. Launch Storage Explorer -> Open 'Settings' -> 'Application -> Regional Settings'.
2. Select 'Deutsch' -> Restart Storage Explorer.
3. Expand one storage account -> Blob Containers.
4. Create a blob container -> Upload one blob.
5. Right click the blob -> Click 'Change Access Tier...'.
6. Check whether the string 'Cold' is localized.
## Expected Experience ##
The string 'Cold' is localized.

## Actual Experience ##
The string 'Cold' is not localized.

| 1.0 | The string 'Cold' is not localized on the 'Update Access Tier' dialog - **Storage Explorer Version:** 1.32.0-dev (93)
**Build Number:** 20230923.2
**Branch:** main
**Platform/OS:** Windows 10/Linux Ubuntu 20.04/MacOS Ventura 13.5.2 (Apple M1 Pro)
**Language:** German
**Architecture:** x64/x64/x64
**How Found:** From running test cases
**Regression From:** Previous release (1.29.2)
## Steps to Reproduce ##
1. Launch Storage Explorer -> Open 'Settings' -> 'Application -> Regional Settings'.
2. Select 'Deutsch' -> Restart Storage Explorer.
3. Expand one storage account -> Blob Containers.
4. Create a blob container -> Upload one blob.
5. Right click the blob -> Click 'Change Access Tier...'.
6. Check whether the string 'Cold' is localized.
## Expected Experience ##
The string 'Cold' is localized.

## Actual Experience ##
The string 'Cold' is not localized.

| non_main | the string cold is not localized on the update access tier dialog storage explorer version dev build number branch main platform os windows linux ubuntu macos ventura apple pro language german architecture how found from running test cases regression from previous release steps to reproduce launch storage explorer open settings application regional settings select deutsch restart storage explorer expand one storage account blob containers create a blob container upload one blob right click the blob click change access tier check whether the string cold is localized expected experience the string cold is localized actual experience the string cold is not localized | 0 |
312,313 | 23,423,169,525 | IssuesEvent | 2022-08-14 01:48:11 | SOFTENG310-Team3/A1-project | https://api.github.com/repos/SOFTENG310-Team3/A1-project | closed | Create contributing guidelines | documentation | Add contributing guidelines documentation so that contributors know how they can contribute | 1.0 | Create contributing guidelines - Add contributing guidelines documentation so that contributors know how they can contribute | non_main | create contributing guidelines add contributing guidelines documentation so that contributors know how they can contribute | 0 |
210,214 | 7,186,922,396 | IssuesEvent | 2018-02-02 01:52:07 | GoogleCloudPlatform/forseti-security | https://api.github.com/repos/GoogleCloudPlatform/forseti-security | closed | Exception on Creating New Inventory | module: inventory priority: p1 status: assigned | Getting this in my dev env, after blowing away all my tables, and then creating a new inventory with simply `forseti inventory create`.
Although the crawler is able to continue, the server doesn't seem to return or completes.
```
ERROR:root:Exception iterating responses: '2018-02-02T01:14:09.62189' has type str, but expected one of: int, long
Traceback (most recent call last):
File "/usr/local/google/home/henryc/.virtualenvs/forseti-security/local/lib/python2.7/site-packages/grpc/_server.py", line 403, in _take_response_from_response_iterator
return next(response_iterator), True
File "/usr/local/google/home/henryc/git/forseti-security/google/cloud/forseti/services/utils.py", line 44, in wrapper
for result in f(*args):
File "/usr/local/google/home/henryc/git/forseti-security/google/cloud/forseti/services/inventory/service.py", line 86, in Create
last_error=repr(progress.last_error))
TypeError: '2018-02-02T01:14:09.62189' has type str, but expected one of: int, long
``` | 1.0 | Exception on Creating New Inventory - Getting this in my dev env, after blowing away all my tables, and then creating a new inventory with simply `forseti inventory create`.
Although the crawler is able to continue, the server doesn't seem to return or completes.
```
ERROR:root:Exception iterating responses: '2018-02-02T01:14:09.62189' has type str, but expected one of: int, long
Traceback (most recent call last):
File "/usr/local/google/home/henryc/.virtualenvs/forseti-security/local/lib/python2.7/site-packages/grpc/_server.py", line 403, in _take_response_from_response_iterator
return next(response_iterator), True
File "/usr/local/google/home/henryc/git/forseti-security/google/cloud/forseti/services/utils.py", line 44, in wrapper
for result in f(*args):
File "/usr/local/google/home/henryc/git/forseti-security/google/cloud/forseti/services/inventory/service.py", line 86, in Create
last_error=repr(progress.last_error))
TypeError: '2018-02-02T01:14:09.62189' has type str, but expected one of: int, long
``` | non_main | exception on creating new inventory getting this in my dev env after blowing away all my tables and then creating a new inventory with simply forseti inventory create although the crawler is able to continue the server doesn t seem to return or completes error root exception iterating responses has type str but expected one of int long traceback most recent call last file usr local google home henryc virtualenvs forseti security local lib site packages grpc server py line in take response from response iterator return next response iterator true file usr local google home henryc git forseti security google cloud forseti services utils py line in wrapper for result in f args file usr local google home henryc git forseti security google cloud forseti services inventory service py line in create last error repr progress last error typeerror has type str but expected one of int long | 0 |
872 | 4,537,741,655 | IssuesEvent | 2016-09-09 02:10:03 | Microsoft/DirectXTex | https://api.github.com/repos/Microsoft/DirectXTex | closed | Code cleanup | maintainence | DirectXTex was one of my first C++11 libraries, and I started it back in the Windows XP / VS 2010 days. As such, it's got a few lingering issues compared to my current coding style and usage:
* Replace ``LPVOID``, ``LPCVOID``, ``LPCWSTR`` with standard types
* Should use ``=delete``
* Use anonymous namespaces instead of static
* Use VS standard 'smart-indent' formatting (otherwise I spend a lot of time fighting it to put back in the spaces)
* Leading ``_`` in identifiers is reserved by the language for use by implementers; should minimize/remove them in case it causes problems with future compiler versions
* Don't need to keep the few conditional paths for Windows XP compat; Windows Vista or later is fine.
* I'm using '0' in a few places where I should be using ``nullptr`` | True | Code cleanup - DirectXTex was one of my first C++11 libraries, and I started it back in the Windows XP / VS 2010 days. As such, it's got a few lingering issues compared to my current coding style and usage:
* Replace ``LPVOID``, ``LPCVOID``, ``LPCWSTR`` with standard types
* Should use ``=delete``
* Use anonymous namespaces instead of static
* Use VS standard 'smart-indent' formatting (otherwise I spend a lot of time fighting it to put back in the spaces)
* Leading ``_`` in identifiers is reserved by the language for use by implementers; should minimize/remove them in case it causes problems with future compiler versions
* Don't need to keep the few conditional paths for Windows XP compat; Windows Vista or later is fine.
* I'm using '0' in a few places where I should be using ``nullptr`` | main | code cleanup directxtex was one of my first c libraries and i started it back in the windows xp vs days as such it s got a few lingering issues compared to my current coding style and usage replace lpvoid lpcvoid lpcwstr with standard types should use delete use anonymous namespaces instead of static use vs standard smart indent formatting otherwise i spend a lot of time fighting it to put back in the spaces leading in identifiers is reserved by the language for use by implementers should minimize remove them in case it causes problems with future compiler versions don t need to keep the few conditional paths for windows xp compat windows vista or later is fine i m using in a few places where i should be using nullptr | 1 |
876 | 4,541,032,413 | IssuesEvent | 2016-09-09 16:25:21 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | junos_package.py: package_version undefined | affects_2.2 bug_report in progress networking P2 waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
junos_package
##### ANSIBLE VERSION
devel
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
https://github.com/ansible/ansible-modules-core/blame/devel/network/junos/junos_package.py#L141
`wants_ver = module.params['version'] or package_version(module)`
I can't find anywhere in the `ansible/ansible` code base where `package_version` is defined
| True | junos_package.py: package_version undefined - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
junos_package
##### ANSIBLE VERSION
devel
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
https://github.com/ansible/ansible-modules-core/blame/devel/network/junos/junos_package.py#L141
`wants_ver = module.params['version'] or package_version(module)`
I can't find anywhere in the `ansible/ansible` code base where `package_version` is defined
| main | junos package py package version undefined issue type bug report component name junos package ansible version devel configuration os environment summary wants ver module params or package version module i can t find anywhere in the ansible ansible code base where package version is defined | 1 |
1,091 | 4,953,034,651 | IssuesEvent | 2016-12-01 13:59:27 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | docker_container never starts containers | affects_2.2 bug_report cloud docker waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`docker_container`
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /usr/src/playbooks/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
```
Linux 37059fd9679f 3.13.0-91-generic #138-Ubuntu SMP Fri Jun 24 17:00:34 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
```
##### SUMMARY
Containers are created with the wrong name, outside network and never start
##### STEPS TO REPRODUCE
Run a playbook that does the following:
<!--- Paste example playbooks or commands between quotes below -->
```
- name: ensure docker network exists
docker_network:
appends: True
connected:
- existing_container
name: my_network
state: present
- name: ensure new container is running
docker_container:
image: mongo:3.2
name: new-container
networks:
- name: my_network
state: started
```
##### EXPECTED RESULTS
A container named `new-container` running, and in the `my_networ` network
##### ACTUAL RESULTS
Container with generic name is `Created` outside network, not running
[`ansible` `-vvvvv` output](https://gist.github.com/gvilarino/5b4c71773dee722005ca3777230f1cda)
[`dockerd -D` output](https://gist.github.com/gvilarino/b8acb2d76dfdc02f133da29766718614)
`docker ps -a` yields:
```
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
c2d470cdcc51 mongo:3.2 "/entrypoint.sh mongo" 21 minutes ago Created stupefied_pasteur
```
According to the docker daemon logs, the daemon never gets the `/start` instruction, just the `/create`
| True | docker_container never starts containers - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`docker_container`
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /usr/src/playbooks/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
```
Linux 37059fd9679f 3.13.0-91-generic #138-Ubuntu SMP Fri Jun 24 17:00:34 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
```
##### SUMMARY
Containers are created with the wrong name, outside network and never start
##### STEPS TO REPRODUCE
Run a playbook that does the following:
<!--- Paste example playbooks or commands between quotes below -->
```
- name: ensure docker network exists
docker_network:
appends: True
connected:
- existing_container
name: my_network
state: present
- name: ensure new container is running
docker_container:
image: mongo:3.2
name: new-container
networks:
- name: my_network
state: started
```
##### EXPECTED RESULTS
A container named `new-container` running, and in the `my_networ` network
##### ACTUAL RESULTS
Container with generic name is `Created` outside network, not running
[`ansible` `-vvvvv` output](https://gist.github.com/gvilarino/5b4c71773dee722005ca3777230f1cda)
[`dockerd -D` output](https://gist.github.com/gvilarino/b8acb2d76dfdc02f133da29766718614)
`docker ps -a` yields:
```
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
c2d470cdcc51 mongo:3.2 "/entrypoint.sh mongo" 21 minutes ago Created stupefied_pasteur
```
According to the docker daemon logs, the daemon never gets the `/start` instruction, just the `/create`
| main | docker container never starts containers issue type bug report component name docker container ansible version ansible config file usr src playbooks ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment linux generic ubuntu smp fri jun utc gnu linux summary containers are created with the wrong name outside network and never start steps to reproduce run a playbook that does the following name ensure docker network exists docker network appends true connected existing container name my network state present name ensure new container is running docker container image mongo name new container networks name my network state started expected results a container named new container running and in the my networ network actual results container with generic name is created outside network not running docker ps a yields container id image command created status ports names mongo entrypoint sh mongo minutes ago created stupefied pasteur according to the docker daemon logs the daemon never gets the start instruction just the create | 1 |
5,138 | 26,196,925,522 | IssuesEvent | 2023-01-03 14:16:20 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Clicking on the row header should select the whole row | type: bug work: frontend status: ready restricted: maintainers | ## Expected behavior
When the user clicks on the row header the entire row should be selected.
| True | Clicking on the row header should select the whole row - ## Expected behavior
When the user clicks on the row header the entire row should be selected.
| main | clicking on the row header should select the whole row expected behavior when the user clicks on the row header the entire row should be selected | 1 |
998 | 4,761,580,079 | IssuesEvent | 2016-10-25 08:42:50 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | blockinfile creates a file without ending newline | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
blockinfile module
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
No overrides
##### OS / ENVIRONMENT
Tried this on Mac OS 10.11.16 (latest El Capitan) and Fedora Linux with the same results.
##### SUMMARY
`blockinfile` module produces a file with no ending newline.
##### STEPS TO REPRODUCE
Put this in `foo.yml`:
```
---
- hosts: all
tasks:
- blockinfile: dest=foo.txt block=TESTING create=yes
```
Now run:
ansible-playbook -i localhost, -c local foo.yml
The results is a file `foo.txt` containing this:
```
# BEGIN ANSIBLE MANAGED BLOCK
TESTING
# END ANSIBLE MANAGED BLOCK
```
with no ending newline.
##### EXPECTED RESULTS
Ending newline.
##### ACTUAL RESULTS
Running the above with `-vvvv`
```
No config file found; using defaults
Loaded callback default of type stdout, v2.0
PLAYBOOK: foo.yml **************************************************************
1 plays in foo.yml
PLAY [all] *********************************************************************
TASK [setup] *******************************************************************
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: ksylvan
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351 `" && echo ansible-tmp-1470543742.93-257915674738351="` echo $HOME/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351 `" ) && sleep 0'
<localhost> PUT /var/folders/dd/mnhldrmj02zdlsf6fjfw_3_80000gn/T/tmpau5YyQ TO /Users/ksylvan/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351/setup
<localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/ksylvan/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351/setup; rm -rf "/Users/ksylvan/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351/" > /dev/null 2>&1 && sleep 0'
ok: [localhost]
TASK [blockinfile] *************************************************************
task path: /private/tmp/X/foo.yml:4
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: ksylvan
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122 `" && echo ansible-tmp-1470543743.61-113322904514122="` echo $HOME/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122 `" ) && sleep 0'
<localhost> PUT /var/folders/dd/mnhldrmj02zdlsf6fjfw_3_80000gn/T/tmpAPf_ky TO /Users/ksylvan/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122/blockinfile
<localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/ksylvan/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122/blockinfile; rm -rf "/Users/ksylvan/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122/" > /dev/null 2>&1 && sleep 0'
changed: [localhost] => {"changed": true, "invocation": {"module_args": {"backup": false, "block": "TESTING", "content": null, "create": true, "delimiter": null, "dest": "foo.txt", "directory_mode": null, "follow": false, "force": null, "group": null, "insertafter": null, "insertbefore": null, "marker": "# {mark} ANSIBLE MANAGED BLOCK", "mode": null, "owner": null, "regexp": null, "remote_src": null, "selevel": null, "serole": null, "setype": null, "seuser": null, "src": null, "state": "present", "validate": null}, "module_name": "blockinfile"}, "msg": "File created"}
PLAY RECAP *********************************************************************
localhost : ok=2 changed=1 unreachable=0 failed=0
```
| True | blockinfile creates a file without ending newline - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
blockinfile module
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
No overrides
##### OS / ENVIRONMENT
Tried this on Mac OS 10.11.16 (latest El Capitan) and Fedora Linux with the same results.
##### SUMMARY
`blockinfile` module produces a file with no ending newline.
##### STEPS TO REPRODUCE
Put this in `foo.yml`:
```
---
- hosts: all
tasks:
- blockinfile: dest=foo.txt block=TESTING create=yes
```
Now run:
ansible-playbook -i localhost, -c local foo.yml
The results is a file `foo.txt` containing this:
```
# BEGIN ANSIBLE MANAGED BLOCK
TESTING
# END ANSIBLE MANAGED BLOCK
```
with no ending newline.
##### EXPECTED RESULTS
Ending newline.
##### ACTUAL RESULTS
Running the above with `-vvvv`
```
No config file found; using defaults
Loaded callback default of type stdout, v2.0
PLAYBOOK: foo.yml **************************************************************
1 plays in foo.yml
PLAY [all] *********************************************************************
TASK [setup] *******************************************************************
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: ksylvan
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351 `" && echo ansible-tmp-1470543742.93-257915674738351="` echo $HOME/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351 `" ) && sleep 0'
<localhost> PUT /var/folders/dd/mnhldrmj02zdlsf6fjfw_3_80000gn/T/tmpau5YyQ TO /Users/ksylvan/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351/setup
<localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/ksylvan/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351/setup; rm -rf "/Users/ksylvan/.ansible/tmp/ansible-tmp-1470543742.93-257915674738351/" > /dev/null 2>&1 && sleep 0'
ok: [localhost]
TASK [blockinfile] *************************************************************
task path: /private/tmp/X/foo.yml:4
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: ksylvan
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122 `" && echo ansible-tmp-1470543743.61-113322904514122="` echo $HOME/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122 `" ) && sleep 0'
<localhost> PUT /var/folders/dd/mnhldrmj02zdlsf6fjfw_3_80000gn/T/tmpAPf_ky TO /Users/ksylvan/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122/blockinfile
<localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/ksylvan/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122/blockinfile; rm -rf "/Users/ksylvan/.ansible/tmp/ansible-tmp-1470543743.61-113322904514122/" > /dev/null 2>&1 && sleep 0'
changed: [localhost] => {"changed": true, "invocation": {"module_args": {"backup": false, "block": "TESTING", "content": null, "create": true, "delimiter": null, "dest": "foo.txt", "directory_mode": null, "follow": false, "force": null, "group": null, "insertafter": null, "insertbefore": null, "marker": "# {mark} ANSIBLE MANAGED BLOCK", "mode": null, "owner": null, "regexp": null, "remote_src": null, "selevel": null, "serole": null, "setype": null, "seuser": null, "src": null, "state": "present", "validate": null}, "module_name": "blockinfile"}, "msg": "File created"}
PLAY RECAP *********************************************************************
localhost : ok=2 changed=1 unreachable=0 failed=0
```
| main | blockinfile creates a file without ending newline issue type bug report component name blockinfile module ansible version ansible config file configured module search path default w o overrides configuration no overrides os environment tried this on mac os latest el capitan and fedora linux with the same results summary blockinfile module produces a file with no ending newline steps to reproduce put this in foo yml hosts all tasks blockinfile dest foo txt block testing create yes now run ansible playbook i localhost c local foo yml the results is a file foo txt containing this begin ansible managed block testing end ansible managed block with no ending newline expected results ending newline actual results running the above with vvvv no config file found using defaults loaded callback default of type stdout playbook foo yml plays in foo yml play task establish local connection for user ksylvan exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders dd t to users ksylvan ansible tmp ansible tmp setup exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users ksylvan ansible tmp ansible tmp setup rm rf users ksylvan ansible tmp ansible tmp dev null sleep ok task task path private tmp x foo yml establish local connection for user ksylvan exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders dd t tmpapf ky to users ksylvan ansible tmp ansible tmp blockinfile exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users ksylvan ansible tmp ansible tmp blockinfile rm rf users ksylvan ansible tmp ansible tmp dev null sleep changed changed true invocation module args backup false block testing content null create true delimiter null dest foo txt directory mode null follow false force null group null insertafter null insertbefore null marker mark ansible managed block mode null owner null regexp null remote src null selevel null serole null setype null seuser null src null state present validate null module name blockinfile msg file created play recap localhost ok changed unreachable failed | 1 |
2,544 | 8,674,758,361 | IssuesEvent | 2018-11-30 08:50:42 | citrusframework/citrus | https://api.github.com/repos/citrusframework/citrus | opened | Scp client builder does not offer autoReadFiles anymore | Prio: High READY Type: Maintainance | Citrus Version
>= 2.7.7
Description
If you upgrade your Citrus version to 2.7.7 or higher, we've a breaking change in the `ScpClientBuilder` which does not support the `autoReadFiles` method anymore. We'll correct this with one of the future releases to ensure effortless version upgrades
API before change
```java
/**
* Sets the auto read files property.
* @param autoReadFiles
* @return
*/
public ScpClientBuilder autoReadFiles(boolean autoReadFiles) {
endpoint.getEndpointConfiguration().setAutoReadFiles(autoReadFiles);
return this;
}
```
API after change: Method not present anymore
Additional information
Issue: #435
Commit: https://github.com/citrusframework/citrus/commit/b64e2e5bf510809d622695a0c399836fe3ee7aab#diff-7b734d11520ab62bc8fe2b60dc0942d4L72
BR,
Sven | True | Scp client builder does not offer autoReadFiles anymore - Citrus Version
>= 2.7.7
Description
If you upgrade your Citrus version to 2.7.7 or higher, we've a breaking change in the `ScpClientBuilder` which does not support the `autoReadFiles` method anymore. We'll correct this with one of the future releases to ensure effortless version upgrades
API before change
```java
/**
* Sets the auto read files property.
* @param autoReadFiles
* @return
*/
public ScpClientBuilder autoReadFiles(boolean autoReadFiles) {
endpoint.getEndpointConfiguration().setAutoReadFiles(autoReadFiles);
return this;
}
```
API after change: Method not present anymore
Additional information
Issue: #435
Commit: https://github.com/citrusframework/citrus/commit/b64e2e5bf510809d622695a0c399836fe3ee7aab#diff-7b734d11520ab62bc8fe2b60dc0942d4L72
BR,
Sven | main | scp client builder does not offer autoreadfiles anymore citrus version description if you upgrade your citrus version to or higher we ve a breaking change in the scpclientbuilder which does not support the autoreadfiles method anymore we ll correct this with one of the future releases to ensure effortless version upgrades api before change java sets the auto read files property param autoreadfiles return public scpclientbuilder autoreadfiles boolean autoreadfiles endpoint getendpointconfiguration setautoreadfiles autoreadfiles return this api after change method not present anymore additional information issue commit br sven | 1 |
116,435 | 24,918,063,352 | IssuesEvent | 2022-10-30 16:38:19 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | [Linux] mixed Interpreter and JIT compiler | area-CodeGen-coreclr | Hi,
OpenJDK is able to just enable interpreter:
```
java -Xint
```
or just enable JIT compilers:
```
java -Xcomp
```
or mixed:
```
java -Xmixed
```
And what about CoreCLR?
1. enable Interpreter build https://github.com/dotnet/coreclr/pull/11252
```
./build.sh debug cmakeargs "-DFEATURE_INTERPRETER=1" skipcrossgen ignorewarnings
```
2. HelloWorld testcase for mixing Interpreter and JIT compiler
* X86
```
export CORE_LIBRARIES=~/dotnet-sdk-3.0.100-linux-x64/shared/Microsoft.NETCore.App/3.0.0
# interpreter
export COMPlus_DumpInterpreterStubs=1
export COMPlus_Interpret=*
export COMPlus_InterpreterJITThreshold=999999
export COMPlus_InterpreterPrintPostMortem=1
export COMPlus_InterpreterDoLoopMethods=1
export COMPlus_TraceInterpreterEntries=1
export COMPlus_TieredCompilation=1
export COMPlus_JITMinOpts=1
export COMPlus_ZapDisable=1
# jit
export COMPlus_JitFunctionTrace=1
./bin/Product/Linux.x64.Debug/corerun ~/HelloWorld.dll
```
* ARM64 with similar environment variables
```
./bin/Product/Linux.arm64.Debug/corerun ~/HelloWorld.dll
```
* MIPS64 ditto
```
./bin/Product/Linux.mips64.Debug/corerun ~/HelloWorld.dll
```
But there are only `Start Jitting` and `Jitted Entry`, no `Generating interpretation stub` at all:
```
...
{ Start Jitting ILStubClass:IL_STUB_PInvoke(long) (MethodHash=3045e69f)
} Jitted Entry 22e at 000000ff`722b1650 method ILStubClass:IL_STUB_PInvoke(long) size 000001d0
Histogram of method executions:
# of execs | # meths (%) | cum % | % cum execs
-------------------------------------------------------
0 | 0 ( nan%) | nan% | nan%
1 | 0 ( nan%) | nan% | nan%
2 | 0 ( nan%) | nan% | nan%
3 | 0 ( nan%) | nan% | nan%
4 | 0 ( nan%) | nan% | nan%
5 | 0 ( nan%) | nan% | nan%
6 | 0 ( nan%) | nan% | nan%
7 | 0 ( nan%) | nan% | nan%
8 | 0 ( nan%) | nan% | nan%
9 | 0 ( nan%) | nan% | nan%
10 | 0 ( nan%) | nan% | nan%
For methods sorted in ascending # of executions order, cumulative % of executions:
Total number of calls from interpreted code: 0.
Also, 0 are intrinsics; 0 of these are not currently handled intrinsically.
Of these, 0 to potential property getters (0 of these dead simple), 0 to setters.
Of the dead simple getter calls, 0 have been short-circuited.
Token resolutions by category:
Category | opportunities | calls | %
---------------------------------------------------
Undefined | 0 | 0 | 0.00%
Constrained | 0 | 0 | 0.00%
NewObj | 0 | 0 | 0.00%
NewArr | 0 | 0 | 0.00%
LdToken | 0 | 0 | 0.00%
LdFtn | 0 | 0 | 0.00%
LdVirtFtn | 0 | 0 | 0.00%
SFldAddr | 0 | 0 | 0.00%
LdElem | 0 | 0 | 0.00%
Call | 0 | 0 | 0.00%
LdObj | 0 | 0 | 0.00%
StObj | 0 | 0 | 0.00%
CpObj | 0 | 0 | 0.00%
InitObj | 0 | 0 | 0.00%
IsInst | 0 | 0 | 0.00%
CastClass | 0 | 0 | 0.00%
MkRefAny | 0 | 0 | 0.00%
RefAnyVal | 0 | 0 | 0.00%
Sizeof | 0 | 0 | 0.00%
StElem | 0 | 0 | 0.00%
Box | 0 | 0 | 0.00%
Unbox | 0 | 0 | 0.00%
UnboxAny | 0 | 0 | 0.00%
LdFld | 0 | 0 | 0.00%
LdFldA | 0 | 0 | 0.00%
StFld | 0 | 0 | 0.00%
FindClass | 0 | 0 | 0.00%
Exception | 0 | 0 | 0.00%
```
Although it is able to hack tieredcompilation https://github.com/dotnet/coreclr/issues/24824#issuecomment-497048982 to force interpreter.
There is SEGFAULT issue:
* X86
```
Generating interpretation stub (# 6 = 0x6, hash = 0x9d983a13) for ILStubClass:IL_STUB_InstantiatingStub.
Generating interpretation stub (# 7 = 0x7, hash = 0x253d4d8d) for System.Collections.Generic.EqualityComparer`1[__Canon][System.__Canon]:.cctor.
Entering method dotnet/coreclr#5 (= 0x5): System.Collections.Generic.EqualityComparer`1[__Canon][System.__Canon]:.cctor().
arguments:
0: nativeint: 140737318125363 (= 0x7ffff5da7f33)
Program received signal SIGSEGV, Segmentation fault.
MethodTable::GetTypeDefRid_NoLogging (this=0x758b48f0758948f8) at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:8161
8161 WORD token = m_wToken;
(gdb) bt
#0 MethodTable::GetTypeDefRid_NoLogging (this=0x758b48f0758948f8)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:8161
dotnet/coreclr#1 0x00007ffff5e6473f in MethodTable::GetTypeDefRid (this=0x758b48f0758948f8)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:8153
dotnet/coreclr#2 0x00007ffff5e64620 in MethodTable::HasSameTypeDefAs (this=0x758b48f0758948f8, pMT=0x7fff7c355c48)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:644
dotnet/coreclr#3 0x00007ffff5e771f0 in MethodTable::GetMethodTableMatchingParentClass (this=
0x7ffff5da7f33 <FrameWithCookie<GCFrame>::FrameWithCookie(OBJECTREF*, unsigned int, int)+67>, pWhichParent=0x7fff7c355c48)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:8361
dotnet/coreclr#4 0x00007ffff6019bf5 in InterpreterMethodInfo::GetPreciseGenericsContext (this=0x6b9620, thisArg=0x0,
genericsCtxtArg=0x7ffff5da7f33 <FrameWithCookie<GCFrame>::FrameWithCookie(OBJECTREF*, unsigned int, int)+67>)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:8857
dotnet/coreclr#5 0x00007ffff6021d9c in Interpreter::GetPreciseGenericsContext (this=0x7ffffffec198)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:1781
dotnet/coreclr#6 0x00007ffff60216da in Interpreter::ResolveToken (this=0x7ffffffec198, resTok=0x7ffffffeb808, token=452984835,
tokenType=CORINFO_TOKENKIND_Ldtoken, rtk=Interpreter::RTK_LdToken)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.hpp:396
dotnet/coreclr#7 0x00007ffff601102f in Interpreter::LdToken (this=0x7ffffffec198)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:6337
dotnet/coreclr#8 0x00007ffff6006b35 in Interpreter::ExecuteMethod (this=0x7ffffffec198, retVal=0x7ffffffec180, pDoJmpCall=0x7ffffffed6ef,
pJmpCallToken=0x7ffffffec17c) at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:2942
dotnet/coreclr#9 0x00007ffff600347c in Interpreter::ExecuteMethodWrapper (interpMethInfo=0x6b9620, directCall=false,
ilArgs=0x7ffffffed990 "3\177\332\365\377\177", stubContext=0x7ffffffeddf0, pDoJmpCall=0x7ffffffed6ef,
pResolvedToken=0x7ffffffed698) at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1589
dotnet/coreclr#10 0x00007ffff6021b02 in Interpreter::InterpretMethodBody (interpMethInfo=0x6b9620, directCall=false,
ilArgs=0x7ffffffed990 "3\177\332\365\377\177", stubContext=0x7ffffffeddf0)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1624
dotnet/coreclr#11 0x00007ffff6002d07 in InterpretMethod (interpMethInfo=0x6b9620, ilArgs=0x7ffffffed990 "3\177\332\365\377\177",
stubContext=0x7ffffffeddf0) at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1745
dotnet/coreclr#12 0x00007fff7c0cd2e4 in ?? ()
dotnet/coreclr#13 0x00007fffffffe0a0 in ?? ()
dotnet/coreclr#14 0x0000000000000000 in ?? ()
```
* ARM64
```
Generating interpretation stub (# 1 = 0x1, hash = 0xd8570543) for System.AppContext:Setup.
Thread 1 "corerun" received signal SIGSEGV, Segmentation fault.
0x0000007fb70b5658 in Interpreter::Interpreter (this=0x7fffffe710, methInfo_=0x0, directCall_=127, ilArgs_=0x0, stubContext_=0x7f3d43ab60, frameMemory=0x7f3d43ab60 "0~D=\177") at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:812
812 {
(gdb) bt
#0 0x0000007fb70b5658 in Interpreter::Interpreter (this=0x7fffffe710, methInfo_=0x0, directCall_=127, ilArgs_=0x0, stubContext_=0x7f3d43ab60, frameMemory=0x7f3d43ab60 "0~D=\177") at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:812
dotnet/coreclr#1 0x0000007fb7098eb4 in Interpreter::ExecuteMethodWrapper (interpMethInfo=0x7fb6f8d4c4 <MetaSig::NextArg()+156>, directCall=false, ilArgs=0x7fffffe2f8 "\020vC", stubContext=0x0, pDoJmpCall=0x7fffffe01c, pResolvedToken=0x7fffffdfc8) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1577
dotnet/coreclr#2 0x0000007fb70b5f7c in Interpreter::InterpretMethodBody (interpMethInfo=0x7fb6f8d4c4 <MetaSig::NextArg()+156>, directCall=false, ilArgs=0x7fffffe2f8 "\020vC", stubContext=0x0) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1624
dotnet/coreclr#3 0x0000007fb709887c in InterpretMethod (interpMethInfo=0x7fb6f8d4c4 <MetaSig::NextArg()+156>, ilArgs=0x7fffffe2f8 "\020vC", stubContext=0x0) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1745
dotnet/coreclr#4 0x0000007f3d19d060 in ?? ()
dotnet/coreclr#5 0x0000000000437610 in ?? ()
```
* MIPS64
```
Generating interpretation stub (# 1 = 0x1, hash = 0xd8570543) for System.AppContext:Setup.
Program received signal SIGSEGV, Segmentation fault.
0x000000fff696b224 in Interpreter::Interpreter (this=<error reading variable: Cannot access memory at address 0xf33332ba70>, methInfo_=<error reading variable: Cannot access memory at address 0xf33332ba68>, directCall_=<error reading variable: Cannot access memory at address 0xf33332ba64>, ilArgs_=<error reading variable: Cannot access memory at address 0xf33332ba58>, stubContext_=<error reading variable: Cannot access memory at address 0xf33332ba50>, frameMemory=<error reading variable: Cannot access memory at address 0xf33332ba48>) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:812
812 {
Missing separate debuginfos, use: debuginfo-install libgcc-4.9.3-11.fc21.loongson.11.mips64el libstdc++-4.9.3-11.fc21.loongson.11.mips64el lttng-ust-2.4.1-3.fc21.loongson.mips64el userspace-rcu-0.8.1-6.fc21.loongson.mips64el xz-libs-5.1.2-14alpha.fc21.loongson.mips64el
(gdb) bt
#0 0x000000fff696b224 in Interpreter::Interpreter (this=<error reading variable: Cannot access memory at address 0xf33332ba70>, methInfo_=<error reading variable: Cannot access memory at address 0xf33332ba68>, directCall_=<error reading variable: Cannot access memory at address 0xf33332ba64>, ilArgs_=<error reading variable: Cannot access memory at address 0xf33332ba58>, stubContext_=<error reading variable: Cannot access memory at address 0xf33332ba50>, frameMemory=<error reading variable: Cannot access memory at address 0xf33332ba48>) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:812
dotnet/coreclr#1 0x000000fff693d75c in Interpreter::ExecuteMethodWrapper (interpMethInfo=<error reading variable: Cannot access memory at address 0xf33332bb38>, directCall=<error reading variable: Cannot access memory at address 0xf33332bb34>, ilArgs=<error reading variable: Cannot access memory at address 0xf33332bb28>, stubContext=<error reading variable: Cannot access memory at address 0xf33332bb20>, pDoJmpCall=<error reading variable: Cannot access memory at address 0xf33332bb18>, pResolvedToken=<error reading variable: Cannot access memory at address 0xf33332bb10>) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1627
dotnet/coreclr#2 0x000000fff696c1b4 in Interpreter::InterpretMethodBody (interpMethInfo=0x1200be190, directCall=false, ilArgs=0xffffffa0b8 "py\003 \001", stubContext=0x0) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1674
dotnet/coreclr#3 0x000000fff693cbec in InterpretMethod (interpMethInfo=0x1200be190, ilArgs=0xffffffa0b8 "py\003 \001", stubContext=0x0) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1795
dotnet/coreclr#4 0x000000fff6c0b608 in CallDescrWorkerInternal () at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/mips64/calldescrworkermips64.S:74
Backtrace stopped: frame did not save the PC
```
How to mix Interpreter and JIT compiler? Please give me some hint.
Thanks,
Leslie Zhai
category:correctness
theme:interpreter
skill-level:intermediate
cost:small | 1.0 | [Linux] mixed Interpreter and JIT compiler - Hi,
OpenJDK is able to just enable interpreter:
```
java -Xint
```
or just enable JIT compilers:
```
java -Xcomp
```
or mixed:
```
java -Xmixed
```
And what about CoreCLR?
1. enable Interpreter build https://github.com/dotnet/coreclr/pull/11252
```
./build.sh debug cmakeargs "-DFEATURE_INTERPRETER=1" skipcrossgen ignorewarnings
```
2. HelloWorld testcase for mixing Interpreter and JIT compiler
* X86
```
export CORE_LIBRARIES=~/dotnet-sdk-3.0.100-linux-x64/shared/Microsoft.NETCore.App/3.0.0
# interpreter
export COMPlus_DumpInterpreterStubs=1
export COMPlus_Interpret=*
export COMPlus_InterpreterJITThreshold=999999
export COMPlus_InterpreterPrintPostMortem=1
export COMPlus_InterpreterDoLoopMethods=1
export COMPlus_TraceInterpreterEntries=1
export COMPlus_TieredCompilation=1
export COMPlus_JITMinOpts=1
export COMPlus_ZapDisable=1
# jit
export COMPlus_JitFunctionTrace=1
./bin/Product/Linux.x64.Debug/corerun ~/HelloWorld.dll
```
* ARM64 with similar environment variables
```
./bin/Product/Linux.arm64.Debug/corerun ~/HelloWorld.dll
```
* MIPS64 ditto
```
./bin/Product/Linux.mips64.Debug/corerun ~/HelloWorld.dll
```
But there are only `Start Jitting` and `Jitted Entry`, no `Generating interpretation stub` at all:
```
...
{ Start Jitting ILStubClass:IL_STUB_PInvoke(long) (MethodHash=3045e69f)
} Jitted Entry 22e at 000000ff`722b1650 method ILStubClass:IL_STUB_PInvoke(long) size 000001d0
Histogram of method executions:
# of execs | # meths (%) | cum % | % cum execs
-------------------------------------------------------
0 | 0 ( nan%) | nan% | nan%
1 | 0 ( nan%) | nan% | nan%
2 | 0 ( nan%) | nan% | nan%
3 | 0 ( nan%) | nan% | nan%
4 | 0 ( nan%) | nan% | nan%
5 | 0 ( nan%) | nan% | nan%
6 | 0 ( nan%) | nan% | nan%
7 | 0 ( nan%) | nan% | nan%
8 | 0 ( nan%) | nan% | nan%
9 | 0 ( nan%) | nan% | nan%
10 | 0 ( nan%) | nan% | nan%
For methods sorted in ascending # of executions order, cumulative % of executions:
Total number of calls from interpreted code: 0.
Also, 0 are intrinsics; 0 of these are not currently handled intrinsically.
Of these, 0 to potential property getters (0 of these dead simple), 0 to setters.
Of the dead simple getter calls, 0 have been short-circuited.
Token resolutions by category:
Category | opportunities | calls | %
---------------------------------------------------
Undefined | 0 | 0 | 0.00%
Constrained | 0 | 0 | 0.00%
NewObj | 0 | 0 | 0.00%
NewArr | 0 | 0 | 0.00%
LdToken | 0 | 0 | 0.00%
LdFtn | 0 | 0 | 0.00%
LdVirtFtn | 0 | 0 | 0.00%
SFldAddr | 0 | 0 | 0.00%
LdElem | 0 | 0 | 0.00%
Call | 0 | 0 | 0.00%
LdObj | 0 | 0 | 0.00%
StObj | 0 | 0 | 0.00%
CpObj | 0 | 0 | 0.00%
InitObj | 0 | 0 | 0.00%
IsInst | 0 | 0 | 0.00%
CastClass | 0 | 0 | 0.00%
MkRefAny | 0 | 0 | 0.00%
RefAnyVal | 0 | 0 | 0.00%
Sizeof | 0 | 0 | 0.00%
StElem | 0 | 0 | 0.00%
Box | 0 | 0 | 0.00%
Unbox | 0 | 0 | 0.00%
UnboxAny | 0 | 0 | 0.00%
LdFld | 0 | 0 | 0.00%
LdFldA | 0 | 0 | 0.00%
StFld | 0 | 0 | 0.00%
FindClass | 0 | 0 | 0.00%
Exception | 0 | 0 | 0.00%
```
Although it is able to hack tieredcompilation https://github.com/dotnet/coreclr/issues/24824#issuecomment-497048982 to force interpreter.
There is SEGFAULT issue:
* X86
```
Generating interpretation stub (# 6 = 0x6, hash = 0x9d983a13) for ILStubClass:IL_STUB_InstantiatingStub.
Generating interpretation stub (# 7 = 0x7, hash = 0x253d4d8d) for System.Collections.Generic.EqualityComparer`1[__Canon][System.__Canon]:.cctor.
Entering method dotnet/coreclr#5 (= 0x5): System.Collections.Generic.EqualityComparer`1[__Canon][System.__Canon]:.cctor().
arguments:
0: nativeint: 140737318125363 (= 0x7ffff5da7f33)
Program received signal SIGSEGV, Segmentation fault.
MethodTable::GetTypeDefRid_NoLogging (this=0x758b48f0758948f8) at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:8161
8161 WORD token = m_wToken;
(gdb) bt
#0 MethodTable::GetTypeDefRid_NoLogging (this=0x758b48f0758948f8)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:8161
dotnet/coreclr#1 0x00007ffff5e6473f in MethodTable::GetTypeDefRid (this=0x758b48f0758948f8)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:8153
dotnet/coreclr#2 0x00007ffff5e64620 in MethodTable::HasSameTypeDefAs (this=0x758b48f0758948f8, pMT=0x7fff7c355c48)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:644
dotnet/coreclr#3 0x00007ffff5e771f0 in MethodTable::GetMethodTableMatchingParentClass (this=
0x7ffff5da7f33 <FrameWithCookie<GCFrame>::FrameWithCookie(OBJECTREF*, unsigned int, int)+67>, pWhichParent=0x7fff7c355c48)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/methodtable.cpp:8361
dotnet/coreclr#4 0x00007ffff6019bf5 in InterpreterMethodInfo::GetPreciseGenericsContext (this=0x6b9620, thisArg=0x0,
genericsCtxtArg=0x7ffff5da7f33 <FrameWithCookie<GCFrame>::FrameWithCookie(OBJECTREF*, unsigned int, int)+67>)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:8857
dotnet/coreclr#5 0x00007ffff6021d9c in Interpreter::GetPreciseGenericsContext (this=0x7ffffffec198)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:1781
dotnet/coreclr#6 0x00007ffff60216da in Interpreter::ResolveToken (this=0x7ffffffec198, resTok=0x7ffffffeb808, token=452984835,
tokenType=CORINFO_TOKENKIND_Ldtoken, rtk=Interpreter::RTK_LdToken)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.hpp:396
dotnet/coreclr#7 0x00007ffff601102f in Interpreter::LdToken (this=0x7ffffffec198)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:6337
dotnet/coreclr#8 0x00007ffff6006b35 in Interpreter::ExecuteMethod (this=0x7ffffffec198, retVal=0x7ffffffec180, pDoJmpCall=0x7ffffffed6ef,
pJmpCallToken=0x7ffffffec17c) at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:2942
dotnet/coreclr#9 0x00007ffff600347c in Interpreter::ExecuteMethodWrapper (interpMethInfo=0x6b9620, directCall=false,
ilArgs=0x7ffffffed990 "3\177\332\365\377\177", stubContext=0x7ffffffeddf0, pDoJmpCall=0x7ffffffed6ef,
pResolvedToken=0x7ffffffed698) at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1589
dotnet/coreclr#10 0x00007ffff6021b02 in Interpreter::InterpretMethodBody (interpMethInfo=0x6b9620, directCall=false,
ilArgs=0x7ffffffed990 "3\177\332\365\377\177", stubContext=0x7ffffffeddf0)
at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1624
dotnet/coreclr#11 0x00007ffff6002d07 in InterpretMethod (interpMethInfo=0x6b9620, ilArgs=0x7ffffffed990 "3\177\332\365\377\177",
stubContext=0x7ffffffeddf0) at /home/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1745
dotnet/coreclr#12 0x00007fff7c0cd2e4 in ?? ()
dotnet/coreclr#13 0x00007fffffffe0a0 in ?? ()
dotnet/coreclr#14 0x0000000000000000 in ?? ()
```
* ARM64
```
Generating interpretation stub (# 1 = 0x1, hash = 0xd8570543) for System.AppContext:Setup.
Thread 1 "corerun" received signal SIGSEGV, Segmentation fault.
0x0000007fb70b5658 in Interpreter::Interpreter (this=0x7fffffe710, methInfo_=0x0, directCall_=127, ilArgs_=0x0, stubContext_=0x7f3d43ab60, frameMemory=0x7f3d43ab60 "0~D=\177") at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:812
812 {
(gdb) bt
#0 0x0000007fb70b5658 in Interpreter::Interpreter (this=0x7fffffe710, methInfo_=0x0, directCall_=127, ilArgs_=0x0, stubContext_=0x7f3d43ab60, frameMemory=0x7f3d43ab60 "0~D=\177") at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:812
dotnet/coreclr#1 0x0000007fb7098eb4 in Interpreter::ExecuteMethodWrapper (interpMethInfo=0x7fb6f8d4c4 <MetaSig::NextArg()+156>, directCall=false, ilArgs=0x7fffffe2f8 "\020vC", stubContext=0x0, pDoJmpCall=0x7fffffe01c, pResolvedToken=0x7fffffdfc8) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1577
dotnet/coreclr#2 0x0000007fb70b5f7c in Interpreter::InterpretMethodBody (interpMethInfo=0x7fb6f8d4c4 <MetaSig::NextArg()+156>, directCall=false, ilArgs=0x7fffffe2f8 "\020vC", stubContext=0x0) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1624
dotnet/coreclr#3 0x0000007fb709887c in InterpretMethod (interpMethInfo=0x7fb6f8d4c4 <MetaSig::NextArg()+156>, ilArgs=0x7fffffe2f8 "\020vC", stubContext=0x0) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1745
dotnet/coreclr#4 0x0000007f3d19d060 in ?? ()
dotnet/coreclr#5 0x0000000000437610 in ?? ()
```
* MIPS64
```
Generating interpretation stub (# 1 = 0x1, hash = 0xd8570543) for System.AppContext:Setup.
Program received signal SIGSEGV, Segmentation fault.
0x000000fff696b224 in Interpreter::Interpreter (this=<error reading variable: Cannot access memory at address 0xf33332ba70>, methInfo_=<error reading variable: Cannot access memory at address 0xf33332ba68>, directCall_=<error reading variable: Cannot access memory at address 0xf33332ba64>, ilArgs_=<error reading variable: Cannot access memory at address 0xf33332ba58>, stubContext_=<error reading variable: Cannot access memory at address 0xf33332ba50>, frameMemory=<error reading variable: Cannot access memory at address 0xf33332ba48>) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:812
812 {
Missing separate debuginfos, use: debuginfo-install libgcc-4.9.3-11.fc21.loongson.11.mips64el libstdc++-4.9.3-11.fc21.loongson.11.mips64el lttng-ust-2.4.1-3.fc21.loongson.mips64el userspace-rcu-0.8.1-6.fc21.loongson.mips64el xz-libs-5.1.2-14alpha.fc21.loongson.mips64el
(gdb) bt
#0 0x000000fff696b224 in Interpreter::Interpreter (this=<error reading variable: Cannot access memory at address 0xf33332ba70>, methInfo_=<error reading variable: Cannot access memory at address 0xf33332ba68>, directCall_=<error reading variable: Cannot access memory at address 0xf33332ba64>, ilArgs_=<error reading variable: Cannot access memory at address 0xf33332ba58>, stubContext_=<error reading variable: Cannot access memory at address 0xf33332ba50>, frameMemory=<error reading variable: Cannot access memory at address 0xf33332ba48>) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.h:812
dotnet/coreclr#1 0x000000fff693d75c in Interpreter::ExecuteMethodWrapper (interpMethInfo=<error reading variable: Cannot access memory at address 0xf33332bb38>, directCall=<error reading variable: Cannot access memory at address 0xf33332bb34>, ilArgs=<error reading variable: Cannot access memory at address 0xf33332bb28>, stubContext=<error reading variable: Cannot access memory at address 0xf33332bb20>, pDoJmpCall=<error reading variable: Cannot access memory at address 0xf33332bb18>, pResolvedToken=<error reading variable: Cannot access memory at address 0xf33332bb10>) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1627
dotnet/coreclr#2 0x000000fff696c1b4 in Interpreter::InterpretMethodBody (interpMethInfo=0x1200be190, directCall=false, ilArgs=0xffffffa0b8 "py\003 \001", stubContext=0x0) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1674
dotnet/coreclr#3 0x000000fff693cbec in InterpretMethod (interpMethInfo=0x1200be190, ilArgs=0xffffffa0b8 "py\003 \001", stubContext=0x0) at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/interpreter.cpp:1795
dotnet/coreclr#4 0x000000fff6c0b608 in CallDescrWorkerInternal () at /home/loongson/zhaixiang/coreclr-mips64-dev/src/vm/mips64/calldescrworkermips64.S:74
Backtrace stopped: frame did not save the PC
```
How to mix Interpreter and JIT compiler? Please give me some hint.
Thanks,
Leslie Zhai
category:correctness
theme:interpreter
skill-level:intermediate
cost:small | non_main | mixed interpreter and jit compiler hi openjdk is able to just enable interpreter java xint or just enable jit compilers java xcomp or mixed java xmixed and what about coreclr enable interpreter build build sh debug cmakeargs dfeature interpreter skipcrossgen ignorewarnings helloworld testcase for mixing interpreter and jit compiler export core libraries dotnet sdk linux shared microsoft netcore app interpreter export complus dumpinterpreterstubs export complus interpret export complus interpreterjitthreshold export complus interpreterprintpostmortem export complus interpreterdoloopmethods export complus traceinterpreterentries export complus tieredcompilation export complus jitminopts export complus zapdisable jit export complus jitfunctiontrace bin product linux debug corerun helloworld dll with similar environment variables bin product linux debug corerun helloworld dll ditto bin product linux debug corerun helloworld dll but there are only start jitting and jitted entry no generating interpretation stub at all start jitting ilstubclass il stub pinvoke long methodhash jitted entry at method ilstubclass il stub pinvoke long size histogram of method executions of execs meths cum cum execs nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan nan for methods sorted in ascending of executions order cumulative of executions total number of calls from interpreted code also are intrinsics of these are not currently handled intrinsically of these to potential property getters of these dead simple to setters of the dead simple getter calls have been short circuited token resolutions by category category opportunities calls undefined constrained newobj newarr ldtoken ldftn ldvirtftn sfldaddr ldelem call ldobj stobj cpobj initobj isinst castclass mkrefany refanyval sizeof stelem box unbox unboxany ldfld ldflda stfld findclass exception although it is able to hack tieredcompilation to force interpreter there is segfault issue generating interpretation stub hash for ilstubclass il stub instantiatingstub generating interpretation stub hash for system collections generic equalitycomparer cctor entering method dotnet coreclr system collections generic equalitycomparer cctor arguments nativeint program received signal sigsegv segmentation fault methodtable gettypedefrid nologging this at home zhaixiang coreclr dev src vm methodtable cpp word token m wtoken gdb bt methodtable gettypedefrid nologging this at home zhaixiang coreclr dev src vm methodtable cpp dotnet coreclr in methodtable gettypedefrid this at home zhaixiang coreclr dev src vm methodtable cpp dotnet coreclr in methodtable hassametypedefas this pmt at home zhaixiang coreclr dev src vm methodtable cpp dotnet coreclr in methodtable getmethodtablematchingparentclass this framewithcookie objectref unsigned int int pwhichparent at home zhaixiang coreclr dev src vm methodtable cpp dotnet coreclr in interpretermethodinfo getprecisegenericscontext this thisarg genericsctxtarg framewithcookie objectref unsigned int int at home zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpreter getprecisegenericscontext this at home zhaixiang coreclr dev src vm interpreter h dotnet coreclr in interpreter resolvetoken this restok token tokentype corinfo tokenkind ldtoken rtk interpreter rtk ldtoken at home zhaixiang coreclr dev src vm interpreter hpp dotnet coreclr in interpreter ldtoken this at home zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpreter executemethod this retval pdojmpcall pjmpcalltoken at home zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpreter executemethodwrapper interpmethinfo directcall false ilargs stubcontext pdojmpcall presolvedtoken at home zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpreter interpretmethodbody interpmethinfo directcall false ilargs stubcontext at home zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpretmethod interpmethinfo ilargs stubcontext at home zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in dotnet coreclr in dotnet coreclr in generating interpretation stub hash for system appcontext setup thread corerun received signal sigsegv segmentation fault in interpreter interpreter this methinfo directcall ilargs stubcontext framememory d at home loongson zhaixiang coreclr dev src vm interpreter h gdb bt in interpreter interpreter this methinfo directcall ilargs stubcontext framememory d at home loongson zhaixiang coreclr dev src vm interpreter h dotnet coreclr in interpreter executemethodwrapper interpmethinfo directcall false ilargs stubcontext pdojmpcall presolvedtoken at home loongson zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpreter interpretmethodbody interpmethinfo directcall false ilargs stubcontext at home loongson zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpretmethod interpmethinfo ilargs stubcontext at home loongson zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in dotnet coreclr in generating interpretation stub hash for system appcontext setup program received signal sigsegv segmentation fault in interpreter interpreter this methinfo directcall ilargs stubcontext framememory at home loongson zhaixiang coreclr dev src vm interpreter h missing separate debuginfos use debuginfo install libgcc loongson libstdc loongson lttng ust loongson userspace rcu loongson xz libs loongson gdb bt in interpreter interpreter this methinfo directcall ilargs stubcontext framememory at home loongson zhaixiang coreclr dev src vm interpreter h dotnet coreclr in interpreter executemethodwrapper interpmethinfo directcall ilargs stubcontext pdojmpcall presolvedtoken at home loongson zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpreter interpretmethodbody interpmethinfo directcall false ilargs py stubcontext at home loongson zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in interpretmethod interpmethinfo ilargs py stubcontext at home loongson zhaixiang coreclr dev src vm interpreter cpp dotnet coreclr in calldescrworkerinternal at home loongson zhaixiang coreclr dev src vm s backtrace stopped frame did not save the pc how to mix interpreter and jit compiler please give me some hint thanks leslie zhai category correctness theme interpreter skill level intermediate cost small | 0 |
105,883 | 16,661,243,076 | IssuesEvent | 2021-06-06 11:09:01 | AlexRogalskiy/weather-time | https://api.github.com/repos/AlexRogalskiy/weather-time | opened | CVE-2020-7753 (High) detected in trim-0.0.1.tgz | security vulnerability | ## CVE-2020-7753 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-0.0.1.tgz</b></p></summary>
<p>Trim string whitespace</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p>
<p>Path to dependency file: weather-time/package.json</p>
<p>Path to vulnerable library: weather-time/node_modules/trim/package.json</p>
<p>
Dependency Hierarchy:
- remark-preset-davidtheclark-0.12.0.tgz (Root Library)
- remark-cli-7.0.1.tgz
- remark-11.0.2.tgz
- remark-parse-7.0.2.tgz
- :x: **trim-0.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/weather-time/commit/559d39620a54998a6f35e7cbea6c60528e5da3b0">559d39620a54998a6f35e7cbea6c60528e5da3b0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim().
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/component/trim/pull/8">https://github.com/component/trim/pull/8</a></p>
<p>Release Date: 2020-10-27</p>
<p>Fix Resolution: trim - 0.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7753 (High) detected in trim-0.0.1.tgz - ## CVE-2020-7753 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-0.0.1.tgz</b></p></summary>
<p>Trim string whitespace</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p>
<p>Path to dependency file: weather-time/package.json</p>
<p>Path to vulnerable library: weather-time/node_modules/trim/package.json</p>
<p>
Dependency Hierarchy:
- remark-preset-davidtheclark-0.12.0.tgz (Root Library)
- remark-cli-7.0.1.tgz
- remark-11.0.2.tgz
- remark-parse-7.0.2.tgz
- :x: **trim-0.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/weather-time/commit/559d39620a54998a6f35e7cbea6c60528e5da3b0">559d39620a54998a6f35e7cbea6c60528e5da3b0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim().
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/component/trim/pull/8">https://github.com/component/trim/pull/8</a></p>
<p>Release Date: 2020-10-27</p>
<p>Fix Resolution: trim - 0.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in trim tgz cve high severity vulnerability vulnerable library trim tgz trim string whitespace library home page a href path to dependency file weather time package json path to vulnerable library weather time node modules trim package json dependency hierarchy remark preset davidtheclark tgz root library remark cli tgz remark tgz remark parse tgz x trim tgz vulnerable library found in head commit a href vulnerability details all versions of package trim are vulnerable to regular expression denial of service redos via trim publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution trim step up your open source security game with whitesource | 0 |
12,184 | 9,632,106,490 | IssuesEvent | 2019-05-15 15:30:11 | OpenLiberty/open-liberty | https://api.github.com/repos/OpenLiberty/open-liberty | opened | Update com.ibm.ws.org.osgi.annotation.versioning to use 1.1 | team:OSGi Infrastructure | Currently, `com.ibm.ws.org.osgi.annotation.versioning` sources `org.osgi.annotation.versioning` 1.0.
This should be updated to `org.osgi.annotation.versioning` 1.1.
For example:
Microprofile projects leverage the use of the `@ProviderType` annotation on packages.
An annotation target that is not provided in 1.0 but rather 1.1
| 1.0 | Update com.ibm.ws.org.osgi.annotation.versioning to use 1.1 - Currently, `com.ibm.ws.org.osgi.annotation.versioning` sources `org.osgi.annotation.versioning` 1.0.
This should be updated to `org.osgi.annotation.versioning` 1.1.
For example:
Microprofile projects leverage the use of the `@ProviderType` annotation on packages.
An annotation target that is not provided in 1.0 but rather 1.1
| non_main | update com ibm ws org osgi annotation versioning to use currently com ibm ws org osgi annotation versioning sources org osgi annotation versioning this should be updated to org osgi annotation versioning for example microprofile projects leverage the use of the providertype annotation on packages an annotation target that is not provided in but rather | 0 |
400,733 | 11,779,899,407 | IssuesEvent | 2020-03-16 18:55:33 | epam/cloud-pipeline | https://api.github.com/repos/epam/cloud-pipeline | opened | Expose `System` parameters to the worker nodes | kind/enhancement priority/high sys/core | **Background**
When a cluster run is started, only a limited number of `well-known` parameters are passed to the workers. This causes issues when the default values are changed for the master (e.g. `COMMON_DIR`).
**Approach**
* Let's introduce a preference `launch.system.parameters.pass` (boolean, default `false`)
* If it is enabled - any system parameter, specified for the master node, shall be set for the workers as well
| 1.0 | Expose `System` parameters to the worker nodes - **Background**
When a cluster run is started, only a limited number of `well-known` parameters are passed to the workers. This causes issues when the default values are changed for the master (e.g. `COMMON_DIR`).
**Approach**
* Let's introduce a preference `launch.system.parameters.pass` (boolean, default `false`)
* If it is enabled - any system parameter, specified for the master node, shall be set for the workers as well
| non_main | expose system parameters to the worker nodes background when a cluster run is started only a limited number of well known parameters are passed to the workers this causes issues when the default values are changed for the master e g common dir approach let s introduce a preference launch system parameters pass boolean default false if it is enabled any system parameter specified for the master node shall be set for the workers as well | 0 |
5,063 | 25,942,604,651 | IssuesEvent | 2022-12-16 20:05:08 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | reopened | Support DLQ for SNS subscriptions | area/event-source contributors/good-first-issue contributors/welcome stage/waiting-for-release maintainer/need-response type/new-property area/sns | **Description:**
SNS subscriptions support DLQs using a so called RedrivePolicy https://docs.aws.amazon.com/sns/latest/dg/sns-configure-dead-letter-queue.html
**Observed result:**
The SNS event type does not allow to define a RedrivePolicy https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#sns
**Expected result:**
Support RedrivePolicy
| True | Support DLQ for SNS subscriptions - **Description:**
SNS subscriptions support DLQs using a so called RedrivePolicy https://docs.aws.amazon.com/sns/latest/dg/sns-configure-dead-letter-queue.html
**Observed result:**
The SNS event type does not allow to define a RedrivePolicy https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#sns
**Expected result:**
Support RedrivePolicy
| main | support dlq for sns subscriptions description sns subscriptions support dlqs using a so called redrivepolicy observed result the sns event type does not allow to define a redrivepolicy expected result support redrivepolicy | 1 |
3,225 | 12,368,705,990 | IssuesEvent | 2020-05-18 14:13:27 | Kashdeya/Tiny-Progressions | https://api.github.com/repos/Kashdeya/Tiny-Progressions | closed | Wub hammer and enchants | Version not Maintainted | Pack : Foolcraft 3
Tiny progressions : tinyprogressions-1.12.2-3.3.29-Release.jar
forge : 1.12 2705
Hello,
It looks like adding enchants to the wub hammer doesn't do anything. I tried fortune III and smelting (from cofh) and they don't work. It works fine on the wub pickaxe though. | True | Wub hammer and enchants - Pack : Foolcraft 3
Tiny progressions : tinyprogressions-1.12.2-3.3.29-Release.jar
forge : 1.12 2705
Hello,
It looks like adding enchants to the wub hammer doesn't do anything. I tried fortune III and smelting (from cofh) and they don't work. It works fine on the wub pickaxe though. | main | wub hammer and enchants pack foolcraft tiny progressions tinyprogressions release jar forge hello it looks like adding enchants to the wub hammer doesn t do anything i tried fortune iii and smelting from cofh and they don t work it works fine on the wub pickaxe though | 1 |
5,838 | 31,017,534,383 | IssuesEvent | 2023-08-10 00:36:27 | bazelbuild/intellij | https://api.github.com/repos/bazelbuild/intellij | opened | Stacktraces in unit test results are not navigatable | type: bug awaiting-maintainer | ### Description of the bug:
Stacktrace looks like just text and there is no way to click on it to get to particular class
<img width="1342" alt="image" src="https://github.com/bazelbuild/intellij/assets/50216138/c2039d8b-28f0-46ee-aaf9-4bffc9b47889">
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Run unit test that fails
### Which Intellij IDE are you using? Please provide the specific version.
2023.2
### What programming languages and tools are you using? Please provide specific versions.
Java
### What Bazel plugin version are you using?
2023.07.18.0.1
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_ | True | Stacktraces in unit test results are not navigatable - ### Description of the bug:
Stacktrace looks like just text and there is no way to click on it to get to particular class
<img width="1342" alt="image" src="https://github.com/bazelbuild/intellij/assets/50216138/c2039d8b-28f0-46ee-aaf9-4bffc9b47889">
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
Run unit test that fails
### Which Intellij IDE are you using? Please provide the specific version.
2023.2
### What programming languages and tools are you using? Please provide specific versions.
Java
### What Bazel plugin version are you using?
2023.07.18.0.1
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_ | main | stacktraces in unit test results are not navigatable description of the bug stacktrace looks like just text and there is no way to click on it to get to particular class img width alt image src what s the simplest easiest way to reproduce this bug please provide a minimal example if possible run unit test that fails which intellij ide are you using please provide the specific version what programming languages and tools are you using please provide specific versions java what bazel plugin version are you using have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response | 1 |
56,121 | 8,051,821,198 | IssuesEvent | 2018-08-01 17:18:31 | rtfd/readthedocs.org | https://api.github.com/repos/rtfd/readthedocs.org | opened | Update sponsors pages | Accepted Needed: documentation | We are mentioning Rackspace (our current host) at:
* https://readthedocs.org/sustainability/
* https://docs.readthedocs.io/en/latest/sponsors.html#past-sponsors
* https://readthedocs.org/
* http://docs.readthedocs.io/en/latest/architecture.html
We should probably update these pages to mention Azure and/or adapt them. | 1.0 | Update sponsors pages - We are mentioning Rackspace (our current host) at:
* https://readthedocs.org/sustainability/
* https://docs.readthedocs.io/en/latest/sponsors.html#past-sponsors
* https://readthedocs.org/
* http://docs.readthedocs.io/en/latest/architecture.html
We should probably update these pages to mention Azure and/or adapt them. | non_main | update sponsors pages we are mentioning rackspace our current host at we should probably update these pages to mention azure and or adapt them | 0 |
1,917 | 2,578,987,018 | IssuesEvent | 2015-02-13 04:58:08 | CMPUT301W15T13/TravelPlanner | https://api.github.com/repos/CMPUT301W15T13/TravelPlanner | closed | 04.02.01 | Testing | As a claimant, I want the category for an expense item to be one of air fare, ground transport, vehicle rental, private automobile, fuel, parking, registration, accommodation, meal, or supplies. | 1.0 | 04.02.01 - As a claimant, I want the category for an expense item to be one of air fare, ground transport, vehicle rental, private automobile, fuel, parking, registration, accommodation, meal, or supplies. | non_main | as a claimant i want the category for an expense item to be one of air fare ground transport vehicle rental private automobile fuel parking registration accommodation meal or supplies | 0 |
1,828 | 6,577,346,514 | IssuesEvent | 2017-09-12 00:16:15 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | template and copy modules should have state parameter | affects_2.1 feature_idea waiting_on_maintainer | ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
copy
template
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
N/A
##### SUMMARY
As was suggested in https://github.com/ansible/ansible/issues/6929, I am opening new issue on this matter
Please add 'state' parameter for template and copy modules, so items could be removed with the same task by using condition or variable override.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Something like this should be possible
```
- name: govern foo template
template: src=foo dest=/etc/foo state="{% if enable_foo == true %}present{% else %}absent{% endif%}"
```
On second thought, file, copy, and template modules could be merged together into file module to make some cool things like:
```
- name: linked template
file: template=foo.conf state=link src=../bar/bar.conf dest=/etc/foo/foo.conf
# makes link /etc/foo/foo.conf > /etc/bar/bar.conf,
# templates /etc/foo/foo.conf from foo.conf effectively writing to /etc/bar/bar.conf.
```
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
```
| True | template and copy modules should have state parameter - ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
copy
template
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
N/A
##### SUMMARY
As was suggested in https://github.com/ansible/ansible/issues/6929, I am opening new issue on this matter
Please add 'state' parameter for template and copy modules, so items could be removed with the same task by using condition or variable override.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Something like this should be possible
```
- name: govern foo template
template: src=foo dest=/etc/foo state="{% if enable_foo == true %}present{% else %}absent{% endif%}"
```
On second thought, file, copy, and template modules could be merged together into file module to make some cool things like:
```
- name: linked template
file: template=foo.conf state=link src=../bar/bar.conf dest=/etc/foo/foo.conf
# makes link /etc/foo/foo.conf > /etc/bar/bar.conf,
# templates /etc/foo/foo.conf from foo.conf effectively writing to /etc/bar/bar.conf.
```
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
```
| main | template and copy modules should have state parameter issue type feature idea component name copy template ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific n a summary as was suggested in i am opening new issue on this matter please add state parameter for template and copy modules so items could be removed with the same task by using condition or variable override steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used expected results something like this should be possible name govern foo template template src foo dest etc foo state if enable foo true present else absent endif on second thought file copy and template modules could be merged together into file module to make some cool things like name linked template file template foo conf state link src bar bar conf dest etc foo foo conf makes link etc foo foo conf etc bar bar conf templates etc foo foo conf from foo conf effectively writing to etc bar bar conf actual results | 1 |
336,027 | 30,114,913,472 | IssuesEvent | 2023-06-30 10:39:29 | vaadin/platform | https://api.github.com/repos/vaadin/platform | closed | run test for platform under dev mode | testing | Currently, platform tests are running in production mode only, we should test everything under two modes.
as we have discovered the issue with running V14+ application on safari 10 under `dev` mode https://github.com/vaadin/flow/issues/6817 . after the issue got fixed, we can
1) update the travis configuration to test under `dev` mode; (**on going**)
2) update platform snapshot version to test under `production` mode(**done**);
3) update platform JDK test running under `production` mode (**done**).
4) Enable safari 10 test after flow issue got solved #956 (**on going**) | 1.0 | run test for platform under dev mode - Currently, platform tests are running in production mode only, we should test everything under two modes.
as we have discovered the issue with running V14+ application on safari 10 under `dev` mode https://github.com/vaadin/flow/issues/6817 . after the issue got fixed, we can
1) update the travis configuration to test under `dev` mode; (**on going**)
2) update platform snapshot version to test under `production` mode(**done**);
3) update platform JDK test running under `production` mode (**done**).
4) Enable safari 10 test after flow issue got solved #956 (**on going**) | non_main | run test for platform under dev mode currently platform tests are running in production mode only we should test everything under two modes as we have discovered the issue with running application on safari under dev mode after the issue got fixed we can update the travis configuration to test under dev mode on going update platform snapshot version to test under production mode done update platform jdk test running under production mode done enable safari test after flow issue got solved on going | 0 |
2,017 | 6,756,313,460 | IssuesEvent | 2017-10-24 06:25:58 | sympy/sympy | https://api.github.com/repos/sympy/sympy | opened | Use __all__ top-level variable in module files | Easy to Fix Maintainability Needs Decision | I have seen `__all__` top-level variable set to list of objects that are imported when we do `from sympy.combinatorics.coset_table import *` I get `bisect_left` which isn't even supposed to be in the module.
Also I think this will reduce import time when doing `from sympy.combinatorics.coset_table import *` (will this also reduce import time for `from sympy import *` after considering the files like `coset_table.py` has public available classes and functions.
I have seen this often used in scikit-learn source code. This will be easy to fix, once someone comments that it is a good decision to have this. | True | Use __all__ top-level variable in module files - I have seen `__all__` top-level variable set to list of objects that are imported when we do `from sympy.combinatorics.coset_table import *` I get `bisect_left` which isn't even supposed to be in the module.
Also I think this will reduce import time when doing `from sympy.combinatorics.coset_table import *` (will this also reduce import time for `from sympy import *` after considering the files like `coset_table.py` has public available classes and functions.
I have seen this often used in scikit-learn source code. This will be easy to fix, once someone comments that it is a good decision to have this. | main | use all top level variable in module files i have seen all top level variable set to list of objects that are imported when we do from sympy combinatorics coset table import i get bisect left which isn t even supposed to be in the module also i think this will reduce import time when doing from sympy combinatorics coset table import will this also reduce import time for from sympy import after considering the files like coset table py has public available classes and functions i have seen this often used in scikit learn source code this will be easy to fix once someone comments that it is a good decision to have this | 1 |
230,184 | 25,443,021,223 | IssuesEvent | 2022-11-24 01:33:48 | Hans-Zamorano-Matamala/mean_entrenamiento | https://api.github.com/repos/Hans-Zamorano-Matamala/mean_entrenamiento | opened | CVE-2022-41940 (High) detected in engine.io-1.8.3.tgz | security vulnerability | ## CVE-2022-41940 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>engine.io-1.8.3.tgz</b></p></summary>
<p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p>
<p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-1.8.3.tgz">https://registry.npmjs.org/engine.io/-/engine.io-1.8.3.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/engine.io/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- socket.io-1.7.3.tgz
- :x: **engine.io-1.8.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Hans-Zamorano-Matamala/mean_entrenamiento/commit/0f094ecc422f26d3138f57e9bfc643b6c44307ca">0f094ecc422f26d3138f57e9bfc643b6c44307ca</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Engine.IO is the implementation of transport-based cross-browser/cross-device bi-directional communication layer for Socket.IO. A specially crafted HTTP request can trigger an uncaught exception on the Engine.IO server, thus killing the Node.js process. This impacts all the users of the engine.io package, including those who uses depending packages like socket.io. There is no known workaround except upgrading to a safe version. There are patches for this issue released in versions 3.6.1 and 6.2.1.
<p>Publish Date: 2022-11-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41940>CVE-2022-41940</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/socketio/engine.io/security/advisories/GHSA-r7qp-cfhv-p84w">https://github.com/socketio/engine.io/security/advisories/GHSA-r7qp-cfhv-p84w</a></p>
<p>Release Date: 2022-11-22</p>
<p>Fix Resolution: engine.io - 3.6.1,6.2.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-41940 (High) detected in engine.io-1.8.3.tgz - ## CVE-2022-41940 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>engine.io-1.8.3.tgz</b></p></summary>
<p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p>
<p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-1.8.3.tgz">https://registry.npmjs.org/engine.io/-/engine.io-1.8.3.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/engine.io/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.7.1.tgz (Root Library)
- socket.io-1.7.3.tgz
- :x: **engine.io-1.8.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Hans-Zamorano-Matamala/mean_entrenamiento/commit/0f094ecc422f26d3138f57e9bfc643b6c44307ca">0f094ecc422f26d3138f57e9bfc643b6c44307ca</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Engine.IO is the implementation of transport-based cross-browser/cross-device bi-directional communication layer for Socket.IO. A specially crafted HTTP request can trigger an uncaught exception on the Engine.IO server, thus killing the Node.js process. This impacts all the users of the engine.io package, including those who uses depending packages like socket.io. There is no known workaround except upgrading to a safe version. There are patches for this issue released in versions 3.6.1 and 6.2.1.
<p>Publish Date: 2022-11-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41940>CVE-2022-41940</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/socketio/engine.io/security/advisories/GHSA-r7qp-cfhv-p84w">https://github.com/socketio/engine.io/security/advisories/GHSA-r7qp-cfhv-p84w</a></p>
<p>Release Date: 2022-11-22</p>
<p>Fix Resolution: engine.io - 3.6.1,6.2.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in engine io tgz cve high severity vulnerability vulnerable library engine io tgz the realtime engine behind socket io provides the foundation of a bidirectional connection between client and server library home page a href path to dependency file client package json path to vulnerable library client node modules engine io package json dependency hierarchy karma tgz root library socket io tgz x engine io tgz vulnerable library found in head commit a href vulnerability details engine io is the implementation of transport based cross browser cross device bi directional communication layer for socket io a specially crafted http request can trigger an uncaught exception on the engine io server thus killing the node js process this impacts all the users of the engine io package including those who uses depending packages like socket io there is no known workaround except upgrading to a safe version there are patches for this issue released in versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution engine io step up your open source security game with mend | 0 |
27,888 | 5,409,646,103 | IssuesEvent | 2017-03-01 05:16:58 | webpack/webpack.js.org | https://api.github.com/repos/webpack/webpack.js.org | opened | add language selection menu | Documentation: In Planning | webpack china doc is ready. Shall we add a language selection menu on top navigation bar?
Best regards,
Lee | 1.0 | add language selection menu - webpack china doc is ready. Shall we add a language selection menu on top navigation bar?
Best regards,
Lee | non_main | add language selection menu webpack china doc is ready shall we add a language selection menu on top navigation bar best regards lee | 0 |
2,043 | 6,893,218,211 | IssuesEvent | 2017-11-23 01:56:09 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | opened | Add secure kext loading information / caveat | awaiting maintainer feedback core discussion | Copied from: https://github.com/caskroom/homebrew-cask/pull/41111#issuecomment-346287577

Related Reports:
- https://github.com/caskroom/homebrew-cask/issues/40755
- https://github.com/caskroom/homebrew-cask/issues/39856
- https://github.com/caskroom/homebrew-cask/issues/39369
- Probably a couple more I've missed
Apple tech note: https://developer.apple.com/library/content/technotes/tn2459/_index.html | True | Add secure kext loading information / caveat - Copied from: https://github.com/caskroom/homebrew-cask/pull/41111#issuecomment-346287577

Related Reports:
- https://github.com/caskroom/homebrew-cask/issues/40755
- https://github.com/caskroom/homebrew-cask/issues/39856
- https://github.com/caskroom/homebrew-cask/issues/39369
- Probably a couple more I've missed
Apple tech note: https://developer.apple.com/library/content/technotes/tn2459/_index.html | main | add secure kext loading information caveat copied from related reports probably a couple more i ve missed apple tech note | 1 |
3,606 | 4,552,868,805 | IssuesEvent | 2016-09-13 01:12:25 | gravitational/teleport | https://api.github.com/repos/gravitational/teleport | closed | TTL of signed certificate far too long | security | Hi,
I was just reading through the code and noticed that the `ValidBefore` field of the `ssh.Certificate` is being set to `time.Now()...UnixNano()`, which is the time period expressed in nanoseconds (https://github.com/gravitational/teleport/blob/master/lib/auth/native/native.go#L145).
According to the documentation I've found, http://cvsweb.openbsd.org/cgi-bin/cvsweb/src/usr.bin/ssh/PROTOCOL.certkeys?annotate=HEAD, `ValidBefore` and `ValidAfter` should be `time in seconds since 1970-01-01 00:00:00` which means that the valid period of the certificates being generated is 1,000,000,000 times greater than what is being asked for. | True | TTL of signed certificate far too long - Hi,
I was just reading through the code and noticed that the `ValidBefore` field of the `ssh.Certificate` is being set to `time.Now()...UnixNano()`, which is the time period expressed in nanoseconds (https://github.com/gravitational/teleport/blob/master/lib/auth/native/native.go#L145).
According to the documentation I've found, http://cvsweb.openbsd.org/cgi-bin/cvsweb/src/usr.bin/ssh/PROTOCOL.certkeys?annotate=HEAD, `ValidBefore` and `ValidAfter` should be `time in seconds since 1970-01-01 00:00:00` which means that the valid period of the certificates being generated is 1,000,000,000 times greater than what is being asked for. | non_main | ttl of signed certificate far too long hi i was just reading through the code and noticed that the validbefore field of the ssh certificate is being set to time now unixnano which is the time period expressed in nanoseconds according to the documentation i ve found validbefore and validafter should be time in seconds since which means that the valid period of the certificates being generated is times greater than what is being asked for | 0 |
1,804 | 6,575,933,684 | IssuesEvent | 2017-09-11 17:53:26 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | mount module no swap support | affects_2.0 feature_idea waiting_on_maintainer | Come back to me asap, please. Were prepared for impl. if OK.
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
mount module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.0.0.2
```
| True | mount module no swap support - Come back to me asap, please. Were prepared for impl. if OK.
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
mount module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.0.0.2
```
| main | mount module no swap support come back to me asap please were prepared for impl if ok issue type feature idea component name mount module ansible version ansible | 1 |
5,236 | 26,552,491,850 | IssuesEvent | 2023-01-20 09:09:41 | bazelbuild/intellij | https://api.github.com/repos/bazelbuild/intellij | closed | Generated files should be added to sources | type: bug product: IntelliJ lang: java topic: project views awaiting-maintainer | I created a generated file, but when I navigate to it in IDE, I navigate to a compiled .class file. Nice to be able to navigate to generated .java file instead. | True | Generated files should be added to sources - I created a generated file, but when I navigate to it in IDE, I navigate to a compiled .class file. Nice to be able to navigate to generated .java file instead. | main | generated files should be added to sources i created a generated file but when i navigate to it in ide i navigate to a compiled class file nice to be able to navigate to generated java file instead | 1 |
4,111 | 19,522,078,194 | IssuesEvent | 2021-12-29 20:40:28 | aws/aws-sam-build-images | https://api.github.com/repos/aws/aws-sam-build-images | closed | Possible to update pip in the public.ecr.aws/sam/build-python3.6 build image? | dependencies python maintainer/need-response | The version of pip installed in the `public.ecr.aws/sam/build-python3.6` image currently is `18.1`. As mentioned in this [`psycopg2` project issue](https://github.com/psycopg/psycopg2/issues/1290), older versions of `pip` (which include `18.1`) fail to install the latest `psycopg2` versions (`2.9` or later).
If we upgrade to the latest `pip` version, `21.1.2`, we are able to install the later `psycopg2` versions without issues. Would it be possible to upgrade the default `pip` version in the `public.ecr.aws/sam/build-python3.6` image to the latest version? | True | Possible to update pip in the public.ecr.aws/sam/build-python3.6 build image? - The version of pip installed in the `public.ecr.aws/sam/build-python3.6` image currently is `18.1`. As mentioned in this [`psycopg2` project issue](https://github.com/psycopg/psycopg2/issues/1290), older versions of `pip` (which include `18.1`) fail to install the latest `psycopg2` versions (`2.9` or later).
If we upgrade to the latest `pip` version, `21.1.2`, we are able to install the later `psycopg2` versions without issues. Would it be possible to upgrade the default `pip` version in the `public.ecr.aws/sam/build-python3.6` image to the latest version? | main | possible to update pip in the public ecr aws sam build build image the version of pip installed in the public ecr aws sam build image currently is as mentioned in this older versions of pip which include fail to install the latest versions or later if we upgrade to the latest pip version we are able to install the later versions without issues would it be possible to upgrade the default pip version in the public ecr aws sam build image to the latest version | 1 |
185,543 | 14,359,662,445 | IssuesEvent | 2020-11-30 15:53:54 | serverless/serverless | https://api.github.com/repos/serverless/serverless | closed | Tests refactor: Move all unit tests to "test/unit" folder | refactor tests | <!-- ⚠️⚠️ Acknowledge ALL below remarks -->
<!-- ⚠️⚠️ Request may not be processed if it doesn't meet outlined criteria -->
<!-- ⚠️⚠️ Search existing issues to avoid creating duplicates -->
<!-- ⚠️⚠️ Plugin enhancements should be proposed at plugin repository, not here -->
<!-- ⚠️⚠️ Answer ALL required questions below -->
<!--
Q1: Describe the problem (use case) that needs to be solved
-->
### Use case description
Removing tests from folders in which we keep implementation will:
- Make implementation folders cleaner and less noisy
- Will allow to easily reduce the package publication size
<!--
Q2: Propose solution (e.g. provide configuration example)
Note: This is optional, remove this section if you do not wish to propose anything at this point
-->
### Proposed solution
- Move all test files one by one into `test/unit` folder, with help of [movejs](https://github.com/medikoo/movejs#readme) utility (it ensures efficient require paths rewrite)
- Update `npm test` command so only `test/unit` folder is search for tests
| 1.0 | Tests refactor: Move all unit tests to "test/unit" folder - <!-- ⚠️⚠️ Acknowledge ALL below remarks -->
<!-- ⚠️⚠️ Request may not be processed if it doesn't meet outlined criteria -->
<!-- ⚠️⚠️ Search existing issues to avoid creating duplicates -->
<!-- ⚠️⚠️ Plugin enhancements should be proposed at plugin repository, not here -->
<!-- ⚠️⚠️ Answer ALL required questions below -->
<!--
Q1: Describe the problem (use case) that needs to be solved
-->
### Use case description
Removing tests from folders in which we keep implementation will:
- Make implementation folders cleaner and less noisy
- Will allow to easily reduce the package publication size
<!--
Q2: Propose solution (e.g. provide configuration example)
Note: This is optional, remove this section if you do not wish to propose anything at this point
-->
### Proposed solution
- Move all test files one by one into `test/unit` folder, with help of [movejs](https://github.com/medikoo/movejs#readme) utility (it ensures efficient require paths rewrite)
- Update `npm test` command so only `test/unit` folder is search for tests
| non_main | tests refactor move all unit tests to test unit folder describe the problem use case that needs to be solved use case description removing tests from folders in which we keep implementation will make implementation folders cleaner and less noisy will allow to easily reduce the package publication size propose solution e g provide configuration example note this is optional remove this section if you do not wish to propose anything at this point proposed solution move all test files one by one into test unit folder with help of utility it ensures efficient require paths rewrite update npm test command so only test unit folder is search for tests | 0 |
3,059 | 11,456,335,758 | IssuesEvent | 2020-02-06 21:00:13 | Datatamer/tamr-client | https://api.github.com/repos/Datatamer/tamr-client | closed | Make some docs changes visible on the stable branch | ⚙️ Maintainers 💬 RFC | # 💬 RFC
Figure out a mechanism for pushing changes to the docs that should become immediately visible.
## 🔦 Context
If we don't want to change our code, but we do want to change the docs that describe that code, how do we do that? We want to change the [**stable** docs](https://tamr-unify-python-client.readthedocs.io/en/stable/), not the [**latest** docs](https://tamr-unify-python-client.readthedocs.io/en/latest/), but we don't want to necessitate a full release to do so.
## 💻 Examples
#58 adds refinements to the Contributor Guide, but those refinements are only visible on **latest** / `0.4.0-dev`. They should be visible on **stable** / `0.3.0` too.
---
We could add a commit to fix the docs both to `master` and to the most recent **stable** release branch. BUT readthedocs uses Github tags for building multiple versions of the docs, so we would need to re-tag after this commit got included? BUT the Github release is tied to the release tag... hmm 🤔
| True | Make some docs changes visible on the stable branch - # 💬 RFC
Figure out a mechanism for pushing changes to the docs that should become immediately visible.
## 🔦 Context
If we don't want to change our code, but we do want to change the docs that describe that code, how do we do that? We want to change the [**stable** docs](https://tamr-unify-python-client.readthedocs.io/en/stable/), not the [**latest** docs](https://tamr-unify-python-client.readthedocs.io/en/latest/), but we don't want to necessitate a full release to do so.
## 💻 Examples
#58 adds refinements to the Contributor Guide, but those refinements are only visible on **latest** / `0.4.0-dev`. They should be visible on **stable** / `0.3.0` too.
---
We could add a commit to fix the docs both to `master` and to the most recent **stable** release branch. BUT readthedocs uses Github tags for building multiple versions of the docs, so we would need to re-tag after this commit got included? BUT the Github release is tied to the release tag... hmm 🤔
| main | make some docs changes visible on the stable branch 💬 rfc figure out a mechanism for pushing changes to the docs that should become immediately visible 🔦 context if we don t want to change our code but we do want to change the docs that describe that code how do we do that we want to change the not the but we don t want to necessitate a full release to do so 💻 examples adds refinements to the contributor guide but those refinements are only visible on latest dev they should be visible on stable too we could add a commit to fix the docs both to master and to the most recent stable release branch but readthedocs uses github tags for building multiple versions of the docs so we would need to re tag after this commit got included but the github release is tied to the release tag hmm 🤔 | 1 |
3,190 | 12,227,659,871 | IssuesEvent | 2020-05-03 16:09:20 | cthit/react-digit-components | https://api.github.com/repos/cthit/react-digit-components | opened | What props that has required doesn't really make any sense anywhere | Type: Maintainence | Right now it feels like there's no real logic to what props are required. Just look at `DigitEditDataCard`. | True | What props that has required doesn't really make any sense anywhere - Right now it feels like there's no real logic to what props are required. Just look at `DigitEditDataCard`. | main | what props that has required doesn t really make any sense anywhere right now it feels like there s no real logic to what props are required just look at digiteditdatacard | 1 |
152,585 | 19,690,143,597 | IssuesEvent | 2022-01-12 05:31:29 | rdharmedran/codeguru | https://api.github.com/repos/rdharmedran/codeguru | opened | CVE-2021-22569 (Medium) detected in protobuf-java-3.2.0.jar | security vulnerability | ## CVE-2021-22569 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>protobuf-java-3.2.0.jar</b></p></summary>
<p>Core Protocol Buffers library. Protocol Buffers are a way of encoding structured data in an
efficient yet extensible format.</p>
<p>Library home page: <a href="https://developers.google.com/protocol-buffers/">https://developers.google.com/protocol-buffers/</a></p>
<p>Path to dependency file: /example/example/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/protobuf/protobuf-java/3.2.0/protobuf-java-3.2.0.jar</p>
<p>
Dependency Hierarchy:
- google-cloud-dataflow-java-sdk-all-2.5.0.jar (Root Library)
- beam-sdks-java-io-google-cloud-platform-2.5.0.jar
- :x: **protobuf-java-3.2.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue in protobuf-java allowed the interleaving of com.google.protobuf.UnknownFieldSet fields in such a way that would be processed out of order. A small malicious payload can occupy the parser for several minutes by creating large numbers of short-lived objects that cause frequent, repeated pauses. We recommend upgrading libraries beyond the vulnerable versions.
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22569>CVE-2021-22569</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-22569 (Medium) detected in protobuf-java-3.2.0.jar - ## CVE-2021-22569 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>protobuf-java-3.2.0.jar</b></p></summary>
<p>Core Protocol Buffers library. Protocol Buffers are a way of encoding structured data in an
efficient yet extensible format.</p>
<p>Library home page: <a href="https://developers.google.com/protocol-buffers/">https://developers.google.com/protocol-buffers/</a></p>
<p>Path to dependency file: /example/example/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/protobuf/protobuf-java/3.2.0/protobuf-java-3.2.0.jar</p>
<p>
Dependency Hierarchy:
- google-cloud-dataflow-java-sdk-all-2.5.0.jar (Root Library)
- beam-sdks-java-io-google-cloud-platform-2.5.0.jar
- :x: **protobuf-java-3.2.0.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue in protobuf-java allowed the interleaving of com.google.protobuf.UnknownFieldSet fields in such a way that would be processed out of order. A small malicious payload can occupy the parser for several minutes by creating large numbers of short-lived objects that cause frequent, repeated pauses. We recommend upgrading libraries beyond the vulnerable versions.
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22569>CVE-2021-22569</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in protobuf java jar cve medium severity vulnerability vulnerable library protobuf java jar core protocol buffers library protocol buffers are a way of encoding structured data in an efficient yet extensible format library home page a href path to dependency file example example pom xml path to vulnerable library home wss scanner repository com google protobuf protobuf java protobuf java jar dependency hierarchy google cloud dataflow java sdk all jar root library beam sdks java io google cloud platform jar x protobuf java jar vulnerable library found in base branch master vulnerability details an issue in protobuf java allowed the interleaving of com google protobuf unknownfieldset fields in such a way that would be processed out of order a small malicious payload can occupy the parser for several minutes by creating large numbers of short lived objects that cause frequent repeated pauses we recommend upgrading libraries beyond the vulnerable versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href step up your open source security game with whitesource | 0 |
3,751 | 15,782,420,618 | IssuesEvent | 2021-04-01 12:46:43 | svengreb/tmpl | https://api.github.com/repos/svengreb/tmpl | closed | From npm to Yarn and back again | context-techstack context-workflow scope-compatibility scope-maintainability scope-stability target-base type-improvement | Some years ago I switched from [npm][] (`v4`) to [Yarn][yarn-v1] (`v1`) mainly because of the fantastic [workspace feature][yarn-docs-ws] for [monorepos][wiki-monorepo] as well as the great performance and UX improvements. This was a good decision and almost every popular and well-known project used to do the same, but with the announcement of [Yarn v2][yarn] (named [“berry“][gh-yarnpkg/berry]) the community got upset about the path the project has taken. Next to this, [npm joined GitHub][gh-blog-npm_joins] back in March 2020 which meant that the development continues in a good direction and is baked by the open source platform itself.
These events, the overall fantastic new npm release version `v7`, including the introduction of [workspaces][npm-docs-ws], and the fact that I never liked the disadvantage of requiring to use an “external“ package manager instead of the one that is bundled with Node, lead to the decision to finally switch back to npm again.
The only drawback is the constraint that the minimum [npm version must be `v7.7.0`][gh-npm/cli-rel-v7.7.0] because this is the first version that comes with workspace support for the `run-script` and `exec` commands through the `--workspace`/`-w` and `--workspaces`/`-ws` CLI flags, e.g. `npm run -w PACKAGE run lint`. The first Node version that ships with npm `v7.7.x` is [`v15.13.0`][node-dist-v15.13.0] which is globally available as of april 1, 2021 (no, it‘s not an april fool :smile:). To ensure that these constraints are met, without only relying on users to read the documentation, both `npm` and `node` will be added to [the `engines` field][npm-docs-pkgjson#engines] of the `package.json` file.
This change also requires updates to all documentations, including the addition of the version constraints, as well as updates to repository template features like the GitHub Action workflows.
The `.yarnrc` file will be replaced by `.npmrc` that includes the `package-lock=false` and `save-exact=false` configurations.
[gh-blog-npm_joins]: https://github.blog/2020-03-16-npm-is-joining-github
[gh-npm/cli-rel-v7.7.0]: https://github.com/npm/cli/releases/tag/v7.7.0
[gh-yarnpkg/berry]: https://github.com/yarnpkg/berry
[node-dist-v15.13.0]: https://nodejs.org/dist/v15.13.0
[npm-docs-pkgjson#engines]: https://docs.npmjs.com/cli/v7/configuring-npm/package-json#engines
[npm-docs-ws]: https://docs.npmjs.com/cli/v7/using-npm/workspaces
[npm]: https://www.npmjs.com
[wiki-monorepo]: https://en.wikipedia.org/wiki/Monorepo
[yarn-v1]: https://classic.yarnpkg.com
[yarn]: https://yarnpkg.com | True | From npm to Yarn and back again - Some years ago I switched from [npm][] (`v4`) to [Yarn][yarn-v1] (`v1`) mainly because of the fantastic [workspace feature][yarn-docs-ws] for [monorepos][wiki-monorepo] as well as the great performance and UX improvements. This was a good decision and almost every popular and well-known project used to do the same, but with the announcement of [Yarn v2][yarn] (named [“berry“][gh-yarnpkg/berry]) the community got upset about the path the project has taken. Next to this, [npm joined GitHub][gh-blog-npm_joins] back in March 2020 which meant that the development continues in a good direction and is baked by the open source platform itself.
These events, the overall fantastic new npm release version `v7`, including the introduction of [workspaces][npm-docs-ws], and the fact that I never liked the disadvantage of requiring to use an “external“ package manager instead of the one that is bundled with Node, lead to the decision to finally switch back to npm again.
The only drawback is the constraint that the minimum [npm version must be `v7.7.0`][gh-npm/cli-rel-v7.7.0] because this is the first version that comes with workspace support for the `run-script` and `exec` commands through the `--workspace`/`-w` and `--workspaces`/`-ws` CLI flags, e.g. `npm run -w PACKAGE run lint`. The first Node version that ships with npm `v7.7.x` is [`v15.13.0`][node-dist-v15.13.0] which is globally available as of april 1, 2021 (no, it‘s not an april fool :smile:). To ensure that these constraints are met, without only relying on users to read the documentation, both `npm` and `node` will be added to [the `engines` field][npm-docs-pkgjson#engines] of the `package.json` file.
This change also requires updates to all documentations, including the addition of the version constraints, as well as updates to repository template features like the GitHub Action workflows.
The `.yarnrc` file will be replaced by `.npmrc` that includes the `package-lock=false` and `save-exact=false` configurations.
[gh-blog-npm_joins]: https://github.blog/2020-03-16-npm-is-joining-github
[gh-npm/cli-rel-v7.7.0]: https://github.com/npm/cli/releases/tag/v7.7.0
[gh-yarnpkg/berry]: https://github.com/yarnpkg/berry
[node-dist-v15.13.0]: https://nodejs.org/dist/v15.13.0
[npm-docs-pkgjson#engines]: https://docs.npmjs.com/cli/v7/configuring-npm/package-json#engines
[npm-docs-ws]: https://docs.npmjs.com/cli/v7/using-npm/workspaces
[npm]: https://www.npmjs.com
[wiki-monorepo]: https://en.wikipedia.org/wiki/Monorepo
[yarn-v1]: https://classic.yarnpkg.com
[yarn]: https://yarnpkg.com | main | from npm to yarn and back again some years ago i switched from to mainly because of the fantastic for as well as the great performance and ux improvements this was a good decision and almost every popular and well known project used to do the same but with the announcement of named the community got upset about the path the project has taken next to this back in march which meant that the development continues in a good direction and is baked by the open source platform itself these events the overall fantastic new npm release version including the introduction of and the fact that i never liked the disadvantage of requiring to use an “external“ package manager instead of the one that is bundled with node lead to the decision to finally switch back to npm again the only drawback is the constraint that the minimum because this is the first version that comes with workspace support for the run script and exec commands through the workspace w and workspaces ws cli flags e g npm run w package run lint the first node version that ships with npm x is which is globally available as of april no it‘s not an april fool smile to ensure that these constraints are met without only relying on users to read the documentation both npm and node will be added to of the package json file this change also requires updates to all documentations including the addition of the version constraints as well as updates to repository template features like the github action workflows the yarnrc file will be replaced by npmrc that includes the package lock false and save exact false configurations | 1 |
366,281 | 10,819,238,305 | IssuesEvent | 2019-11-08 13:59:06 | luna/enso | https://api.github.com/repos/luna/enso | closed | Implement the Runtime Cache | Category: Interpreter Category: Performance Category: RTS Change: Non-Breaking Difficulty: Core Contributor Priority: Low Type: Enhancement | ### Summary
Given it's so crucial for performance when working with large amounts of data, a complete design for the IVC (#148) means that it's ready to be implemented. This task deals with the process of implementing the caching mechanism, dependency tracking and analysis, and eviction strategy.
### Value
Recomputation after changes in large projects will be faster through a reduction in the average amount of computation that needs to be performed after a change.
### Specification
- Implement the design for caching, first implementing the actual value storage.
- Then implement the dependency analysis and tracking.
- Finally, implement the cache eviction strategy.
### Acceptance Criteria & Test Cases
- IVC has been implemented and results in meaningful performance improvements.
- It also results in no regressions in the uncached general case performance.
| 1.0 | Implement the Runtime Cache - ### Summary
Given it's so crucial for performance when working with large amounts of data, a complete design for the IVC (#148) means that it's ready to be implemented. This task deals with the process of implementing the caching mechanism, dependency tracking and analysis, and eviction strategy.
### Value
Recomputation after changes in large projects will be faster through a reduction in the average amount of computation that needs to be performed after a change.
### Specification
- Implement the design for caching, first implementing the actual value storage.
- Then implement the dependency analysis and tracking.
- Finally, implement the cache eviction strategy.
### Acceptance Criteria & Test Cases
- IVC has been implemented and results in meaningful performance improvements.
- It also results in no regressions in the uncached general case performance.
| non_main | implement the runtime cache summary given it s so crucial for performance when working with large amounts of data a complete design for the ivc means that it s ready to be implemented this task deals with the process of implementing the caching mechanism dependency tracking and analysis and eviction strategy value recomputation after changes in large projects will be faster through a reduction in the average amount of computation that needs to be performed after a change specification implement the design for caching first implementing the actual value storage then implement the dependency analysis and tracking finally implement the cache eviction strategy acceptance criteria test cases ivc has been implemented and results in meaningful performance improvements it also results in no regressions in the uncached general case performance | 0 |
404 | 3,461,820,634 | IssuesEvent | 2015-12-20 12:15:26 | KazDragon/terminalpp | https://api.github.com/repos/KazDragon/terminalpp | closed | Consider migrating from CppUnit | in progress Maintainability | Due to a bug, CppUnit renders Clang's memory sanitizer useless. It would be useful to keep the option open. | True | Consider migrating from CppUnit - Due to a bug, CppUnit renders Clang's memory sanitizer useless. It would be useful to keep the option open. | main | consider migrating from cppunit due to a bug cppunit renders clang s memory sanitizer useless it would be useful to keep the option open | 1 |
3,778 | 15,897,019,576 | IssuesEvent | 2021-04-11 19:29:41 | microsoft/DirectXTex | https://api.github.com/repos/microsoft/DirectXTex | closed | CreateStreamOnHGlobal use is not recommended, find suitable replacement | maintainence | The ``CreateStreamOnHGlobal`` function has a number of performance issues, so it's recommended that you avoid using it. I need a memory IStream for the ``SaveToWICMemory`` functionality.
* For UWP I make use of ``InMemoryRandomAccessStream``
* For Windows Desktop on Windows 8 or later, I use ``SHCreateMemStream``
For other cases, I'm still falling back to ``CreateStreamOnHGlobal``. The best solution for other platforms may in fact be to write my own implementation of IStream. | True | CreateStreamOnHGlobal use is not recommended, find suitable replacement - The ``CreateStreamOnHGlobal`` function has a number of performance issues, so it's recommended that you avoid using it. I need a memory IStream for the ``SaveToWICMemory`` functionality.
* For UWP I make use of ``InMemoryRandomAccessStream``
* For Windows Desktop on Windows 8 or later, I use ``SHCreateMemStream``
For other cases, I'm still falling back to ``CreateStreamOnHGlobal``. The best solution for other platforms may in fact be to write my own implementation of IStream. | main | createstreamonhglobal use is not recommended find suitable replacement the createstreamonhglobal function has a number of performance issues so it s recommended that you avoid using it i need a memory istream for the savetowicmemory functionality for uwp i make use of inmemoryrandomaccessstream for windows desktop on windows or later i use shcreatememstream for other cases i m still falling back to createstreamonhglobal the best solution for other platforms may in fact be to write my own implementation of istream | 1 |
743,581 | 25,905,923,814 | IssuesEvent | 2022-12-15 10:12:31 | banco-alimentar/alimentestaideia.pt | https://api.github.com/repos/banco-alimentar/alimentestaideia.pt | closed | Referral code is not working in some cases | bug very.high.priority | **Describe the bug**
Same people have reported that using the referral code, the overview page is not updated.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to [Campaign page](https://www.alimentestaideia.pt/Referral/Overview/EDPDEZ22)
2. On another browser, in anonymous mode complete a donation using [this link](https://www.alimentestaideia.pt/Donation?referral=EDPDEZ22&utm_campaign=Alimente%20esta%20Ideia%20Banco%20|%20Alimentar%20Contra%20a%20Fome&utm_medium=email&utm_source=internal_newsletter)
3. Complete donation
4. Check if campaign page updated with your new donation
**Expected behavior**
The campaign overview page should update reflecting the new donation
| 1.0 | Referral code is not working in some cases - **Describe the bug**
Same people have reported that using the referral code, the overview page is not updated.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to [Campaign page](https://www.alimentestaideia.pt/Referral/Overview/EDPDEZ22)
2. On another browser, in anonymous mode complete a donation using [this link](https://www.alimentestaideia.pt/Donation?referral=EDPDEZ22&utm_campaign=Alimente%20esta%20Ideia%20Banco%20|%20Alimentar%20Contra%20a%20Fome&utm_medium=email&utm_source=internal_newsletter)
3. Complete donation
4. Check if campaign page updated with your new donation
**Expected behavior**
The campaign overview page should update reflecting the new donation
| non_main | referral code is not working in some cases describe the bug same people have reported that using the referral code the overview page is not updated to reproduce steps to reproduce the behavior go to on another browser in anonymous mode complete a donation using complete donation check if campaign page updated with your new donation expected behavior the campaign overview page should update reflecting the new donation | 0 |
2,828 | 10,141,099,911 | IssuesEvent | 2019-08-03 10:48:53 | chocolatey-community/chocolatey-package-requests | https://api.github.com/repos/chocolatey-community/chocolatey-package-requests | closed | RFP - mitmproxy | Status: Available For Maintainer(s) | mitmproxy is an interactive TLS-capable intercepting HTTP proxy for penetration testers and software developers.
https://mitmproxy.org/ | True | RFP - mitmproxy - mitmproxy is an interactive TLS-capable intercepting HTTP proxy for penetration testers and software developers.
https://mitmproxy.org/ | main | rfp mitmproxy mitmproxy is an interactive tls capable intercepting http proxy for penetration testers and software developers | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.