Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,519 | 6,572,207,154 | IssuesEvent | 2017-09-11 00:02:27 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | s3_bucket fails when loading JSON policy from a template | affects_2.1 aws bug_report cloud in progress waiting_on_maintainer | ##### Issue Type:
- Bug Report
##### Plugin Name:
s3_bucket.py
##### Ansible Version:
```
ansible 2.1.0
config file =
configured module search path = Default w/o overrides
```
##### Ansible Configuration:
N/A
##### Environment:
N/A
##### Summary:
Loading an S3 bucket policy from a file results in failure due to various silent conversions performed by the lookup function, ansible core, and the s3_bucket function itself.
##### Steps To Reproduce:
For bugs, please show exactly how to reproduce the problem. For new
features, show how the feature would be used.
Playbook:
```
vars:
domain: "example.com"
envname: "stage"
aws_account_number: "12345678"
region: "us-west-2"
s3_bucket_name: "{{domain | regex_replace('\\.', '-')}}-{{envname}}-{{aws_account_number}}"
elb_principal_mappings:
us-east-1: 127311923021
us-west-2: 797873946194
us-west-1: 027434742980
eu-west-1: 156460612806
eu-central-1: 054676820928
ap-southeast-1: 114774131450
ap-northeast-1: 582318560864
ap-southeast-2: 783225319266
ap-northeast-2: 600734575887
sa-east-1: 507241528517
- name: Create S3 asset bucket
s3_bucket:
name: "{{ s3_bucket_name }}"
region: "{{ region }}"
policy: "{{ lookup('template', './s3_bucket_policy.json.j2', convert_data=False) }}"
register: site_s3_bucket
```
Template (Could be anything but just for completeness):
```
{
"Id": "AllowELBWriteAccess",
"Version": "2012-10-17",
"Statement": [
{
"Sid": "Stmt1454671534294",
"Action": [
"s3:PutObject"
],
"Effect": "Allow",
"Resource": "arn:aws:s3:::{{domain | regex_replace('\.', '-')}}-{{envname}}-{{aws_account_number}}/accesslogs/AWSLogs/{{ aws_account_number }}/*",
"Principal": {
"AWS": [
"{{ elb_principal_mappings[region] }}"
]
}
}
]
}
```
##### Expected Results:
An S3 bucket is created
##### Actual Results:
Failure incorrectly implying that that the JSON is invalid.
```
TASK [Create S3 asset bucket] **************************************************
fatal: [127.0.0.1]: FAILED! => {"changed": false, "failed": true, "msg": "Policies must be valid JSON and the first byte must be '{'"}
```
| True | s3_bucket fails when loading JSON policy from a template - ##### Issue Type:
- Bug Report
##### Plugin Name:
s3_bucket.py
##### Ansible Version:
```
ansible 2.1.0
config file =
configured module search path = Default w/o overrides
```
##### Ansible Configuration:
N/A
##### Environment:
N/A
##### Summary:
Loading an S3 bucket policy from a file results in failure due to various silent conversions performed by the lookup function, ansible core, and the s3_bucket function itself.
##### Steps To Reproduce:
For bugs, please show exactly how to reproduce the problem. For new
features, show how the feature would be used.
Playbook:
```
vars:
domain: "example.com"
envname: "stage"
aws_account_number: "12345678"
region: "us-west-2"
s3_bucket_name: "{{domain | regex_replace('\\.', '-')}}-{{envname}}-{{aws_account_number}}"
elb_principal_mappings:
us-east-1: 127311923021
us-west-2: 797873946194
us-west-1: 027434742980
eu-west-1: 156460612806
eu-central-1: 054676820928
ap-southeast-1: 114774131450
ap-northeast-1: 582318560864
ap-southeast-2: 783225319266
ap-northeast-2: 600734575887
sa-east-1: 507241528517
- name: Create S3 asset bucket
s3_bucket:
name: "{{ s3_bucket_name }}"
region: "{{ region }}"
policy: "{{ lookup('template', './s3_bucket_policy.json.j2', convert_data=False) }}"
register: site_s3_bucket
```
Template (Could be anything but just for completeness):
```
{
"Id": "AllowELBWriteAccess",
"Version": "2012-10-17",
"Statement": [
{
"Sid": "Stmt1454671534294",
"Action": [
"s3:PutObject"
],
"Effect": "Allow",
"Resource": "arn:aws:s3:::{{domain | regex_replace('\.', '-')}}-{{envname}}-{{aws_account_number}}/accesslogs/AWSLogs/{{ aws_account_number }}/*",
"Principal": {
"AWS": [
"{{ elb_principal_mappings[region] }}"
]
}
}
]
}
```
##### Expected Results:
An S3 bucket is created
##### Actual Results:
Failure incorrectly implying that that the JSON is invalid.
```
TASK [Create S3 asset bucket] **************************************************
fatal: [127.0.0.1]: FAILED! => {"changed": false, "failed": true, "msg": "Policies must be valid JSON and the first byte must be '{'"}
```
| main | bucket fails when loading json policy from a template issue type bug report plugin name bucket py ansible version ansible config file configured module search path default w o overrides ansible configuration n a environment n a summary loading an bucket policy from a file results in failure due to various silent conversions performed by the lookup function ansible core and the bucket function itself steps to reproduce for bugs please show exactly how to reproduce the problem for new features show how the feature would be used playbook vars domain example com envname stage aws account number region us west bucket name domain regex replace envname aws account number elb principal mappings us east us west us west eu west eu central ap southeast ap northeast ap southeast ap northeast sa east name create asset bucket bucket name bucket name region region policy lookup template bucket policy json convert data false register site bucket template could be anything but just for completeness id allowelbwriteaccess version statement sid action putobject effect allow resource arn aws domain regex replace envname aws account number accesslogs awslogs aws account number principal aws elb principal mappings expected results an bucket is created actual results failure incorrectly implying that that the json is invalid task fatal failed changed false failed true msg policies must be valid json and the first byte must be | 1 |
262,244 | 19,768,800,754 | IssuesEvent | 2022-01-17 07:42:39 | kubernetes-sigs/descheduler | https://api.github.com/repos/kubernetes-sigs/descheduler | closed | Docs around autohealing are misleading | lifecycle/rotten kind/documentation | The [docs around autohealing](https://github.com/kubernetes-sigs/descheduler/blob/master/docs/user-guide.md#autoheal-node-problems) are a bit misleading in my opinion.
They link off to Node Problem Detector, claiming that `Node Problem Detector can detect specific Node problems and taint any Nodes which have those problems.`. In fact, NPD doesn't do any tainting. It's the `TaintNodeByCondition` feature of the node controller that takes _some_ conditions and turns them in to taints. However this only works for the default node conditions: `PIDPressure`, `MemoryPressure`, `DiskPressure`, `Ready`, and some cloud provider specific conditions.
There is an [open PR](https://github.com/kubernetes/node-problem-detector/pull/565) on NPD that wants to add this tainting behaviour, but the maintainers seem to think it shouldn't be NPD that does the tainting.
The effect is that the autoheal cycle describe doesn't actually work, at least not for custom conditions. It would be wonderful if it did, because it's quite a compelling outcome, and it would be amazing if it were offered exclusively in terms of Kubernetes first party tooling.
At the very least, we should change the wording in the docs to make it clear that NPD doesn't really participate in the autohealing. At most, I'm hoping that by raising this issue, the fact that this cycle doesn't currently work as intended can get a little more visibility. My guess at possible solutions:
* Merge the PR linked above, so that NPD creates taints
* Extend the node controller to add condition based taints for _all_ conditions, including custom ones
* Create a new project to convert conditions in to taints
* Add a new strategy in this repo that allows for descheduling based on Conditions | 1.0 | Docs around autohealing are misleading - The [docs around autohealing](https://github.com/kubernetes-sigs/descheduler/blob/master/docs/user-guide.md#autoheal-node-problems) are a bit misleading in my opinion.
They link off to Node Problem Detector, claiming that `Node Problem Detector can detect specific Node problems and taint any Nodes which have those problems.`. In fact, NPD doesn't do any tainting. It's the `TaintNodeByCondition` feature of the node controller that takes _some_ conditions and turns them in to taints. However this only works for the default node conditions: `PIDPressure`, `MemoryPressure`, `DiskPressure`, `Ready`, and some cloud provider specific conditions.
There is an [open PR](https://github.com/kubernetes/node-problem-detector/pull/565) on NPD that wants to add this tainting behaviour, but the maintainers seem to think it shouldn't be NPD that does the tainting.
The effect is that the autoheal cycle describe doesn't actually work, at least not for custom conditions. It would be wonderful if it did, because it's quite a compelling outcome, and it would be amazing if it were offered exclusively in terms of Kubernetes first party tooling.
At the very least, we should change the wording in the docs to make it clear that NPD doesn't really participate in the autohealing. At most, I'm hoping that by raising this issue, the fact that this cycle doesn't currently work as intended can get a little more visibility. My guess at possible solutions:
* Merge the PR linked above, so that NPD creates taints
* Extend the node controller to add condition based taints for _all_ conditions, including custom ones
* Create a new project to convert conditions in to taints
* Add a new strategy in this repo that allows for descheduling based on Conditions | non_main | docs around autohealing are misleading the are a bit misleading in my opinion they link off to node problem detector claiming that node problem detector can detect specific node problems and taint any nodes which have those problems in fact npd doesn t do any tainting it s the taintnodebycondition feature of the node controller that takes some conditions and turns them in to taints however this only works for the default node conditions pidpressure memorypressure diskpressure ready and some cloud provider specific conditions there is an on npd that wants to add this tainting behaviour but the maintainers seem to think it shouldn t be npd that does the tainting the effect is that the autoheal cycle describe doesn t actually work at least not for custom conditions it would be wonderful if it did because it s quite a compelling outcome and it would be amazing if it were offered exclusively in terms of kubernetes first party tooling at the very least we should change the wording in the docs to make it clear that npd doesn t really participate in the autohealing at most i m hoping that by raising this issue the fact that this cycle doesn t currently work as intended can get a little more visibility my guess at possible solutions merge the pr linked above so that npd creates taints extend the node controller to add condition based taints for all conditions including custom ones create a new project to convert conditions in to taints add a new strategy in this repo that allows for descheduling based on conditions | 0 |
3,662 | 14,942,823,689 | IssuesEvent | 2021-01-25 21:55:46 | flag-camp-2020-t3/Spare4Fun-Server | https://api.github.com/repos/flag-camp-2020-t3/Spare4Fun-Server | opened | Dao Throw Exception & captured by exception handler | backend-v2.0 critical enhancement maintain / refactor | branch: ```refactor-dao-throw-exception-exception-handler``` | True | Dao Throw Exception & captured by exception handler - branch: ```refactor-dao-throw-exception-exception-handler``` | main | dao throw exception captured by exception handler branch refactor dao throw exception exception handler | 1 |
4,390 | 22,497,682,559 | IssuesEvent | 2022-06-23 09:00:59 | zoj613/polyagamma | https://api.github.com/repos/zoj613/polyagamma | closed | MAINT: Get full precision for 32 bit floating point random values. | good first issue maintainance random_polyagamma | A numpy [issue](https://github.com/numpy/numpy/issues/17478) showed that numpy's formula for generating 32bit random uniform values was waiting bits. There is a PR that fixes this which was merge in https://github.com/numpy/numpy/pull/20314.
It states
```
The formula to convert a 32 bit random integer to a random float32,
(next_uint32(bitgen_state) >> 9) * (1.0f / 8388608.0f)
shifts by one bit too many, resulting in uniform float32 samples always
having a 0 in the least significant bit. The formula is corrected to
(next_uint32(bitgen_state) >> 8) * (1.0f / 16777216.0f)
Occurrences of the incorrect formula in numpy/random/tests/test_direct.py
were also corrected.
```
Since the project uses this formula at https://github.com/zoj613/polyagamma/blob/fca4a8bce462066803c888708e2f109d1991c00c/src/pgm_macros.h#L54-L56
it is worth updating to get the full precision of 32bit float because they are used extensively in the accept-rejection steps of the samplers. Im positive this slight adjustment does not affect the results nor the performance but I think it's worth updating the formula. | True | MAINT: Get full precision for 32 bit floating point random values. - A numpy [issue](https://github.com/numpy/numpy/issues/17478) showed that numpy's formula for generating 32bit random uniform values was waiting bits. There is a PR that fixes this which was merge in https://github.com/numpy/numpy/pull/20314.
It states
```
The formula to convert a 32 bit random integer to a random float32,
(next_uint32(bitgen_state) >> 9) * (1.0f / 8388608.0f)
shifts by one bit too many, resulting in uniform float32 samples always
having a 0 in the least significant bit. The formula is corrected to
(next_uint32(bitgen_state) >> 8) * (1.0f / 16777216.0f)
Occurrences of the incorrect formula in numpy/random/tests/test_direct.py
were also corrected.
```
Since the project uses this formula at https://github.com/zoj613/polyagamma/blob/fca4a8bce462066803c888708e2f109d1991c00c/src/pgm_macros.h#L54-L56
it is worth updating to get the full precision of 32bit float because they are used extensively in the accept-rejection steps of the samplers. Im positive this slight adjustment does not affect the results nor the performance but I think it's worth updating the formula. | main | maint get full precision for bit floating point random values a numpy showed that numpy s formula for generating random uniform values was waiting bits there is a pr that fixes this which was merge in it states the formula to convert a bit random integer to a random next bitgen state shifts by one bit too many resulting in uniform samples always having a in the least significant bit the formula is corrected to next bitgen state occurrences of the incorrect formula in numpy random tests test direct py were also corrected since the project uses this formula at it is worth updating to get the full precision of float because they are used extensively in the accept rejection steps of the samplers im positive this slight adjustment does not affect the results nor the performance but i think it s worth updating the formula | 1 |
5,214 | 26,464,344,104 | IssuesEvent | 2023-01-16 21:18:37 | bazelbuild/intellij | https://api.github.com/repos/bazelbuild/intellij | closed | Flag --incompatible_disable_starlark_host_transitions will break Android Studio Plugin in Bazel 7.0 | type: bug product: Android Studio topic: bazel awaiting-maintainer | Incompatible flag `--incompatible_disable_starlark_host_transitions` will be enabled by default in the next major release (Bazel 7.0), thus breaking Android Studio Plugin. Please migrate to fix this and unblock the flip of this flag.
The flag is documented here: [bazelbuild/bazel#17032](https://github.com/bazelbuild/bazel/issues/17032).
Please check the following CI builds for build and test results:
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc33-4a0d-bfba-a9d0f36f4e1c)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc36-44c7-9a05-b49d4a1c32f5)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc2c-44c5-ba2f-d20d0c3515e1)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc3c-4a48-a7bc-5c475690a11b)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc39-4303-8caa-95e9935cab6d)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc3f-4bda-9815-991ef8b9d7ad)
Never heard of incompatible flags before? We have [documentation](https://docs.bazel.build/versions/master/backward-compatibility.html) that explains everything.
If you have any questions, please file an issue in https://github.com/bazelbuild/continuous-integration. | True | Flag --incompatible_disable_starlark_host_transitions will break Android Studio Plugin in Bazel 7.0 - Incompatible flag `--incompatible_disable_starlark_host_transitions` will be enabled by default in the next major release (Bazel 7.0), thus breaking Android Studio Plugin. Please migrate to fix this and unblock the flip of this flag.
The flag is documented here: [bazelbuild/bazel#17032](https://github.com/bazelbuild/bazel/issues/17032).
Please check the following CI builds for build and test results:
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc33-4a0d-bfba-a9d0f36f4e1c)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc36-44c7-9a05-b49d4a1c32f5)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc2c-44c5-ba2f-d20d0c3515e1)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc3c-4a48-a7bc-5c475690a11b)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc39-4303-8caa-95e9935cab6d)
- [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc3f-4bda-9815-991ef8b9d7ad)
Never heard of incompatible flags before? We have [documentation](https://docs.bazel.build/versions/master/backward-compatibility.html) that explains everything.
If you have any questions, please file an issue in https://github.com/bazelbuild/continuous-integration. | main | flag incompatible disable starlark host transitions will break android studio plugin in bazel incompatible flag incompatible disable starlark host transitions will be enabled by default in the next major release bazel thus breaking android studio plugin please migrate to fix this and unblock the flip of this flag the flag is documented here please check the following ci builds for build and test results never heard of incompatible flags before we have that explains everything if you have any questions please file an issue in | 1 |
507,049 | 14,679,153,479 | IssuesEvent | 2020-12-31 06:04:31 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | closed | Remove constraints from ScrollView in profile_reset_pin_activity | Priority: Nice-to-have Type: Improvement good first issue | Remove constraints from ScrollView in profile_reset_pin_activity both landscape and portrait xml files.
https://github.com/oppia/oppia-android/blob/2511715a4770cda65df593cf6821b17b7f8f3d28/app/src/main/res/layout/profile_reset_pin_activity.xml#L55
| 1.0 | Remove constraints from ScrollView in profile_reset_pin_activity - Remove constraints from ScrollView in profile_reset_pin_activity both landscape and portrait xml files.
https://github.com/oppia/oppia-android/blob/2511715a4770cda65df593cf6821b17b7f8f3d28/app/src/main/res/layout/profile_reset_pin_activity.xml#L55
| non_main | remove constraints from scrollview in profile reset pin activity remove constraints from scrollview in profile reset pin activity both landscape and portrait xml files | 0 |
46,944 | 24,794,622,649 | IssuesEvent | 2022-10-24 16:10:55 | iree-org/iree | https://api.github.com/repos/iree-org/iree | closed | Add `hal_inline` dialect/module for tiny environments. | runtime performance ⚡ | For environments where the execution model is known to be exclusively local and inline (embedded systems) we can have a paired down HAL that pretty much only contains executable support. The idea is to still use HAL executable translation in the compiler but lowering the stream dialect to a new lightweight dialect that pretty much only manages executables and dispatches. Most of the local/ implementation of the executable loader and the loaders themselves have no dependencies on command buffers, allocators, buffers, or devices and can be cleanly pulled into the module without bringing in the bulk of the HAL API.
It's debatable whether the allocator/buffer stuff should be included - that would allow the coming allocator types to be reused but at the cost of additional non-user-controllable overheads. Since the local executables take byte spans all buffers could just be `iree_vm_buffer_t` which is already compiled in and available for use - and since they take a custom `iree_allocator_t` it's still possible for hosting applications to manage memory however they want.
Executables themselves will still be injected on the module when created same as today, allowing for dynamic, static, vmvx, etc executables to be run this way. This allows us to separate the execution model from the deployment model at the cost of a few vtables.
Outline:
* [x] Add `hal_inline` dialect with basic ops:
* [x] `hal_inline.executable.create`
* [x] `hal_inline.executable.dispatch`
* [x] `hal_inline.executable_layout.create`? (still need this to reuse loaders/libraries)
* [x] Add `--execution-mode=` iree-compile flag to switch between `hal-async` and `hal-inline` (or w/e)
* [x] Have a new `iree-hal-inline-transformation-pipeline` that still performs interface materialization and executable translation but otherwise lowers `stream` itself
* [x] Add `iree/modules/hal_inline` runtime module that links directly against the `iree/hal/local/` libraries
* [x] Build a runner tool that uses the inline module (or make iree-run-module/etc always support it with a flag)
(could also call this the `inline` dialect or something - it's still technically a HAL though as the executables being called are abstracted across hardware - can have CPU/FPGA/DSP/etc) | True | Add `hal_inline` dialect/module for tiny environments. - For environments where the execution model is known to be exclusively local and inline (embedded systems) we can have a paired down HAL that pretty much only contains executable support. The idea is to still use HAL executable translation in the compiler but lowering the stream dialect to a new lightweight dialect that pretty much only manages executables and dispatches. Most of the local/ implementation of the executable loader and the loaders themselves have no dependencies on command buffers, allocators, buffers, or devices and can be cleanly pulled into the module without bringing in the bulk of the HAL API.
It's debatable whether the allocator/buffer stuff should be included - that would allow the coming allocator types to be reused but at the cost of additional non-user-controllable overheads. Since the local executables take byte spans all buffers could just be `iree_vm_buffer_t` which is already compiled in and available for use - and since they take a custom `iree_allocator_t` it's still possible for hosting applications to manage memory however they want.
Executables themselves will still be injected on the module when created same as today, allowing for dynamic, static, vmvx, etc executables to be run this way. This allows us to separate the execution model from the deployment model at the cost of a few vtables.
Outline:
* [x] Add `hal_inline` dialect with basic ops:
* [x] `hal_inline.executable.create`
* [x] `hal_inline.executable.dispatch`
* [x] `hal_inline.executable_layout.create`? (still need this to reuse loaders/libraries)
* [x] Add `--execution-mode=` iree-compile flag to switch between `hal-async` and `hal-inline` (or w/e)
* [x] Have a new `iree-hal-inline-transformation-pipeline` that still performs interface materialization and executable translation but otherwise lowers `stream` itself
* [x] Add `iree/modules/hal_inline` runtime module that links directly against the `iree/hal/local/` libraries
* [x] Build a runner tool that uses the inline module (or make iree-run-module/etc always support it with a flag)
(could also call this the `inline` dialect or something - it's still technically a HAL though as the executables being called are abstracted across hardware - can have CPU/FPGA/DSP/etc) | non_main | add hal inline dialect module for tiny environments for environments where the execution model is known to be exclusively local and inline embedded systems we can have a paired down hal that pretty much only contains executable support the idea is to still use hal executable translation in the compiler but lowering the stream dialect to a new lightweight dialect that pretty much only manages executables and dispatches most of the local implementation of the executable loader and the loaders themselves have no dependencies on command buffers allocators buffers or devices and can be cleanly pulled into the module without bringing in the bulk of the hal api it s debatable whether the allocator buffer stuff should be included that would allow the coming allocator types to be reused but at the cost of additional non user controllable overheads since the local executables take byte spans all buffers could just be iree vm buffer t which is already compiled in and available for use and since they take a custom iree allocator t it s still possible for hosting applications to manage memory however they want executables themselves will still be injected on the module when created same as today allowing for dynamic static vmvx etc executables to be run this way this allows us to separate the execution model from the deployment model at the cost of a few vtables outline add hal inline dialect with basic ops hal inline executable create hal inline executable dispatch hal inline executable layout create still need this to reuse loaders libraries add execution mode iree compile flag to switch between hal async and hal inline or w e have a new iree hal inline transformation pipeline that still performs interface materialization and executable translation but otherwise lowers stream itself add iree modules hal inline runtime module that links directly against the iree hal local libraries build a runner tool that uses the inline module or make iree run module etc always support it with a flag could also call this the inline dialect or something it s still technically a hal though as the executables being called are abstracted across hardware can have cpu fpga dsp etc | 0 |
14,714 | 3,419,369,800 | IssuesEvent | 2015-12-08 09:26:03 | centreon/centreon | https://api.github.com/repos/centreon/centreon | closed | Delete a poller doesn't delete associated Centreon Broker configuration | BetaTest Kind/Bug Status/Solved | When I delete a poller the associated Centreon Broker configuration is disabled but not deleted.
to have a clean configuration it would be good to remove the entire configuration of Centreon Broker
Regards, | 1.0 | Delete a poller doesn't delete associated Centreon Broker configuration - When I delete a poller the associated Centreon Broker configuration is disabled but not deleted.
to have a clean configuration it would be good to remove the entire configuration of Centreon Broker
Regards, | non_main | delete a poller doesn t delete associated centreon broker configuration when i delete a poller the associated centreon broker configuration is disabled but not deleted to have a clean configuration it would be good to remove the entire configuration of centreon broker regards | 0 |
239,613 | 7,799,878,792 | IssuesEvent | 2018-06-09 01:34:37 | tine20/Tine-2.0-Open-Source-Groupware-and-CRM | https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM | closed | 0005764:
convert $_folder->cache_uidvalidity to integer (in DB) | Felamimail Mantis high priority | **Reported by pschuele on 20 Feb 2012 09:00**
convert $_folder->cache_uidvalidity to integer (in DB)
- as imap_uidvalidity is an integer, too and there are issues regarding comparison between the two when using postgresql
**Additional information:** http://www.tine20.org/forum/viewtopic.php?f=10&t=10508
| 1.0 | 0005764:
convert $_folder->cache_uidvalidity to integer (in DB) - **Reported by pschuele on 20 Feb 2012 09:00**
convert $_folder->cache_uidvalidity to integer (in DB)
- as imap_uidvalidity is an integer, too and there are issues regarding comparison between the two when using postgresql
**Additional information:** http://www.tine20.org/forum/viewtopic.php?f=10&t=10508
| non_main | convert folder cache uidvalidity to integer in db reported by pschuele on feb convert folder gt cache uidvalidity to integer in db as imap uidvalidity is an integer too and there are issues regarding comparison between the two when using postgresql additional information | 0 |
4,752 | 24,509,600,788 | IssuesEvent | 2022-10-10 19:55:50 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | The columns endpoint results in a 500, possibly metadata related | type: bug work: backend status: ready restricted: maintainers | ## Description
Endpoint: `http://localhost:8000/api/db/v0/tables/<table_id>/columns/`
```
Environment:
Request Method: GET
Request URL: http://localhost:8000/api/db/v0/tables/5/columns/?limit=500
Django Version: 3.1.14
Python Version: 3.9.8
Installed Applications:
['django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'django_filters',
'django_property_filter',
'mathesar']
Installed Middleware:
['django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware']
Traceback (most recent call last):
File "/code/mathesar/models/base.py", line 675, in __getattribute__
return super().__getattribute__(name)
During handling of the above exception ('Column' object has no attribute 'primary_key'), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__
return self._index[key]
The above exception ('nspname') was the direct cause of the following exception:
File "/code/mathesar/models/base.py", line 675, in __getattribute__
return super().__getattribute__(name)
File "/code/mathesar/models/base.py", line 696, in _sa_column
return self.table.sa_columns[self.name]
File "/code/mathesar/models/base.py", line 362, in sa_columns
return self._enriched_column_sa_table.columns
File "/code/mathesar/models/base.py", line 350, in _enriched_column_sa_table
table=self._sa_table,
File "/code/mathesar/state/cached_property.py", line 62, in __get__
new_value = self.original_get_fn(instance)
File "/code/mathesar/models/base.py", line 331, in _sa_table
sa_table = reflect_table_from_oid(
File "/code/db/tables/operations/select.py", line 23, in reflect_table_from_oid
tables = reflect_tables_from_oids([oid], engine, metadata=metadata, connection_to_use=connection_to_use)
File "/code/db/tables/operations/select.py", line 29, in reflect_tables_from_oids
get_map_of_table_oid_to_schema_name_and_table_name(
File "/code/db/tables/operations/select.py", line 59, in get_map_of_table_oid_to_schema_name_and_table_name
select(pg_namespace.c.nspname, pg_class.c.relname, pg_class.c.oid)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__
util.raise_(AttributeError(key), replace_context=err)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
During handling of the above exception (nspname), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__
return self._index[key]
The above exception ('nspname') was the direct cause of the following exception:
File "/code/mathesar/models/base.py", line 675, in __getattribute__
return super().__getattribute__(name)
File "/code/mathesar/models/base.py", line 696, in _sa_column
return self.table.sa_columns[self.name]
File "/code/mathesar/models/base.py", line 362, in sa_columns
return self._enriched_column_sa_table.columns
File "/code/mathesar/models/base.py", line 350, in _enriched_column_sa_table
table=self._sa_table,
File "/code/mathesar/state/cached_property.py", line 62, in __get__
new_value = self.original_get_fn(instance)
File "/code/mathesar/models/base.py", line 331, in _sa_table
sa_table = reflect_table_from_oid(
File "/code/db/tables/operations/select.py", line 23, in reflect_table_from_oid
tables = reflect_tables_from_oids([oid], engine, metadata=metadata, connection_to_use=connection_to_use)
File "/code/db/tables/operations/select.py", line 29, in reflect_tables_from_oids
get_map_of_table_oid_to_schema_name_and_table_name(
File "/code/db/tables/operations/select.py", line 59, in get_map_of_table_oid_to_schema_name_and_table_name
select(pg_namespace.c.nspname, pg_class.c.relname, pg_class.c.oid)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__
util.raise_(AttributeError(key), replace_context=err)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
During handling of the above exception (nspname), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 826, in __getattr__
return getattr(self.comparator, key)
The above exception ('Comparator' object has no attribute '_sa_column') was the direct cause of the following exception:
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view
return view_func(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view
return self.dispatch(request, *args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch
response = self.handle_exception(exc)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception
response = exception_handler(exc, context)
File "/code/mathesar/exception_handlers.py", line 55, in mathesar_exception_handler
raise exc
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch
response = handler(request, *args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/mixins.py", line 38, in list
queryset = self.filter_queryset(self.get_queryset())
File "/code/mathesar/api/db/viewsets/columns.py", line 33, in get_queryset
queryset = Column.objects.filter(table=self.kwargs['table_pk']).order_by('attnum')
File "/usr/local/lib/python3.9/site-packages/django/db/models/manager.py", line 85, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/code/mathesar/models/base.py", line 70, in get_queryset
make_sure_initial_reflection_happened()
File "/code/mathesar/state/base.py", line 8, in make_sure_initial_reflection_happened
reset_reflection()
File "/code/mathesar/state/base.py", line 27, in reset_reflection
_trigger_django_model_reflection()
File "/code/mathesar/state/base.py", line 31, in _trigger_django_model_reflection
reflect_db_objects(metadata=get_cached_metadata())
File "/code/mathesar/state/django.py", line 44, in reflect_db_objects
reflect_columns_from_tables(tables, metadata=metadata)
File "/code/mathesar/state/django.py", line 123, in reflect_columns_from_tables
models._compute_preview_template(table)
File "/code/mathesar/models/base.py", line 859, in _compute_preview_template
if column.primary_key:
File "/code/mathesar/models/base.py", line 681, in __getattribute__
return getattr(self._sa_column, name)
File "/code/mathesar/models/base.py", line 681, in __getattribute__
return getattr(self._sa_column, name)
File "/code/mathesar/models/base.py", line 681, in __getattribute__
return getattr(self._sa_column, name)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 828, in __getattr__
util.raise_(
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
Exception Type: AttributeError at /api/db/v0/tables/5/columns/
Exception Value: Neither 'MathesarColumn' object nor 'Comparator' object has an attribute '_sa_column'
```
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
## To Reproduce
<!-- How can we recreate this bug? Please try to provide a Minimal, Complete, and Verifiable (http://stackoverflow.com/help/mcve) example if code-related. -->
## Environment
- OS: (_eg._ macOS 10.14.6; Fedora 32)
- Browser: (_eg._ Safari; Firefox)
- Browser Version: (_eg._ 13; 73)
- Other info:
## Additional context
<!-- Add any other context about the problem or screenshots here. -->
| True | The columns endpoint results in a 500, possibly metadata related - ## Description
Endpoint: `http://localhost:8000/api/db/v0/tables/<table_id>/columns/`
```
Environment:
Request Method: GET
Request URL: http://localhost:8000/api/db/v0/tables/5/columns/?limit=500
Django Version: 3.1.14
Python Version: 3.9.8
Installed Applications:
['django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'django_filters',
'django_property_filter',
'mathesar']
Installed Middleware:
['django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware']
Traceback (most recent call last):
File "/code/mathesar/models/base.py", line 675, in __getattribute__
return super().__getattribute__(name)
During handling of the above exception ('Column' object has no attribute 'primary_key'), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__
return self._index[key]
The above exception ('nspname') was the direct cause of the following exception:
File "/code/mathesar/models/base.py", line 675, in __getattribute__
return super().__getattribute__(name)
File "/code/mathesar/models/base.py", line 696, in _sa_column
return self.table.sa_columns[self.name]
File "/code/mathesar/models/base.py", line 362, in sa_columns
return self._enriched_column_sa_table.columns
File "/code/mathesar/models/base.py", line 350, in _enriched_column_sa_table
table=self._sa_table,
File "/code/mathesar/state/cached_property.py", line 62, in __get__
new_value = self.original_get_fn(instance)
File "/code/mathesar/models/base.py", line 331, in _sa_table
sa_table = reflect_table_from_oid(
File "/code/db/tables/operations/select.py", line 23, in reflect_table_from_oid
tables = reflect_tables_from_oids([oid], engine, metadata=metadata, connection_to_use=connection_to_use)
File "/code/db/tables/operations/select.py", line 29, in reflect_tables_from_oids
get_map_of_table_oid_to_schema_name_and_table_name(
File "/code/db/tables/operations/select.py", line 59, in get_map_of_table_oid_to_schema_name_and_table_name
select(pg_namespace.c.nspname, pg_class.c.relname, pg_class.c.oid)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__
util.raise_(AttributeError(key), replace_context=err)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
During handling of the above exception (nspname), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__
return self._index[key]
The above exception ('nspname') was the direct cause of the following exception:
File "/code/mathesar/models/base.py", line 675, in __getattribute__
return super().__getattribute__(name)
File "/code/mathesar/models/base.py", line 696, in _sa_column
return self.table.sa_columns[self.name]
File "/code/mathesar/models/base.py", line 362, in sa_columns
return self._enriched_column_sa_table.columns
File "/code/mathesar/models/base.py", line 350, in _enriched_column_sa_table
table=self._sa_table,
File "/code/mathesar/state/cached_property.py", line 62, in __get__
new_value = self.original_get_fn(instance)
File "/code/mathesar/models/base.py", line 331, in _sa_table
sa_table = reflect_table_from_oid(
File "/code/db/tables/operations/select.py", line 23, in reflect_table_from_oid
tables = reflect_tables_from_oids([oid], engine, metadata=metadata, connection_to_use=connection_to_use)
File "/code/db/tables/operations/select.py", line 29, in reflect_tables_from_oids
get_map_of_table_oid_to_schema_name_and_table_name(
File "/code/db/tables/operations/select.py", line 59, in get_map_of_table_oid_to_schema_name_and_table_name
select(pg_namespace.c.nspname, pg_class.c.relname, pg_class.c.oid)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__
util.raise_(AttributeError(key), replace_context=err)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
During handling of the above exception (nspname), another exception occurred:
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 826, in __getattr__
return getattr(self.comparator, key)
The above exception ('Comparator' object has no attribute '_sa_column') was the direct cause of the following exception:
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view
return view_func(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view
return self.dispatch(request, *args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch
response = self.handle_exception(exc)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception
response = exception_handler(exc, context)
File "/code/mathesar/exception_handlers.py", line 55, in mathesar_exception_handler
raise exc
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch
response = handler(request, *args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/mixins.py", line 38, in list
queryset = self.filter_queryset(self.get_queryset())
File "/code/mathesar/api/db/viewsets/columns.py", line 33, in get_queryset
queryset = Column.objects.filter(table=self.kwargs['table_pk']).order_by('attnum')
File "/usr/local/lib/python3.9/site-packages/django/db/models/manager.py", line 85, in manager_method
return getattr(self.get_queryset(), name)(*args, **kwargs)
File "/code/mathesar/models/base.py", line 70, in get_queryset
make_sure_initial_reflection_happened()
File "/code/mathesar/state/base.py", line 8, in make_sure_initial_reflection_happened
reset_reflection()
File "/code/mathesar/state/base.py", line 27, in reset_reflection
_trigger_django_model_reflection()
File "/code/mathesar/state/base.py", line 31, in _trigger_django_model_reflection
reflect_db_objects(metadata=get_cached_metadata())
File "/code/mathesar/state/django.py", line 44, in reflect_db_objects
reflect_columns_from_tables(tables, metadata=metadata)
File "/code/mathesar/state/django.py", line 123, in reflect_columns_from_tables
models._compute_preview_template(table)
File "/code/mathesar/models/base.py", line 859, in _compute_preview_template
if column.primary_key:
File "/code/mathesar/models/base.py", line 681, in __getattribute__
return getattr(self._sa_column, name)
File "/code/mathesar/models/base.py", line 681, in __getattribute__
return getattr(self._sa_column, name)
File "/code/mathesar/models/base.py", line 681, in __getattribute__
return getattr(self._sa_column, name)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 828, in __getattr__
util.raise_(
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
Exception Type: AttributeError at /api/db/v0/tables/5/columns/
Exception Value: Neither 'MathesarColumn' object nor 'Comparator' object has an attribute '_sa_column'
```
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
## To Reproduce
<!-- How can we recreate this bug? Please try to provide a Minimal, Complete, and Verifiable (http://stackoverflow.com/help/mcve) example if code-related. -->
## Environment
- OS: (_eg._ macOS 10.14.6; Fedora 32)
- Browser: (_eg._ Safari; Firefox)
- Browser Version: (_eg._ 13; 73)
- Other info:
## Additional context
<!-- Add any other context about the problem or screenshots here. -->
| main | the columns endpoint results in a possibly metadata related description endpoint environment request method get request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware traceback most recent call last file code mathesar models base py line in getattribute return super getattribute name during handling of the above exception column object has no attribute primary key another exception occurred file usr local lib site packages sqlalchemy sql base py line in getattr return self index the above exception nspname was the direct cause of the following exception file code mathesar models base py line in getattribute return super getattribute name file code mathesar models base py line in sa column return self table sa columns file code mathesar models base py line in sa columns return self enriched column sa table columns file code mathesar models base py line in enriched column sa table table self sa table file code mathesar state cached property py line in get new value self original get fn instance file code mathesar models base py line in sa table sa table reflect table from oid file code db tables operations select py line in reflect table from oid tables reflect tables from oids engine metadata metadata connection to use connection to use file code db tables operations select py line in reflect tables from oids get map of table oid to schema name and table name file code db tables operations select py line in get map of table oid to schema name and table name select pg namespace c nspname pg class c relname pg class c oid file usr local lib site packages sqlalchemy sql base py line in getattr util raise attributeerror key replace context err file usr local lib site packages sqlalchemy util compat py line in raise raise exception during handling of the above exception nspname another exception occurred file usr local lib site packages sqlalchemy sql base py line in getattr return self index the above exception nspname was the direct cause of the following exception file code mathesar models base py line in getattribute return super getattribute name file code mathesar models base py line in sa column return self table sa columns file code mathesar models base py line in sa columns return self enriched column sa table columns file code mathesar models base py line in enriched column sa table table self sa table file code mathesar state cached property py line in get new value self original get fn instance file code mathesar models base py line in sa table sa table reflect table from oid file code db tables operations select py line in reflect table from oid tables reflect tables from oids engine metadata metadata connection to use connection to use file code db tables operations select py line in reflect tables from oids get map of table oid to schema name and table name file code db tables operations select py line in get map of table oid to schema name and table name select pg namespace c nspname pg class c relname pg class c oid file usr local lib site packages sqlalchemy sql base py line in getattr util raise attributeerror key replace context err file usr local lib site packages sqlalchemy util compat py line in raise raise exception during handling of the above exception nspname another exception occurred file usr local lib site packages sqlalchemy sql elements py line in getattr return getattr self comparator key the above exception comparator object has no attribute sa column was the direct cause of the following exception file usr local lib site packages django core handlers exception py line in inner response get response request file usr local lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr local lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file usr local lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file usr local lib site packages rest framework views py line in dispatch response self handle exception exc file usr local lib site packages rest framework views py line in handle exception response exception handler exc context file code mathesar exception handlers py line in mathesar exception handler raise exc file usr local lib site packages rest framework views py line in dispatch response handler request args kwargs file usr local lib site packages rest framework mixins py line in list queryset self filter queryset self get queryset file code mathesar api db viewsets columns py line in get queryset queryset column objects filter table self kwargs order by attnum file usr local lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file code mathesar models base py line in get queryset make sure initial reflection happened file code mathesar state base py line in make sure initial reflection happened reset reflection file code mathesar state base py line in reset reflection trigger django model reflection file code mathesar state base py line in trigger django model reflection reflect db objects metadata get cached metadata file code mathesar state django py line in reflect db objects reflect columns from tables tables metadata metadata file code mathesar state django py line in reflect columns from tables models compute preview template table file code mathesar models base py line in compute preview template if column primary key file code mathesar models base py line in getattribute return getattr self sa column name file code mathesar models base py line in getattribute return getattr self sa column name file code mathesar models base py line in getattribute return getattr self sa column name file usr local lib site packages sqlalchemy sql elements py line in getattr util raise file usr local lib site packages sqlalchemy util compat py line in raise raise exception exception type attributeerror at api db tables columns exception value neither mathesarcolumn object nor comparator object has an attribute sa column expected behavior to reproduce environment os eg macos fedora browser eg safari firefox browser version eg other info additional context | 1 |
130,908 | 18,213,855,857 | IssuesEvent | 2021-09-30 00:00:49 | ghc-dev/Margaret-Rose | https://api.github.com/repos/ghc-dev/Margaret-Rose | opened | CVE-2021-21295 (Medium) detected in netty-codec-http-4.1.39.Final.jar | security vulnerability | ## CVE-2021-21295 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http-4.1.39.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: Margaret-Rose/build.gradle</p>
<p>Path to vulnerable library: /caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **netty-codec-http-4.1.39.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Margaret-Rose/commit/c4e02f67dd4676db950425e7cac2d1a3f3883f24">c4e02f67dd4676db950425e7cac2d1a3f3883f24</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel's pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`.
<p>Publish Date: 2021-03-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21295>CVE-2021-21295</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-wm47-8v5p-wjpj">https://github.com/advisories/GHSA-wm47-8v5p-wjpj</a></p>
<p>Release Date: 2021-03-09</p>
<p>Fix Resolution: io.netty:netty-all:4.1.60;io.netty:netty-codec-http:4.1.60;io.netty:netty-codec-http2:4.1.60</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http","packageVersion":"4.1.39.Final","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"io.netty:netty-codec-http:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.60;io.netty:netty-codec-http:4.1.60;io.netty:netty-codec-http2:4.1.60"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21295","vulnerabilityDetails":"Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers \u0026 clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel\u0027s pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21295","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-21295 (Medium) detected in netty-codec-http-4.1.39.Final.jar - ## CVE-2021-21295 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http-4.1.39.Final.jar</b></p></summary>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: Margaret-Rose/build.gradle</p>
<p>Path to vulnerable library: /caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **netty-codec-http-4.1.39.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Margaret-Rose/commit/c4e02f67dd4676db950425e7cac2d1a3f3883f24">c4e02f67dd4676db950425e7cac2d1a3f3883f24</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel's pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`.
<p>Publish Date: 2021-03-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21295>CVE-2021-21295</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-wm47-8v5p-wjpj">https://github.com/advisories/GHSA-wm47-8v5p-wjpj</a></p>
<p>Release Date: 2021-03-09</p>
<p>Fix Resolution: io.netty:netty-all:4.1.60;io.netty:netty-codec-http:4.1.60;io.netty:netty-codec-http2:4.1.60</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http","packageVersion":"4.1.39.Final","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"io.netty:netty-codec-http:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.60;io.netty:netty-codec-http:4.1.60;io.netty:netty-codec-http2:4.1.60"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21295","vulnerabilityDetails":"Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers \u0026 clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel\u0027s pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21295","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_main | cve medium detected in netty codec http final jar cve medium severity vulnerability vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file margaret rose build gradle path to vulnerable library caches modules files io netty netty codec http final netty codec http final jar dependency hierarchy x netty codec http final jar vulnerable library found in head commit a href found in base branch master vulnerability details netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty io netty netty codec before version final there is a vulnerability that enables request smuggling if a content length header is present in the original http request the field is not validated by as it is propagated up this is fine as long as the request is not proxied through as http if the request comes in as an http stream gets converted into the http domain objects httprequest httpcontent etc via and then sent up to the child channel s pipeline and proxied through a remote peer as http this may result in request smuggling in a proxy case users may assume the content length is validated somehow which is not the case if the request is forwarded to a backend channel that is a http connection the content length now has meaning and needs to be checked an attacker can smuggle requests inside the body as it gets downgraded from http to http for an example attack refer to the linked github advisory users are only affected if all of this is true or is used is used to convert to http objects and these http objects are forwarded to another remote peer this has been patched in final as a workaround the user can do the validation by themselves by implementing a custom channelinboundhandler that is put in the channelpipeline behind publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all io netty netty codec http io netty netty codec isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree io netty netty codec http final isminimumfixversionavailable true minimumfixversion io netty netty all io netty netty codec http io netty netty codec basebranches vulnerabilityidentifier cve vulnerabilitydetails netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty io netty netty codec before version final there is a vulnerability that enables request smuggling if a content length header is present in the original http request the field is not validated by as it is propagated up this is fine as long as the request is not proxied through as http if the request comes in as an http stream gets converted into the http domain objects httprequest httpcontent etc via and then sent up to the child channel pipeline and proxied through a remote peer as http this may result in request smuggling in a proxy case users may assume the content length is validated somehow which is not the case if the request is forwarded to a backend channel that is a http connection the content length now has meaning and needs to be checked an attacker can smuggle requests inside the body as it gets downgraded from http to http for an example attack refer to the linked github advisory users are only affected if all of this is true or is used is used to convert to http objects and these http objects are forwarded to another remote peer this has been patched in final as a workaround the user can do the validation by themselves by implementing a custom channelinboundhandler that is put in the channelpipeline behind vulnerabilityurl | 0 |
5,252 | 26,581,338,965 | IssuesEvent | 2023-01-22 13:49:14 | albertlauncher/plugins | https://api.github.com/repos/albertlauncher/plugins | closed | Applications plugins cannot find pycharm | Maintainer wanted | If I would like to try to find `pycharm`, it doesn't show anything

desktop file looks like
```
[Desktop Entry]
Type=Application
Name=PyCharm Professional Edition
Icon=pycharm
Comment=Python IDE for Professional Developers.
Exec=pycharm %f
Terminal=false
Categories=Development;IDE;Python;
StartupNotify=true
StartupWMClass=jetbrains-pycharm
```
`Charm` (case sensitive) search doesn't work either. However it is possible to find by typing `pycharm` instead (or even `professional`). Fuzzy search option has no effect

| True | Applications plugins cannot find pycharm - If I would like to try to find `pycharm`, it doesn't show anything

desktop file looks like
```
[Desktop Entry]
Type=Application
Name=PyCharm Professional Edition
Icon=pycharm
Comment=Python IDE for Professional Developers.
Exec=pycharm %f
Terminal=false
Categories=Development;IDE;Python;
StartupNotify=true
StartupWMClass=jetbrains-pycharm
```
`Charm` (case sensitive) search doesn't work either. However it is possible to find by typing `pycharm` instead (or even `professional`). Fuzzy search option has no effect

| main | applications plugins cannot find pycharm if i would like to try to find pycharm it doesn t show anything desktop file looks like type application name pycharm professional edition icon pycharm comment python ide for professional developers exec pycharm f terminal false categories development ide python startupnotify true startupwmclass jetbrains pycharm charm case sensitive search doesn t work either however it is possible to find by typing pycharm instead or even professional fuzzy search option has no effect | 1 |
3,756 | 15,788,683,637 | IssuesEvent | 2021-04-01 21:12:24 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | Add a new property to component Dropdown to allow specifying the min-width of the` .bx--list-box__menu` class | status: needs triage 🕵️♀️ status: waiting for maintainer response 💬 type: enhancement 💡 |
### Summary
Add a new property to component Dropdown to allow specifying the min-width of the` .bx--list-box__menu` class
inside Dropdown so we can see full text (and not truncated text) of all values in the options list.
Clarify if you are asking for design, development, or both design and
development.
### Justification
Without this property, consumer of Dropdown has to use a specific className for their Dropdown and add a css to specify the min-width of the class ` .bx--list-box__menu` in the scope of the Dropdown class , that means having a specific css per Dropdown.
### Desired UX and success metrics
<!--alex disable failure-->
Describe the full user experience for this feature. Also define the metrics by
which we can measure success/failure for the user.
<!--alex enable failure-->
### "Must have" functionality
Highlight any "must have" needs and functionality for the request.
This should not be a full list of functionality; the Carbon team will work with
you to define functionality based on the desired UX.
### Specific timeline issues / requests
It's a UX requirement for Cognos Analytics release 11.2.1 (scheduled probably for early next year)
<!--alex disable period-->
Do you want this work within a specific time period? Is it related to an
upcoming release?
<!--alex enable period-->
_NB: The Carbon team will try to work with your timeline, but it's not
guaranteed. The earlier you make a request in advance of a desired delivery
date, the better!_
### Available extra resources
What resources do you have to assist this effort?
_Carbon is a collaborative system. We encourage teams to build components and
submit them for integration as either add-ons or core components._
| True | Add a new property to component Dropdown to allow specifying the min-width of the` .bx--list-box__menu` class -
### Summary
Add a new property to component Dropdown to allow specifying the min-width of the` .bx--list-box__menu` class
inside Dropdown so we can see full text (and not truncated text) of all values in the options list.
Clarify if you are asking for design, development, or both design and
development.
### Justification
Without this property, consumer of Dropdown has to use a specific className for their Dropdown and add a css to specify the min-width of the class ` .bx--list-box__menu` in the scope of the Dropdown class , that means having a specific css per Dropdown.
### Desired UX and success metrics
<!--alex disable failure-->
Describe the full user experience for this feature. Also define the metrics by
which we can measure success/failure for the user.
<!--alex enable failure-->
### "Must have" functionality
Highlight any "must have" needs and functionality for the request.
This should not be a full list of functionality; the Carbon team will work with
you to define functionality based on the desired UX.
### Specific timeline issues / requests
It's a UX requirement for Cognos Analytics release 11.2.1 (scheduled probably for early next year)
<!--alex disable period-->
Do you want this work within a specific time period? Is it related to an
upcoming release?
<!--alex enable period-->
_NB: The Carbon team will try to work with your timeline, but it's not
guaranteed. The earlier you make a request in advance of a desired delivery
date, the better!_
### Available extra resources
What resources do you have to assist this effort?
_Carbon is a collaborative system. We encourage teams to build components and
submit them for integration as either add-ons or core components._
| main | add a new property to component dropdown to allow specifying the min width of the bx list box menu class summary add a new property to component dropdown to allow specifying the min width of the bx list box menu class inside dropdown so we can see full text and not truncated text of all values in the options list clarify if you are asking for design development or both design and development justification without this property consumer of dropdown has to use a specific classname for their dropdown and add a css to specify the min width of the class bx list box menu in the scope of the dropdown class that means having a specific css per dropdown desired ux and success metrics describe the full user experience for this feature also define the metrics by which we can measure success failure for the user must have functionality highlight any must have needs and functionality for the request this should not be a full list of functionality the carbon team will work with you to define functionality based on the desired ux specific timeline issues requests it s a ux requirement for cognos analytics release scheduled probably for early next year do you want this work within a specific time period is it related to an upcoming release nb the carbon team will try to work with your timeline but it s not guaranteed the earlier you make a request in advance of a desired delivery date the better available extra resources what resources do you have to assist this effort carbon is a collaborative system we encourage teams to build components and submit them for integration as either add ons or core components | 1 |
1,385 | 6,011,672,681 | IssuesEvent | 2017-06-06 15:39:12 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | win_chocolatey: upgrade feature doesnt seem to work | affects_2.2 bug_report waiting_on_maintainer windows | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
win_chocolatey
##### ANSIBLE VERSION
```
$ ansible --version
ansible 2.2.0 (devel 58b9f637a1) last updated 2016/08/24 09:07:32 (GMT +100)
lib/ansible/modules/core: (detached HEAD 368ca738fa) last updated 2016/08/23 20:15:23 (GMT +100)
lib/ansible/modules/extras: (detached HEAD 0749ce6faa) last updated 2016/08/23 20:15:24 (GMT +100)
config file = /home/id/ansible/ansible.cfg
configured module search path = ['./library']
```
##### OS / ENVIRONMENT
win10 pro anniversary update Build 14393
##### SUMMARY
Upgrade feature calls `choco install` instead of `choco upgrade`
##### STEPS TO REPRODUCE
```
- name: Upgrade choco_pkgs
win_chocolatey:
name: "{{item}}"
upgrade: true # <----- no effect
state: present
allow_empty_checksums: true
with_items: "{{ choco_pkgs_auto }}"
ignore_errors: yes
```
##### EXPECTED RESULTS
```
choco.exe upgrade -dv -y --allow-empty-checksums "{{ item }}"
```
##### ACTUAL RESULTS
```
choco.exe install -dv -y --allow-empty-checksums "{{ item }}"
```
For example as seen here whenever the command errors out:
```
task path: /home/id/ansible/windows/roles/5_pkgs_upgrade/tasks/main.yml:72
changed: [brix-3205] => {"changed": true, "rc": 0, "stderr": "", "stdout": "upgrading\r\npkgs...\r\n", "stdout_lines": ["upgrading", "pkgs..."]}
____________________________________________
< TASK [5_pkgs_upgrade : Upgrade choco_pkgs] >
--------------------------------------------
\ ^__^
\ (oo)\_______
(__)\ )\/\
||----w |
|| ||
task path: /home/id/ansible/windows/roles/5_pkgs_upgrade/tasks/main.yml:76
ok: [brix-3205] => (item=windirstat) => {"changed": false, "item": "windirstat"}
ok: [brix-3205] => (item=realtemp) => {"changed": false, "item": "realtemp"}
ok: [brix-3205] => (item=f.lux) => {"changed": false, "item": "f.lux"}
ok: [brix-3205] => (item=procmon) => {"changed": false, "item": "procmon"}
ok: [brix-3205] => (item=crystaldiskinfo) => {"changed": false, "item": "crystaldiskinfo"}
ok: [brix-3205] => (item=crystaldiskmark) => {"changed": false, "item": "crystaldiskmark"}
failed: [brix-3205] (item=easybcd) => {"changed": false, "choco_error_cmd": "choco.exe install -dv -y easybcd --allow-empty-checksums", "choco_error_log": "Chocolatey v0.10.0 Chocolatey is running on Windows v 10.0.14393.0 Attempting to delete file \"C:/ProgramData/chocolatey/choco.exe.old\". Attempting to delete file \"C:\\ProgramData\\chocolatey\\choco.exe.old\". Command line: \"C:\\ProgramData\\chocolatey\\choco.exe\" install -dv -y easybcd --allow-empty-checksums Received arguments: install -dv -y easybcd --allow-empty-checksums RemovePendingPackagesTask is now ready and waiting for PreRunMessage. Sending message 'PreRunMessage' out if there are subscribers... [Pending] Removing all pending packages that should not be considered installed... The source 'https://chocolatey.org/api/v2/' evaluated to a 'normal' source type NOTE: Hiding sensitive configuration data! Please double and triple check to be sure no sensitive data is shown, especially if copying
```
| True | win_chocolatey: upgrade feature doesnt seem to work - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
win_chocolatey
##### ANSIBLE VERSION
```
$ ansible --version
ansible 2.2.0 (devel 58b9f637a1) last updated 2016/08/24 09:07:32 (GMT +100)
lib/ansible/modules/core: (detached HEAD 368ca738fa) last updated 2016/08/23 20:15:23 (GMT +100)
lib/ansible/modules/extras: (detached HEAD 0749ce6faa) last updated 2016/08/23 20:15:24 (GMT +100)
config file = /home/id/ansible/ansible.cfg
configured module search path = ['./library']
```
##### OS / ENVIRONMENT
win10 pro anniversary update Build 14393
##### SUMMARY
Upgrade feature calls `choco install` instead of `choco upgrade`
##### STEPS TO REPRODUCE
```
- name: Upgrade choco_pkgs
win_chocolatey:
name: "{{item}}"
upgrade: true # <----- no effect
state: present
allow_empty_checksums: true
with_items: "{{ choco_pkgs_auto }}"
ignore_errors: yes
```
##### EXPECTED RESULTS
```
choco.exe upgrade -dv -y --allow-empty-checksums "{{ item }}"
```
##### ACTUAL RESULTS
```
choco.exe install -dv -y --allow-empty-checksums "{{ item }}"
```
For example as seen here whenever the command errors out:
```
task path: /home/id/ansible/windows/roles/5_pkgs_upgrade/tasks/main.yml:72
changed: [brix-3205] => {"changed": true, "rc": 0, "stderr": "", "stdout": "upgrading\r\npkgs...\r\n", "stdout_lines": ["upgrading", "pkgs..."]}
____________________________________________
< TASK [5_pkgs_upgrade : Upgrade choco_pkgs] >
--------------------------------------------
\ ^__^
\ (oo)\_______
(__)\ )\/\
||----w |
|| ||
task path: /home/id/ansible/windows/roles/5_pkgs_upgrade/tasks/main.yml:76
ok: [brix-3205] => (item=windirstat) => {"changed": false, "item": "windirstat"}
ok: [brix-3205] => (item=realtemp) => {"changed": false, "item": "realtemp"}
ok: [brix-3205] => (item=f.lux) => {"changed": false, "item": "f.lux"}
ok: [brix-3205] => (item=procmon) => {"changed": false, "item": "procmon"}
ok: [brix-3205] => (item=crystaldiskinfo) => {"changed": false, "item": "crystaldiskinfo"}
ok: [brix-3205] => (item=crystaldiskmark) => {"changed": false, "item": "crystaldiskmark"}
failed: [brix-3205] (item=easybcd) => {"changed": false, "choco_error_cmd": "choco.exe install -dv -y easybcd --allow-empty-checksums", "choco_error_log": "Chocolatey v0.10.0 Chocolatey is running on Windows v 10.0.14393.0 Attempting to delete file \"C:/ProgramData/chocolatey/choco.exe.old\". Attempting to delete file \"C:\\ProgramData\\chocolatey\\choco.exe.old\". Command line: \"C:\\ProgramData\\chocolatey\\choco.exe\" install -dv -y easybcd --allow-empty-checksums Received arguments: install -dv -y easybcd --allow-empty-checksums RemovePendingPackagesTask is now ready and waiting for PreRunMessage. Sending message 'PreRunMessage' out if there are subscribers... [Pending] Removing all pending packages that should not be considered installed... The source 'https://chocolatey.org/api/v2/' evaluated to a 'normal' source type NOTE: Hiding sensitive configuration data! Please double and triple check to be sure no sensitive data is shown, especially if copying
```
| main | win chocolatey upgrade feature doesnt seem to work issue type bug report component name win chocolatey ansible version ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file home id ansible ansible cfg configured module search path os environment pro anniversary update build summary upgrade feature calls choco install instead of choco upgrade steps to reproduce name upgrade choco pkgs win chocolatey name item upgrade true no effect state present allow empty checksums true with items choco pkgs auto ignore errors yes expected results choco exe upgrade dv y allow empty checksums item actual results choco exe install dv y allow empty checksums item for example as seen here whenever the command errors out task path home id ansible windows roles pkgs upgrade tasks main yml changed changed true rc stderr stdout upgrading r npkgs r n stdout lines oo w task path home id ansible windows roles pkgs upgrade tasks main yml ok item windirstat changed false item windirstat ok item realtemp changed false item realtemp ok item f lux changed false item f lux ok item procmon changed false item procmon ok item crystaldiskinfo changed false item crystaldiskinfo ok item crystaldiskmark changed false item crystaldiskmark failed item easybcd changed false choco error cmd choco exe install dv y easybcd allow empty checksums choco error log chocolatey chocolatey is running on windows v attempting to delete file c programdata chocolatey choco exe old attempting to delete file c programdata chocolatey choco exe old command line c programdata chocolatey choco exe install dv y easybcd allow empty checksums received arguments install dv y easybcd allow empty checksums removependingpackagestask is now ready and waiting for prerunmessage sending message prerunmessage out if there are subscribers removing all pending packages that should not be considered installed the source evaluated to a normal source type note hiding sensitive configuration data please double and triple check to be sure no sensitive data is shown especially if copying | 1 |
131,333 | 12,481,158,896 | IssuesEvent | 2020-05-29 21:49:45 | syl20bnr/spacemacs | https://api.github.com/repos/syl20bnr/spacemacs | closed | Explain the basics of Org-mode for beginners | - Forum - Documentation ✏ Feature request Org stale | All documentation attached with Spacemacs is in org-files. Alas, knowing how to browse and read org files is necessary for a person to get a basic knowledge about Spacemacs.
But currently it's doesn't explain the very basics of org-mode. The only reason that I know how to expand the headlines and show their content is that I have read Org-modes' info files in an earlier adventure with Emacs.
I propose that the basics of browsing and reading org files should be added besides the Evil and Emacs tutorial in the "Quick Help" box on the start buffer. A relevant headline could be "How to read Spacemacs' documentation in Emacs".
| 1.0 | Explain the basics of Org-mode for beginners - All documentation attached with Spacemacs is in org-files. Alas, knowing how to browse and read org files is necessary for a person to get a basic knowledge about Spacemacs.
But currently it's doesn't explain the very basics of org-mode. The only reason that I know how to expand the headlines and show their content is that I have read Org-modes' info files in an earlier adventure with Emacs.
I propose that the basics of browsing and reading org files should be added besides the Evil and Emacs tutorial in the "Quick Help" box on the start buffer. A relevant headline could be "How to read Spacemacs' documentation in Emacs".
| non_main | explain the basics of org mode for beginners all documentation attached with spacemacs is in org files alas knowing how to browse and read org files is necessary for a person to get a basic knowledge about spacemacs but currently it s doesn t explain the very basics of org mode the only reason that i know how to expand the headlines and show their content is that i have read org modes info files in an earlier adventure with emacs i propose that the basics of browsing and reading org files should be added besides the evil and emacs tutorial in the quick help box on the start buffer a relevant headline could be how to read spacemacs documentation in emacs | 0 |
345,778 | 10,372,990,992 | IssuesEvent | 2019-09-09 05:45:59 | alexakasanjeev/magento_react_native | https://api.github.com/repos/alexakasanjeev/magento_react_native | opened | Product Detail Page UI Change List | Priority: Medium Status: On Hold Type: Enhancement | **Is your feature request related to a problem? Please describe.**
Product Detail page has lots of unimplemented feature, which are essential
1. [ ] `WebView` height displaying product description should be dynamic, currently it is set statically
2. [ ] Use `type_id === 'configurable'` logic to check it has options, everywhere in product detail page
3. [ ] No message is shown when user hit add to cart button, show success or error message accordingly
4. [ ] No check in `configurable` type product to disable certain options which are not available
> Example: suppose in size `s` color `red` is not available, so when user select size `s`, `red` option should be disabled.
5. [ ] No check written to check whether product is out of stock or not, if out of stock, disable `add-to-cart` button
6. [ ] Add input box to let user enter quantity of that product for cart currently defaults to 1, and not more then available quantity
**Describe the solution you'd like**
Implement above list, in Product Detail page
| 1.0 | Product Detail Page UI Change List - **Is your feature request related to a problem? Please describe.**
Product Detail page has lots of unimplemented feature, which are essential
1. [ ] `WebView` height displaying product description should be dynamic, currently it is set statically
2. [ ] Use `type_id === 'configurable'` logic to check it has options, everywhere in product detail page
3. [ ] No message is shown when user hit add to cart button, show success or error message accordingly
4. [ ] No check in `configurable` type product to disable certain options which are not available
> Example: suppose in size `s` color `red` is not available, so when user select size `s`, `red` option should be disabled.
5. [ ] No check written to check whether product is out of stock or not, if out of stock, disable `add-to-cart` button
6. [ ] Add input box to let user enter quantity of that product for cart currently defaults to 1, and not more then available quantity
**Describe the solution you'd like**
Implement above list, in Product Detail page
| non_main | product detail page ui change list is your feature request related to a problem please describe product detail page has lots of unimplemented feature which are essential webview height displaying product description should be dynamic currently it is set statically use type id configurable logic to check it has options everywhere in product detail page no message is shown when user hit add to cart button show success or error message accordingly no check in configurable type product to disable certain options which are not available example suppose in size s color red is not available so when user select size s red option should be disabled no check written to check whether product is out of stock or not if out of stock disable add to cart button add input box to let user enter quantity of that product for cart currently defaults to and not more then available quantity describe the solution you d like implement above list in product detail page | 0 |
48,632 | 12,225,260,065 | IssuesEvent | 2020-05-03 04:01:18 | Autodesk/arnold-usd | https://api.github.com/repos/Autodesk/arnold-usd | closed | Update testsuite scripts to match arnold | bug build | **Describe the bug**
The arnold-usd testsuite scripts have slightly derived from arnold core ones, which is causing issues when running the testsuite from arnold. The parameter `resave` should be called again `resaved`, and we should check whether it's a string or a boolean. When set to true, we assume the scene has to be resaved to .ass. This way, both test scripts will be similar again | 1.0 | Update testsuite scripts to match arnold - **Describe the bug**
The arnold-usd testsuite scripts have slightly derived from arnold core ones, which is causing issues when running the testsuite from arnold. The parameter `resave` should be called again `resaved`, and we should check whether it's a string or a boolean. When set to true, we assume the scene has to be resaved to .ass. This way, both test scripts will be similar again | non_main | update testsuite scripts to match arnold describe the bug the arnold usd testsuite scripts have slightly derived from arnold core ones which is causing issues when running the testsuite from arnold the parameter resave should be called again resaved and we should check whether it s a string or a boolean when set to true we assume the scene has to be resaved to ass this way both test scripts will be similar again | 0 |
18 | 2,515,212,688 | IssuesEvent | 2015-01-15 17:07:08 | simplesamlphp/simplesamlphp | https://api.github.com/repos/simplesamlphp/simplesamlphp | opened | Create an AuthMemCookie module | enhancement low maintainability | This should be on its own standalone repository and installable through composer. Since the functionality is already there, it's just about moving `www/authmemcookie.php` and `lib/SimpleSAML/AuthMemCookie.php` out and creating the additional module structure. | True | Create an AuthMemCookie module - This should be on its own standalone repository and installable through composer. Since the functionality is already there, it's just about moving `www/authmemcookie.php` and `lib/SimpleSAML/AuthMemCookie.php` out and creating the additional module structure. | main | create an authmemcookie module this should be on its own standalone repository and installable through composer since the functionality is already there it s just about moving www authmemcookie php and lib simplesaml authmemcookie php out and creating the additional module structure | 1 |
1,584 | 6,572,353,642 | IssuesEvent | 2017-09-11 01:39:15 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | [ecs_service_facts] value of details must be one of: true,false, got: True | affects_2.1 aws bug_report cloud waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`ecs_service_facts`
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file = /srv/code/ops/ansible/ansible.cfg
configured module search path = ['./library']
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
OSX
##### SUMMARY
Running this playbook (with `details: true`):
```
- hosts: localhost
connection: local
gather_facts: false
tasks:
- ecs_service_facts:
region: "us-west-1"
cluster: "my-cluster"
service: "my-service"
details: true
```
##### EXPECTED RESULTS
Detailed result as mentioned in [docs](http://docs.ansible.com/ansible/ecs_service_facts_module.html).
##### ACTUAL RESULTS
```
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"cluster": "mtdev-microservices-cluster", "details": "True", "region": "us-west-2", "service": "mtdev-threat-service", "validate_certs": true}, "module_name": "ecs_service_facts"}, "msg": "value of details must be one of: true,false, got: True"}
```
| True | [ecs_service_facts] value of details must be one of: true,false, got: True - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`ecs_service_facts`
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file = /srv/code/ops/ansible/ansible.cfg
configured module search path = ['./library']
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
OSX
##### SUMMARY
Running this playbook (with `details: true`):
```
- hosts: localhost
connection: local
gather_facts: false
tasks:
- ecs_service_facts:
region: "us-west-1"
cluster: "my-cluster"
service: "my-service"
details: true
```
##### EXPECTED RESULTS
Detailed result as mentioned in [docs](http://docs.ansible.com/ansible/ecs_service_facts_module.html).
##### ACTUAL RESULTS
```
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"cluster": "mtdev-microservices-cluster", "details": "True", "region": "us-west-2", "service": "mtdev-threat-service", "validate_certs": true}, "module_name": "ecs_service_facts"}, "msg": "value of details must be one of: true,false, got: True"}
```
| main | value of details must be one of true false got true issue type bug report component name ecs service facts ansible version ansible config file srv code ops ansible ansible cfg configured module search path configuration n a os environment osx summary running this playbook with details true hosts localhost connection local gather facts false tasks ecs service facts region us west cluster my cluster service my service details true expected results detailed result as mentioned in actual results fatal failed changed false failed true invocation module args cluster mtdev microservices cluster details true region us west service mtdev threat service validate certs true module name ecs service facts msg value of details must be one of true false got true | 1 |
88,777 | 17,663,608,706 | IssuesEvent | 2021-08-22 02:12:31 | adventuregamestudio/ags | https://api.github.com/repos/adventuregamestudio/ags | closed | INFO: a variant of utf8 support for translations | type: information context: unicode | A while ago @mgambrell has pointed me his work on hacking in utf-8 string support for translations. It's based on modifying alfont source to use parts of the **musl** library to convert strings.
Here's the commit (and maybe there are more commits around): https://github.com/RatalaikaGames/ags/commit/a00e7d933458d8360c17cf84c8cfbf5125111c3b
I am leaving this here only for the reference if someone would like to look into this at some point, because afaik the solution was not applied universally to whole engine, and it's hard to tell which nuances may be met when fully moving to utf8 support in both editor and engine (especially if you care about loading old games into new interpreter). | 1.0 | INFO: a variant of utf8 support for translations - A while ago @mgambrell has pointed me his work on hacking in utf-8 string support for translations. It's based on modifying alfont source to use parts of the **musl** library to convert strings.
Here's the commit (and maybe there are more commits around): https://github.com/RatalaikaGames/ags/commit/a00e7d933458d8360c17cf84c8cfbf5125111c3b
I am leaving this here only for the reference if someone would like to look into this at some point, because afaik the solution was not applied universally to whole engine, and it's hard to tell which nuances may be met when fully moving to utf8 support in both editor and engine (especially if you care about loading old games into new interpreter). | non_main | info a variant of support for translations a while ago mgambrell has pointed me his work on hacking in utf string support for translations it s based on modifying alfont source to use parts of the musl library to convert strings here s the commit and maybe there are more commits around i am leaving this here only for the reference if someone would like to look into this at some point because afaik the solution was not applied universally to whole engine and it s hard to tell which nuances may be met when fully moving to support in both editor and engine especially if you care about loading old games into new interpreter | 0 |
4,440 | 23,067,911,648 | IssuesEvent | 2022-07-25 15:20:25 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [Bug]: MeterChart generates error in the console | type: bug 🐛 status: needs triage 🕵️♀️ status: waiting for maintainer response 💬 | ### Package
carbon-components
### Browser
_No response_
### Package version
v11.0
### React version
_No response_
### Description
When using a meter chart on a page, the console produces the following error:
Error: Invalid value for <circle> attribute cy="calc(1em / 2)"
I have to remove the chart from my design, because we have quite a few of these charts on the page. It produces many errors.
### Reproduction/example
https://carbondesignsystem.com/data-visualization/simple-charts/#meter
### Steps to reproduce
Please open example page.
Inspect
Switch to the console... and you will see the error message.
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | True | [Bug]: MeterChart generates error in the console - ### Package
carbon-components
### Browser
_No response_
### Package version
v11.0
### React version
_No response_
### Description
When using a meter chart on a page, the console produces the following error:
Error: Invalid value for <circle> attribute cy="calc(1em / 2)"
I have to remove the chart from my design, because we have quite a few of these charts on the page. It produces many errors.
### Reproduction/example
https://carbondesignsystem.com/data-visualization/simple-charts/#meter
### Steps to reproduce
Please open example page.
Inspect
Switch to the console... and you will see the error message.
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | main | meterchart generates error in the console package carbon components browser no response package version react version no response description when using a meter chart on a page the console produces the following error error invalid value for attribute cy calc i have to remove the chart from my design because we have quite a few of these charts on the page it produces many errors reproduction example steps to reproduce please open example page inspect switch to the console and you will see the error message code of conduct i agree to follow this project s i checked the for duplicate problems | 1 |
460,889 | 13,219,917,163 | IssuesEvent | 2020-08-17 11:24:54 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | almudi.org - "Secure connection Failed error message is displayed | browser-focus-geckoview engine-gecko priority-normal severity-critical type-unsupported-tls | <!-- @browser: Firefox Mobile 71.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://almudi.org/calendario-liturgico/meditacion/56/domingo-1-de-adviento-ciclo-a?acm=29397_2119
**Browser / Version**: Firefox Mobile 71.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: conexión segura fallida
**Steps to Reproduce**:
No se ha podido verificar autenticidad del sitio
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | almudi.org - "Secure connection Failed error message is displayed - <!-- @browser: Firefox Mobile 71.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://almudi.org/calendario-liturgico/meditacion/56/domingo-1-de-adviento-ciclo-a?acm=29397_2119
**Browser / Version**: Firefox Mobile 71.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: conexión segura fallida
**Steps to Reproduce**:
No se ha podido verificar autenticidad del sitio
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | almudi org secure connection failed error message is displayed url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description conexión segura fallida steps to reproduce no se ha podido verificar autenticidad del sitio browser configuration none from with ❤️ | 0 |
3,641 | 14,730,765,981 | IssuesEvent | 2021-01-06 13:44:34 | AMYMEME/re-cycle-app | https://api.github.com/repos/AMYMEME/re-cycle-app | closed | firebase 통합환경 만들기 | maintain | # Database
찾아보니까 Firebase에 DB로 쓸 수 있는게 있는데, 원래 Google Storage를 포함해서 실시간 DB, firestore가 있음
## Google Storage
구글 스토리지는 90일 300$이용이나 for firebase용도 따로 있는 것 같은데,
java지원이 안되고, 우리가 갖고 있는 데이터에 적합하지 않아보임.
이미지같은 큰 바이너리 데이터를 저장할 때 가장 유용할 것 같음
## realtime DB
NoSQL 형태, Android, iOS, 자바스크립트 SDK로 연동할 수 있음
따라서 클라이언트 쪽에서는 잘 모르겠는데, 백엔드 쪽에서는 잘 모르겠음
## firestore
NoSQL 형태. 백엔드 쪽에도 적합해 보임
안드로이드(자바, 코틀린 모두 지원), iOS, 노드JS, 스프링, 파이썬, golang 모두 지원하고
우리가 저장할 데이터가 그렇게 큰 편이 아니어서 적합해보임 | True | firebase 통합환경 만들기 - # Database
찾아보니까 Firebase에 DB로 쓸 수 있는게 있는데, 원래 Google Storage를 포함해서 실시간 DB, firestore가 있음
## Google Storage
구글 스토리지는 90일 300$이용이나 for firebase용도 따로 있는 것 같은데,
java지원이 안되고, 우리가 갖고 있는 데이터에 적합하지 않아보임.
이미지같은 큰 바이너리 데이터를 저장할 때 가장 유용할 것 같음
## realtime DB
NoSQL 형태, Android, iOS, 자바스크립트 SDK로 연동할 수 있음
따라서 클라이언트 쪽에서는 잘 모르겠는데, 백엔드 쪽에서는 잘 모르겠음
## firestore
NoSQL 형태. 백엔드 쪽에도 적합해 보임
안드로이드(자바, 코틀린 모두 지원), iOS, 노드JS, 스프링, 파이썬, golang 모두 지원하고
우리가 저장할 데이터가 그렇게 큰 편이 아니어서 적합해보임 | main | firebase 통합환경 만들기 database 찾아보니까 firebase에 db로 쓸 수 있는게 있는데 원래 google storage를 포함해서 실시간 db firestore가 있음 google storage 구글 스토리지는 이용이나 for firebase용도 따로 있는 것 같은데 java지원이 안되고 우리가 갖고 있는 데이터에 적합하지 않아보임 이미지같은 큰 바이너리 데이터를 저장할 때 가장 유용할 것 같음 realtime db nosql 형태 android ios 자바스크립트 sdk로 연동할 수 있음 따라서 클라이언트 쪽에서는 잘 모르겠는데 백엔드 쪽에서는 잘 모르겠음 firestore nosql 형태 백엔드 쪽에도 적합해 보임 안드로이드 자바 코틀린 모두 지원 ios 노드js 스프링 파이썬 golang 모두 지원하고 우리가 저장할 데이터가 그렇게 큰 편이 아니어서 적합해보임 | 1 |
3,013 | 11,140,133,870 | IssuesEvent | 2019-12-21 11:49:39 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | terraform plan_file custom path broken | affects_2.8 bug cloud has_pr module needs_maintainer needs_triage python3 support:community | <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
https://github.com/ansible/ansible/blob/devel/lib/ansible/modules/cloud/misc/terraform.py#L353 expects the plan_file to be a filename not a path to a file.inside of project_path. Works with relative paths but not with absolute paths. Lucky for me I use a path inside project_path but it worked before with paths outside too.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
terraform
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible 2.8.1
config file = /home/adam/.ansible.cfg
configured module search path = ['/home/adam/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /home/adam/.local/lib/python3.6/site-packages/ansible
executable location = /home/adam/.local/bin/ansible
python version = 3.6.8 (default, Jan 14 2019, 11:02:34) [GCC 8.0.1 20180414 (experimental) [trunk revision 259383]]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
<!--- Paste example playbooks or commands between quotes below -->
```yaml
use absolute path for plan_file
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
uses the filepath provided
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes -->
```paste below
TASK [run terraform apply] ***************************************************************************************************************************************************************************************************************************************************************
fatal: [localhost]: FAILED! => {"changed": false, "msg": "Could not find plan_file \"/home/.../plan.tfplan\", check the path and try again."}
```
| True | terraform plan_file custom path broken - <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
https://github.com/ansible/ansible/blob/devel/lib/ansible/modules/cloud/misc/terraform.py#L353 expects the plan_file to be a filename not a path to a file.inside of project_path. Works with relative paths but not with absolute paths. Lucky for me I use a path inside project_path but it worked before with paths outside too.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
terraform
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible 2.8.1
config file = /home/adam/.ansible.cfg
configured module search path = ['/home/adam/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /home/adam/.local/lib/python3.6/site-packages/ansible
executable location = /home/adam/.local/bin/ansible
python version = 3.6.8 (default, Jan 14 2019, 11:02:34) [GCC 8.0.1 20180414 (experimental) [trunk revision 259383]]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
<!--- Paste example playbooks or commands between quotes below -->
```yaml
use absolute path for plan_file
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
uses the filepath provided
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes -->
```paste below
TASK [run terraform apply] ***************************************************************************************************************************************************************************************************************************************************************
fatal: [localhost]: FAILED! => {"changed": false, "msg": "Could not find plan_file \"/home/.../plan.tfplan\", check the path and try again."}
```
| main | terraform plan file custom path broken summary expects the plan file to be a filename not a path to a file inside of project path works with relative paths but not with absolute paths lucky for me i use a path inside project path but it worked before with paths outside too issue type bug report component name terraform ansible version paste below ansible config file home adam ansible cfg configured module search path ansible python module location home adam local lib site packages ansible executable location home adam local bin ansible python version default jan configuration paste below os environment steps to reproduce yaml use absolute path for plan file expected results uses the filepath provided actual results paste below task fatal failed changed false msg could not find plan file home plan tfplan check the path and try again | 1 |
5,524 | 27,616,562,954 | IssuesEvent | 2023-03-09 19:51:52 | microsoft/mu_plus | https://api.github.com/repos/microsoft/mu_plus | closed | SetCurrentTextString() in EditBox from SimpleUiToolkit not working bug | type:bug complexity:easy state:needs-maintainer-feedback | This function leaves the box empty ant text is never set. Just add this:
StrnCpyS(this->m_EditBoxDisplayText, sizeof(this->m_EditBoxDisplayText) / sizeof(CHAR16), NewTextString, (UIT_EDITBOX_MAX_STRING_LENGTH - 1));
after StrnCpyS (this->m_EditBoxText...
on the line 322 in function SetCurrentTextString() in MsGraphics\Library\SimpleUIToolKit\EditBox.c to make it work.
| True | SetCurrentTextString() in EditBox from SimpleUiToolkit not working bug - This function leaves the box empty ant text is never set. Just add this:
StrnCpyS(this->m_EditBoxDisplayText, sizeof(this->m_EditBoxDisplayText) / sizeof(CHAR16), NewTextString, (UIT_EDITBOX_MAX_STRING_LENGTH - 1));
after StrnCpyS (this->m_EditBoxText...
on the line 322 in function SetCurrentTextString() in MsGraphics\Library\SimpleUIToolKit\EditBox.c to make it work.
| main | setcurrenttextstring in editbox from simpleuitoolkit not working bug this function leaves the box empty ant text is never set just add this strncpys this m editboxdisplaytext sizeof this m editboxdisplaytext sizeof newtextstring uit editbox max string length after strncpys this m editboxtext on the line in function setcurrenttextstring in msgraphics library simpleuitoolkit editbox c to make it work | 1 |
5,623 | 28,133,658,972 | IssuesEvent | 2023-04-01 05:34:58 | beefproject/beef | https://api.github.com/repos/beefproject/beef | closed | Install Script: Remove --without test develop from ./install | Maintainability Install | https://github.com/beefproject/beef/blob/1ae320c3bc9db748dff7e53fe919cdef80089062/install#L240
is deprecated and should be removed
bundle not requires a local config file to store this information
| True | Install Script: Remove --without test develop from ./install - https://github.com/beefproject/beef/blob/1ae320c3bc9db748dff7e53fe919cdef80089062/install#L240
is deprecated and should be removed
bundle not requires a local config file to store this information
| main | install script remove without test develop from install is deprecated and should be removed bundle not requires a local config file to store this information | 1 |
204,864 | 15,953,303,083 | IssuesEvent | 2021-04-15 12:17:01 | srijan-sivakumar/redant | https://api.github.com/repos/srijan-sivakumar/redant | opened | Creating a readme for writing test cases | documentation enhancement | Create a readme that can guide anyone to integrate with the ops and write their own tests. | 1.0 | Creating a readme for writing test cases - Create a readme that can guide anyone to integrate with the ops and write their own tests. | non_main | creating a readme for writing test cases create a readme that can guide anyone to integrate with the ops and write their own tests | 0 |
1,786 | 6,575,879,768 | IssuesEvent | 2017-09-11 17:41:12 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Adding Namespace as an editable parameter for the docker_login module. | affects_2.1 cloud docker feature_idea waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
ansible-modules-core/cloud/docker/docker_login.py
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Currently, I'm not using the docker_login module, I am exporting the docker login details via terminal.
DOCKERCLOUD_USER=username
DOCKERCLOUD_PASS=password
DOCKERCLOUD_NAMESPACE=organization
-->
##### OS / ENVIRONMENT
<!---
Ubuntu 16.04.01 LTS
-->
##### SUMMARY
I am trying to log into my organization using the docker_login module. Docker Hub and Cloud now both have Organizations which improve your ability to control who can create, edit or delete Docker Hub repositories.
##### STEPS TO REPRODUCE
<!---
I would just specify DOCKERCLOUD_NAMESPACE in the config file that docker would have by default.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Log into DockerHub
docker_login:
username: docker
password: rekcod
email: docker@docker.io
namespace: docker_organization
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
I would expect that Docker now logs into my personal account but uses the organization that I am linked to and not my own personal account for pushing images or doing any tasks Docker related on the machine I used the docker_login module. This way I can effectively work with multiple teams that have their own repositories and I can effectively deploy Docker Images that are private from different teams.
##### ACTUAL RESULTS
This isn't currently possible.
| True | Adding Namespace as an editable parameter for the docker_login module. - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
ansible-modules-core/cloud/docker/docker_login.py
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Currently, I'm not using the docker_login module, I am exporting the docker login details via terminal.
DOCKERCLOUD_USER=username
DOCKERCLOUD_PASS=password
DOCKERCLOUD_NAMESPACE=organization
-->
##### OS / ENVIRONMENT
<!---
Ubuntu 16.04.01 LTS
-->
##### SUMMARY
I am trying to log into my organization using the docker_login module. Docker Hub and Cloud now both have Organizations which improve your ability to control who can create, edit or delete Docker Hub repositories.
##### STEPS TO REPRODUCE
<!---
I would just specify DOCKERCLOUD_NAMESPACE in the config file that docker would have by default.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Log into DockerHub
docker_login:
username: docker
password: rekcod
email: docker@docker.io
namespace: docker_organization
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
I would expect that Docker now logs into my personal account but uses the organization that I am linked to and not my own personal account for pushing images or doing any tasks Docker related on the machine I used the docker_login module. This way I can effectively work with multiple teams that have their own repositories and I can effectively deploy Docker Images that are private from different teams.
##### ACTUAL RESULTS
This isn't currently possible.
| main | adding namespace as an editable parameter for the docker login module issue type feature idea component name ansible modules core cloud docker docker login py ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration currently i m not using the docker login module i am exporting the docker login details via terminal dockercloud user username dockercloud pass password dockercloud namespace organization os environment ubuntu lts summary i am trying to log into my organization using the docker login module docker hub and cloud now both have organizations which improve your ability to control who can create edit or delete docker hub repositories steps to reproduce i would just specify dockercloud namespace in the config file that docker would have by default name log into dockerhub docker login username docker password rekcod email docker docker io namespace docker organization expected results i would expect that docker now logs into my personal account but uses the organization that i am linked to and not my own personal account for pushing images or doing any tasks docker related on the machine i used the docker login module this way i can effectively work with multiple teams that have their own repositories and i can effectively deploy docker images that are private from different teams actual results this isn t currently possible | 1 |
1,185 | 5,100,927,191 | IssuesEvent | 2017-01-04 14:01:28 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | opened | minimal install uses scipy | maintainability | Our minimal install now pulls scipy as a dependency. This comes from the griddataformats conda package (which depends on scipy for remapping). To solve this griddataformats would have to be installed from pip for the minimal build.
Or we just allow scipy as a dependency.
| True | minimal install uses scipy - Our minimal install now pulls scipy as a dependency. This comes from the griddataformats conda package (which depends on scipy for remapping). To solve this griddataformats would have to be installed from pip for the minimal build.
Or we just allow scipy as a dependency.
| main | minimal install uses scipy our minimal install now pulls scipy as a dependency this comes from the griddataformats conda package which depends on scipy for remapping to solve this griddataformats would have to be installed from pip for the minimal build or we just allow scipy as a dependency | 1 |
121,074 | 10,149,402,262 | IssuesEvent | 2019-08-05 15:08:53 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: tpchbench/tpchVec/nodes=3/cpu=4/sf=1 failed | C-test-failure O-roachtest O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/cfdaadc3514e7e8660f6c009ba159fdfd604f0a8
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=tpchbench/tpchVec/nodes=3/cpu=4/sf=1 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1409070&tab=buildLog
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190727-1409070/tpchbench/tpchVec/nodes=3/cpu=4/sf=1/run_1
test_runner.go:706: test timed out (10h0m0s)
tpchbench.go:119,cluster.go:2069,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1564208378-06-n4cpu4:4 -- ./workload run querybench --db=tpch --concurrency=1 --query-file=tpchVec --num-runs=3 --max-ops=27 --vectorized=true {pgurl:1-3} --histograms=perf/stats.json --histograms-max-latency=8m20s returned:
stderr:
stdout:
TH AND l_returnflag = 'R' AND c_nationkey = n_nationkey GROUP BY c_custkey, c_name, c_acctbal, c_phone, n_name, c_address, c_comment ORDER BY revenue DESC LIMIT 20
9h57m15s 0 0.0 0.0 0.0 0.0 0.0 0.0 8: SELECT ps_partkey, sum(ps_supplycost * ps_availqty::float) AS value FROM partsupp, supplier, nation WHERE ps_suppkey = s_suppkey AND s_nationkey = n_nationkey AND n_name = 'GERMANY' GROUP BY ps_partkey HAVING sum(ps_supplycost * ps_availqty::float) > ( SELECT sum(ps_supplycost * ps_availqty::float) * 0.0001 FROM partsupp, supplier, nation WHERE ps_suppkey = s_suppkey AND s_nationkey = n_nationkey AND n_name = 'GERMANY') ORDER BY value DESC
9h57m15s 0 0.0 0.0 0.0 0.0 0.0 0.0 9: SELECT sum(l_extendedprice) / 7.0 AS avg_yearly FROM lineitem, part WHERE p_partkey = l_partkey AND p_brand = 'Brand#23' AND p_container = 'MED BOX' AND l_quantity < ( SELECT 0.2 * avg(l_quantity) FROM lineitem WHERE l_partkey = p_partkey)
: signal: killed
cluster.go:2090,tpchbench.go:123,tpchbench.go:244,test_runner.go:691: Goexit() was called
``` | 2.0 | roachtest: tpchbench/tpchVec/nodes=3/cpu=4/sf=1 failed - SHA: https://github.com/cockroachdb/cockroach/commits/cfdaadc3514e7e8660f6c009ba159fdfd604f0a8
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=tpchbench/tpchVec/nodes=3/cpu=4/sf=1 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1409070&tab=buildLog
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190727-1409070/tpchbench/tpchVec/nodes=3/cpu=4/sf=1/run_1
test_runner.go:706: test timed out (10h0m0s)
tpchbench.go:119,cluster.go:2069,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1564208378-06-n4cpu4:4 -- ./workload run querybench --db=tpch --concurrency=1 --query-file=tpchVec --num-runs=3 --max-ops=27 --vectorized=true {pgurl:1-3} --histograms=perf/stats.json --histograms-max-latency=8m20s returned:
stderr:
stdout:
TH AND l_returnflag = 'R' AND c_nationkey = n_nationkey GROUP BY c_custkey, c_name, c_acctbal, c_phone, n_name, c_address, c_comment ORDER BY revenue DESC LIMIT 20
9h57m15s 0 0.0 0.0 0.0 0.0 0.0 0.0 8: SELECT ps_partkey, sum(ps_supplycost * ps_availqty::float) AS value FROM partsupp, supplier, nation WHERE ps_suppkey = s_suppkey AND s_nationkey = n_nationkey AND n_name = 'GERMANY' GROUP BY ps_partkey HAVING sum(ps_supplycost * ps_availqty::float) > ( SELECT sum(ps_supplycost * ps_availqty::float) * 0.0001 FROM partsupp, supplier, nation WHERE ps_suppkey = s_suppkey AND s_nationkey = n_nationkey AND n_name = 'GERMANY') ORDER BY value DESC
9h57m15s 0 0.0 0.0 0.0 0.0 0.0 0.0 9: SELECT sum(l_extendedprice) / 7.0 AS avg_yearly FROM lineitem, part WHERE p_partkey = l_partkey AND p_brand = 'Brand#23' AND p_container = 'MED BOX' AND l_quantity < ( SELECT 0.2 * avg(l_quantity) FROM lineitem WHERE l_partkey = p_partkey)
: signal: killed
cluster.go:2090,tpchbench.go:123,tpchbench.go:244,test_runner.go:691: Goexit() was called
``` | non_main | roachtest tpchbench tpchvec nodes cpu sf failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests tpchbench tpchvec nodes cpu sf pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts tpchbench tpchvec nodes cpu sf run test runner go test timed out tpchbench go cluster go errgroup go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run querybench db tpch concurrency query file tpchvec num runs max ops vectorized true pgurl histograms perf stats json histograms max latency returned stderr stdout th and l returnflag r and c nationkey n nationkey group by c custkey c name c acctbal c phone n name c address c comment order by revenue desc limit select ps partkey sum ps supplycost ps availqty float as value from partsupp supplier nation where ps suppkey s suppkey and s nationkey n nationkey and n name germany group by ps partkey having sum ps supplycost ps availqty float select sum ps supplycost ps availqty float from partsupp supplier nation where ps suppkey s suppkey and s nationkey n nationkey and n name germany order by value desc select sum l extendedprice as avg yearly from lineitem part where p partkey l partkey and p brand brand and p container med box and l quantity select avg l quantity from lineitem where l partkey p partkey signal killed cluster go tpchbench go tpchbench go test runner go goexit was called | 0 |
51,390 | 13,635,116,181 | IssuesEvent | 2020-09-25 01:56:37 | nasifimtiazohi/openmrs-module-reporting-1.20.0 | https://api.github.com/repos/nasifimtiazohi/openmrs-module-reporting-1.20.0 | opened | CVE-2018-19362 (High) detected in jackson-databind-2.9.0.jar | security vulnerability | ## CVE-2018-19362 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: openmrs-module-reporting-1.20.0/api-2.2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- openmrs-api-2.2.0.jar (Root Library)
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-module-reporting-1.20.0/commit/43757d56a9ab9f7202e297fea95f1861af41888c">43757d56a9ab9f7202e297fea95f1861af41888c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19362>CVE-2018-19362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-19362 (High) detected in jackson-databind-2.9.0.jar - ## CVE-2018-19362 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: openmrs-module-reporting-1.20.0/api-2.2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- openmrs-api-2.2.0.jar (Root Library)
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-module-reporting-1.20.0/commit/43757d56a9ab9f7202e297fea95f1861af41888c">43757d56a9ab9f7202e297fea95f1861af41888c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19362>CVE-2018-19362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file openmrs module reporting api pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy openmrs api jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before might allow attackers to have unspecified impact by leveraging failure to block the jboss common core class from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
4,275 | 21,476,657,365 | IssuesEvent | 2022-04-26 14:11:16 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | opened | Long-term solution for import sorting | maintainability proposal | ## Is your feature request related to a problem? ##
As [discussed](https://github.com/MDAnalysis/mdanalysis/pull/3644#issuecomment-1100836585) in #3644 I am always frustrated when I see unsorted imports 🙃. Especially, for a new import, it is not clear where to put it.
## Describe the solution you'd like ##
Run `isort` with options (to be discussed) for the complete repo
```toml
line_length = 80
indent = 4
multi_line_output = 8 # Vertical Hanging Indent Bracket
include_trailing_comma = true
lines_after_imports = 2
known_first_party = "MDAnalysis"
```
and add a command like `isort --verbose --check-only --diff` to the CI.
## Describe alternatives you've considered ##
Leave everything as it is.
| True | Long-term solution for import sorting - ## Is your feature request related to a problem? ##
As [discussed](https://github.com/MDAnalysis/mdanalysis/pull/3644#issuecomment-1100836585) in #3644 I am always frustrated when I see unsorted imports 🙃. Especially, for a new import, it is not clear where to put it.
## Describe the solution you'd like ##
Run `isort` with options (to be discussed) for the complete repo
```toml
line_length = 80
indent = 4
multi_line_output = 8 # Vertical Hanging Indent Bracket
include_trailing_comma = true
lines_after_imports = 2
known_first_party = "MDAnalysis"
```
and add a command like `isort --verbose --check-only --diff` to the CI.
## Describe alternatives you've considered ##
Leave everything as it is.
| main | long term solution for import sorting is your feature request related to a problem as in i am always frustrated when i see unsorted imports 🙃 especially for a new import it is not clear where to put it describe the solution you d like run isort with options to be discussed for the complete repo toml line length indent multi line output vertical hanging indent bracket include trailing comma true lines after imports known first party mdanalysis and add a command like isort verbose check only diff to the ci describe alternatives you ve considered leave everything as it is | 1 |
159,690 | 20,085,892,819 | IssuesEvent | 2022-02-05 01:08:01 | AkshayMukkavilli/Tensorflow | https://api.github.com/repos/AkshayMukkavilli/Tensorflow | opened | CVE-2021-41207 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl | security vulnerability | ## CVE-2021-41207 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /Tensorflow/src/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. In affected versions the implementation of `ParallelConcat` misses some input validation and can produce a division by 0. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-11-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41207>CVE-2021-41207</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-7v94-64hj-m82h">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-7v94-64hj-m82h</a></p>
<p>Release Date: 2021-11-05</p>
<p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-41207 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-41207 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /Tensorflow/src/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. In affected versions the implementation of `ParallelConcat` misses some input validation and can produce a division by 0. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-11-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41207>CVE-2021-41207</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-7v94-64hj-m82h">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-7v94-64hj-m82h</a></p>
<p>Release Date: 2021-11-05</p>
<p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file tensorflow src requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source platform for machine learning in affected versions the implementation of parallelconcat misses some input validation and can produce a division by the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource | 0 |
306,675 | 23,168,976,601 | IssuesEvent | 2022-07-30 12:02:13 | swarm-game/swarm | https://api.github.com/repos/swarm-game/swarm | closed | Documentation for how to author scenarios | Z-Feature C-Low Hanging Fruit S-Critical Z-Documentation | This is potentially related to #346 and #347, but is fundamentally separate. We need some documentation somewhere (e.g. on the [wiki](https://github.com/swarm-game/swarm/wiki)?) explaining the format of the `.yaml` files used to describe scenarios and what all the different options mean. | 1.0 | Documentation for how to author scenarios - This is potentially related to #346 and #347, but is fundamentally separate. We need some documentation somewhere (e.g. on the [wiki](https://github.com/swarm-game/swarm/wiki)?) explaining the format of the `.yaml` files used to describe scenarios and what all the different options mean. | non_main | documentation for how to author scenarios this is potentially related to and but is fundamentally separate we need some documentation somewhere e g on the explaining the format of the yaml files used to describe scenarios and what all the different options mean | 0 |
808 | 4,425,771,330 | IssuesEvent | 2016-08-16 16:20:22 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | AttributeError: 'DockerManager' object has no attribute 'client' | bug_report cloud docker waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
docker module
##### ANSIBLE VERSION
```
ansible 2.1.1.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### STEPS TO REPRODUCE
```
- name: run the site in a docker container
docker:
name: app_test
env_file: /opt/app/env.conf
publish_all_ports: yes
cap_add:
- "SYS_PTRACE"
tty: yes
detach: yes
volumes: "/usr/share/GeoIP/GeoLiteCity.dat:/usr/share/GeoIP/GeoLiteCity.dat"
image: "app_test:{{ RUBY_SEMVER }}"
state: started
when: RUBY_SEMVER is defined
```
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 1975, in <module>
main()
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 1912, in main
manager = DockerManager(module)
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 749, in __init__
self.environment = self.get_environment(env, env_file)
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 895, in get_environment
self.ensure_capability('env_file')
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 870, in ensure_capability
self._check_capabilities()
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 853, in _check_capabilities
api_version = self.client.version()['ApiVersion']
AttributeError: 'DockerManager' object has no attribute 'client'
fatal: [srv-1]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "docker"}, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 1975, in <module>\n main()\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 1912, in main\n manager = DockerManager(module)\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 749, in __init__\n self.environment = self.get_environment(env, env_file)\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 895, in get_environment\n self.ensure_capability('env_file')\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 870, in ensure_capability\n self._check_capabilities()\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 853, in _check_capabilities\n api_version = self.client.version()['ApiVersion']\nAttributeError: 'DockerManager' object has no attribute 'client'\n", "module_stdout": "", "msg": "MODULE FAILURE", "parsed": false}
```
| True | AttributeError: 'DockerManager' object has no attribute 'client' - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
docker module
##### ANSIBLE VERSION
```
ansible 2.1.1.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### STEPS TO REPRODUCE
```
- name: run the site in a docker container
docker:
name: app_test
env_file: /opt/app/env.conf
publish_all_ports: yes
cap_add:
- "SYS_PTRACE"
tty: yes
detach: yes
volumes: "/usr/share/GeoIP/GeoLiteCity.dat:/usr/share/GeoIP/GeoLiteCity.dat"
image: "app_test:{{ RUBY_SEMVER }}"
state: started
when: RUBY_SEMVER is defined
```
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 1975, in <module>
main()
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 1912, in main
manager = DockerManager(module)
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 749, in __init__
self.environment = self.get_environment(env, env_file)
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 895, in get_environment
self.ensure_capability('env_file')
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 870, in ensure_capability
self._check_capabilities()
File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 853, in _check_capabilities
api_version = self.client.version()['ApiVersion']
AttributeError: 'DockerManager' object has no attribute 'client'
fatal: [srv-1]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "docker"}, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 1975, in <module>\n main()\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 1912, in main\n manager = DockerManager(module)\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 749, in __init__\n self.environment = self.get_environment(env, env_file)\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 895, in get_environment\n self.ensure_capability('env_file')\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 870, in ensure_capability\n self._check_capabilities()\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 853, in _check_capabilities\n api_version = self.client.version()['ApiVersion']\nAttributeError: 'DockerManager' object has no attribute 'client'\n", "module_stdout": "", "msg": "MODULE FAILURE", "parsed": false}
```
| main | attributeerror dockermanager object has no attribute client issue type bug report component name docker module ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables steps to reproduce name run the site in a docker container docker name app test env file opt app env conf publish all ports yes cap add sys ptrace tty yes detach yes volumes usr share geoip geolitecity dat usr share geoip geolitecity dat image app test ruby semver state started when ruby semver is defined actual results an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible cbwbjg ansible module docker py line in main file tmp ansible cbwbjg ansible module docker py line in main manager dockermanager module file tmp ansible cbwbjg ansible module docker py line in init self environment self get environment env env file file tmp ansible cbwbjg ansible module docker py line in get environment self ensure capability env file file tmp ansible cbwbjg ansible module docker py line in ensure capability self check capabilities file tmp ansible cbwbjg ansible module docker py line in check capabilities api version self client version attributeerror dockermanager object has no attribute client fatal failed changed false failed true invocation module name docker module stderr traceback most recent call last n file tmp ansible cbwbjg ansible module docker py line in n main n file tmp ansible cbwbjg ansible module docker py line in main n manager dockermanager module n file tmp ansible cbwbjg ansible module docker py line in init n self environment self get environment env env file n file tmp ansible cbwbjg ansible module docker py line in get environment n self ensure capability env file n file tmp ansible cbwbjg ansible module docker py line in ensure capability n self check capabilities n file tmp ansible cbwbjg ansible module docker py line in check capabilities n api version self client version nattributeerror dockermanager object has no attribute client n module stdout msg module failure parsed false | 1 |
4,711 | 24,270,835,605 | IssuesEvent | 2022-09-28 10:07:43 | mozilla/foundation.mozilla.org | https://api.github.com/repos/mozilla/foundation.mozilla.org | closed | SEO | incorrect page found in sitemap.xml | engineering Maintain | The Same [page identified](https://foundation.mozilla.org/en/blog/why-our-tech-community-is-most-excited-about-mozfest-2022/) with a 5XX code is listed here as well. However, using multiple sitemap validator tools, they were unable to validate the XML sitemap for the domain foundation.mozilla.org - additional attention should be paid to the output for the SITEMAP to ensure it follows the standard with all necessary XML structures. | True | SEO | incorrect page found in sitemap.xml - The Same [page identified](https://foundation.mozilla.org/en/blog/why-our-tech-community-is-most-excited-about-mozfest-2022/) with a 5XX code is listed here as well. However, using multiple sitemap validator tools, they were unable to validate the XML sitemap for the domain foundation.mozilla.org - additional attention should be paid to the output for the SITEMAP to ensure it follows the standard with all necessary XML structures. | main | seo incorrect page found in sitemap xml the same with a code is listed here as well however using multiple sitemap validator tools they were unable to validate the xml sitemap for the domain foundation mozilla org additional attention should be paid to the output for the sitemap to ensure it follows the standard with all necessary xml structures | 1 |
4,747 | 24,489,629,195 | IssuesEvent | 2022-10-09 22:15:22 | rustsec/advisory-db | https://api.github.com/repos/rustsec/advisory-db | closed | `num-format` Status | Unmaintained | _3,232,716 downloads, ~8k a day_
Last release was over three years ago
It is using the old version of itoa: https://github.com/rustsec/advisory-db/issues/1404
Ralf was helpful to ping earlier: https://github.com/bcmyers/num-format/issues/29
Maintenance status was asked on 9 Jan 2022
https://github.com/bcmyers/num-format/issues/27
@bcmyers - I wonder if people should be still using this crate today and whether it would be helpful to bump the itoa dependency | True | `num-format` Status - _3,232,716 downloads, ~8k a day_
Last release was over three years ago
It is using the old version of itoa: https://github.com/rustsec/advisory-db/issues/1404
Ralf was helpful to ping earlier: https://github.com/bcmyers/num-format/issues/29
Maintenance status was asked on 9 Jan 2022
https://github.com/bcmyers/num-format/issues/27
@bcmyers - I wonder if people should be still using this crate today and whether it would be helpful to bump the itoa dependency | main | num format status downloads a day last release was over three years ago it is using the old version of itoa ralf was helpful to ping earlier maintenance status was asked on jan bcmyers i wonder if people should be still using this crate today and whether it would be helpful to bump the itoa dependency | 1 |
1,052 | 4,863,765,703 | IssuesEvent | 2016-11-14 16:14:18 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | template src does not work for roles | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
Module template
##### ANSIBLE VERSION
```
ansible 2.1.1.0
```
##### CONFIGURATION
```
[defaults]
inventory = ./hosts.ini
library = ./library
forks = 50
gathering = smart
roles_path = ./roles
vault_password_file = /xxx/ansible_vault_password.txt
fact_caching = jsonfile
fact_caching_connection = /var/cache/ansible-facts
fact_caching_timeout = 86400
var_compression_level = 5
module_compression = 'ZIP_DEFLATED'
[privilege_escalation]
[paramiko_connection]
[ssh_connection]
pipelining = True
[accelerate]
[selinux]
[colors]
```
##### OS / ENVIRONMENT
Ubuntu 14.04.5 LTS
##### SUMMARY
The module `template` does not search in the `files` directory of a role as it is done by the `copy` module.
##### STEPS TO REPRODUCE
I have a file layout as the best practice documentation recommends:
site.yml
roles/facts/vars/main.yml
roles/facts/tasks/main.yml
roles/facts/tasks/facts_file.yml
roles/facts/files/bash.j2
My `site.yml` includes the role `facts` and `tasks/main.yml` includes `facts_file.yml`. In `facts_file.yml` I use the template module to transfer files to a remote system with the following `src` attribute in a loop:
src: "{{item.shell}}.j2"
which expands to
src: bash.j2
But when I run my `site.yml` I get the error:
> IOError: [Errno 2] No such file or directory: u'/home/ziemann/ansible/bash.j2'
The file is searched in Ansible's base directory but not in the `files` directory of the role.
##### EXPECTED RESULTS
I expect the template module to act on the `src` attribute in the same way as the `copy` module.
##### ACTUAL RESULTS
The template module and copy module work in different ways.
##### BTW
It seems to me that it is a design error, that there are two different copy modules: one with template expansion and one without. It might be better to merge them together.
##### Workaround?
How can I build the path to the file based on the role by myself? | True | template src does not work for roles - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
Module template
##### ANSIBLE VERSION
```
ansible 2.1.1.0
```
##### CONFIGURATION
```
[defaults]
inventory = ./hosts.ini
library = ./library
forks = 50
gathering = smart
roles_path = ./roles
vault_password_file = /xxx/ansible_vault_password.txt
fact_caching = jsonfile
fact_caching_connection = /var/cache/ansible-facts
fact_caching_timeout = 86400
var_compression_level = 5
module_compression = 'ZIP_DEFLATED'
[privilege_escalation]
[paramiko_connection]
[ssh_connection]
pipelining = True
[accelerate]
[selinux]
[colors]
```
##### OS / ENVIRONMENT
Ubuntu 14.04.5 LTS
##### SUMMARY
The module `template` does not search in the `files` directory of a role as it is done by the `copy` module.
##### STEPS TO REPRODUCE
I have a file layout as the best practice documentation recommends:
site.yml
roles/facts/vars/main.yml
roles/facts/tasks/main.yml
roles/facts/tasks/facts_file.yml
roles/facts/files/bash.j2
My `site.yml` includes the role `facts` and `tasks/main.yml` includes `facts_file.yml`. In `facts_file.yml` I use the template module to transfer files to a remote system with the following `src` attribute in a loop:
src: "{{item.shell}}.j2"
which expands to
src: bash.j2
But when I run my `site.yml` I get the error:
> IOError: [Errno 2] No such file or directory: u'/home/ziemann/ansible/bash.j2'
The file is searched in Ansible's base directory but not in the `files` directory of the role.
##### EXPECTED RESULTS
I expect the template module to act on the `src` attribute in the same way as the `copy` module.
##### ACTUAL RESULTS
The template module and copy module work in different ways.
##### BTW
It seems to me that it is a design error, that there are two different copy modules: one with template expansion and one without. It might be better to merge them together.
##### Workaround?
How can I build the path to the file based on the role by myself? | main | template src does not work for roles issue type bug report component name module template ansible version ansible configuration inventory hosts ini library library forks gathering smart roles path roles vault password file xxx ansible vault password txt fact caching jsonfile fact caching connection var cache ansible facts fact caching timeout var compression level module compression zip deflated pipelining true os environment ubuntu lts summary the module template does not search in the files directory of a role as it is done by the copy module steps to reproduce i have a file layout as the best practice documentation recommends site yml roles facts vars main yml roles facts tasks main yml roles facts tasks facts file yml roles facts files bash my site yml includes the role facts and tasks main yml includes facts file yml in facts file yml i use the template module to transfer files to a remote system with the following src attribute in a loop src item shell which expands to src bash but when i run my site yml i get the error ioerror no such file or directory u home ziemann ansible bash the file is searched in ansible s base directory but not in the files directory of the role expected results i expect the template module to act on the src attribute in the same way as the copy module actual results the template module and copy module work in different ways btw it seems to me that it is a design error that there are two different copy modules one with template expansion and one without it might be better to merge them together workaround how can i build the path to the file based on the role by myself | 1 |
3,363 | 13,033,183,014 | IssuesEvent | 2020-07-28 06:20:00 | OpenRefine/OpenRefine | https://api.github.com/repos/OpenRefine/OpenRefine | closed | Update Vicino for N-gram clusterer bug fix | bug maintainability | We've got a bug fix for the SIMILE Vicino N-gram clusterer sitting at https://github.com/OpenRefine/simile-vicino/commit/e9e9eda18bf905f5a0ee6c04cc6a1b48d621b8c0 which never got published. We should publish a new version with the bug fix and update OpenRefine to use it.
We could use this opportunity to clean up the OpenRefine dependencies a little by:
- [x] switching to the official `secondstring` dependency from the original author instead of publishing our own (resolved in https://github.com/OpenRefine/simile-vicino/pull/1/
- [x] switching vicino bzip2 dependency to use Apache commons-compress instead of ant-tools, as we've done for OpenRefine (resolved in https://github.com/OpenRefine/simile-vicino/pull/1/)
- [x] updating to arithcode-1.2. This is a very minor release, but it includes tests, which is a plus. (pending in https://github.com/OpenRefine/simile-vicino/pull/2/)
- [x] moving the `secondstring` and `arithcode` dependencies from OpenRefine to https://github.com/OpenRefine/simile-vicino
| True | Update Vicino for N-gram clusterer bug fix - We've got a bug fix for the SIMILE Vicino N-gram clusterer sitting at https://github.com/OpenRefine/simile-vicino/commit/e9e9eda18bf905f5a0ee6c04cc6a1b48d621b8c0 which never got published. We should publish a new version with the bug fix and update OpenRefine to use it.
We could use this opportunity to clean up the OpenRefine dependencies a little by:
- [x] switching to the official `secondstring` dependency from the original author instead of publishing our own (resolved in https://github.com/OpenRefine/simile-vicino/pull/1/
- [x] switching vicino bzip2 dependency to use Apache commons-compress instead of ant-tools, as we've done for OpenRefine (resolved in https://github.com/OpenRefine/simile-vicino/pull/1/)
- [x] updating to arithcode-1.2. This is a very minor release, but it includes tests, which is a plus. (pending in https://github.com/OpenRefine/simile-vicino/pull/2/)
- [x] moving the `secondstring` and `arithcode` dependencies from OpenRefine to https://github.com/OpenRefine/simile-vicino
| main | update vicino for n gram clusterer bug fix we ve got a bug fix for the simile vicino n gram clusterer sitting at which never got published we should publish a new version with the bug fix and update openrefine to use it we could use this opportunity to clean up the openrefine dependencies a little by switching to the official secondstring dependency from the original author instead of publishing our own resolved in switching vicino dependency to use apache commons compress instead of ant tools as we ve done for openrefine resolved in updating to arithcode this is a very minor release but it includes tests which is a plus pending in moving the secondstring and arithcode dependencies from openrefine to | 1 |
272 | 3,040,294,501 | IssuesEvent | 2015-08-07 14:41:09 | simplesamlphp/simplesamlphp | https://api.github.com/repos/simplesamlphp/simplesamlphp | closed | Extract the metaedit module out of the repository | enhancement low maintainability | It should get its own repository and allow installation through composer. | True | Extract the metaedit module out of the repository - It should get its own repository and allow installation through composer. | main | extract the metaedit module out of the repository it should get its own repository and allow installation through composer | 1 |
3,080 | 2,536,590,014 | IssuesEvent | 2015-01-26 15:09:28 | JMurk/Snowblower_Issues | https://api.github.com/repos/JMurk/Snowblower_Issues | opened | Snow - Color Blind Accessibility | enhancement low priority | **Future Enhancement**
Per Amy, there is a color blind user. He will not be able to distinguish anything on the map as-is now. We should develop something that allows him to be able to use the map effectively. | 1.0 | Snow - Color Blind Accessibility - **Future Enhancement**
Per Amy, there is a color blind user. He will not be able to distinguish anything on the map as-is now. We should develop something that allows him to be able to use the map effectively. | non_main | snow color blind accessibility future enhancement per amy there is a color blind user he will not be able to distinguish anything on the map as is now we should develop something that allows him to be able to use the map effectively | 0 |
417,349 | 28,110,369,032 | IssuesEvent | 2023-03-31 06:36:33 | euph00/ped | https://api.github.com/repos/euph00/ped | opened | command summary table in UG is out of date | type.DocumentationBug severity.Low | delete command suspected to refer to delete_patient
find command suspected to refer to find_patient
find_details command suspected to refer to find_patient details
etc


<!--session: 1680241960933-9b72ee5c-285f-4f6d-9a4f-a1ade050869a-->
<!--Version: Web v3.4.7--> | 1.0 | command summary table in UG is out of date - delete command suspected to refer to delete_patient
find command suspected to refer to find_patient
find_details command suspected to refer to find_patient details
etc


<!--session: 1680241960933-9b72ee5c-285f-4f6d-9a4f-a1ade050869a-->
<!--Version: Web v3.4.7--> | non_main | command summary table in ug is out of date delete command suspected to refer to delete patient find command suspected to refer to find patient find details command suspected to refer to find patient details etc | 0 |
646,847 | 21,076,654,515 | IssuesEvent | 2022-04-02 08:34:32 | SE701-T1/backend | https://api.github.com/repos/SE701-T1/backend | closed | Fix CI action workflow | Priority: Low Status: Review Needed Type: Bug | **Describe the task that needs to be done.**
<!-- *(If this issue is about a bug, please describe the problem and steps to reproduce the issue. You can also include screenshots of any stack traces, or any other supporting images).* -->
The CI workflow for GitHub actions is not being triggered for execution in all of the desired cases. Considering the jobs in the CI workflow, it could be better to have the CI triggered for execution by all push and pull request actions regardless of branch and to have a manually triggered workflow dispatch in case there is an error that is not related to the source code.
**Describe how a solution to your proposed task might look like (and any alternatives considered).**
A proposed solution is to remove the specified branch names from the pull_request event trigger, and to add push and workflow_dispatch event triggers for any branch:
```
on:
push:
pull_request:
workflow_dispatch:
```
**Notes**
| 1.0 | Fix CI action workflow - **Describe the task that needs to be done.**
<!-- *(If this issue is about a bug, please describe the problem and steps to reproduce the issue. You can also include screenshots of any stack traces, or any other supporting images).* -->
The CI workflow for GitHub actions is not being triggered for execution in all of the desired cases. Considering the jobs in the CI workflow, it could be better to have the CI triggered for execution by all push and pull request actions regardless of branch and to have a manually triggered workflow dispatch in case there is an error that is not related to the source code.
**Describe how a solution to your proposed task might look like (and any alternatives considered).**
A proposed solution is to remove the specified branch names from the pull_request event trigger, and to add push and workflow_dispatch event triggers for any branch:
```
on:
push:
pull_request:
workflow_dispatch:
```
**Notes**
| non_main | fix ci action workflow describe the task that needs to be done the ci workflow for github actions is not being triggered for execution in all of the desired cases considering the jobs in the ci workflow it could be better to have the ci triggered for execution by all push and pull request actions regardless of branch and to have a manually triggered workflow dispatch in case there is an error that is not related to the source code describe how a solution to your proposed task might look like and any alternatives considered a proposed solution is to remove the specified branch names from the pull request event trigger and to add push and workflow dispatch event triggers for any branch on push pull request workflow dispatch notes | 0 |
176,171 | 28,039,053,342 | IssuesEvent | 2023-03-28 17:03:48 | pulumi/pulumi-docker | https://api.github.com/repos/pulumi/pulumi-docker | closed | docker:index:Image could not open dockerfile at relative path Dockerfile | kind/bug resolution/by-design | ### What happened?
Upgraded from `pulumi-docker` `4.0.0` to `4.1.0`.
`pulumi up` unexpectedly errors:
```
Diagnostics:
docker:index:Image (my-app:latest):
error: could not open dockerfile at relative path Dockerfile. Try setting `dockerfile` to "/Users/shed/Repos/my-repo/projects/my-app/Dockerfile"
```
### Expected Behavior
`pulumi up` to locate `Dockerfile` relative to my project directory.
### Steps to reproduce
My pulumi resource looks something like:
```python
image = docker.Image(
"my_app:latest",
build=docker.DockerBuildArgs(
context="/Users/shed/Repos/my-repo/projects/my-app/",
dockerfile="Dockerfile",
args={"BUILDKIT_INLINE_CACHE": "1"}, # may not be necessary?
cache_from=docker.CacheFromArgs(
images=[
pulumi.Output.concat(repo.repository_url, ":", "latest")
]
),
platform="linux/amd64",
),
registry=registry,
image_name=pulumi.Output.concat(self.repo.repository_url, ":", "latest"),
)
```
There is a file `/Users/shed/Repos/my-repo/projects/my-app/Dockerfile`.
`context` is set to the directory and `dockerfile` is explicitly set to `Dockerfile`.
### Output of `pulumi about`
I don't want to copy/paste this from a work project. I'll try to provide specific info as requested if relevant.
### Additional context
Split from #566
### Contributing
Vote on this issue by adding a 👍 reaction.
To contribute a fix for this issue, leave a comment (and link to your pull request, if you've opened one already).
| 1.0 | docker:index:Image could not open dockerfile at relative path Dockerfile - ### What happened?
Upgraded from `pulumi-docker` `4.0.0` to `4.1.0`.
`pulumi up` unexpectedly errors:
```
Diagnostics:
docker:index:Image (my-app:latest):
error: could not open dockerfile at relative path Dockerfile. Try setting `dockerfile` to "/Users/shed/Repos/my-repo/projects/my-app/Dockerfile"
```
### Expected Behavior
`pulumi up` to locate `Dockerfile` relative to my project directory.
### Steps to reproduce
My pulumi resource looks something like:
```python
image = docker.Image(
"my_app:latest",
build=docker.DockerBuildArgs(
context="/Users/shed/Repos/my-repo/projects/my-app/",
dockerfile="Dockerfile",
args={"BUILDKIT_INLINE_CACHE": "1"}, # may not be necessary?
cache_from=docker.CacheFromArgs(
images=[
pulumi.Output.concat(repo.repository_url, ":", "latest")
]
),
platform="linux/amd64",
),
registry=registry,
image_name=pulumi.Output.concat(self.repo.repository_url, ":", "latest"),
)
```
There is a file `/Users/shed/Repos/my-repo/projects/my-app/Dockerfile`.
`context` is set to the directory and `dockerfile` is explicitly set to `Dockerfile`.
### Output of `pulumi about`
I don't want to copy/paste this from a work project. I'll try to provide specific info as requested if relevant.
### Additional context
Split from #566
### Contributing
Vote on this issue by adding a 👍 reaction.
To contribute a fix for this issue, leave a comment (and link to your pull request, if you've opened one already).
| non_main | docker index image could not open dockerfile at relative path dockerfile what happened upgraded from pulumi docker to pulumi up unexpectedly errors diagnostics docker index image my app latest error could not open dockerfile at relative path dockerfile try setting dockerfile to users shed repos my repo projects my app dockerfile expected behavior pulumi up to locate dockerfile relative to my project directory steps to reproduce my pulumi resource looks something like python image docker image my app latest build docker dockerbuildargs context users shed repos my repo projects my app dockerfile dockerfile args buildkit inline cache may not be necessary cache from docker cachefromargs images pulumi output concat repo repository url latest platform linux registry registry image name pulumi output concat self repo repository url latest there is a file users shed repos my repo projects my app dockerfile context is set to the directory and dockerfile is explicitly set to dockerfile output of pulumi about i don t want to copy paste this from a work project i ll try to provide specific info as requested if relevant additional context split from contributing vote on this issue by adding a 👍 reaction to contribute a fix for this issue leave a comment and link to your pull request if you ve opened one already | 0 |
91,265 | 15,856,386,602 | IssuesEvent | 2021-04-08 02:13:23 | n-devs/NodeJSControUI | https://api.github.com/repos/n-devs/NodeJSControUI | opened | CVE-2021-23337 (High) detected in lodash-4.17.5.tgz | security vulnerability | ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.5.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz</a></p>
<p>Path to dependency file: /NodeJSControUI/package.json</p>
<p>Path to vulnerable library: NodeJSControUI/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- jsdom-11.6.2.tgz (Root Library)
- request-promise-native-1.0.5.tgz
- request-promise-core-1.1.1.tgz
- :x: **lodash-4.17.5.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23337 (High) detected in lodash-4.17.5.tgz - ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.5.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz</a></p>
<p>Path to dependency file: /NodeJSControUI/package.json</p>
<p>Path to vulnerable library: NodeJSControUI/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- jsdom-11.6.2.tgz (Root Library)
- request-promise-native-1.0.5.tgz
- request-promise-core-1.1.1.tgz
- :x: **lodash-4.17.5.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file nodejscontroui package json path to vulnerable library nodejscontroui node modules lodash package json dependency hierarchy jsdom tgz root library request promise native tgz request promise core tgz x lodash tgz vulnerable library vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource | 0 |
5,707 | 8,367,726,522 | IssuesEvent | 2018-10-04 13:05:10 | SAEONData/ckanext-metadata | https://api.github.com/repos/SAEONData/ckanext-metadata | closed | Metadata record uniqueness | requirement | An attempt to create a metadata record with the same DOI and download link as an existing one, should be processed as an update to that existing record.
A match on DOI but not on download link, or vice versa, should be interpreted as an error. | 1.0 | Metadata record uniqueness - An attempt to create a metadata record with the same DOI and download link as an existing one, should be processed as an update to that existing record.
A match on DOI but not on download link, or vice versa, should be interpreted as an error. | non_main | metadata record uniqueness an attempt to create a metadata record with the same doi and download link as an existing one should be processed as an update to that existing record a match on doi but not on download link or vice versa should be interpreted as an error | 0 |
344,070 | 10,339,838,277 | IssuesEvent | 2019-09-03 20:20:30 | knative/docs | https://api.github.com/repos/knative/docs | closed | Lacking guidance on nuances of writing controllers | kind/new-docs-needed priority/1 | I'd like us to write down or link to some of the accumulated wisdom and best practices learned about how to write Kubernetes controllers. This will make onboarding new contributors easier (see https://github.com/knative/eventing/pull/308#discussion_r207280167) and help keep old contributors honest (and informed) about why we do things a certain way.
In particular, I think these questions need well-supported answers:
- When should status be updated during reconcile?
- What counts as a reconcile error (i.e., what errors should requeue the resource)?
- When is it ok to update the spec of a resource?
- Can I update a different resource during reconcile?
When these answers already exist in https://github.com/kubernetes/community/tree/master/contributors/devel (https://github.com/kubernetes/community/blob/master/contributors/devel/api-conventions.md and https://github.com/kubernetes/community/blob/master/contributors/devel/controllers.md probably have some great info), we should link to them, since those documents are hard to find otherwise.
/cc @pmorie @mattmoor @dprotaso @evankanderson @n3wscott
| 1.0 | Lacking guidance on nuances of writing controllers - I'd like us to write down or link to some of the accumulated wisdom and best practices learned about how to write Kubernetes controllers. This will make onboarding new contributors easier (see https://github.com/knative/eventing/pull/308#discussion_r207280167) and help keep old contributors honest (and informed) about why we do things a certain way.
In particular, I think these questions need well-supported answers:
- When should status be updated during reconcile?
- What counts as a reconcile error (i.e., what errors should requeue the resource)?
- When is it ok to update the spec of a resource?
- Can I update a different resource during reconcile?
When these answers already exist in https://github.com/kubernetes/community/tree/master/contributors/devel (https://github.com/kubernetes/community/blob/master/contributors/devel/api-conventions.md and https://github.com/kubernetes/community/blob/master/contributors/devel/controllers.md probably have some great info), we should link to them, since those documents are hard to find otherwise.
/cc @pmorie @mattmoor @dprotaso @evankanderson @n3wscott
| non_main | lacking guidance on nuances of writing controllers i d like us to write down or link to some of the accumulated wisdom and best practices learned about how to write kubernetes controllers this will make onboarding new contributors easier see and help keep old contributors honest and informed about why we do things a certain way in particular i think these questions need well supported answers when should status be updated during reconcile what counts as a reconcile error i e what errors should requeue the resource when is it ok to update the spec of a resource can i update a different resource during reconcile when these answers already exist in and probably have some great info we should link to them since those documents are hard to find otherwise cc pmorie mattmoor dprotaso evankanderson | 0 |
44,999 | 18,294,573,972 | IssuesEvent | 2021-10-05 19:02:43 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Update CTMA GIS Data / Add Automation to Process | Type: Data Service: Geo Workgroup: ATD | CMTA updated their data again in TX Open Data Portal, update all layers in GISDM, them AGOL. | 1.0 | Update CTMA GIS Data / Add Automation to Process - CMTA updated their data again in TX Open Data Portal, update all layers in GISDM, them AGOL. | non_main | update ctma gis data add automation to process cmta updated their data again in tx open data portal update all layers in gisdm them agol | 0 |
295,400 | 22,215,131,647 | IssuesEvent | 2022-06-08 00:17:34 | aws/aws-sdk-go-v2 | https://api.github.com/repos/aws/aws-sdk-go-v2 | opened | README badges are not searchable with ctrl + F | documentation needs-triage | ### Describe the issue
Reopening #1293.
If I do ctrl + F on the landing page and search for "migration" it doesn't find the badge in the results, because the badge is an image not text. This makes it hard to find the migration guide unless you already know it's a badge (so in essence https://github.com/aws/aws-sdk-go-v2/issues/1289 isn't truly resolved).
### Links
https://github.com/aws/aws-sdk-go-v2/blob/main/README.md
### AWS Go SDK V2 Module Versions Used
n/a | 1.0 | README badges are not searchable with ctrl + F - ### Describe the issue
Reopening #1293.
If I do ctrl + F on the landing page and search for "migration" it doesn't find the badge in the results, because the badge is an image not text. This makes it hard to find the migration guide unless you already know it's a badge (so in essence https://github.com/aws/aws-sdk-go-v2/issues/1289 isn't truly resolved).
### Links
https://github.com/aws/aws-sdk-go-v2/blob/main/README.md
### AWS Go SDK V2 Module Versions Used
n/a | non_main | readme badges are not searchable with ctrl f describe the issue reopening if i do ctrl f on the landing page and search for migration it doesn t find the badge in the results because the badge is an image not text this makes it hard to find the migration guide unless you already know it s a badge so in essence isn t truly resolved links aws go sdk module versions used n a | 0 |
3,913 | 17,480,659,881 | IssuesEvent | 2021-08-09 01:13:23 | DLR-RM/rl-baselines3-zoo | https://api.github.com/repos/DLR-RM/rl-baselines3-zoo | closed | "MySQLdb._exceptions.OperationalError: (1054, "Unknown column 'infe0' in 'field list'") | Maintainers on vacation | I've recently been using RL Baselines Zoo and I'm getting an error message that I'm not able to comprehend, and is almost certainly a bug. I've attached the giant traceback as a txt file to this. I'd be willing to fix it if I actually understood what was happening because this is holding up moderately important work. After some Googling, the issue appears to be somehow related to the inf showing up in the logs: https://github.com/pandas-dev/pandas/issues/34431
[magical_error.txt](https://github.com/DLR-RM/rl-baselines3-zoo/files/6919284/magical_error.txt)
| True | "MySQLdb._exceptions.OperationalError: (1054, "Unknown column 'infe0' in 'field list'") - I've recently been using RL Baselines Zoo and I'm getting an error message that I'm not able to comprehend, and is almost certainly a bug. I've attached the giant traceback as a txt file to this. I'd be willing to fix it if I actually understood what was happening because this is holding up moderately important work. After some Googling, the issue appears to be somehow related to the inf showing up in the logs: https://github.com/pandas-dev/pandas/issues/34431
[magical_error.txt](https://github.com/DLR-RM/rl-baselines3-zoo/files/6919284/magical_error.txt)
| main | mysqldb exceptions operationalerror unknown column in field list i ve recently been using rl baselines zoo and i m getting an error message that i m not able to comprehend and is almost certainly a bug i ve attached the giant traceback as a txt file to this i d be willing to fix it if i actually understood what was happening because this is holding up moderately important work after some googling the issue appears to be somehow related to the inf showing up in the logs | 1 |
4,281 | 21,527,840,627 | IssuesEvent | 2022-04-28 20:24:07 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Error when trying to load tables | type: enhancement work: frontend status: ready restricted: maintainers | ## Reproduce
1. Go to `http://localhost:8000/mathesar_tables/2/?t=W1tdLG51bGxd` to load the "astronomy" schema.
1. Switch to the "public" schema.
1. Observe a 500 response from `/api/db/v0/tables/` with error code 4999 and the following message:
> "Got KeyError when attempting to get a value for field `input` on serializer `BooleanDisplayOptionSerializer`.\nThe serializer field might be named incorrectly and not match any attribute or key on the `dict` instance.\nOriginal exception text was: 'input'."
| True | Error when trying to load tables - ## Reproduce
1. Go to `http://localhost:8000/mathesar_tables/2/?t=W1tdLG51bGxd` to load the "astronomy" schema.
1. Switch to the "public" schema.
1. Observe a 500 response from `/api/db/v0/tables/` with error code 4999 and the following message:
> "Got KeyError when attempting to get a value for field `input` on serializer `BooleanDisplayOptionSerializer`.\nThe serializer field might be named incorrectly and not match any attribute or key on the `dict` instance.\nOriginal exception text was: 'input'."
| main | error when trying to load tables reproduce go to to load the astronomy schema switch to the public schema observe a response from api db tables with error code and the following message got keyerror when attempting to get a value for field input on serializer booleandisplayoptionserializer nthe serializer field might be named incorrectly and not match any attribute or key on the dict instance noriginal exception text was input | 1 |
27,514 | 5,357,755,517 | IssuesEvent | 2017-02-20 19:38:36 | webgme/webgme | https://api.github.com/repos/webgme/webgme | opened | Revise API Documentation | Documentation Minor REST API | - Bearer security scheme should be explained.
- Fix wrong/outdated responses
- Add security where applicable
| 1.0 | Revise API Documentation - - Bearer security scheme should be explained.
- Fix wrong/outdated responses
- Add security where applicable
| non_main | revise api documentation bearer security scheme should be explained fix wrong outdated responses add security where applicable | 0 |
2,500 | 8,655,458,140 | IssuesEvent | 2018-11-27 16:00:19 | codestation/qcma | https://api.github.com/repos/codestation/qcma | closed | no ID3 tags shown for music | unmaintained | I cant use the album list or anything other than all songs even though id3 tags are verry orderly
| True | no ID3 tags shown for music - I cant use the album list or anything other than all songs even though id3 tags are verry orderly
| main | no tags shown for music i cant use the album list or anything other than all songs even though tags are verry orderly | 1 |
190,044 | 15,215,956,697 | IssuesEvent | 2021-02-17 14:59:16 | fduquesne/poker-planning | https://api.github.com/repos/fduquesne/poker-planning | opened | Changer le README.md | documentation | Changer le contenu du README.md pour le spécifier à notre application.
Y placer aussi les https://shields.io/ | 1.0 | Changer le README.md - Changer le contenu du README.md pour le spécifier à notre application.
Y placer aussi les https://shields.io/ | non_main | changer le readme md changer le contenu du readme md pour le spécifier à notre application y placer aussi les | 0 |
100,352 | 8,736,849,846 | IssuesEvent | 2018-12-11 20:45:28 | Princeton-CDH/ppa-django | https://api.github.com/repos/Princeton-CDH/ppa-django | closed | As an admin, I want to suppress items from the site so that I can pull content that should not be included or was wrongly added as I am going through and assigning collections to archive volumes. | awaiting testing | ## Notes for testing
- [x] digitized works now have a status field; default is public, you can set manually to suppressed
- [x] should see an indicator on the list view if something is public or suppressed
- [x] should be able to filter the list view on status
- [x] status should be included in CSV export
- when you set a record to suppressed the data should be deleted from the hathitrust pairtree data so we don't actually import and index it again (not sure how you can test this; you could ask us to run the hathi import script on the source id?)
- [x] if you try to switch a suppressed record back to public, you should get a validation error because it's not yet supported
## Notes for development
We don't want to actually delete the record from the database; we'll want to keep a stub at least, to indicate the record was removed and track the history.
- [x] add a status field; options public/suppressed, default to public
- [x] make editable in admin
- [x] display status in the admin list view so removed items are obvious; also configure as a filter.
- [x] Include status field in CSV export
- [x] when status is changed to suppressed, delete rsync data so it won't be re-added/indexed on a full import
- [x] don't allow un-suppressing items (validation? pre-save hook?)
### out of scope
- We may eventually want a bulk removal option, but consider that out of scope for now.
- Supporting "un-suppress" logic is out of scope for now.
| 1.0 | As an admin, I want to suppress items from the site so that I can pull content that should not be included or was wrongly added as I am going through and assigning collections to archive volumes. - ## Notes for testing
- [x] digitized works now have a status field; default is public, you can set manually to suppressed
- [x] should see an indicator on the list view if something is public or suppressed
- [x] should be able to filter the list view on status
- [x] status should be included in CSV export
- when you set a record to suppressed the data should be deleted from the hathitrust pairtree data so we don't actually import and index it again (not sure how you can test this; you could ask us to run the hathi import script on the source id?)
- [x] if you try to switch a suppressed record back to public, you should get a validation error because it's not yet supported
## Notes for development
We don't want to actually delete the record from the database; we'll want to keep a stub at least, to indicate the record was removed and track the history.
- [x] add a status field; options public/suppressed, default to public
- [x] make editable in admin
- [x] display status in the admin list view so removed items are obvious; also configure as a filter.
- [x] Include status field in CSV export
- [x] when status is changed to suppressed, delete rsync data so it won't be re-added/indexed on a full import
- [x] don't allow un-suppressing items (validation? pre-save hook?)
### out of scope
- We may eventually want a bulk removal option, but consider that out of scope for now.
- Supporting "un-suppress" logic is out of scope for now.
| non_main | as an admin i want to suppress items from the site so that i can pull content that should not be included or was wrongly added as i am going through and assigning collections to archive volumes notes for testing digitized works now have a status field default is public you can set manually to suppressed should see an indicator on the list view if something is public or suppressed should be able to filter the list view on status status should be included in csv export when you set a record to suppressed the data should be deleted from the hathitrust pairtree data so we don t actually import and index it again not sure how you can test this you could ask us to run the hathi import script on the source id if you try to switch a suppressed record back to public you should get a validation error because it s not yet supported notes for development we don t want to actually delete the record from the database we ll want to keep a stub at least to indicate the record was removed and track the history add a status field options public suppressed default to public make editable in admin display status in the admin list view so removed items are obvious also configure as a filter include status field in csv export when status is changed to suppressed delete rsync data so it won t be re added indexed on a full import don t allow un suppressing items validation pre save hook out of scope we may eventually want a bulk removal option but consider that out of scope for now supporting un suppress logic is out of scope for now | 0 |
5,452 | 27,288,408,339 | IssuesEvent | 2023-02-23 15:01:55 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Handle 404 pages throughout the app | type: bug work: frontend status: ready restricted: maintainers | ## Description
* Client side routing: A non existent url results in a blank screen, or has obscure errors.
* Server side routing: A non existent url entered on the browser results in a 404, which returns a django 404 page.
## Expected behavior
* 404s should be clearly presented to the user in all valid cases.
* The 404 page should be consistent in appearance in both server side & client side routing. | True | Handle 404 pages throughout the app - ## Description
* Client side routing: A non existent url results in a blank screen, or has obscure errors.
* Server side routing: A non existent url entered on the browser results in a 404, which returns a django 404 page.
## Expected behavior
* 404s should be clearly presented to the user in all valid cases.
* The 404 page should be consistent in appearance in both server side & client side routing. | main | handle pages throughout the app description client side routing a non existent url results in a blank screen or has obscure errors server side routing a non existent url entered on the browser results in a which returns a django page expected behavior should be clearly presented to the user in all valid cases the page should be consistent in appearance in both server side client side routing | 1 |
2,534 | 12,228,232,895 | IssuesEvent | 2020-05-03 18:31:32 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | ARM template not successfully deploying | Pri2 automation/svc cxp product-question triaged update-management/subsvc | Hi there, I'm having an issue with the ARM file on this page.
It fails deployment, it creates the automation account, creates the workspace, links the automation account with the workspace. All successfully.
Finally it tries to perform some sort of update on the automation account which appears in the deployment list as "Updates(workspacename)" and returns an error "Not Found" and "Bad Request". It's a bit vague.
The powershell error isn't very helpful either:
New-AzResourceGroupDeployment : 15:17:41 - Resource Microsoft.OperationsManagement/solutions 'Updates(workspacename)'
failed with message '{
"error": {
"code": "BadRequest",
"message": ""
}
}'
At line:1 char:1
If I can provide any more information please let me know.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: c0123bad-47d3-a6bc-95b4-95aea681ad95
* Version Independent ID: acaf27f7-1d41-954d-f4f2-5065ffd896d6
* Content: [Use Azure Resource Manager templates to onboard Update Management](https://docs.microsoft.com/en-us/azure/automation/automation-update-management-deploy-template#feedback)
* Content Source: [articles/automation/automation-update-management-deploy-template.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-update-management-deploy-template.md)
* Service: **automation**
* Sub-service: **update-management**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte** | 1.0 | ARM template not successfully deploying - Hi there, I'm having an issue with the ARM file on this page.
It fails deployment, it creates the automation account, creates the workspace, links the automation account with the workspace. All successfully.
Finally it tries to perform some sort of update on the automation account which appears in the deployment list as "Updates(workspacename)" and returns an error "Not Found" and "Bad Request". It's a bit vague.
The powershell error isn't very helpful either:
New-AzResourceGroupDeployment : 15:17:41 - Resource Microsoft.OperationsManagement/solutions 'Updates(workspacename)'
failed with message '{
"error": {
"code": "BadRequest",
"message": ""
}
}'
At line:1 char:1
If I can provide any more information please let me know.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: c0123bad-47d3-a6bc-95b4-95aea681ad95
* Version Independent ID: acaf27f7-1d41-954d-f4f2-5065ffd896d6
* Content: [Use Azure Resource Manager templates to onboard Update Management](https://docs.microsoft.com/en-us/azure/automation/automation-update-management-deploy-template#feedback)
* Content Source: [articles/automation/automation-update-management-deploy-template.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-update-management-deploy-template.md)
* Service: **automation**
* Sub-service: **update-management**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte** | non_main | arm template not successfully deploying hi there i m having an issue with the arm file on this page it fails deployment it creates the automation account creates the workspace links the automation account with the workspace all successfully finally it tries to perform some sort of update on the automation account which appears in the deployment list as updates workspacename and returns an error not found and bad request it s a bit vague the powershell error isn t very helpful either new azresourcegroupdeployment resource microsoft operationsmanagement solutions updates workspacename failed with message error code badrequest message at line char if i can provide any more information please let me know document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service update management github login mgoedtel microsoft alias magoedte | 0 |
3,948 | 17,873,129,737 | IssuesEvent | 2021-09-06 19:41:32 | antigenomics/vdjdb-db | https://api.github.com/repos/antigenomics/vdjdb-db | opened | TRAJ24*01 vs TRAJ24*02 | maintainance formatting&proofreading | Reported by Jamie Heather
About 2/3 of the ~400 human TCR entries reportedly making use of TRAJ24*01 are actually probably using TRAJ24*02. There's a two nt mismatch between them that corresponds to a two AA FGXG-proximal mismatch:
```
TTDSWGKFEFGAGTQVVVTP TRAJ24*01
TTDSWGKLQFGAGTQVVVTP TRAJ24*02
^^
```
It'd take a lot of deletion to make *02 into *01 it's pretty likely that most of the '-WGKLQF' CDR3s (which is the majority) are actually *02. | True | TRAJ24*01 vs TRAJ24*02 - Reported by Jamie Heather
About 2/3 of the ~400 human TCR entries reportedly making use of TRAJ24*01 are actually probably using TRAJ24*02. There's a two nt mismatch between them that corresponds to a two AA FGXG-proximal mismatch:
```
TTDSWGKFEFGAGTQVVVTP TRAJ24*01
TTDSWGKLQFGAGTQVVVTP TRAJ24*02
^^
```
It'd take a lot of deletion to make *02 into *01 it's pretty likely that most of the '-WGKLQF' CDR3s (which is the majority) are actually *02. | main | vs reported by jamie heather about of the human tcr entries reportedly making use of are actually probably using there s a two nt mismatch between them that corresponds to a two aa fgxg proximal mismatch ttdswgkfefgagtqvvvtp ttdswgklqfgagtqvvvtp it d take a lot of deletion to make into it s pretty likely that most of the wgklqf which is the majority are actually | 1 |
4,847 | 24,973,526,603 | IssuesEvent | 2022-11-02 04:54:57 | rthadur/bazel | https://api.github.com/repos/rthadur/bazel | closed | Test 3 | awaiting-review awaiting-user-response more data needed awaiting-maintainer | ### Description
Test 3
### Issue Type
_No response_
### Operating System
_No response_
### Coral Device
_No response_
### Other Devices
_No response_
### Programming Language
_No response_
### Relevant Log Output
_No response_ | True | Test 3 - ### Description
Test 3
### Issue Type
_No response_
### Operating System
_No response_
### Coral Device
_No response_
### Other Devices
_No response_
### Programming Language
_No response_
### Relevant Log Output
_No response_ | main | test description test issue type no response operating system no response coral device no response other devices no response programming language no response relevant log output no response | 1 |
963 | 4,706,289,311 | IssuesEvent | 2016-10-13 16:42:58 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | ansible-modules-core/network/ - Code review | affects_2.2 bug_report in progress networking P1 waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
eos_facts
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0 (devel 70e63ddf6c) last updated 2016/09/15 10:17:19 (GMT +100)
lib/ansible/modules/core: (devel 683e5e4d1a) last updated 2016/09/15 10:17:22 (GMT +100)
lib/ansible/modules/extras: (devel 170adf16bd) last updated 2016/09/15 10:17:23 (GMT +100)
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
I've raised one issue to track all the issues found rather than having a fairly bitty chain of tickets. If it's easier for you to raise different PRs to address the issues found I'm no issue with that - whatever is easiest for you.
I'm wondering if for items we are happy with we should add ignore markers in, as shown here
http://stackoverflow.com/questions/28829236/is-it-possible-to-ignore-one-single-specific-line-with-pylint
```
pylint -E network/*/*
No config file found, using default configuration
************* Module ansible.modules.core.network.nxos.nxos_hsrp
E:402,41: Undefined variable 'module' (undefined-variable)
************* Module ansible.modules.core.network.nxos.nxos_interface
E:147,19: Instance of 'list' has no 'split' member (no-member)
E:535,56: Undefined variable 'command' (undefined-variable)
E:581,13: Undefined variable 'get_module' (undefined-variable)
************* Module ansible.modules.core.network.nxos.nxos_static_route
E:158,23: Instance of 'CustomNetworkConfig' has no 'to_lines' member (no-member)
E:295,26: Instance of 'list' has no 'split' member (no-member)
E:402,66: Using variable 'address' before assignment (used-before-assignment)
************* Module ansible.modules.core.network.nxos.nxos_switchport
E:486,56: Undefined variable 'command' (undefined-variable)
E:527,13: Undefined variable 'get_module' (undefined-variable)
************* Module ansible.modules.core.network.nxos.nxos_vrf
E:501,52: Undefined variable 'cmds' (undefined-variable)
```
##### STEPS TO REPRODUCE
##### EXPECTED RESULTS
##### ACTUAL RESULTS
| True | ansible-modules-core/network/ - Code review - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
eos_facts
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0 (devel 70e63ddf6c) last updated 2016/09/15 10:17:19 (GMT +100)
lib/ansible/modules/core: (devel 683e5e4d1a) last updated 2016/09/15 10:17:22 (GMT +100)
lib/ansible/modules/extras: (devel 170adf16bd) last updated 2016/09/15 10:17:23 (GMT +100)
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
I've raised one issue to track all the issues found rather than having a fairly bitty chain of tickets. If it's easier for you to raise different PRs to address the issues found I'm no issue with that - whatever is easiest for you.
I'm wondering if for items we are happy with we should add ignore markers in, as shown here
http://stackoverflow.com/questions/28829236/is-it-possible-to-ignore-one-single-specific-line-with-pylint
```
pylint -E network/*/*
No config file found, using default configuration
************* Module ansible.modules.core.network.nxos.nxos_hsrp
E:402,41: Undefined variable 'module' (undefined-variable)
************* Module ansible.modules.core.network.nxos.nxos_interface
E:147,19: Instance of 'list' has no 'split' member (no-member)
E:535,56: Undefined variable 'command' (undefined-variable)
E:581,13: Undefined variable 'get_module' (undefined-variable)
************* Module ansible.modules.core.network.nxos.nxos_static_route
E:158,23: Instance of 'CustomNetworkConfig' has no 'to_lines' member (no-member)
E:295,26: Instance of 'list' has no 'split' member (no-member)
E:402,66: Using variable 'address' before assignment (used-before-assignment)
************* Module ansible.modules.core.network.nxos.nxos_switchport
E:486,56: Undefined variable 'command' (undefined-variable)
E:527,13: Undefined variable 'get_module' (undefined-variable)
************* Module ansible.modules.core.network.nxos.nxos_vrf
E:501,52: Undefined variable 'cmds' (undefined-variable)
```
##### STEPS TO REPRODUCE
##### EXPECTED RESULTS
##### ACTUAL RESULTS
| main | ansible modules core network code review issue type bug report component name eos facts ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt configuration os environment summary i ve raised one issue to track all the issues found rather than having a fairly bitty chain of tickets if it s easier for you to raise different prs to address the issues found i m no issue with that whatever is easiest for you i m wondering if for items we are happy with we should add ignore markers in as shown here pylint e network no config file found using default configuration module ansible modules core network nxos nxos hsrp e undefined variable module undefined variable module ansible modules core network nxos nxos interface e instance of list has no split member no member e undefined variable command undefined variable e undefined variable get module undefined variable module ansible modules core network nxos nxos static route e instance of customnetworkconfig has no to lines member no member e instance of list has no split member no member e using variable address before assignment used before assignment module ansible modules core network nxos nxos switchport e undefined variable command undefined variable e undefined variable get module undefined variable module ansible modules core network nxos nxos vrf e undefined variable cmds undefined variable steps to reproduce expected results actual results | 1 |
1,915 | 6,577,706,212 | IssuesEvent | 2017-09-12 02:44:55 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | AWS Route53 limited in options for setting up routes (weighted latency not available) | affects_2.0 aws cloud feature_idea waiting_on_maintainer | ##### Issue Type:
- Feature Idea
##### Plugin Name:
- cloud/amazon/route53
##### Ansible Version:
ansible 2.0.0.2
config file = /home/db/.ansible.cfg
configured module search path = Default w/o overrides
##### Ansible Configuration:
None
##### Environment:
Ubuntu 12.04 running on a VirtualBox VM.
##### Summary:
I have manually built some infrastructure in AWS and I want to automate the build of that with Ansible. This manually built infrastructure uses weighted latency based routing. However, I cannot read back the information fully from the facts nor is there any option for setting this up.
Speaking with @defionscode at Ansiblefest 2016 last Thursday, I was asked to raise this issue for a fix.
##### Steps To Reproduce:
Not a bug - feature request.
##### Expected Results:
Simply trying to get the results back of the current routes to reflect the configuration available through the console.
##### Actual Results:
Not applicable.
| True | AWS Route53 limited in options for setting up routes (weighted latency not available) - ##### Issue Type:
- Feature Idea
##### Plugin Name:
- cloud/amazon/route53
##### Ansible Version:
ansible 2.0.0.2
config file = /home/db/.ansible.cfg
configured module search path = Default w/o overrides
##### Ansible Configuration:
None
##### Environment:
Ubuntu 12.04 running on a VirtualBox VM.
##### Summary:
I have manually built some infrastructure in AWS and I want to automate the build of that with Ansible. This manually built infrastructure uses weighted latency based routing. However, I cannot read back the information fully from the facts nor is there any option for setting this up.
Speaking with @defionscode at Ansiblefest 2016 last Thursday, I was asked to raise this issue for a fix.
##### Steps To Reproduce:
Not a bug - feature request.
##### Expected Results:
Simply trying to get the results back of the current routes to reflect the configuration available through the console.
##### Actual Results:
Not applicable.
| main | aws limited in options for setting up routes weighted latency not available issue type feature idea plugin name cloud amazon ansible version ansible config file home db ansible cfg configured module search path default w o overrides ansible configuration none environment ubuntu running on a virtualbox vm summary i have manually built some infrastructure in aws and i want to automate the build of that with ansible this manually built infrastructure uses weighted latency based routing however i cannot read back the information fully from the facts nor is there any option for setting this up speaking with defionscode at ansiblefest last thursday i was asked to raise this issue for a fix steps to reproduce not a bug feature request expected results simply trying to get the results back of the current routes to reflect the configuration available through the console actual results not applicable | 1 |
214,824 | 24,120,463,020 | IssuesEvent | 2022-09-20 18:13:17 | microsoft/Microsoft365DSC | https://api.github.com/repos/microsoft/Microsoft365DSC | closed | Security and Compliance settings - SCLabelPolicy workload | Enhancement Security & Compliance Center | Wondering if there are any plans to include the "Auto-labelling" policies as part of this tool?

| True | Security and Compliance settings - SCLabelPolicy workload - Wondering if there are any plans to include the "Auto-labelling" policies as part of this tool?

| non_main | security and compliance settings sclabelpolicy workload wondering if there are any plans to include the auto labelling policies as part of this tool | 0 |
202,058 | 23,053,921,543 | IssuesEvent | 2022-07-25 01:17:05 | Xi0ngfei/e-mart-backend | https://api.github.com/repos/Xi0ngfei/e-mart-backend | opened | spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar: 1 vulnerabilities (highest severity is: 9.1) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /emart-eureka-service/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/5.0.3/woodstox-core-5.0.3.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Xi0ngfei/e-mart-backend/commit/9e2cdf0fabfba0aa30b3a80420cea42d1b714754">9e2cdf0fabfba0aa30b3a80420cea42d1b714754</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [WS-2018-0629](https://github.com/FasterXML/woodstox/commit/7937f97c638ef8afd385ebf4a675a9b096ccdd57) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | woodstox-core-5.0.3.jar | Transitive | 2.2.0.RELEASE | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0629</summary>
### Vulnerable Library - <b>woodstox-core-5.0.3.jar</b></p>
<p>Woodstox is a high-performance XML processor that
implements Stax (JSR-173), SAX2 and Stax2 APIs</p>
<p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p>
<p>Path to dependency file: /emart-eureka-service/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/5.0.3/woodstox-core-5.0.3.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar (Root Library)
- spring-cloud-netflix-eureka-server-2.1.0.RELEASE.jar
- jackson-dataformat-xml-2.9.7.jar
- :x: **woodstox-core-5.0.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Xi0ngfei/e-mart-backend/commit/9e2cdf0fabfba0aa30b3a80420cea42d1b714754">9e2cdf0fabfba0aa30b3a80420cea42d1b714754</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The woodstox-core package is vulnerable to improper restriction of XXE reference.
<p>Publish Date: 2018-08-23
<p>URL: <a href=https://github.com/FasterXML/woodstox/commit/7937f97c638ef8afd385ebf4a675a9b096ccdd57>WS-2018-0629</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-08-23</p>
<p>Fix Resolution (com.fasterxml.woodstox:woodstox-core): 5.2.1</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-netflix-eureka-server): 2.2.0.RELEASE</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar: 1 vulnerabilities (highest severity is: 9.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /emart-eureka-service/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/5.0.3/woodstox-core-5.0.3.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/Xi0ngfei/e-mart-backend/commit/9e2cdf0fabfba0aa30b3a80420cea42d1b714754">9e2cdf0fabfba0aa30b3a80420cea42d1b714754</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [WS-2018-0629](https://github.com/FasterXML/woodstox/commit/7937f97c638ef8afd385ebf4a675a9b096ccdd57) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | woodstox-core-5.0.3.jar | Transitive | 2.2.0.RELEASE | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0629</summary>
### Vulnerable Library - <b>woodstox-core-5.0.3.jar</b></p>
<p>Woodstox is a high-performance XML processor that
implements Stax (JSR-173), SAX2 and Stax2 APIs</p>
<p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p>
<p>Path to dependency file: /emart-eureka-service/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/5.0.3/woodstox-core-5.0.3.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar (Root Library)
- spring-cloud-netflix-eureka-server-2.1.0.RELEASE.jar
- jackson-dataformat-xml-2.9.7.jar
- :x: **woodstox-core-5.0.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Xi0ngfei/e-mart-backend/commit/9e2cdf0fabfba0aa30b3a80420cea42d1b714754">9e2cdf0fabfba0aa30b3a80420cea42d1b714754</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The woodstox-core package is vulnerable to improper restriction of XXE reference.
<p>Publish Date: 2018-08-23
<p>URL: <a href=https://github.com/FasterXML/woodstox/commit/7937f97c638ef8afd385ebf4a675a9b096ccdd57>WS-2018-0629</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2018-08-23</p>
<p>Fix Resolution (com.fasterxml.woodstox:woodstox-core): 5.2.1</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-netflix-eureka-server): 2.2.0.RELEASE</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_main | spring cloud starter netflix eureka server release jar vulnerabilities highest severity is vulnerable library spring cloud starter netflix eureka server release jar path to dependency file emart eureka service pom xml path to vulnerable library home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high woodstox core jar transitive release details ws vulnerable library woodstox core jar woodstox is a high performance xml processor that implements stax jsr and apis library home page a href path to dependency file emart eureka service pom xml path to vulnerable library home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar dependency hierarchy spring cloud starter netflix eureka server release jar root library spring cloud netflix eureka server release jar jackson dataformat xml jar x woodstox core jar vulnerable library found in head commit a href found in base branch master vulnerability details the woodstox core package is vulnerable to improper restriction of xxe reference publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml woodstox woodstox core direct dependency fix resolution org springframework cloud spring cloud starter netflix eureka server release step up your open source security game with mend | 0 |
176,398 | 28,087,475,263 | IssuesEvent | 2023-03-30 10:46:14 | audacity/audacity | https://api.github.com/repos/audacity/audacity | reopened | Inconsistency with Linear dB view at default track size | bug Design / UX | ### Bug description
**Default track size**
When using the new **Linear (dB)** view in the Vertical Scale with stereo tracks at **_default_** size, there is an odd **-3dB** shown only in the uppper half of the right channel. The left channel and the lower half of the right channel do not have this **-3dB** labelling. This looks odd and unbalanced:

**Track size reduced slightly from default**
If you reduce the height of the stereo track a little then you get a balnced Vertical Scale display:

**Track size expanded slightly from default**
If you expand it slightly more than the defaul stereo track size teh you get a different unbalanced view with the -3db labelling in just the upper half of both tracks:

**Track size expanded a lot from default**
Expanding the track size further (image below is full default window size) gives a nicely balanced view:

### Steps to reproduce
1. clear audaicty settings folder
2. launch 3.3.0 alpha
3. record a stero track - or import one
4. right-click in the Vertical Scale and select **Linear (dB)**
5. Observe: only the upper half of the right channel has a **-3db** marker
### Expected behavior
Consistency.
If **-3dB** is to be shown, then it should be shown in both halves of both channels
### Actual behavior
See above
### Audacity Version
3.3.0 alpha - audacity-win-3.3.0-alpha-20230310+9b9b935-x64-msvc2022
### Operating system
Windows 11 - but assume all OS
### Additional context
This is on my smaller Zurich W11 HP Probook with a 13.5 inch screen.
_I can retest this next week back in Manchester on my 17 inch W10 HP Envy - and on my Macbook Pro_
**Importance**
Although this looks like a trivial display/cosmetic issue, it is somewhat important as it mars an otherwise useful and nice new feature added by GSoC22 participant @micpap25 | 1.0 | Inconsistency with Linear dB view at default track size - ### Bug description
**Default track size**
When using the new **Linear (dB)** view in the Vertical Scale with stereo tracks at **_default_** size, there is an odd **-3dB** shown only in the uppper half of the right channel. The left channel and the lower half of the right channel do not have this **-3dB** labelling. This looks odd and unbalanced:

**Track size reduced slightly from default**
If you reduce the height of the stereo track a little then you get a balnced Vertical Scale display:

**Track size expanded slightly from default**
If you expand it slightly more than the defaul stereo track size teh you get a different unbalanced view with the -3db labelling in just the upper half of both tracks:

**Track size expanded a lot from default**
Expanding the track size further (image below is full default window size) gives a nicely balanced view:

### Steps to reproduce
1. clear audaicty settings folder
2. launch 3.3.0 alpha
3. record a stero track - or import one
4. right-click in the Vertical Scale and select **Linear (dB)**
5. Observe: only the upper half of the right channel has a **-3db** marker
### Expected behavior
Consistency.
If **-3dB** is to be shown, then it should be shown in both halves of both channels
### Actual behavior
See above
### Audacity Version
3.3.0 alpha - audacity-win-3.3.0-alpha-20230310+9b9b935-x64-msvc2022
### Operating system
Windows 11 - but assume all OS
### Additional context
This is on my smaller Zurich W11 HP Probook with a 13.5 inch screen.
_I can retest this next week back in Manchester on my 17 inch W10 HP Envy - and on my Macbook Pro_
**Importance**
Although this looks like a trivial display/cosmetic issue, it is somewhat important as it mars an otherwise useful and nice new feature added by GSoC22 participant @micpap25 | non_main | inconsistency with linear db view at default track size bug description default track size when using the new linear db view in the vertical scale with stereo tracks at default size there is an odd shown only in the uppper half of the right channel the left channel and the lower half of the right channel do not have this labelling this looks odd and unbalanced track size reduced slightly from default if you reduce the height of the stereo track a little then you get a balnced vertical scale display track size expanded slightly from default if you expand it slightly more than the defaul stereo track size teh you get a different unbalanced view with the labelling in just the upper half of both tracks track size expanded a lot from default expanding the track size further image below is full default window size gives a nicely balanced view steps to reproduce clear audaicty settings folder launch alpha record a stero track or import one right click in the vertical scale and select linear db observe only the upper half of the right channel has a marker expected behavior consistency if is to be shown then it should be shown in both halves of both channels actual behavior see above audacity version alpha audacity win alpha operating system windows but assume all os additional context this is on my smaller zurich hp probook with a inch screen i can retest this next week back in manchester on my inch hp envy and on my macbook pro importance although this looks like a trivial display cosmetic issue it is somewhat important as it mars an otherwise useful and nice new feature added by participant | 0 |
799,290 | 28,303,979,374 | IssuesEvent | 2023-04-10 09:06:49 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | chrome.google.com - see bug description | browser-chrome priority-critical | <!-- @browser: Chrome 111.0.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/120632 -->
**URL**: https://chrome.google.com/webstore/detail/mozilla-rally/bahhehaddofgkccippmjcecepdakppme/related
**Browser / Version**: Chrome 111.0.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Firefox
**Problem type**: Something else
**Description**: Cant install Rally extension in chrome
**Steps to Reproduce**:
Cant install Mozilla Rally on Chrome browser even when logged in to google account
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/4/64757b61-aed8-4301-a1f2-3949ae03be55.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | chrome.google.com - see bug description - <!-- @browser: Chrome 111.0.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/120632 -->
**URL**: https://chrome.google.com/webstore/detail/mozilla-rally/bahhehaddofgkccippmjcecepdakppme/related
**Browser / Version**: Chrome 111.0.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Firefox
**Problem type**: Something else
**Description**: Cant install Rally extension in chrome
**Steps to Reproduce**:
Cant install Mozilla Rally on Chrome browser even when logged in to google account
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/4/64757b61-aed8-4301-a1f2-3949ae03be55.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | chrome google com see bug description url browser version chrome operating system windows tested another browser yes firefox problem type something else description cant install rally extension in chrome steps to reproduce cant install mozilla rally on chrome browser even when logged in to google account view the screenshot img alt screenshot src browser configuration none from with ❤️ | 0 |
42,359 | 6,975,511,753 | IssuesEvent | 2017-12-12 07:27:06 | php-deal/framework | https://api.github.com/repos/php-deal/framework | closed | Library name in Scrutinizer and Packagist | documentation enhancement | To be consistent, scrutinizer-ci.com/g/**lisachenko/php-deal** and packagist.org/packages/**lisachenko/php-deal** should be renamed or a new project should be created :) | 1.0 | Library name in Scrutinizer and Packagist - To be consistent, scrutinizer-ci.com/g/**lisachenko/php-deal** and packagist.org/packages/**lisachenko/php-deal** should be renamed or a new project should be created :) | non_main | library name in scrutinizer and packagist to be consistent scrutinizer ci com g lisachenko php deal and packagist org packages lisachenko php deal should be renamed or a new project should be created | 0 |
3,418 | 13,182,092,276 | IssuesEvent | 2020-08-12 15:15:16 | duo-labs/cloudmapper | https://api.github.com/repos/duo-labs/cloudmapper | closed | Network visualization: Filter by subnets | map unmaintained_functionality | It would be great to have a flag that enables to filter the prepare command output w.r.t. to the subnet(s) just like VPC. | True | Network visualization: Filter by subnets - It would be great to have a flag that enables to filter the prepare command output w.r.t. to the subnet(s) just like VPC. | main | network visualization filter by subnets it would be great to have a flag that enables to filter the prepare command output w r t to the subnet s just like vpc | 1 |
95,467 | 10,880,769,452 | IssuesEvent | 2019-11-17 13:31:10 | DiscipleTools/disciple-tools-theme | https://api.github.com/repos/DiscipleTools/disciple-tools-theme | closed | GULP browser sync optimization Improvement | bug documentation |
Developers who desire to improve the DiscipleTools theme, may request for live reloading from a local development standpoint. This needs to be fixed internally, and also documented within the README.md for other developers to use.
GULP, and JavaScript coding is required.
| 1.0 | GULP browser sync optimization Improvement -
Developers who desire to improve the DiscipleTools theme, may request for live reloading from a local development standpoint. This needs to be fixed internally, and also documented within the README.md for other developers to use.
GULP, and JavaScript coding is required.
| non_main | gulp browser sync optimization improvement developers who desire to improve the discipletools theme may request for live reloading from a local development standpoint this needs to be fixed internally and also documented within the readme md for other developers to use gulp and javascript coding is required | 0 |
2,328 | 8,345,162,625 | IssuesEvent | 2018-10-01 00:10:32 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Airlocks use static types for tool interaction instead of the tool_behaviour system | Consistency Issue Maintainability/Hinders improvements | https://github.com/tgstation/tgstation/blob/master/code/game/machinery/doors/door.dm#L183
Which means they're not weldable with custom welders such as, currently, plasma cutters, and not crowbarable by anything but crowbars and fireaxes. | True | Airlocks use static types for tool interaction instead of the tool_behaviour system - https://github.com/tgstation/tgstation/blob/master/code/game/machinery/doors/door.dm#L183
Which means they're not weldable with custom welders such as, currently, plasma cutters, and not crowbarable by anything but crowbars and fireaxes. | main | airlocks use static types for tool interaction instead of the tool behaviour system which means they re not weldable with custom welders such as currently plasma cutters and not crowbarable by anything but crowbars and fireaxes | 1 |
3,699 | 15,098,796,194 | IssuesEvent | 2021-02-08 00:18:49 | pypiserver/pypiserver | https://api.github.com/repos/pypiserver/pypiserver | opened | Speed up test-pypy job in GH actions | difficulty.easy good first issue type.Maintainance | The test-pypy job is _much_ slower than the others, largely because its `pip install` on the "Install dependencies" step takes two orders of magnitude longer due to its needing to compile wheels.
Investigate whether we can cache the built wheels for pypy to avoid needing to rebuild them every time. | True | Speed up test-pypy job in GH actions - The test-pypy job is _much_ slower than the others, largely because its `pip install` on the "Install dependencies" step takes two orders of magnitude longer due to its needing to compile wheels.
Investigate whether we can cache the built wheels for pypy to avoid needing to rebuild them every time. | main | speed up test pypy job in gh actions the test pypy job is much slower than the others largely because its pip install on the install dependencies step takes two orders of magnitude longer due to its needing to compile wheels investigate whether we can cache the built wheels for pypy to avoid needing to rebuild them every time | 1 |
2,111 | 7,176,917,820 | IssuesEvent | 2018-01-31 11:44:11 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | opened | Analyzer should report "wrong" kinds of exceptions being thrown | analyzer feature maintainability | The analyzer should report that there are exception thrown (created) which are not acceptable to be created.
Such exceptions are:
- System.NullReferenceException
- System.Exception
- System.FatalEngineException | True | Analyzer should report "wrong" kinds of exceptions being thrown - The analyzer should report that there are exception thrown (created) which are not acceptable to be created.
Such exceptions are:
- System.NullReferenceException
- System.Exception
- System.FatalEngineException | main | analyzer should report wrong kinds of exceptions being thrown the analyzer should report that there are exception thrown created which are not acceptable to be created such exceptions are system nullreferenceexception system exception system fatalengineexception | 1 |
4,399 | 22,588,082,467 | IssuesEvent | 2022-06-28 16:59:35 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [Bug]: Using appendTo with DatePicker causes carriage returns to not work in sibling textarea | type: bug 🐛 status: needs triage 🕵️♀️ status: waiting for maintainer response 💬 | ### Package
carbon-components-react
### Browser
Chrome
### Package version
7.57.1
### React version
16.14.0
### Description
Some time ago, I reported an issue that when the calendar popup was open, it didn't scroll with the input field. https://github.com/carbon-design-system/carbon/issues/4158. The appendTo prop had been broken, and when it was fixed, I set it to the immediate child of the scroll element per the suggestion given for https://github.com/carbon-design-system/carbon/issues/4818.
At some point later, we realized that the TextArea that was in the same "appendTo" as the DatePicker was getting it's carriage returns swallowed by flatpickr. Flatpickr had a PR that at first looked like it might fix it https://github.com/flatpickr/flatpickr/issues/2054 but it has since been merged and it doesn't help.
I started to write a defect against Flatpickr but found that I couldn't reproduce the scrolling issue when there's no appendTo, and it doesn't have the carriage return issue when the appendTo is set to a child of the scroll container.
Note that my teammate wrote a similar issue here. https://github.com/carbon-design-system/carbon/issues/11640 However, I think this is a more straight-forward use case.
### Reproduction/example
https://9qiwkg.csb.app/
### Steps to reproduce
Put your cursor in the text area field and hit Enter. The cursor does not move to the next line.
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | True | [Bug]: Using appendTo with DatePicker causes carriage returns to not work in sibling textarea - ### Package
carbon-components-react
### Browser
Chrome
### Package version
7.57.1
### React version
16.14.0
### Description
Some time ago, I reported an issue that when the calendar popup was open, it didn't scroll with the input field. https://github.com/carbon-design-system/carbon/issues/4158. The appendTo prop had been broken, and when it was fixed, I set it to the immediate child of the scroll element per the suggestion given for https://github.com/carbon-design-system/carbon/issues/4818.
At some point later, we realized that the TextArea that was in the same "appendTo" as the DatePicker was getting it's carriage returns swallowed by flatpickr. Flatpickr had a PR that at first looked like it might fix it https://github.com/flatpickr/flatpickr/issues/2054 but it has since been merged and it doesn't help.
I started to write a defect against Flatpickr but found that I couldn't reproduce the scrolling issue when there's no appendTo, and it doesn't have the carriage return issue when the appendTo is set to a child of the scroll container.
Note that my teammate wrote a similar issue here. https://github.com/carbon-design-system/carbon/issues/11640 However, I think this is a more straight-forward use case.
### Reproduction/example
https://9qiwkg.csb.app/
### Steps to reproduce
Put your cursor in the text area field and hit Enter. The cursor does not move to the next line.
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | main | using appendto with datepicker causes carriage returns to not work in sibling textarea package carbon components react browser chrome package version react version description some time ago i reported an issue that when the calendar popup was open it didn t scroll with the input field the appendto prop had been broken and when it was fixed i set it to the immediate child of the scroll element per the suggestion given for at some point later we realized that the textarea that was in the same appendto as the datepicker was getting it s carriage returns swallowed by flatpickr flatpickr had a pr that at first looked like it might fix it but it has since been merged and it doesn t help i started to write a defect against flatpickr but found that i couldn t reproduce the scrolling issue when there s no appendto and it doesn t have the carriage return issue when the appendto is set to a child of the scroll container note that my teammate wrote a similar issue here however i think this is a more straight forward use case reproduction example steps to reproduce put your cursor in the text area field and hit enter the cursor does not move to the next line code of conduct i agree to follow this project s i checked the for duplicate problems | 1 |
341,541 | 30,591,655,455 | IssuesEvent | 2023-07-21 17:36:17 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix tensor.test_torch_instance_arccos | PyTorch Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-failure-red></a>
| 1.0 | Fix tensor.test_torch_instance_arccos - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-failure-red></a>
| non_main | fix tensor test torch instance arccos tensorflow a href src jax a href src numpy a href src torch a href src paddle a href src | 0 |
3,042 | 11,277,788,696 | IssuesEvent | 2020-01-15 04:13:00 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | terraform plan does not provide output for 2.9.0 | affects_2.9 bug cloud has_pr module needs_maintainer needs_triage support:community | ##### SUMMARY
Running terraform plan from ansible 2.9.0 does not provide expected output in stdout. This may be similar to issue #46589.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
terraform
##### ANSIBLE VERSION
```
# ansible --version
ansible 2.9.0
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/dist-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.15+ (default, Oct 7 2019, 17:39:04) [GCC 7.4.0]
```
##### CONFIGURATION
```
# ansible-config dump --only-changed
ALLOW_WORLD_READABLE_TMPFILES(/etc/ansible/ansible.cfg) = True
ANSIBLE_NOCOWS(/etc/ansible/ansible.cfg) = True
DEFAULT_GATHERING(/etc/ansible/ansible.cfg) = explicit
DEFAULT_STDOUT_CALLBACK(/etc/ansible/ansible.cfg) = debug
HOST_KEY_CHECKING(/etc/ansible/ansible.cfg) = False
INTERPRETER_PYTHON(/etc/ansible/ansible.cfg) = auto_silent
```
##### OS / ENVIRONMENT
- Ubuntu 18.04.3 LTS
```
# uname -a
Linux 57f42cc031d6 4.15.0-66-generic #75-Ubuntu SMP Tue Oct 1 05:24:09 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
# terraform --version
Terraform v0.12.13
# pip freeze
ansible==2.9.0
asn1crypto==0.24.0
cryptography==2.1.4
enum34==1.1.6
httplib2==0.9.2
idna==2.6
ipaddress==1.0.17
Jinja2==2.10
keyring==10.6.0
keyrings.alt==3.0
MarkupSafe==1.0
paramiko==2.0.0
pyasn1==0.4.2
pycrypto==2.6.1
pygobject==3.26.1
pyxdg==0.25
PyYAML==3.12
SecretStorage==2.3.1
six==1.11.0
# pip3 freeze
asn1crypto==0.24.0
awscli==1.16.272
boto==2.49.0
boto3==1.10.8
botocore==1.13.8
chardet==3.0.4
colorama==0.4.1
cryptography==2.1.4
docutils==0.15.2
idna==2.6
jmespath==0.9.4
keyring==10.6.0
keyrings.alt==3.0
pyasn1==0.4.7
pycrypto==2.6.1
pygobject==3.26.1
python-apt==1.6.4
python-dateutil==2.8.1
python-debian==0.1.32
pyxdg==0.25
PyYAML==5.1.2
rsa==3.4.2
s3transfer==0.2.1
SecretStorage==2.3.1
six==1.11.0
unattended-upgrades==0.1
urllib3==1.25.6
virtualenv==15.1.0
```
##### STEPS TO REPRODUCE
Run terraform plan from ansible
```
############
# Run terraform plan
- name: Run terraform plan for VPC resources
terraform:
state: planned
project_path: "{{ fileTerraformWorkingPath }}"
plan_file: "{{ fileTerraformWorkingPath }}/plan.tfplan"
force_init: yes
backend_config:
region: "{{ nameAWSRegion }}"
register: vpc_tf_stack
############
# Print information about the base VPC
- name: Display everything with terraform
debug:
var: vpc_tf_stack
- name: Display all terraform output for VPC
debug:
var: vpc_tf_stack.stdout
```
##### EXPECTED RESULTS
Terraform plan stdout is shown.
##### ACTUAL RESULTS
Nothing is shown in stdout.
```
TASK [planTerraform : Display everything with terraform] *************************************************************************************************************************************
ok: [127.0.0.1] => {
"vpc_tf_stack": {
"changed": false,
"command": "/usr/bin/terraform plan -input=false -no-color -detailed-exitcode -out /tmp/terraform-20191104141619238338988/plan.tfplan /tmp/terraform-20191104141619238338988/plan.tfplan",
"failed": false,
"outputs": {
"idAdminHostedZone": {
"sensitive": false,
"type": "string",
"value": "Z23423423423423423423"
},
"idExternalSecurityGroup": {
"sensitive": false,
"type": "string",
"value": "sg-12312312312312312"
},
"idIGW": {
"sensitive": false,
"type": "string",
"value": "igw-12312312312312312"
},
"idLocalHostedZone": {
"sensitive": false,
"type": "string",
"value": "Z12312312312312312312"
},
"idPublicRouteTable": {
"sensitive": false,
"type": "string",
"value": "rtb-12312312312312312"
},
"idVPC": {
"sensitive": false,
"type": "string",
"value": "vpc-12312312312312312"
}
},
"state": "planned",
"stderr": "",
"stderr_lines": [],
"stdout": "",
"stdout_lines": [],
"workspace": "default"
}
}
TASK [planTerraform : Display all terraform output for VPC] **********************************************************************************************************************************
ok: [127.0.0.1] => {
"vpc_tf_stack.stdout": ""
}
```
| True | terraform plan does not provide output for 2.9.0 - ##### SUMMARY
Running terraform plan from ansible 2.9.0 does not provide expected output in stdout. This may be similar to issue #46589.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
terraform
##### ANSIBLE VERSION
```
# ansible --version
ansible 2.9.0
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/dist-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.15+ (default, Oct 7 2019, 17:39:04) [GCC 7.4.0]
```
##### CONFIGURATION
```
# ansible-config dump --only-changed
ALLOW_WORLD_READABLE_TMPFILES(/etc/ansible/ansible.cfg) = True
ANSIBLE_NOCOWS(/etc/ansible/ansible.cfg) = True
DEFAULT_GATHERING(/etc/ansible/ansible.cfg) = explicit
DEFAULT_STDOUT_CALLBACK(/etc/ansible/ansible.cfg) = debug
HOST_KEY_CHECKING(/etc/ansible/ansible.cfg) = False
INTERPRETER_PYTHON(/etc/ansible/ansible.cfg) = auto_silent
```
##### OS / ENVIRONMENT
- Ubuntu 18.04.3 LTS
```
# uname -a
Linux 57f42cc031d6 4.15.0-66-generic #75-Ubuntu SMP Tue Oct 1 05:24:09 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
# terraform --version
Terraform v0.12.13
# pip freeze
ansible==2.9.0
asn1crypto==0.24.0
cryptography==2.1.4
enum34==1.1.6
httplib2==0.9.2
idna==2.6
ipaddress==1.0.17
Jinja2==2.10
keyring==10.6.0
keyrings.alt==3.0
MarkupSafe==1.0
paramiko==2.0.0
pyasn1==0.4.2
pycrypto==2.6.1
pygobject==3.26.1
pyxdg==0.25
PyYAML==3.12
SecretStorage==2.3.1
six==1.11.0
# pip3 freeze
asn1crypto==0.24.0
awscli==1.16.272
boto==2.49.0
boto3==1.10.8
botocore==1.13.8
chardet==3.0.4
colorama==0.4.1
cryptography==2.1.4
docutils==0.15.2
idna==2.6
jmespath==0.9.4
keyring==10.6.0
keyrings.alt==3.0
pyasn1==0.4.7
pycrypto==2.6.1
pygobject==3.26.1
python-apt==1.6.4
python-dateutil==2.8.1
python-debian==0.1.32
pyxdg==0.25
PyYAML==5.1.2
rsa==3.4.2
s3transfer==0.2.1
SecretStorage==2.3.1
six==1.11.0
unattended-upgrades==0.1
urllib3==1.25.6
virtualenv==15.1.0
```
##### STEPS TO REPRODUCE
Run terraform plan from ansible
```
############
# Run terraform plan
- name: Run terraform plan for VPC resources
terraform:
state: planned
project_path: "{{ fileTerraformWorkingPath }}"
plan_file: "{{ fileTerraformWorkingPath }}/plan.tfplan"
force_init: yes
backend_config:
region: "{{ nameAWSRegion }}"
register: vpc_tf_stack
############
# Print information about the base VPC
- name: Display everything with terraform
debug:
var: vpc_tf_stack
- name: Display all terraform output for VPC
debug:
var: vpc_tf_stack.stdout
```
##### EXPECTED RESULTS
Terraform plan stdout is shown.
##### ACTUAL RESULTS
Nothing is shown in stdout.
```
TASK [planTerraform : Display everything with terraform] *************************************************************************************************************************************
ok: [127.0.0.1] => {
"vpc_tf_stack": {
"changed": false,
"command": "/usr/bin/terraform plan -input=false -no-color -detailed-exitcode -out /tmp/terraform-20191104141619238338988/plan.tfplan /tmp/terraform-20191104141619238338988/plan.tfplan",
"failed": false,
"outputs": {
"idAdminHostedZone": {
"sensitive": false,
"type": "string",
"value": "Z23423423423423423423"
},
"idExternalSecurityGroup": {
"sensitive": false,
"type": "string",
"value": "sg-12312312312312312"
},
"idIGW": {
"sensitive": false,
"type": "string",
"value": "igw-12312312312312312"
},
"idLocalHostedZone": {
"sensitive": false,
"type": "string",
"value": "Z12312312312312312312"
},
"idPublicRouteTable": {
"sensitive": false,
"type": "string",
"value": "rtb-12312312312312312"
},
"idVPC": {
"sensitive": false,
"type": "string",
"value": "vpc-12312312312312312"
}
},
"state": "planned",
"stderr": "",
"stderr_lines": [],
"stdout": "",
"stdout_lines": [],
"workspace": "default"
}
}
TASK [planTerraform : Display all terraform output for VPC] **********************************************************************************************************************************
ok: [127.0.0.1] => {
"vpc_tf_stack.stdout": ""
}
```
| main | terraform plan does not provide output for summary running terraform plan from ansible does not provide expected output in stdout this may be similar to issue issue type bug report component name terraform ansible version ansible version ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib dist packages ansible executable location usr bin ansible python version default oct configuration ansible config dump only changed allow world readable tmpfiles etc ansible ansible cfg true ansible nocows etc ansible ansible cfg true default gathering etc ansible ansible cfg explicit default stdout callback etc ansible ansible cfg debug host key checking etc ansible ansible cfg false interpreter python etc ansible ansible cfg auto silent os environment ubuntu lts uname a linux generic ubuntu smp tue oct utc gnu linux terraform version terraform pip freeze ansible cryptography idna ipaddress keyring keyrings alt markupsafe paramiko pycrypto pygobject pyxdg pyyaml secretstorage six freeze awscli boto botocore chardet colorama cryptography docutils idna jmespath keyring keyrings alt pycrypto pygobject python apt python dateutil python debian pyxdg pyyaml rsa secretstorage six unattended upgrades virtualenv steps to reproduce run terraform plan from ansible run terraform plan name run terraform plan for vpc resources terraform state planned project path fileterraformworkingpath plan file fileterraformworkingpath plan tfplan force init yes backend config region nameawsregion register vpc tf stack print information about the base vpc name display everything with terraform debug var vpc tf stack name display all terraform output for vpc debug var vpc tf stack stdout expected results terraform plan stdout is shown actual results nothing is shown in stdout task ok vpc tf stack changed false command usr bin terraform plan input false no color detailed exitcode out tmp terraform plan tfplan tmp terraform plan tfplan failed false outputs idadminhostedzone sensitive false type string value idexternalsecuritygroup sensitive false type string value sg idigw sensitive false type string value igw idlocalhostedzone sensitive false type string value idpublicroutetable sensitive false type string value rtb idvpc sensitive false type string value vpc state planned stderr stderr lines stdout stdout lines workspace default task ok vpc tf stack stdout | 1 |
640 | 4,156,813,608 | IssuesEvent | 2016-06-16 19:11:51 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | closed | Code Search: Change API Endpoint | Maintainer Input Requested | The Spice isn't working because the API endpoint has changed to:
`https://searchcode.com/api/jsonp_search_IV/?q=php&callback=myCallback`
from
`https://searchcode.com/api/jsonp_codesearch_I/?q=php&callback=myCallback`
Please double check if the data has changed, i.e., if the Spice is showing up properly or not as well!
Thanks!
---
Maintainer: @boyter
IA Page: https://duck.co/ia/view/code_search | True | Code Search: Change API Endpoint - The Spice isn't working because the API endpoint has changed to:
`https://searchcode.com/api/jsonp_search_IV/?q=php&callback=myCallback`
from
`https://searchcode.com/api/jsonp_codesearch_I/?q=php&callback=myCallback`
Please double check if the data has changed, i.e., if the Spice is showing up properly or not as well!
Thanks!
---
Maintainer: @boyter
IA Page: https://duck.co/ia/view/code_search | main | code search change api endpoint the spice isn t working because the api endpoint has changed to from please double check if the data has changed i e if the spice is showing up properly or not as well thanks maintainer boyter ia page | 1 |
2,319 | 8,303,638,825 | IssuesEvent | 2018-09-21 18:15:57 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | deprecate start/stop/step in AnalysisBase.__init__() | Component-Analysis deprecation maintainability | In preparation for #1463 where we want analysis classes to use `start`, `stop`, `step` in the `run()` method only, we need to deprecate the use in
- [x] `analysis.base.AnalysisBase.__init__`: https://github.com/MDAnalysis/mdanalysis/blob/1a1b2f1442d52d1216b53b32c4c42c3b47368d38/package/MDAnalysis/analysis/base.py#L92
- custom analysis classes (`git grep 'start=.*stop='`)
- [x] density.py
- [x] diffusionmap.py (clean-up only, already uses AnalysisBase)
- [x] hbonds/hbond_analysis.py
- [x] hbonds/wbridge_analysis.py
- [x] rms.py (removed the *deprecation* for start/stop/step in `run()`!) | True | deprecate start/stop/step in AnalysisBase.__init__() - In preparation for #1463 where we want analysis classes to use `start`, `stop`, `step` in the `run()` method only, we need to deprecate the use in
- [x] `analysis.base.AnalysisBase.__init__`: https://github.com/MDAnalysis/mdanalysis/blob/1a1b2f1442d52d1216b53b32c4c42c3b47368d38/package/MDAnalysis/analysis/base.py#L92
- custom analysis classes (`git grep 'start=.*stop='`)
- [x] density.py
- [x] diffusionmap.py (clean-up only, already uses AnalysisBase)
- [x] hbonds/hbond_analysis.py
- [x] hbonds/wbridge_analysis.py
- [x] rms.py (removed the *deprecation* for start/stop/step in `run()`!) | main | deprecate start stop step in analysisbase init in preparation for where we want analysis classes to use start stop step in the run method only we need to deprecate the use in analysis base analysisbase init custom analysis classes git grep start stop density py diffusionmap py clean up only already uses analysisbase hbonds hbond analysis py hbonds wbridge analysis py rms py removed the deprecation for start stop step in run | 1 |
208,265 | 16,108,320,435 | IssuesEvent | 2021-04-27 17:37:09 | uber/h3 | https://api.github.com/repos/uber/h3 | closed | Projected coordinate system | documentation | For those functions related to latitude and longitude, is the projected coordinate system used as WGS84 or something? | 1.0 | Projected coordinate system - For those functions related to latitude and longitude, is the projected coordinate system used as WGS84 or something? | non_main | projected coordinate system for those functions related to latitude and longitude is the projected coordinate system used as or something | 0 |
2,601 | 8,837,878,921 | IssuesEvent | 2019-01-05 10:50:33 | OXIDprojects/oxid-module-internals | https://api.github.com/repos/OXIDprojects/oxid-module-internals | closed | duplicate code | bug maintainability pull request | As Maintainer I like to remove the effort in maintaining and to increase the quality by removeing duplicate code.
- I found duplicate code in CheckConsistency.php | True | duplicate code - As Maintainer I like to remove the effort in maintaining and to increase the quality by removeing duplicate code.
- I found duplicate code in CheckConsistency.php | main | duplicate code as maintainer i like to remove the effort in maintaining and to increase the quality by removeing duplicate code i found duplicate code in checkconsistency php | 1 |
5,054 | 25,888,706,525 | IssuesEvent | 2022-12-14 16:17:32 | deislabs/spiderlightning | https://api.github.com/repos/deislabs/spiderlightning | closed | `slight` panics when the config path isn't a slightfile. | 🐛 bug 🚧 maintainer issue | **Description of the bug**
When running slight with a config path that isn't a slightfile, `slight` CLI will panic. I found that there are various causes to the issues
1. path does not exist
2. path isn't a slightfile
3. slightfile is mis-configured.
**To Reproduce**
```
> ./target/release/slight -c ./examples/http-demo/random_slightfile.toml run -m ./examples/http-demo/target/wasm32-wasi/release/http-demo.wasm
thread 'main' panicked at 'called `Option::unwrap()` on a `None` value', slight/src/commands/run.rs:115:34
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
```
**Additional context**
| True | `slight` panics when the config path isn't a slightfile. - **Description of the bug**
When running slight with a config path that isn't a slightfile, `slight` CLI will panic. I found that there are various causes to the issues
1. path does not exist
2. path isn't a slightfile
3. slightfile is mis-configured.
**To Reproduce**
```
> ./target/release/slight -c ./examples/http-demo/random_slightfile.toml run -m ./examples/http-demo/target/wasm32-wasi/release/http-demo.wasm
thread 'main' panicked at 'called `Option::unwrap()` on a `None` value', slight/src/commands/run.rs:115:34
note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace
```
**Additional context**
| main | slight panics when the config path isn t a slightfile description of the bug when running slight with a config path that isn t a slightfile slight cli will panic i found that there are various causes to the issues path does not exist path isn t a slightfile slightfile is mis configured to reproduce target release slight c examples http demo random slightfile toml run m examples http demo target wasi release http demo wasm thread main panicked at called option unwrap on a none value slight src commands run rs note run with rust backtrace environment variable to display a backtrace additional context | 1 |
4,788 | 24,628,591,726 | IssuesEvent | 2022-10-16 20:38:37 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Destroy TabularData in RecordSelector, TablePage, and TableWidget to prevent Memory leaks | type: bug work: frontend status: ready restricted: maintainers | ## Description
* There are a lot of places (eg.,RecordSelector, TablePage, and TableWidget) where TabularData object is created reactively, like this:
```
$: tabularData = new TabularData({
id: table.id,
abstractTypesMap,
meta,
});
```
This creates a new instance each time any of the arguments get updated, and is never destroyed, leading to memory leaks. Each TabularData instance should be destroyed before the next instance is created.
If possible, we should consider the following improvements:
* The TabularData instance creation is costly and should be kept to a minimum, if possible it should not be made reactive.
* We should update a TabularData's property instead of recreating if it is possible. | True | Destroy TabularData in RecordSelector, TablePage, and TableWidget to prevent Memory leaks - ## Description
* There are a lot of places (eg.,RecordSelector, TablePage, and TableWidget) where TabularData object is created reactively, like this:
```
$: tabularData = new TabularData({
id: table.id,
abstractTypesMap,
meta,
});
```
This creates a new instance each time any of the arguments get updated, and is never destroyed, leading to memory leaks. Each TabularData instance should be destroyed before the next instance is created.
If possible, we should consider the following improvements:
* The TabularData instance creation is costly and should be kept to a minimum, if possible it should not be made reactive.
* We should update a TabularData's property instead of recreating if it is possible. | main | destroy tabulardata in recordselector tablepage and tablewidget to prevent memory leaks description there are a lot of places eg recordselector tablepage and tablewidget where tabulardata object is created reactively like this tabulardata new tabulardata id table id abstracttypesmap meta this creates a new instance each time any of the arguments get updated and is never destroyed leading to memory leaks each tabulardata instance should be destroyed before the next instance is created if possible we should consider the following improvements the tabulardata instance creation is costly and should be kept to a minimum if possible it should not be made reactive we should update a tabulardata s property instead of recreating if it is possible | 1 |
499,632 | 14,474,846,021 | IssuesEvent | 2020-12-10 00:12:37 | Dynna/PBL | https://api.github.com/repos/Dynna/PBL | closed | Insufficient Logging and Monitoring | Backend/PHP/Laravel Priority: HIGH Security | - [x] When it comes to your application and server, log everything, including failed login attempts and password resets.
- [ ] Laravel comes with Monolog out of the box. You can even integrate it with a third party logging service like Papertrail and receive alerts for specific log events. | 1.0 | Insufficient Logging and Monitoring - - [x] When it comes to your application and server, log everything, including failed login attempts and password resets.
- [ ] Laravel comes with Monolog out of the box. You can even integrate it with a third party logging service like Papertrail and receive alerts for specific log events. | non_main | insufficient logging and monitoring when it comes to your application and server log everything including failed login attempts and password resets laravel comes with monolog out of the box you can even integrate it with a third party logging service like papertrail and receive alerts for specific log events | 0 |
332,387 | 10,092,319,707 | IssuesEvent | 2019-07-26 16:21:13 | cBioPortal/cbioportal | https://api.github.com/repos/cBioPortal/cbioportal | closed | Co-expression service performance | api critical priority | Following request can take 20-60 seconds to complete. It's causing some tests to fail.
/api/molecular-profiles/co-expressions/fetch?molecularProfileIdA=coadread_tcga_pub_rna_seq_mrna&molecularProfileIdB=coadread_tcga_pub_rna_seq_mrna&threshold=0
Odd that after our first round of optimization, I don't remember it taking so long. perhaps something else happened since then? | 1.0 | Co-expression service performance - Following request can take 20-60 seconds to complete. It's causing some tests to fail.
/api/molecular-profiles/co-expressions/fetch?molecularProfileIdA=coadread_tcga_pub_rna_seq_mrna&molecularProfileIdB=coadread_tcga_pub_rna_seq_mrna&threshold=0
Odd that after our first round of optimization, I don't remember it taking so long. perhaps something else happened since then? | non_main | co expression service performance following request can take seconds to complete it s causing some tests to fail api molecular profiles co expressions fetch molecularprofileida coadread tcga pub rna seq mrna molecularprofileidb coadread tcga pub rna seq mrna threshold odd that after our first round of optimization i don t remember it taking so long perhaps something else happened since then | 0 |
463,230 | 13,261,967,788 | IssuesEvent | 2020-08-20 20:51:48 | googleapis/google-api-php-client | https://api.github.com/repos/googleapis/google-api-php-client | closed | Basic Example in README.md does not work | priority: p2 type: bug type: docs | In the current README.md the following basic example is provided, however this is not correct, I suspect it's a bit outdated.
```php
// include your composer dependencies
require_once 'vendor/autoload.php';
$client = new Google_Client();
$client->setApplicationName("Client_Library_Examples");
$client->setDeveloperKey("YOUR_APP_KEY");
$service = new Google_Service_Books($client);
$optParams = array('filter' => 'free-ebooks');
$results = $service->volumes->listVolumes('Henry David Thoreau', $optParams);
foreach ($results->getItems() as $item) {
echo $item['volumeInfo']['title'], "<br /> \n";
}
```
I changed the following lines to make it work, once I noticed my IDE telling me that listVolumes only expects 1 parameter.
```
$optParams = array(
'filter' => 'free-ebooks',
'q' => 'Henry David Thoreau',
);
$results = $service->volumes->listVolumes( $optParams);
```
| 1.0 | Basic Example in README.md does not work - In the current README.md the following basic example is provided, however this is not correct, I suspect it's a bit outdated.
```php
// include your composer dependencies
require_once 'vendor/autoload.php';
$client = new Google_Client();
$client->setApplicationName("Client_Library_Examples");
$client->setDeveloperKey("YOUR_APP_KEY");
$service = new Google_Service_Books($client);
$optParams = array('filter' => 'free-ebooks');
$results = $service->volumes->listVolumes('Henry David Thoreau', $optParams);
foreach ($results->getItems() as $item) {
echo $item['volumeInfo']['title'], "<br /> \n";
}
```
I changed the following lines to make it work, once I noticed my IDE telling me that listVolumes only expects 1 parameter.
```
$optParams = array(
'filter' => 'free-ebooks',
'q' => 'Henry David Thoreau',
);
$results = $service->volumes->listVolumes( $optParams);
```
| non_main | basic example in readme md does not work in the current readme md the following basic example is provided however this is not correct i suspect it s a bit outdated php include your composer dependencies require once vendor autoload php client new google client client setapplicationname client library examples client setdeveloperkey your app key service new google service books client optparams array filter free ebooks results service volumes listvolumes henry david thoreau optparams foreach results getitems as item echo item n i changed the following lines to make it work once i noticed my ide telling me that listvolumes only expects parameter optparams array filter free ebooks q henry david thoreau results service volumes listvolumes optparams | 0 |
2,561 | 8,709,008,829 | IssuesEvent | 2018-12-06 12:42:59 | arcticicestudio/nord-docs | https://api.github.com/repos/arcticicestudio/nord-docs | opened | Core HTML element atoms | context-api context-ui scope-configurability scope-maintainability status-tracking type-epic | <p align="center"><img src="https://user-images.githubusercontent.com/7836623/49325616-f1053900-f545-11e8-9725-0b8f5af8d1e2.png" width="20%" /></p>
> Associated epic: #63
To achieve a consistent and uniform style and layout, the basic HTML elements, like e.g. a `<h1>` or `<ul>` should be used through a React component. These components render to the base HTML element they represent, but will apply styles, behavior and layout properties to ensure they match the project's design guidelines instead of using default browser configurations that might be even differ for each user agent.
This allows to use base HTML elements with all the advantages of React and JS without worrying about different render output.
This collection issue tracks the implementations of all the different individual React components. They are all based on the awesome MDN [HTML elements reference][mdn-html-el-ref] documentation and will use the same categorization.
## [Inline Text Semantics][mdn-html-el-ref-inl-txt-sem]
### `A`
> Implementation: #70
Represents the `<a>` HTML element (or anchor element). This is a special dynamic and _failsafe_ component since it'll internally use the [Gatsby Link][gatsby-link] to route within the site (internal links) while also being able to link to external data.
This will be handled through a utility function to conditionally render based on the passed target URL (internal & external).
[gatsby-link]: https://www.gatsbyjs.org/docs/gatsby-link
[mdn-html-el-ref]: https://developer.mozilla.org/en-US/docs/Web/HTML/Element
[mdn-html-el-ref-inl-txt-sem]: https://developer.mozilla.org/en-US/docs/Web/HTML/Element#Inline_text_semantics
| True | Core HTML element atoms - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/49325616-f1053900-f545-11e8-9725-0b8f5af8d1e2.png" width="20%" /></p>
> Associated epic: #63
To achieve a consistent and uniform style and layout, the basic HTML elements, like e.g. a `<h1>` or `<ul>` should be used through a React component. These components render to the base HTML element they represent, but will apply styles, behavior and layout properties to ensure they match the project's design guidelines instead of using default browser configurations that might be even differ for each user agent.
This allows to use base HTML elements with all the advantages of React and JS without worrying about different render output.
This collection issue tracks the implementations of all the different individual React components. They are all based on the awesome MDN [HTML elements reference][mdn-html-el-ref] documentation and will use the same categorization.
## [Inline Text Semantics][mdn-html-el-ref-inl-txt-sem]
### `A`
> Implementation: #70
Represents the `<a>` HTML element (or anchor element). This is a special dynamic and _failsafe_ component since it'll internally use the [Gatsby Link][gatsby-link] to route within the site (internal links) while also being able to link to external data.
This will be handled through a utility function to conditionally render based on the passed target URL (internal & external).
[gatsby-link]: https://www.gatsbyjs.org/docs/gatsby-link
[mdn-html-el-ref]: https://developer.mozilla.org/en-US/docs/Web/HTML/Element
[mdn-html-el-ref-inl-txt-sem]: https://developer.mozilla.org/en-US/docs/Web/HTML/Element#Inline_text_semantics
| main | core html element atoms associated epic to achieve a consistent and uniform style and layout the basic html elements like e g a or should be used through a react component these components render to the base html element they represent but will apply styles behavior and layout properties to ensure they match the project s design guidelines instead of using default browser configurations that might be even differ for each user agent this allows to use base html elements with all the advantages of react and js without worrying about different render output this collection issue tracks the implementations of all the different individual react components they are all based on the awesome mdn documentation and will use the same categorization a implementation represents the html element or anchor element this is a special dynamic and failsafe component since it ll internally use the to route within the site internal links while also being able to link to external data this will be handled through a utility function to conditionally render based on the passed target url internal external | 1 |
750,362 | 26,198,979,086 | IssuesEvent | 2023-01-03 15:48:41 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.amd.com - site is not usable | browser-firefox priority-normal engine-gecko | <!-- @browser: Firefox 108.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:108.0) Gecko/20100101 Firefox/108.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/116181 -->
**URL**: https://www.amd.com/en/direct-buy/
**Browser / Version**: Firefox 108.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
I can't seem to add to cart on FF but can do so on Chrome.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.amd.com - site is not usable - <!-- @browser: Firefox 108.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:108.0) Gecko/20100101 Firefox/108.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/116181 -->
**URL**: https://www.amd.com/en/direct-buy/
**Browser / Version**: Firefox 108.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
I can't seem to add to cart on FF but can do so on Chrome.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description buttons or links not working steps to reproduce i can t seem to add to cart on ff but can do so on chrome browser configuration none from with ❤️ | 0 |
3,812 | 16,541,533,819 | IssuesEvent | 2021-05-27 17:26:45 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | reopened | [TableToolbarSearch] icon overlay on text | status: needs triage 🕵️♀️ status: waiting for maintainer response 💬 type: bug 🐛 | <!-- Feel free to remove sections that aren't relevant.
## Title line template: [Title]: Brief description
-->
## What package(s) are you using?
<!--
Add an x in one of the options below, for example:
- [x] package name
-->
- [x] `carbon-components`
- [x] `carbon-components-react`
## Detailed description
> Describe in detail the issue you're having.
The table search has text overlaying the search icon
<img width="854" alt="Screen Shot 2021-05-06 at 12 29 35 PM" src="https://user-images.githubusercontent.com/17915105/117342895-5bcea300-ae69-11eb-815e-530596b30bb1.png">
This bx-input in search.scss set a padding which broke this.
<img width="560" alt="Screen Shot 2021-05-06 at 12 55 37 PM" src="https://user-images.githubusercontent.com/17915105/117343914-6c334d80-ae6a-11eb-8820-5cacdfa199b9.png">
> Is this issue related to a specific component?
TableToolbarSearch
> What did you expect to happen? What happened instead? What would you like to
> see changed?
The text not overlay on the icon
> What browser are you working in?
Chrome
> What version of the Carbon Design System are you using?
we used "carbon-components": "10.33.0", "carbon-components-react": "7.33.0",
> What offering/product do you work on? Any pressing ship or release dates we
> should be aware of?
IBM Cloud
## Steps to reproduce the issue
It's complicated to copy the whole table in sandbox. you have to import the search.scss
> Please create a reduced test case in CodeSandbox
>
> - Style and vanilla JS:
> https://codesandbox.io/s/github/carbon-design-system/carbon/tree/main/packages/components/examples/codesandbox
> - React:
> https://codesandbox.io/s/github/carbon-design-system/carbon/tree/main/packages/react/examples/codesandbox
https://codesandbox.io/s/pensive-butterfly-miqbx?file=/src/index.js
## Additional information
- Screenshots or code
- Notes
| True | [TableToolbarSearch] icon overlay on text - <!-- Feel free to remove sections that aren't relevant.
## Title line template: [Title]: Brief description
-->
## What package(s) are you using?
<!--
Add an x in one of the options below, for example:
- [x] package name
-->
- [x] `carbon-components`
- [x] `carbon-components-react`
## Detailed description
> Describe in detail the issue you're having.
The table search has text overlaying the search icon
<img width="854" alt="Screen Shot 2021-05-06 at 12 29 35 PM" src="https://user-images.githubusercontent.com/17915105/117342895-5bcea300-ae69-11eb-815e-530596b30bb1.png">
This bx-input in search.scss set a padding which broke this.
<img width="560" alt="Screen Shot 2021-05-06 at 12 55 37 PM" src="https://user-images.githubusercontent.com/17915105/117343914-6c334d80-ae6a-11eb-8820-5cacdfa199b9.png">
> Is this issue related to a specific component?
TableToolbarSearch
> What did you expect to happen? What happened instead? What would you like to
> see changed?
The text not overlay on the icon
> What browser are you working in?
Chrome
> What version of the Carbon Design System are you using?
we used "carbon-components": "10.33.0", "carbon-components-react": "7.33.0",
> What offering/product do you work on? Any pressing ship or release dates we
> should be aware of?
IBM Cloud
## Steps to reproduce the issue
It's complicated to copy the whole table in sandbox. you have to import the search.scss
> Please create a reduced test case in CodeSandbox
>
> - Style and vanilla JS:
> https://codesandbox.io/s/github/carbon-design-system/carbon/tree/main/packages/components/examples/codesandbox
> - React:
> https://codesandbox.io/s/github/carbon-design-system/carbon/tree/main/packages/react/examples/codesandbox
https://codesandbox.io/s/pensive-butterfly-miqbx?file=/src/index.js
## Additional information
- Screenshots or code
- Notes
| main | icon overlay on text feel free to remove sections that aren t relevant title line template brief description what package s are you using add an x in one of the options below for example package name carbon components carbon components react detailed description describe in detail the issue you re having the table search has text overlaying the search icon img width alt screen shot at pm src this bx input in search scss set a padding which broke this img width alt screen shot at pm src is this issue related to a specific component tabletoolbarsearch what did you expect to happen what happened instead what would you like to see changed the text not overlay on the icon what browser are you working in chrome what version of the carbon design system are you using we used carbon components carbon components react what offering product do you work on any pressing ship or release dates we should be aware of ibm cloud steps to reproduce the issue it s complicated to copy the whole table in sandbox you have to import the search scss please create a reduced test case in codesandbox style and vanilla js react additional information screenshots or code notes | 1 |
2,235 | 7,875,825,733 | IssuesEvent | 2018-06-25 21:48:27 | react-navigation/react-navigation | https://api.github.com/repos/react-navigation/react-navigation | closed | Shifting and focused boolean with createMaterialBottomTabNavigator | needs response from maintainer | ### Current Behavior
My aim is to make a MaterialBottomTabNavigator which changes the icon for the selected tab, among 5 tabs, however, the focused boolean appears to be broken when shifting is enabled.
Code example:
```js
import React from 'react';
import { createMaterialBottomTabNavigator } from 'react-navigation-material-bottom-tabs';
class Screen1 extends React.Component {
static navigationOptions = {
tabBarIcon: ({ focused }) => {
console.log('Screen1', focused);
},
activeTintColor: 'blue',
};
render() {
return null;
}
}
class Screen2 extends React.Component {
static navigationOptions = {
tabBarIcon: ({ focused }) => {
console.log('Screen2', focused);
},
activeTintColor: 'red',
};
render() {
return null;
}
}
const MaterialBottomTabNavigator = createMaterialBottomTabNavigator(
{
Screen1: Screen1,
Screen2: Screen2,
},
{
shifting: true,
}
);
export default class App extends React.Component {
render() {
return <MaterialBottomTabNavigator />;
}
}
```
Console shows:

However, if shifting is changed to false, the focused boolean works as expected, although the tabBarIcon still appears to be referenced twice:

### Expected Behavior
- The focused variable should work consistently between shifting enabled or disabled
### How to reproduce
- To see the error run the above code and compare console logs between shifting: true, and shifting: false
### Your Environment
| software | version
| ---------------- | -------
| react-navigation | ^2.0.1
| react-native | 0.55.4
| node | v8.11.2
| yarn | 1.6.0
| True | Shifting and focused boolean with createMaterialBottomTabNavigator - ### Current Behavior
My aim is to make a MaterialBottomTabNavigator which changes the icon for the selected tab, among 5 tabs, however, the focused boolean appears to be broken when shifting is enabled.
Code example:
```js
import React from 'react';
import { createMaterialBottomTabNavigator } from 'react-navigation-material-bottom-tabs';
class Screen1 extends React.Component {
static navigationOptions = {
tabBarIcon: ({ focused }) => {
console.log('Screen1', focused);
},
activeTintColor: 'blue',
};
render() {
return null;
}
}
class Screen2 extends React.Component {
static navigationOptions = {
tabBarIcon: ({ focused }) => {
console.log('Screen2', focused);
},
activeTintColor: 'red',
};
render() {
return null;
}
}
const MaterialBottomTabNavigator = createMaterialBottomTabNavigator(
{
Screen1: Screen1,
Screen2: Screen2,
},
{
shifting: true,
}
);
export default class App extends React.Component {
render() {
return <MaterialBottomTabNavigator />;
}
}
```
Console shows:

However, if shifting is changed to false, the focused boolean works as expected, although the tabBarIcon still appears to be referenced twice:

### Expected Behavior
- The focused variable should work consistently between shifting enabled or disabled
### How to reproduce
- To see the error run the above code and compare console logs between shifting: true, and shifting: false
### Your Environment
| software | version
| ---------------- | -------
| react-navigation | ^2.0.1
| react-native | 0.55.4
| node | v8.11.2
| yarn | 1.6.0
| main | shifting and focused boolean with creatematerialbottomtabnavigator current behavior my aim is to make a materialbottomtabnavigator which changes the icon for the selected tab among tabs however the focused boolean appears to be broken when shifting is enabled code example js import react from react import creatematerialbottomtabnavigator from react navigation material bottom tabs class extends react component static navigationoptions tabbaricon focused console log focused activetintcolor blue render return null class extends react component static navigationoptions tabbaricon focused console log focused activetintcolor red render return null const materialbottomtabnavigator creatematerialbottomtabnavigator shifting true export default class app extends react component render return console shows however if shifting is changed to false the focused boolean works as expected although the tabbaricon still appears to be referenced twice expected behavior the focused variable should work consistently between shifting enabled or disabled how to reproduce to see the error run the above code and compare console logs between shifting true and shifting false your environment software version react navigation react native node yarn | 1 |
331,944 | 29,174,493,486 | IssuesEvent | 2023-05-19 06:39:27 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | DISABLED test_build_tuple_unpack (__main__.StaticDefaultDynamicShapesMiscTests) | triaged module: flaky-tests skipped module: dynamo | Platforms: linux, mac, macos
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_build_tuple_unpack&suite=StaticDefaultDynamicShapesMiscTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/undefined).
Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_build_tuple_unpack`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `dynamo/test_dynamic_shapes.py` or `dynamo/test_dynamic_shapes.py` | 1.0 | DISABLED test_build_tuple_unpack (__main__.StaticDefaultDynamicShapesMiscTests) - Platforms: linux, mac, macos
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_build_tuple_unpack&suite=StaticDefaultDynamicShapesMiscTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/undefined).
Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_build_tuple_unpack`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `dynamo/test_dynamic_shapes.py` or `dynamo/test_dynamic_shapes.py` | non_main | disabled test build tuple unpack main staticdefaultdynamicshapesmisctests platforms linux mac macos this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test build tuple unpack there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path dynamo test dynamic shapes py or dynamo test dynamic shapes py | 0 |
609,957 | 18,891,254,947 | IssuesEvent | 2021-11-15 13:28:49 | ZeligsoftDev/CX4CBDDS | https://api.github.com/repos/ZeligsoftDev/CX4CBDDS | opened | PSDD Example cannot have IDL generated by CX Extensions plugin | bug Priority: Major NGC: Approved To Work | # **Issue and tracking information**
### Developer's time Estimated effort to fix (hours):
### Developer's Actual time spent on fix (hours)
# **Issue reporter to provide a detailed description of the issue in the space below**
The PSDD example can be validated and generate IDL manually, however, the automatic tooling to do batch validations and generations fail on the model with the following message:
`Value 'org.eclipse.uml2.uml.internal.impl.ClassImpl@63aeac6d (name: PSDD_Listen, visibility: <unset>) (isLeaf: false, isAbstract: false, isFinalSpecialization: false) (isActive: false)' is not legal. (file:/home/user/workspace-papyrus/SNA_Core/PSDDPubSub/ModelFiles/PSDDPubSub.uml, -1, -1)`
| 1.0 | PSDD Example cannot have IDL generated by CX Extensions plugin - # **Issue and tracking information**
### Developer's time Estimated effort to fix (hours):
### Developer's Actual time spent on fix (hours)
# **Issue reporter to provide a detailed description of the issue in the space below**
The PSDD example can be validated and generate IDL manually, however, the automatic tooling to do batch validations and generations fail on the model with the following message:
`Value 'org.eclipse.uml2.uml.internal.impl.ClassImpl@63aeac6d (name: PSDD_Listen, visibility: <unset>) (isLeaf: false, isAbstract: false, isFinalSpecialization: false) (isActive: false)' is not legal. (file:/home/user/workspace-papyrus/SNA_Core/PSDDPubSub/ModelFiles/PSDDPubSub.uml, -1, -1)`
| non_main | psdd example cannot have idl generated by cx extensions plugin issue and tracking information developer s time estimated effort to fix hours developer s actual time spent on fix hours issue reporter to provide a detailed description of the issue in the space below the psdd example can be validated and generate idl manually however the automatic tooling to do batch validations and generations fail on the model with the following message value org eclipse uml internal impl classimpl name psdd listen visibility isleaf false isabstract false isfinalspecialization false isactive false is not legal file home user workspace papyrus sna core psddpubsub modelfiles psddpubsub uml | 0 |
1,805 | 6,575,934,355 | IssuesEvent | 2017-09-11 17:53:35 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Feature Idea - add xattr support in FILE & STAT module. (chattr extended attributes in FS) | affects_2.2 feature_idea waiting_on_maintainer | ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
file
stat
##### ANSIBLE VERSION
ansible 2.2.0 (devel c9a5b1c555) last updated 2016/06/02 15:42:56 (GMT +200)
lib/ansible/modules/core: (detached HEAD ca4365b644) last updated 2016/06/02 15:43:14 (GMT +200)
lib/ansible/modules/extras: (detached HEAD b0aec50b9a) last updated 2016/06/02 15:43:15 (GMT +200)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
##### OS / ENVIRONMENT
LSB Version: :core-4.1-amd64:core-4.1-noarch
Distributor ID: RedHatEnterpriseServer
Description: Red Hat Enterprise Linux Server release 7.2 (Maipo)
Release: 7.2
Codename: Maipo
##### SUMMARY
It would be useful add xattr (extended attributes) native support in file and stat modules.
Currently, I have to use shell module (chattr and lsattr) to get xttr setting from files.
##### EXPECTED RESULTS
Implementation idea:
file module: Could be added an additional parameter named xattr or chattr.
stat modue: Could be added an addtional field named stat.xattr
| True | Feature Idea - add xattr support in FILE & STAT module. (chattr extended attributes in FS) - ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
file
stat
##### ANSIBLE VERSION
ansible 2.2.0 (devel c9a5b1c555) last updated 2016/06/02 15:42:56 (GMT +200)
lib/ansible/modules/core: (detached HEAD ca4365b644) last updated 2016/06/02 15:43:14 (GMT +200)
lib/ansible/modules/extras: (detached HEAD b0aec50b9a) last updated 2016/06/02 15:43:15 (GMT +200)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
##### OS / ENVIRONMENT
LSB Version: :core-4.1-amd64:core-4.1-noarch
Distributor ID: RedHatEnterpriseServer
Description: Red Hat Enterprise Linux Server release 7.2 (Maipo)
Release: 7.2
Codename: Maipo
##### SUMMARY
It would be useful add xattr (extended attributes) native support in file and stat modules.
Currently, I have to use shell module (chattr and lsattr) to get xttr setting from files.
##### EXPECTED RESULTS
Implementation idea:
file module: Could be added an additional parameter named xattr or chattr.
stat modue: Could be added an addtional field named stat.xattr
| main | feature idea add xattr support in file stat module chattr extended attributes in fs issue type feature idea component name file stat ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file etc ansible ansible cfg configured module search path default w o overrides os environment lsb version core core noarch distributor id redhatenterpriseserver description red hat enterprise linux server release maipo release codename maipo summary it would be useful add xattr extended attributes native support in file and stat modules currently i have to use shell module chattr and lsattr to get xttr setting from files expected results implementation idea file module could be added an additional parameter named xattr or chattr stat modue could be added an addtional field named stat xattr | 1 |
4,451 | 23,151,971,751 | IssuesEvent | 2022-07-29 09:12:27 | OpenRefine/OpenRefine | https://api.github.com/repos/OpenRefine/OpenRefine | closed | Java linting not enforced on source files | bug maintainability java | The linting of our Java source files seem to work only for our test files, not our main source files.
For instance, a lot of our source files have mixed tabs and spaces.
This seems to be due to the non-standard source directories we are using.
We should add those source directories in the POM files to make sure they are linted. This will require fixing the corresponding linting issues, and rebase / merge a lot of pull requests… Sorry about that, I should have checked that when I introduced the linting in the first place! | True | Java linting not enforced on source files - The linting of our Java source files seem to work only for our test files, not our main source files.
For instance, a lot of our source files have mixed tabs and spaces.
This seems to be due to the non-standard source directories we are using.
We should add those source directories in the POM files to make sure they are linted. This will require fixing the corresponding linting issues, and rebase / merge a lot of pull requests… Sorry about that, I should have checked that when I introduced the linting in the first place! | main | java linting not enforced on source files the linting of our java source files seem to work only for our test files not our main source files for instance a lot of our source files have mixed tabs and spaces this seems to be due to the non standard source directories we are using we should add those source directories in the pom files to make sure they are linted this will require fixing the corresponding linting issues and rebase merge a lot of pull requests… sorry about that i should have checked that when i introduced the linting in the first place | 1 |
5,090 | 26,006,280,933 | IssuesEvent | 2022-12-20 19:42:58 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Implement custom icons for column extracting and moving | type: enhancement work: frontend status: ready restricted: maintainers | Our Figma design specifies custom icons for these actions, but we don't have them yet.
| Figma | App |
| -- | -- |
|  |  |
| True | Implement custom icons for column extracting and moving - Our Figma design specifies custom icons for these actions, but we don't have them yet.
| Figma | App |
| -- | -- |
|  |  |
| main | implement custom icons for column extracting and moving our figma design specifies custom icons for these actions but we don t have them yet figma app | 1 |
33,960 | 7,312,656,995 | IssuesEvent | 2018-02-28 21:44:11 | jccastillo0007/eFacturaT | https://api.github.com/repos/jccastillo0007/eFacturaT | closed | Optibelt - corregir el status de la facturas de este año, con base a la fecha de emisión y condiciones de pago | bug defect | Seguimos atorados con optibelt.
Si bien aparentemente se corrigió el tema de la fecha de vencimiento... pero eso va de la mano con el status...
si la fecha aún no se alcanza, entonces es vigente...
si la fecha ya se alcanzó y no han pagado es vencida...
y cuando se paga su status es pagada...
ya están muy nerviosos con el tema, termina el 2o mes del año, y aún no puden tener estable la aplicación... | 1.0 | Optibelt - corregir el status de la facturas de este año, con base a la fecha de emisión y condiciones de pago - Seguimos atorados con optibelt.
Si bien aparentemente se corrigió el tema de la fecha de vencimiento... pero eso va de la mano con el status...
si la fecha aún no se alcanza, entonces es vigente...
si la fecha ya se alcanzó y no han pagado es vencida...
y cuando se paga su status es pagada...
ya están muy nerviosos con el tema, termina el 2o mes del año, y aún no puden tener estable la aplicación... | non_main | optibelt corregir el status de la facturas de este año con base a la fecha de emisión y condiciones de pago seguimos atorados con optibelt si bien aparentemente se corrigió el tema de la fecha de vencimiento pero eso va de la mano con el status si la fecha aún no se alcanza entonces es vigente si la fecha ya se alcanzó y no han pagado es vencida y cuando se paga su status es pagada ya están muy nerviosos con el tema termina el mes del año y aún no puden tener estable la aplicación | 0 |
3,700 | 15,099,385,021 | IssuesEvent | 2021-02-08 02:24:34 | pypiserver/pypiserver | https://api.github.com/repos/pypiserver/pypiserver | closed | Update CI to push to Docker Hub directly | type.Maintainance | In order to support multiple arch builds easily (see #364) and to isolate ourselves from the truly awful interface that is Docker Hub, I would like to see if we can update the CI to push there directly, rather than relying on webhooks and tag introspection by Docker Hub. The documentation [here](https://docs.docker.com/ci-cd/github-actions/#push-tagged-versions-to-docker-hub) certainly suggests that it should be possible. | True | Update CI to push to Docker Hub directly - In order to support multiple arch builds easily (see #364) and to isolate ourselves from the truly awful interface that is Docker Hub, I would like to see if we can update the CI to push there directly, rather than relying on webhooks and tag introspection by Docker Hub. The documentation [here](https://docs.docker.com/ci-cd/github-actions/#push-tagged-versions-to-docker-hub) certainly suggests that it should be possible. | main | update ci to push to docker hub directly in order to support multiple arch builds easily see and to isolate ourselves from the truly awful interface that is docker hub i would like to see if we can update the ci to push there directly rather than relying on webhooks and tag introspection by docker hub the documentation certainly suggests that it should be possible | 1 |
1,444 | 6,265,849,430 | IssuesEvent | 2017-07-16 20:51:45 | enterprisemediawiki/meza | https://api.github.com/repos/enterprisemediawiki/meza | opened | Use role remote-mysqldump for backups and replication | critical: stability difficulty: hard important: maintainability important: performance | Currently roles [dump-db-wikis](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/dump-db-wikis/tasks/main.yml#L20) and [backup-db-wikis](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/backup-db-wikis/tasks/main.yml#L29) dump SQL files on master then move them through the controller to their final location. Likewise, when setting up replication [db-master dumps SQL then the file is moved to controller and then to replica](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/database/tasks/replication.yml#L91). Instead use the role `remote-mysqldump` to dump these directly. Also setup tests of replica server.
- [ ] Use remote-mysqldump during backups
- [ ] Use remote-mysqldump during replication setup
- [ ] Setup tests of replication in Docker (Travis) | True | Use role remote-mysqldump for backups and replication - Currently roles [dump-db-wikis](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/dump-db-wikis/tasks/main.yml#L20) and [backup-db-wikis](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/backup-db-wikis/tasks/main.yml#L29) dump SQL files on master then move them through the controller to their final location. Likewise, when setting up replication [db-master dumps SQL then the file is moved to controller and then to replica](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/database/tasks/replication.yml#L91). Instead use the role `remote-mysqldump` to dump these directly. Also setup tests of replica server.
- [ ] Use remote-mysqldump during backups
- [ ] Use remote-mysqldump during replication setup
- [ ] Setup tests of replication in Docker (Travis) | main | use role remote mysqldump for backups and replication currently roles and dump sql files on master then move them through the controller to their final location likewise when setting up replication instead use the role remote mysqldump to dump these directly also setup tests of replica server use remote mysqldump during backups use remote mysqldump during replication setup setup tests of replication in docker travis | 1 |
66,161 | 8,883,374,734 | IssuesEvent | 2019-01-14 15:33:02 | LycheeOrg/Lychee | https://api.github.com/repos/LycheeOrg/Lychee | opened | Doc: add Couldron link to wiki | Documentation Low Priority | ### Detailed description of the problem
from #164 :
> We are trying to provide an update for the Lychee Cloudron app. The current version is 3.1.6 and the new version will be 3.2.8-fixed. Since this is for Cloudron, the built-in updater is disabled, since on Cloudron apps run on a read-only filesystem. For reference, our app packaging code can be found at https://git.cloudron.io/cloudron/lychee-app
Maybe adding a pointer to this in the Wiki would be a nice idea. :) | 1.0 | Doc: add Couldron link to wiki - ### Detailed description of the problem
from #164 :
> We are trying to provide an update for the Lychee Cloudron app. The current version is 3.1.6 and the new version will be 3.2.8-fixed. Since this is for Cloudron, the built-in updater is disabled, since on Cloudron apps run on a read-only filesystem. For reference, our app packaging code can be found at https://git.cloudron.io/cloudron/lychee-app
Maybe adding a pointer to this in the Wiki would be a nice idea. :) | non_main | doc add couldron link to wiki detailed description of the problem from we are trying to provide an update for the lychee cloudron app the current version is and the new version will be fixed since this is for cloudron the built in updater is disabled since on cloudron apps run on a read only filesystem for reference our app packaging code can be found at maybe adding a pointer to this in the wiki would be a nice idea | 0 |
1,049 | 4,861,978,851 | IssuesEvent | 2016-11-14 10:44:04 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | Blockinfile overwrites symlink with file if changed | affects_2.1 bug_report waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
blockinfile
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file = /Users/vince/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
ansible.cfg mentioned above:
```
[defaults]
roles_path = ./roles:/etc/ansible/roles
forks=25
retry_files_save_path = ./.retry/
```
##### OS / ENVIRONMENT
I'm running ansible on OSX El Capitain (10.11.6) and managing localhost
I'm also running ansible on debian jessie with ansible 2.2.0 and the same config (as localhost)
##### SUMMARY
blockinfile overwrites the file with the new version if the old version does not have the block.
The problem arises when the file is a symlink. The symlink is overwritten with the file version of the linked file with the new block added. (happens on both OSs/ansible versions)
The links remains in existence if the content does not need to be not changed
##### STEPS TO REPRODUCE
To test:
put the following in `~/test_file`:
```
this is a regular file
```
then symlink `~/test_file` as `~/test_link` like so:
```
$ ln -s ~/test_file ~/test_link
```
Then run the following playbook
``` yaml
- hosts: localhost
become: no
tasks:
- blockinfile:
block: |
hostsum="$(hostname | cksum | awk '{print $1}')";
export hostnum="$((hostsum%253))";
export hostcolor="$(tput setaf $hostnum)"
dest: ~/test_link
marker: "# hostcolor {mark} ANSIBLE MANAGED BLOCK"
```
##### EXPECTED RESULTS
note the filetype `l` of the link and the filesize of `228` of the linked file
```
$ ls -laF ~/test_link ~/test_file
-rw-r--r-- 1 vince vince 228 Sep 28 10:37 /home/vince/test_file
lrwxrwxrwx 1 vince vince 21 Sep 28 10:32 /home/vince/test_link -> /home/vince/test_file
```
I also expect both files to have the same content:
```
$ diff -s ~/test_file ~/test_link
Files /home/vince/test_file and /home/vince/test_link are identical
```
##### ACTUAL RESULTS
```
$ ls -laF ~/test_link ~/test_file
-rw-r--r-- 1 vince vince 23 Sep 28 10:39 /home/vince/test_file
-rw-r--r-- 1 vince vince 228 Sep 28 10:39 /home/vince/test_link
```
```
$ diff -s ~/test_file ~/test_link
1a2,6
> # hostcolor BEGIN ANSIBLE MANAGED BLOCK
> hostsum="$(hostname | cksum | awk '{print $1}')";
> export hostnum="$((hostsum%253))";
> export hostcolor="$(tput setaf $hostnum)"
> # hostcolor END ANSIBLE MANAGED `BLOCK`
```
If I can find the time I'll do some additional research, but this should be enough to get started 😉
| True | Blockinfile overwrites symlink with file if changed - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
blockinfile
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file = /Users/vince/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
ansible.cfg mentioned above:
```
[defaults]
roles_path = ./roles:/etc/ansible/roles
forks=25
retry_files_save_path = ./.retry/
```
##### OS / ENVIRONMENT
I'm running ansible on OSX El Capitain (10.11.6) and managing localhost
I'm also running ansible on debian jessie with ansible 2.2.0 and the same config (as localhost)
##### SUMMARY
blockinfile overwrites the file with the new version if the old version does not have the block.
The problem arises when the file is a symlink. The symlink is overwritten with the file version of the linked file with the new block added. (happens on both OSs/ansible versions)
The links remains in existence if the content does not need to be not changed
##### STEPS TO REPRODUCE
To test:
put the following in `~/test_file`:
```
this is a regular file
```
then symlink `~/test_file` as `~/test_link` like so:
```
$ ln -s ~/test_file ~/test_link
```
Then run the following playbook
``` yaml
- hosts: localhost
become: no
tasks:
- blockinfile:
block: |
hostsum="$(hostname | cksum | awk '{print $1}')";
export hostnum="$((hostsum%253))";
export hostcolor="$(tput setaf $hostnum)"
dest: ~/test_link
marker: "# hostcolor {mark} ANSIBLE MANAGED BLOCK"
```
##### EXPECTED RESULTS
note the filetype `l` of the link and the filesize of `228` of the linked file
```
$ ls -laF ~/test_link ~/test_file
-rw-r--r-- 1 vince vince 228 Sep 28 10:37 /home/vince/test_file
lrwxrwxrwx 1 vince vince 21 Sep 28 10:32 /home/vince/test_link -> /home/vince/test_file
```
I also expect both files to have the same content:
```
$ diff -s ~/test_file ~/test_link
Files /home/vince/test_file and /home/vince/test_link are identical
```
##### ACTUAL RESULTS
```
$ ls -laF ~/test_link ~/test_file
-rw-r--r-- 1 vince vince 23 Sep 28 10:39 /home/vince/test_file
-rw-r--r-- 1 vince vince 228 Sep 28 10:39 /home/vince/test_link
```
```
$ diff -s ~/test_file ~/test_link
1a2,6
> # hostcolor BEGIN ANSIBLE MANAGED BLOCK
> hostsum="$(hostname | cksum | awk '{print $1}')";
> export hostnum="$((hostsum%253))";
> export hostcolor="$(tput setaf $hostnum)"
> # hostcolor END ANSIBLE MANAGED `BLOCK`
```
If I can find the time I'll do some additional research, but this should be enough to get started 😉
| main | blockinfile overwrites symlink with file if changed issue type bug report component name blockinfile ansible version ansible config file users vince ansible ansible cfg configured module search path default w o overrides configuration ansible cfg mentioned above roles path roles etc ansible roles forks retry files save path retry os environment i m running ansible on osx el capitain and managing localhost i m also running ansible on debian jessie with ansible and the same config as localhost summary blockinfile overwrites the file with the new version if the old version does not have the block the problem arises when the file is a symlink the symlink is overwritten with the file version of the linked file with the new block added happens on both oss ansible versions the links remains in existence if the content does not need to be not changed steps to reproduce to test put the following in test file this is a regular file then symlink test file as test link like so ln s test file test link then run the following playbook yaml hosts localhost become no tasks blockinfile block hostsum hostname cksum awk print export hostnum hostsum export hostcolor tput setaf hostnum dest test link marker hostcolor mark ansible managed block expected results note the filetype l of the link and the filesize of of the linked file ls laf test link test file rw r r vince vince sep home vince test file lrwxrwxrwx vince vince sep home vince test link home vince test file i also expect both files to have the same content diff s test file test link files home vince test file and home vince test link are identical actual results ls laf test link test file rw r r vince vince sep home vince test file rw r r vince vince sep home vince test link diff s test file test link hostcolor begin ansible managed block hostsum hostname cksum awk print export hostnum hostsum export hostcolor tput setaf hostnum hostcolor end ansible managed block if i can find the time i ll do some additional research but this should be enough to get started 😉 | 1 |
358,471 | 25,193,184,021 | IssuesEvent | 2022-11-12 06:42:27 | crushten/github_workflow_repo | https://api.github.com/repos/crushten/github_workflow_repo | closed | Centralized the workflows | documentation | Would like to centralized all workflows here. Will need to make some sort of structure and then can use that in other projects.
https://docs.github.com/en/actions/using-workflows/reusing-workflows
https://github.blog/changelog/2020-06-23-github-actions-workflow-templates/
https://github.blog/2022-02-10-using-reusable-workflows-github-actions/
https://github.blog/changelog/2021-08-25-github-actions-reduce-duplication-with-action-composition/
https://docs.github.com/en/actions/creating-actions/creating-a-composite-action
https://betterprogramming.pub/how-to-use-github-actions-reusable-workflow-8604e8cbf258
https://dev.to/n3wt0n/avoid-duplication-github-actions-reusable-workflows-3ae8 | 1.0 | Centralized the workflows - Would like to centralized all workflows here. Will need to make some sort of structure and then can use that in other projects.
https://docs.github.com/en/actions/using-workflows/reusing-workflows
https://github.blog/changelog/2020-06-23-github-actions-workflow-templates/
https://github.blog/2022-02-10-using-reusable-workflows-github-actions/
https://github.blog/changelog/2021-08-25-github-actions-reduce-duplication-with-action-composition/
https://docs.github.com/en/actions/creating-actions/creating-a-composite-action
https://betterprogramming.pub/how-to-use-github-actions-reusable-workflow-8604e8cbf258
https://dev.to/n3wt0n/avoid-duplication-github-actions-reusable-workflows-3ae8 | non_main | centralized the workflows would like to centralized all workflows here will need to make some sort of structure and then can use that in other projects | 0 |
3,945 | 17,793,408,091 | IssuesEvent | 2021-08-31 18:59:14 | tgstation/tgstation-server | https://api.github.com/repos/tgstation/tgstation-server | opened | Separate deployment jobs from CI suite and trigger via workflow_dispatch | CI/CD Maintainability Issue | So when we desperately need a build we can force it out
IDEALLY THOUGH: we'd fix the spurious failures in regular CI and re-parallelize it | True | Separate deployment jobs from CI suite and trigger via workflow_dispatch - So when we desperately need a build we can force it out
IDEALLY THOUGH: we'd fix the spurious failures in regular CI and re-parallelize it | main | separate deployment jobs from ci suite and trigger via workflow dispatch so when we desperately need a build we can force it out ideally though we d fix the spurious failures in regular ci and re parallelize it | 1 |
3,070 | 11,584,438,428 | IssuesEvent | 2020-02-22 17:20:43 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | opened | Conditional expressions (? : ) should be short | Area: analyzer Area: maintainability feature | The `Condition`, `WhenTrue` and `WhenFalse` parts of a conditional expression should be short, they should not exceed 20 characters.
The reason is: The longer the parts are, the harder the conditional expression is to read. | True | Conditional expressions (? : ) should be short - The `Condition`, `WhenTrue` and `WhenFalse` parts of a conditional expression should be short, they should not exceed 20 characters.
The reason is: The longer the parts are, the harder the conditional expression is to read. | main | conditional expressions should be short the condition whentrue and whenfalse parts of a conditional expression should be short they should not exceed characters the reason is the longer the parts are the harder the conditional expression is to read | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.