Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
1,519
6,572,207,154
IssuesEvent
2017-09-11 00:02:27
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
s3_bucket fails when loading JSON policy from a template
affects_2.1 aws bug_report cloud in progress waiting_on_maintainer
##### Issue Type: - Bug Report ##### Plugin Name: s3_bucket.py ##### Ansible Version: ``` ansible 2.1.0 config file = configured module search path = Default w/o overrides ``` ##### Ansible Configuration: N/A ##### Environment: N/A ##### Summary: Loading an S3 bucket policy from a file results in failure due to various silent conversions performed by the lookup function, ansible core, and the s3_bucket function itself. ##### Steps To Reproduce: For bugs, please show exactly how to reproduce the problem. For new features, show how the feature would be used. Playbook: ``` vars: domain: "example.com" envname: "stage" aws_account_number: "12345678" region: "us-west-2" s3_bucket_name: "{{domain | regex_replace('\\.', '-')}}-{{envname}}-{{aws_account_number}}" elb_principal_mappings: us-east-1: 127311923021 us-west-2: 797873946194 us-west-1: 027434742980 eu-west-1: 156460612806 eu-central-1: 054676820928 ap-southeast-1: 114774131450 ap-northeast-1: 582318560864 ap-southeast-2: 783225319266 ap-northeast-2: 600734575887 sa-east-1: 507241528517 - name: Create S3 asset bucket s3_bucket: name: "{{ s3_bucket_name }}" region: "{{ region }}" policy: "{{ lookup('template', './s3_bucket_policy.json.j2', convert_data=False) }}" register: site_s3_bucket ``` Template (Could be anything but just for completeness): ``` { "Id": "AllowELBWriteAccess", "Version": "2012-10-17", "Statement": [ { "Sid": "Stmt1454671534294", "Action": [ "s3:PutObject" ], "Effect": "Allow", "Resource": "arn:aws:s3:::{{domain | regex_replace('\.', '-')}}-{{envname}}-{{aws_account_number}}/accesslogs/AWSLogs/{{ aws_account_number }}/*", "Principal": { "AWS": [ "{{ elb_principal_mappings[region] }}" ] } } ] } ``` ##### Expected Results: An S3 bucket is created ##### Actual Results: Failure incorrectly implying that that the JSON is invalid. ``` TASK [Create S3 asset bucket] ************************************************** fatal: [127.0.0.1]: FAILED! => {"changed": false, "failed": true, "msg": "Policies must be valid JSON and the first byte must be '{'"} ```
True
s3_bucket fails when loading JSON policy from a template - ##### Issue Type: - Bug Report ##### Plugin Name: s3_bucket.py ##### Ansible Version: ``` ansible 2.1.0 config file = configured module search path = Default w/o overrides ``` ##### Ansible Configuration: N/A ##### Environment: N/A ##### Summary: Loading an S3 bucket policy from a file results in failure due to various silent conversions performed by the lookup function, ansible core, and the s3_bucket function itself. ##### Steps To Reproduce: For bugs, please show exactly how to reproduce the problem. For new features, show how the feature would be used. Playbook: ``` vars: domain: "example.com" envname: "stage" aws_account_number: "12345678" region: "us-west-2" s3_bucket_name: "{{domain | regex_replace('\\.', '-')}}-{{envname}}-{{aws_account_number}}" elb_principal_mappings: us-east-1: 127311923021 us-west-2: 797873946194 us-west-1: 027434742980 eu-west-1: 156460612806 eu-central-1: 054676820928 ap-southeast-1: 114774131450 ap-northeast-1: 582318560864 ap-southeast-2: 783225319266 ap-northeast-2: 600734575887 sa-east-1: 507241528517 - name: Create S3 asset bucket s3_bucket: name: "{{ s3_bucket_name }}" region: "{{ region }}" policy: "{{ lookup('template', './s3_bucket_policy.json.j2', convert_data=False) }}" register: site_s3_bucket ``` Template (Could be anything but just for completeness): ``` { "Id": "AllowELBWriteAccess", "Version": "2012-10-17", "Statement": [ { "Sid": "Stmt1454671534294", "Action": [ "s3:PutObject" ], "Effect": "Allow", "Resource": "arn:aws:s3:::{{domain | regex_replace('\.', '-')}}-{{envname}}-{{aws_account_number}}/accesslogs/AWSLogs/{{ aws_account_number }}/*", "Principal": { "AWS": [ "{{ elb_principal_mappings[region] }}" ] } } ] } ``` ##### Expected Results: An S3 bucket is created ##### Actual Results: Failure incorrectly implying that that the JSON is invalid. ``` TASK [Create S3 asset bucket] ************************************************** fatal: [127.0.0.1]: FAILED! => {"changed": false, "failed": true, "msg": "Policies must be valid JSON and the first byte must be '{'"} ```
main
bucket fails when loading json policy from a template issue type bug report plugin name bucket py ansible version ansible config file configured module search path default w o overrides ansible configuration n a environment n a summary loading an bucket policy from a file results in failure due to various silent conversions performed by the lookup function ansible core and the bucket function itself steps to reproduce for bugs please show exactly how to reproduce the problem for new features show how the feature would be used playbook vars domain example com envname stage aws account number region us west bucket name domain regex replace envname aws account number elb principal mappings us east us west us west eu west eu central ap southeast ap northeast ap southeast ap northeast sa east name create asset bucket bucket name bucket name region region policy lookup template bucket policy json convert data false register site bucket template could be anything but just for completeness id allowelbwriteaccess version statement sid action putobject effect allow resource arn aws domain regex replace envname aws account number accesslogs awslogs aws account number principal aws elb principal mappings expected results an bucket is created actual results failure incorrectly implying that that the json is invalid task fatal failed changed false failed true msg policies must be valid json and the first byte must be
1
262,244
19,768,800,754
IssuesEvent
2022-01-17 07:42:39
kubernetes-sigs/descheduler
https://api.github.com/repos/kubernetes-sigs/descheduler
closed
Docs around autohealing are misleading
lifecycle/rotten kind/documentation
The [docs around autohealing](https://github.com/kubernetes-sigs/descheduler/blob/master/docs/user-guide.md#autoheal-node-problems) are a bit misleading in my opinion. They link off to Node Problem Detector, claiming that `Node Problem Detector can detect specific Node problems and taint any Nodes which have those problems.`. In fact, NPD doesn't do any tainting. It's the `TaintNodeByCondition` feature of the node controller that takes _some_ conditions and turns them in to taints. However this only works for the default node conditions: `PIDPressure`, `MemoryPressure`, `DiskPressure`, `Ready`, and some cloud provider specific conditions. There is an [open PR](https://github.com/kubernetes/node-problem-detector/pull/565) on NPD that wants to add this tainting behaviour, but the maintainers seem to think it shouldn't be NPD that does the tainting. The effect is that the autoheal cycle describe doesn't actually work, at least not for custom conditions. It would be wonderful if it did, because it's quite a compelling outcome, and it would be amazing if it were offered exclusively in terms of Kubernetes first party tooling. At the very least, we should change the wording in the docs to make it clear that NPD doesn't really participate in the autohealing. At most, I'm hoping that by raising this issue, the fact that this cycle doesn't currently work as intended can get a little more visibility. My guess at possible solutions: * Merge the PR linked above, so that NPD creates taints * Extend the node controller to add condition based taints for _all_ conditions, including custom ones * Create a new project to convert conditions in to taints * Add a new strategy in this repo that allows for descheduling based on Conditions
1.0
Docs around autohealing are misleading - The [docs around autohealing](https://github.com/kubernetes-sigs/descheduler/blob/master/docs/user-guide.md#autoheal-node-problems) are a bit misleading in my opinion. They link off to Node Problem Detector, claiming that `Node Problem Detector can detect specific Node problems and taint any Nodes which have those problems.`. In fact, NPD doesn't do any tainting. It's the `TaintNodeByCondition` feature of the node controller that takes _some_ conditions and turns them in to taints. However this only works for the default node conditions: `PIDPressure`, `MemoryPressure`, `DiskPressure`, `Ready`, and some cloud provider specific conditions. There is an [open PR](https://github.com/kubernetes/node-problem-detector/pull/565) on NPD that wants to add this tainting behaviour, but the maintainers seem to think it shouldn't be NPD that does the tainting. The effect is that the autoheal cycle describe doesn't actually work, at least not for custom conditions. It would be wonderful if it did, because it's quite a compelling outcome, and it would be amazing if it were offered exclusively in terms of Kubernetes first party tooling. At the very least, we should change the wording in the docs to make it clear that NPD doesn't really participate in the autohealing. At most, I'm hoping that by raising this issue, the fact that this cycle doesn't currently work as intended can get a little more visibility. My guess at possible solutions: * Merge the PR linked above, so that NPD creates taints * Extend the node controller to add condition based taints for _all_ conditions, including custom ones * Create a new project to convert conditions in to taints * Add a new strategy in this repo that allows for descheduling based on Conditions
non_main
docs around autohealing are misleading the are a bit misleading in my opinion they link off to node problem detector claiming that node problem detector can detect specific node problems and taint any nodes which have those problems in fact npd doesn t do any tainting it s the taintnodebycondition feature of the node controller that takes some conditions and turns them in to taints however this only works for the default node conditions pidpressure memorypressure diskpressure ready and some cloud provider specific conditions there is an on npd that wants to add this tainting behaviour but the maintainers seem to think it shouldn t be npd that does the tainting the effect is that the autoheal cycle describe doesn t actually work at least not for custom conditions it would be wonderful if it did because it s quite a compelling outcome and it would be amazing if it were offered exclusively in terms of kubernetes first party tooling at the very least we should change the wording in the docs to make it clear that npd doesn t really participate in the autohealing at most i m hoping that by raising this issue the fact that this cycle doesn t currently work as intended can get a little more visibility my guess at possible solutions merge the pr linked above so that npd creates taints extend the node controller to add condition based taints for all conditions including custom ones create a new project to convert conditions in to taints add a new strategy in this repo that allows for descheduling based on conditions
0
3,662
14,942,823,689
IssuesEvent
2021-01-25 21:55:46
flag-camp-2020-t3/Spare4Fun-Server
https://api.github.com/repos/flag-camp-2020-t3/Spare4Fun-Server
opened
Dao Throw Exception & captured by exception handler
backend-v2.0 critical enhancement maintain / refactor
branch: ```refactor-dao-throw-exception-exception-handler```
True
Dao Throw Exception & captured by exception handler - branch: ```refactor-dao-throw-exception-exception-handler```
main
dao throw exception captured by exception handler branch refactor dao throw exception exception handler
1
4,390
22,497,682,559
IssuesEvent
2022-06-23 09:00:59
zoj613/polyagamma
https://api.github.com/repos/zoj613/polyagamma
closed
MAINT: Get full precision for 32 bit floating point random values.
good first issue maintainance random_polyagamma
A numpy [issue](https://github.com/numpy/numpy/issues/17478) showed that numpy's formula for generating 32bit random uniform values was waiting bits. There is a PR that fixes this which was merge in https://github.com/numpy/numpy/pull/20314. It states ``` The formula to convert a 32 bit random integer to a random float32, (next_uint32(bitgen_state) >> 9) * (1.0f / 8388608.0f) shifts by one bit too many, resulting in uniform float32 samples always having a 0 in the least significant bit. The formula is corrected to (next_uint32(bitgen_state) >> 8) * (1.0f / 16777216.0f) Occurrences of the incorrect formula in numpy/random/tests/test_direct.py were also corrected. ``` Since the project uses this formula at https://github.com/zoj613/polyagamma/blob/fca4a8bce462066803c888708e2f109d1991c00c/src/pgm_macros.h#L54-L56 it is worth updating to get the full precision of 32bit float because they are used extensively in the accept-rejection steps of the samplers. Im positive this slight adjustment does not affect the results nor the performance but I think it's worth updating the formula.
True
MAINT: Get full precision for 32 bit floating point random values. - A numpy [issue](https://github.com/numpy/numpy/issues/17478) showed that numpy's formula for generating 32bit random uniform values was waiting bits. There is a PR that fixes this which was merge in https://github.com/numpy/numpy/pull/20314. It states ``` The formula to convert a 32 bit random integer to a random float32, (next_uint32(bitgen_state) >> 9) * (1.0f / 8388608.0f) shifts by one bit too many, resulting in uniform float32 samples always having a 0 in the least significant bit. The formula is corrected to (next_uint32(bitgen_state) >> 8) * (1.0f / 16777216.0f) Occurrences of the incorrect formula in numpy/random/tests/test_direct.py were also corrected. ``` Since the project uses this formula at https://github.com/zoj613/polyagamma/blob/fca4a8bce462066803c888708e2f109d1991c00c/src/pgm_macros.h#L54-L56 it is worth updating to get the full precision of 32bit float because they are used extensively in the accept-rejection steps of the samplers. Im positive this slight adjustment does not affect the results nor the performance but I think it's worth updating the formula.
main
maint get full precision for bit floating point random values a numpy showed that numpy s formula for generating random uniform values was waiting bits there is a pr that fixes this which was merge in it states the formula to convert a bit random integer to a random next bitgen state shifts by one bit too many resulting in uniform samples always having a in the least significant bit the formula is corrected to next bitgen state occurrences of the incorrect formula in numpy random tests test direct py were also corrected since the project uses this formula at it is worth updating to get the full precision of float because they are used extensively in the accept rejection steps of the samplers im positive this slight adjustment does not affect the results nor the performance but i think it s worth updating the formula
1
5,214
26,464,344,104
IssuesEvent
2023-01-16 21:18:37
bazelbuild/intellij
https://api.github.com/repos/bazelbuild/intellij
closed
Flag --incompatible_disable_starlark_host_transitions will break Android Studio Plugin in Bazel 7.0
type: bug product: Android Studio topic: bazel awaiting-maintainer
Incompatible flag `--incompatible_disable_starlark_host_transitions` will be enabled by default in the next major release (Bazel 7.0), thus breaking Android Studio Plugin. Please migrate to fix this and unblock the flip of this flag. The flag is documented here: [bazelbuild/bazel#17032](https://github.com/bazelbuild/bazel/issues/17032). Please check the following CI builds for build and test results: - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc33-4a0d-bfba-a9d0f36f4e1c) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc36-44c7-9a05-b49d4a1c32f5) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc2c-44c5-ba2f-d20d0c3515e1) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc3c-4a48-a7bc-5c475690a11b) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc39-4303-8caa-95e9935cab6d) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc3f-4bda-9815-991ef8b9d7ad) Never heard of incompatible flags before? We have [documentation](https://docs.bazel.build/versions/master/backward-compatibility.html) that explains everything. If you have any questions, please file an issue in https://github.com/bazelbuild/continuous-integration.
True
Flag --incompatible_disable_starlark_host_transitions will break Android Studio Plugin in Bazel 7.0 - Incompatible flag `--incompatible_disable_starlark_host_transitions` will be enabled by default in the next major release (Bazel 7.0), thus breaking Android Studio Plugin. Please migrate to fix this and unblock the flip of this flag. The flag is documented here: [bazelbuild/bazel#17032](https://github.com/bazelbuild/bazel/issues/17032). Please check the following CI builds for build and test results: - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc33-4a0d-bfba-a9d0f36f4e1c) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc36-44c7-9a05-b49d4a1c32f5) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc2c-44c5-ba2f-d20d0c3515e1) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc3c-4a48-a7bc-5c475690a11b) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc39-4303-8caa-95e9935cab6d) - [Ubuntu 18.04 OpenJDK 11](https://buildkite.com/bazel/bazelisk-plus-incompatible-flags/builds/1365#0185154a-dc3f-4bda-9815-991ef8b9d7ad) Never heard of incompatible flags before? We have [documentation](https://docs.bazel.build/versions/master/backward-compatibility.html) that explains everything. If you have any questions, please file an issue in https://github.com/bazelbuild/continuous-integration.
main
flag incompatible disable starlark host transitions will break android studio plugin in bazel incompatible flag incompatible disable starlark host transitions will be enabled by default in the next major release bazel thus breaking android studio plugin please migrate to fix this and unblock the flip of this flag the flag is documented here please check the following ci builds for build and test results never heard of incompatible flags before we have that explains everything if you have any questions please file an issue in
1
507,049
14,679,153,479
IssuesEvent
2020-12-31 06:04:31
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
closed
Remove constraints from ScrollView in profile_reset_pin_activity
Priority: Nice-to-have Type: Improvement good first issue
Remove constraints from ScrollView in profile_reset_pin_activity both landscape and portrait xml files. https://github.com/oppia/oppia-android/blob/2511715a4770cda65df593cf6821b17b7f8f3d28/app/src/main/res/layout/profile_reset_pin_activity.xml#L55
1.0
Remove constraints from ScrollView in profile_reset_pin_activity - Remove constraints from ScrollView in profile_reset_pin_activity both landscape and portrait xml files. https://github.com/oppia/oppia-android/blob/2511715a4770cda65df593cf6821b17b7f8f3d28/app/src/main/res/layout/profile_reset_pin_activity.xml#L55
non_main
remove constraints from scrollview in profile reset pin activity remove constraints from scrollview in profile reset pin activity both landscape and portrait xml files
0
46,944
24,794,622,649
IssuesEvent
2022-10-24 16:10:55
iree-org/iree
https://api.github.com/repos/iree-org/iree
closed
Add `hal_inline` dialect/module for tiny environments.
runtime performance ⚡
For environments where the execution model is known to be exclusively local and inline (embedded systems) we can have a paired down HAL that pretty much only contains executable support. The idea is to still use HAL executable translation in the compiler but lowering the stream dialect to a new lightweight dialect that pretty much only manages executables and dispatches. Most of the local/ implementation of the executable loader and the loaders themselves have no dependencies on command buffers, allocators, buffers, or devices and can be cleanly pulled into the module without bringing in the bulk of the HAL API. It's debatable whether the allocator/buffer stuff should be included - that would allow the coming allocator types to be reused but at the cost of additional non-user-controllable overheads. Since the local executables take byte spans all buffers could just be `iree_vm_buffer_t` which is already compiled in and available for use - and since they take a custom `iree_allocator_t` it's still possible for hosting applications to manage memory however they want. Executables themselves will still be injected on the module when created same as today, allowing for dynamic, static, vmvx, etc executables to be run this way. This allows us to separate the execution model from the deployment model at the cost of a few vtables. Outline: * [x] Add `hal_inline` dialect with basic ops: * [x] `hal_inline.executable.create` * [x] `hal_inline.executable.dispatch` * [x] `hal_inline.executable_layout.create`? (still need this to reuse loaders/libraries) * [x] Add `--execution-mode=` iree-compile flag to switch between `hal-async` and `hal-inline` (or w/e) * [x] Have a new `iree-hal-inline-transformation-pipeline` that still performs interface materialization and executable translation but otherwise lowers `stream` itself * [x] Add `iree/modules/hal_inline` runtime module that links directly against the `iree/hal/local/` libraries * [x] Build a runner tool that uses the inline module (or make iree-run-module/etc always support it with a flag) (could also call this the `inline` dialect or something - it's still technically a HAL though as the executables being called are abstracted across hardware - can have CPU/FPGA/DSP/etc)
True
Add `hal_inline` dialect/module for tiny environments. - For environments where the execution model is known to be exclusively local and inline (embedded systems) we can have a paired down HAL that pretty much only contains executable support. The idea is to still use HAL executable translation in the compiler but lowering the stream dialect to a new lightweight dialect that pretty much only manages executables and dispatches. Most of the local/ implementation of the executable loader and the loaders themselves have no dependencies on command buffers, allocators, buffers, or devices and can be cleanly pulled into the module without bringing in the bulk of the HAL API. It's debatable whether the allocator/buffer stuff should be included - that would allow the coming allocator types to be reused but at the cost of additional non-user-controllable overheads. Since the local executables take byte spans all buffers could just be `iree_vm_buffer_t` which is already compiled in and available for use - and since they take a custom `iree_allocator_t` it's still possible for hosting applications to manage memory however they want. Executables themselves will still be injected on the module when created same as today, allowing for dynamic, static, vmvx, etc executables to be run this way. This allows us to separate the execution model from the deployment model at the cost of a few vtables. Outline: * [x] Add `hal_inline` dialect with basic ops: * [x] `hal_inline.executable.create` * [x] `hal_inline.executable.dispatch` * [x] `hal_inline.executable_layout.create`? (still need this to reuse loaders/libraries) * [x] Add `--execution-mode=` iree-compile flag to switch between `hal-async` and `hal-inline` (or w/e) * [x] Have a new `iree-hal-inline-transformation-pipeline` that still performs interface materialization and executable translation but otherwise lowers `stream` itself * [x] Add `iree/modules/hal_inline` runtime module that links directly against the `iree/hal/local/` libraries * [x] Build a runner tool that uses the inline module (or make iree-run-module/etc always support it with a flag) (could also call this the `inline` dialect or something - it's still technically a HAL though as the executables being called are abstracted across hardware - can have CPU/FPGA/DSP/etc)
non_main
add hal inline dialect module for tiny environments for environments where the execution model is known to be exclusively local and inline embedded systems we can have a paired down hal that pretty much only contains executable support the idea is to still use hal executable translation in the compiler but lowering the stream dialect to a new lightweight dialect that pretty much only manages executables and dispatches most of the local implementation of the executable loader and the loaders themselves have no dependencies on command buffers allocators buffers or devices and can be cleanly pulled into the module without bringing in the bulk of the hal api it s debatable whether the allocator buffer stuff should be included that would allow the coming allocator types to be reused but at the cost of additional non user controllable overheads since the local executables take byte spans all buffers could just be iree vm buffer t which is already compiled in and available for use and since they take a custom iree allocator t it s still possible for hosting applications to manage memory however they want executables themselves will still be injected on the module when created same as today allowing for dynamic static vmvx etc executables to be run this way this allows us to separate the execution model from the deployment model at the cost of a few vtables outline add hal inline dialect with basic ops hal inline executable create hal inline executable dispatch hal inline executable layout create still need this to reuse loaders libraries add execution mode iree compile flag to switch between hal async and hal inline or w e have a new iree hal inline transformation pipeline that still performs interface materialization and executable translation but otherwise lowers stream itself add iree modules hal inline runtime module that links directly against the iree hal local libraries build a runner tool that uses the inline module or make iree run module etc always support it with a flag could also call this the inline dialect or something it s still technically a hal though as the executables being called are abstracted across hardware can have cpu fpga dsp etc
0
14,714
3,419,369,800
IssuesEvent
2015-12-08 09:26:03
centreon/centreon
https://api.github.com/repos/centreon/centreon
closed
Delete a poller doesn't delete associated Centreon Broker configuration
BetaTest Kind/Bug Status/Solved
When I delete a poller the associated Centreon Broker configuration is disabled but not deleted. to have a clean configuration it would be good to remove the entire configuration of Centreon Broker Regards,
1.0
Delete a poller doesn't delete associated Centreon Broker configuration - When I delete a poller the associated Centreon Broker configuration is disabled but not deleted. to have a clean configuration it would be good to remove the entire configuration of Centreon Broker Regards,
non_main
delete a poller doesn t delete associated centreon broker configuration when i delete a poller the associated centreon broker configuration is disabled but not deleted to have a clean configuration it would be good to remove the entire configuration of centreon broker regards
0
239,613
7,799,878,792
IssuesEvent
2018-06-09 01:34:37
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
closed
0005764: convert $_folder->cache_uidvalidity to integer (in DB)
Felamimail Mantis high priority
**Reported by pschuele on 20 Feb 2012 09:00** convert $_folder->cache_uidvalidity to integer (in DB) - as imap_uidvalidity is an integer, too and there are issues regarding comparison between the two when using postgresql **Additional information:** http://www.tine20.org/forum/viewtopic.php?f=10&t=10508
1.0
0005764: convert $_folder->cache_uidvalidity to integer (in DB) - **Reported by pschuele on 20 Feb 2012 09:00** convert $_folder->cache_uidvalidity to integer (in DB) - as imap_uidvalidity is an integer, too and there are issues regarding comparison between the two when using postgresql **Additional information:** http://www.tine20.org/forum/viewtopic.php?f=10&t=10508
non_main
convert folder cache uidvalidity to integer in db reported by pschuele on feb convert folder gt cache uidvalidity to integer in db as imap uidvalidity is an integer too and there are issues regarding comparison between the two when using postgresql additional information
0
4,752
24,509,600,788
IssuesEvent
2022-10-10 19:55:50
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
The columns endpoint results in a 500, possibly metadata related
type: bug work: backend status: ready restricted: maintainers
## Description Endpoint: `http://localhost:8000/api/db/v0/tables/<table_id>/columns/` ``` Environment: Request Method: GET Request URL: http://localhost:8000/api/db/v0/tables/5/columns/?limit=500 Django Version: 3.1.14 Python Version: 3.9.8 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware'] Traceback (most recent call last): File "/code/mathesar/models/base.py", line 675, in __getattribute__ return super().__getattribute__(name) During handling of the above exception ('Column' object has no attribute 'primary_key'), another exception occurred: File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__ return self._index[key] The above exception ('nspname') was the direct cause of the following exception: File "/code/mathesar/models/base.py", line 675, in __getattribute__ return super().__getattribute__(name) File "/code/mathesar/models/base.py", line 696, in _sa_column return self.table.sa_columns[self.name] File "/code/mathesar/models/base.py", line 362, in sa_columns return self._enriched_column_sa_table.columns File "/code/mathesar/models/base.py", line 350, in _enriched_column_sa_table table=self._sa_table, File "/code/mathesar/state/cached_property.py", line 62, in __get__ new_value = self.original_get_fn(instance) File "/code/mathesar/models/base.py", line 331, in _sa_table sa_table = reflect_table_from_oid( File "/code/db/tables/operations/select.py", line 23, in reflect_table_from_oid tables = reflect_tables_from_oids([oid], engine, metadata=metadata, connection_to_use=connection_to_use) File "/code/db/tables/operations/select.py", line 29, in reflect_tables_from_oids get_map_of_table_oid_to_schema_name_and_table_name( File "/code/db/tables/operations/select.py", line 59, in get_map_of_table_oid_to_schema_name_and_table_name select(pg_namespace.c.nspname, pg_class.c.relname, pg_class.c.oid) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__ util.raise_(AttributeError(key), replace_context=err) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception During handling of the above exception (nspname), another exception occurred: File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__ return self._index[key] The above exception ('nspname') was the direct cause of the following exception: File "/code/mathesar/models/base.py", line 675, in __getattribute__ return super().__getattribute__(name) File "/code/mathesar/models/base.py", line 696, in _sa_column return self.table.sa_columns[self.name] File "/code/mathesar/models/base.py", line 362, in sa_columns return self._enriched_column_sa_table.columns File "/code/mathesar/models/base.py", line 350, in _enriched_column_sa_table table=self._sa_table, File "/code/mathesar/state/cached_property.py", line 62, in __get__ new_value = self.original_get_fn(instance) File "/code/mathesar/models/base.py", line 331, in _sa_table sa_table = reflect_table_from_oid( File "/code/db/tables/operations/select.py", line 23, in reflect_table_from_oid tables = reflect_tables_from_oids([oid], engine, metadata=metadata, connection_to_use=connection_to_use) File "/code/db/tables/operations/select.py", line 29, in reflect_tables_from_oids get_map_of_table_oid_to_schema_name_and_table_name( File "/code/db/tables/operations/select.py", line 59, in get_map_of_table_oid_to_schema_name_and_table_name select(pg_namespace.c.nspname, pg_class.c.relname, pg_class.c.oid) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__ util.raise_(AttributeError(key), replace_context=err) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception During handling of the above exception (nspname), another exception occurred: File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 826, in __getattr__ return getattr(self.comparator, key) The above exception ('Comparator' object has no attribute '_sa_column') was the direct cause of the following exception: File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 55, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/mixins.py", line 38, in list queryset = self.filter_queryset(self.get_queryset()) File "/code/mathesar/api/db/viewsets/columns.py", line 33, in get_queryset queryset = Column.objects.filter(table=self.kwargs['table_pk']).order_by('attnum') File "/usr/local/lib/python3.9/site-packages/django/db/models/manager.py", line 85, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/code/mathesar/models/base.py", line 70, in get_queryset make_sure_initial_reflection_happened() File "/code/mathesar/state/base.py", line 8, in make_sure_initial_reflection_happened reset_reflection() File "/code/mathesar/state/base.py", line 27, in reset_reflection _trigger_django_model_reflection() File "/code/mathesar/state/base.py", line 31, in _trigger_django_model_reflection reflect_db_objects(metadata=get_cached_metadata()) File "/code/mathesar/state/django.py", line 44, in reflect_db_objects reflect_columns_from_tables(tables, metadata=metadata) File "/code/mathesar/state/django.py", line 123, in reflect_columns_from_tables models._compute_preview_template(table) File "/code/mathesar/models/base.py", line 859, in _compute_preview_template if column.primary_key: File "/code/mathesar/models/base.py", line 681, in __getattribute__ return getattr(self._sa_column, name) File "/code/mathesar/models/base.py", line 681, in __getattribute__ return getattr(self._sa_column, name) File "/code/mathesar/models/base.py", line 681, in __getattribute__ return getattr(self._sa_column, name) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 828, in __getattr__ util.raise_( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception Exception Type: AttributeError at /api/db/v0/tables/5/columns/ Exception Value: Neither 'MathesarColumn' object nor 'Comparator' object has an attribute '_sa_column' ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## To Reproduce <!-- How can we recreate this bug? Please try to provide a Minimal, Complete, and Verifiable (http://stackoverflow.com/help/mcve) example if code-related. --> ## Environment - OS: (_eg._ macOS 10.14.6; Fedora 32) - Browser: (_eg._ Safari; Firefox) - Browser Version: (_eg._ 13; 73) - Other info: ## Additional context <!-- Add any other context about the problem or screenshots here. -->
True
The columns endpoint results in a 500, possibly metadata related - ## Description Endpoint: `http://localhost:8000/api/db/v0/tables/<table_id>/columns/` ``` Environment: Request Method: GET Request URL: http://localhost:8000/api/db/v0/tables/5/columns/?limit=500 Django Version: 3.1.14 Python Version: 3.9.8 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware'] Traceback (most recent call last): File "/code/mathesar/models/base.py", line 675, in __getattribute__ return super().__getattribute__(name) During handling of the above exception ('Column' object has no attribute 'primary_key'), another exception occurred: File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__ return self._index[key] The above exception ('nspname') was the direct cause of the following exception: File "/code/mathesar/models/base.py", line 675, in __getattribute__ return super().__getattribute__(name) File "/code/mathesar/models/base.py", line 696, in _sa_column return self.table.sa_columns[self.name] File "/code/mathesar/models/base.py", line 362, in sa_columns return self._enriched_column_sa_table.columns File "/code/mathesar/models/base.py", line 350, in _enriched_column_sa_table table=self._sa_table, File "/code/mathesar/state/cached_property.py", line 62, in __get__ new_value = self.original_get_fn(instance) File "/code/mathesar/models/base.py", line 331, in _sa_table sa_table = reflect_table_from_oid( File "/code/db/tables/operations/select.py", line 23, in reflect_table_from_oid tables = reflect_tables_from_oids([oid], engine, metadata=metadata, connection_to_use=connection_to_use) File "/code/db/tables/operations/select.py", line 29, in reflect_tables_from_oids get_map_of_table_oid_to_schema_name_and_table_name( File "/code/db/tables/operations/select.py", line 59, in get_map_of_table_oid_to_schema_name_and_table_name select(pg_namespace.c.nspname, pg_class.c.relname, pg_class.c.oid) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__ util.raise_(AttributeError(key), replace_context=err) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception During handling of the above exception (nspname), another exception occurred: File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__ return self._index[key] The above exception ('nspname') was the direct cause of the following exception: File "/code/mathesar/models/base.py", line 675, in __getattribute__ return super().__getattribute__(name) File "/code/mathesar/models/base.py", line 696, in _sa_column return self.table.sa_columns[self.name] File "/code/mathesar/models/base.py", line 362, in sa_columns return self._enriched_column_sa_table.columns File "/code/mathesar/models/base.py", line 350, in _enriched_column_sa_table table=self._sa_table, File "/code/mathesar/state/cached_property.py", line 62, in __get__ new_value = self.original_get_fn(instance) File "/code/mathesar/models/base.py", line 331, in _sa_table sa_table = reflect_table_from_oid( File "/code/db/tables/operations/select.py", line 23, in reflect_table_from_oid tables = reflect_tables_from_oids([oid], engine, metadata=metadata, connection_to_use=connection_to_use) File "/code/db/tables/operations/select.py", line 29, in reflect_tables_from_oids get_map_of_table_oid_to_schema_name_and_table_name( File "/code/db/tables/operations/select.py", line 59, in get_map_of_table_oid_to_schema_name_and_table_name select(pg_namespace.c.nspname, pg_class.c.relname, pg_class.c.oid) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__ util.raise_(AttributeError(key), replace_context=err) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception During handling of the above exception (nspname), another exception occurred: File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 826, in __getattr__ return getattr(self.comparator, key) The above exception ('Comparator' object has no attribute '_sa_column') was the direct cause of the following exception: File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 55, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/mixins.py", line 38, in list queryset = self.filter_queryset(self.get_queryset()) File "/code/mathesar/api/db/viewsets/columns.py", line 33, in get_queryset queryset = Column.objects.filter(table=self.kwargs['table_pk']).order_by('attnum') File "/usr/local/lib/python3.9/site-packages/django/db/models/manager.py", line 85, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/code/mathesar/models/base.py", line 70, in get_queryset make_sure_initial_reflection_happened() File "/code/mathesar/state/base.py", line 8, in make_sure_initial_reflection_happened reset_reflection() File "/code/mathesar/state/base.py", line 27, in reset_reflection _trigger_django_model_reflection() File "/code/mathesar/state/base.py", line 31, in _trigger_django_model_reflection reflect_db_objects(metadata=get_cached_metadata()) File "/code/mathesar/state/django.py", line 44, in reflect_db_objects reflect_columns_from_tables(tables, metadata=metadata) File "/code/mathesar/state/django.py", line 123, in reflect_columns_from_tables models._compute_preview_template(table) File "/code/mathesar/models/base.py", line 859, in _compute_preview_template if column.primary_key: File "/code/mathesar/models/base.py", line 681, in __getattribute__ return getattr(self._sa_column, name) File "/code/mathesar/models/base.py", line 681, in __getattribute__ return getattr(self._sa_column, name) File "/code/mathesar/models/base.py", line 681, in __getattribute__ return getattr(self._sa_column, name) File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/elements.py", line 828, in __getattr__ util.raise_( File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_ raise exception Exception Type: AttributeError at /api/db/v0/tables/5/columns/ Exception Value: Neither 'MathesarColumn' object nor 'Comparator' object has an attribute '_sa_column' ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> ## To Reproduce <!-- How can we recreate this bug? Please try to provide a Minimal, Complete, and Verifiable (http://stackoverflow.com/help/mcve) example if code-related. --> ## Environment - OS: (_eg._ macOS 10.14.6; Fedora 32) - Browser: (_eg._ Safari; Firefox) - Browser Version: (_eg._ 13; 73) - Other info: ## Additional context <!-- Add any other context about the problem or screenshots here. -->
main
the columns endpoint results in a possibly metadata related description endpoint environment request method get request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware traceback most recent call last file code mathesar models base py line in getattribute return super getattribute name during handling of the above exception column object has no attribute primary key another exception occurred file usr local lib site packages sqlalchemy sql base py line in getattr return self index the above exception nspname was the direct cause of the following exception file code mathesar models base py line in getattribute return super getattribute name file code mathesar models base py line in sa column return self table sa columns file code mathesar models base py line in sa columns return self enriched column sa table columns file code mathesar models base py line in enriched column sa table table self sa table file code mathesar state cached property py line in get new value self original get fn instance file code mathesar models base py line in sa table sa table reflect table from oid file code db tables operations select py line in reflect table from oid tables reflect tables from oids engine metadata metadata connection to use connection to use file code db tables operations select py line in reflect tables from oids get map of table oid to schema name and table name file code db tables operations select py line in get map of table oid to schema name and table name select pg namespace c nspname pg class c relname pg class c oid file usr local lib site packages sqlalchemy sql base py line in getattr util raise attributeerror key replace context err file usr local lib site packages sqlalchemy util compat py line in raise raise exception during handling of the above exception nspname another exception occurred file usr local lib site packages sqlalchemy sql base py line in getattr return self index the above exception nspname was the direct cause of the following exception file code mathesar models base py line in getattribute return super getattribute name file code mathesar models base py line in sa column return self table sa columns file code mathesar models base py line in sa columns return self enriched column sa table columns file code mathesar models base py line in enriched column sa table table self sa table file code mathesar state cached property py line in get new value self original get fn instance file code mathesar models base py line in sa table sa table reflect table from oid file code db tables operations select py line in reflect table from oid tables reflect tables from oids engine metadata metadata connection to use connection to use file code db tables operations select py line in reflect tables from oids get map of table oid to schema name and table name file code db tables operations select py line in get map of table oid to schema name and table name select pg namespace c nspname pg class c relname pg class c oid file usr local lib site packages sqlalchemy sql base py line in getattr util raise attributeerror key replace context err file usr local lib site packages sqlalchemy util compat py line in raise raise exception during handling of the above exception nspname another exception occurred file usr local lib site packages sqlalchemy sql elements py line in getattr return getattr self comparator key the above exception comparator object has no attribute sa column was the direct cause of the following exception file usr local lib site packages django core handlers exception py line in inner response get response request file usr local lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr local lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file usr local lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file usr local lib site packages rest framework views py line in dispatch response self handle exception exc file usr local lib site packages rest framework views py line in handle exception response exception handler exc context file code mathesar exception handlers py line in mathesar exception handler raise exc file usr local lib site packages rest framework views py line in dispatch response handler request args kwargs file usr local lib site packages rest framework mixins py line in list queryset self filter queryset self get queryset file code mathesar api db viewsets columns py line in get queryset queryset column objects filter table self kwargs order by attnum file usr local lib site packages django db models manager py line in manager method return getattr self get queryset name args kwargs file code mathesar models base py line in get queryset make sure initial reflection happened file code mathesar state base py line in make sure initial reflection happened reset reflection file code mathesar state base py line in reset reflection trigger django model reflection file code mathesar state base py line in trigger django model reflection reflect db objects metadata get cached metadata file code mathesar state django py line in reflect db objects reflect columns from tables tables metadata metadata file code mathesar state django py line in reflect columns from tables models compute preview template table file code mathesar models base py line in compute preview template if column primary key file code mathesar models base py line in getattribute return getattr self sa column name file code mathesar models base py line in getattribute return getattr self sa column name file code mathesar models base py line in getattribute return getattr self sa column name file usr local lib site packages sqlalchemy sql elements py line in getattr util raise file usr local lib site packages sqlalchemy util compat py line in raise raise exception exception type attributeerror at api db tables columns exception value neither mathesarcolumn object nor comparator object has an attribute sa column expected behavior to reproduce environment os eg macos fedora browser eg safari firefox browser version eg other info additional context
1
130,908
18,213,855,857
IssuesEvent
2021-09-30 00:00:49
ghc-dev/Margaret-Rose
https://api.github.com/repos/ghc-dev/Margaret-Rose
opened
CVE-2021-21295 (Medium) detected in netty-codec-http-4.1.39.Final.jar
security vulnerability
## CVE-2021-21295 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http-4.1.39.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p> <p>Path to dependency file: Margaret-Rose/build.gradle</p> <p>Path to vulnerable library: /caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar</p> <p> Dependency Hierarchy: - :x: **netty-codec-http-4.1.39.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Margaret-Rose/commit/c4e02f67dd4676db950425e7cac2d1a3f3883f24">c4e02f67dd4676db950425e7cac2d1a3f3883f24</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel's pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`. <p>Publish Date: 2021-03-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21295>CVE-2021-21295</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-wm47-8v5p-wjpj">https://github.com/advisories/GHSA-wm47-8v5p-wjpj</a></p> <p>Release Date: 2021-03-09</p> <p>Fix Resolution: io.netty:netty-all:4.1.60;io.netty:netty-codec-http:4.1.60;io.netty:netty-codec-http2:4.1.60</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http","packageVersion":"4.1.39.Final","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"io.netty:netty-codec-http:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.60;io.netty:netty-codec-http:4.1.60;io.netty:netty-codec-http2:4.1.60"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21295","vulnerabilityDetails":"Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers \u0026 clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel\u0027s pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21295","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-21295 (Medium) detected in netty-codec-http-4.1.39.Final.jar - ## CVE-2021-21295 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-codec-http-4.1.39.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p> <p>Path to dependency file: Margaret-Rose/build.gradle</p> <p>Path to vulnerable library: /caches/modules-2/files-2.1/io.netty/netty-codec-http/4.1.39.Final/732d06961162e27fa3ae5989541c4460853745d3/netty-codec-http-4.1.39.Final.jar</p> <p> Dependency Hierarchy: - :x: **netty-codec-http-4.1.39.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Margaret-Rose/commit/c4e02f67dd4676db950425e7cac2d1a3f3883f24">c4e02f67dd4676db950425e7cac2d1a3f3883f24</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel's pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`. <p>Publish Date: 2021-03-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21295>CVE-2021-21295</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-wm47-8v5p-wjpj">https://github.com/advisories/GHSA-wm47-8v5p-wjpj</a></p> <p>Release Date: 2021-03-09</p> <p>Fix Resolution: io.netty:netty-all:4.1.60;io.netty:netty-codec-http:4.1.60;io.netty:netty-codec-http2:4.1.60</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.netty","packageName":"netty-codec-http","packageVersion":"4.1.39.Final","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"io.netty:netty-codec-http:4.1.39.Final","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.netty:netty-all:4.1.60;io.netty:netty-codec-http:4.1.60;io.netty:netty-codec-http2:4.1.60"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-21295","vulnerabilityDetails":"Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers \u0026 clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel\u0027s pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21295","cvss3Severity":"medium","cvss3Score":"5.9","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in netty codec http final jar cve medium severity vulnerability vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file margaret rose build gradle path to vulnerable library caches modules files io netty netty codec http final netty codec http final jar dependency hierarchy x netty codec http final jar vulnerable library found in head commit a href found in base branch master vulnerability details netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty io netty netty codec before version final there is a vulnerability that enables request smuggling if a content length header is present in the original http request the field is not validated by as it is propagated up this is fine as long as the request is not proxied through as http if the request comes in as an http stream gets converted into the http domain objects httprequest httpcontent etc via and then sent up to the child channel s pipeline and proxied through a remote peer as http this may result in request smuggling in a proxy case users may assume the content length is validated somehow which is not the case if the request is forwarded to a backend channel that is a http connection the content length now has meaning and needs to be checked an attacker can smuggle requests inside the body as it gets downgraded from http to http for an example attack refer to the linked github advisory users are only affected if all of this is true or is used is used to convert to http objects and these http objects are forwarded to another remote peer this has been patched in final as a workaround the user can do the validation by themselves by implementing a custom channelinboundhandler that is put in the channelpipeline behind publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all io netty netty codec http io netty netty codec isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree io netty netty codec http final isminimumfixversionavailable true minimumfixversion io netty netty all io netty netty codec http io netty netty codec basebranches vulnerabilityidentifier cve vulnerabilitydetails netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty io netty netty codec before version final there is a vulnerability that enables request smuggling if a content length header is present in the original http request the field is not validated by as it is propagated up this is fine as long as the request is not proxied through as http if the request comes in as an http stream gets converted into the http domain objects httprequest httpcontent etc via and then sent up to the child channel pipeline and proxied through a remote peer as http this may result in request smuggling in a proxy case users may assume the content length is validated somehow which is not the case if the request is forwarded to a backend channel that is a http connection the content length now has meaning and needs to be checked an attacker can smuggle requests inside the body as it gets downgraded from http to http for an example attack refer to the linked github advisory users are only affected if all of this is true or is used is used to convert to http objects and these http objects are forwarded to another remote peer this has been patched in final as a workaround the user can do the validation by themselves by implementing a custom channelinboundhandler that is put in the channelpipeline behind vulnerabilityurl
0
5,252
26,581,338,965
IssuesEvent
2023-01-22 13:49:14
albertlauncher/plugins
https://api.github.com/repos/albertlauncher/plugins
closed
Applications plugins cannot find pycharm
Maintainer wanted
If I would like to try to find `pycharm`, it doesn't show anything ![изображение](https://user-images.githubusercontent.com/2799891/210841297-c8baeb8f-b73d-4b61-88b4-19a79584f046.png) desktop file looks like ``` [Desktop Entry] Type=Application Name=PyCharm Professional Edition Icon=pycharm Comment=Python IDE for Professional Developers. Exec=pycharm %f Terminal=false Categories=Development;IDE;Python; StartupNotify=true StartupWMClass=jetbrains-pycharm ``` `Charm` (case sensitive) search doesn't work either. However it is possible to find by typing `pycharm` instead (or even `professional`). Fuzzy search option has no effect ![изображение](https://user-images.githubusercontent.com/2799891/210841805-c3af380b-2034-49a1-9571-a18f7d76d38c.png)
True
Applications plugins cannot find pycharm - If I would like to try to find `pycharm`, it doesn't show anything ![изображение](https://user-images.githubusercontent.com/2799891/210841297-c8baeb8f-b73d-4b61-88b4-19a79584f046.png) desktop file looks like ``` [Desktop Entry] Type=Application Name=PyCharm Professional Edition Icon=pycharm Comment=Python IDE for Professional Developers. Exec=pycharm %f Terminal=false Categories=Development;IDE;Python; StartupNotify=true StartupWMClass=jetbrains-pycharm ``` `Charm` (case sensitive) search doesn't work either. However it is possible to find by typing `pycharm` instead (or even `professional`). Fuzzy search option has no effect ![изображение](https://user-images.githubusercontent.com/2799891/210841805-c3af380b-2034-49a1-9571-a18f7d76d38c.png)
main
applications plugins cannot find pycharm if i would like to try to find pycharm it doesn t show anything desktop file looks like type application name pycharm professional edition icon pycharm comment python ide for professional developers exec pycharm f terminal false categories development ide python startupnotify true startupwmclass jetbrains pycharm charm case sensitive search doesn t work either however it is possible to find by typing pycharm instead or even professional fuzzy search option has no effect
1
3,756
15,788,683,637
IssuesEvent
2021-04-01 21:12:24
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
Add a new property to component Dropdown to allow specifying the min-width of the` .bx--list-box__menu` class
status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬 type: enhancement 💡
### Summary Add a new property to component Dropdown to allow specifying the min-width of the` .bx--list-box__menu` class inside Dropdown so we can see full text (and not truncated text) of all values in the options list. Clarify if you are asking for design, development, or both design and development. ### Justification Without this property, consumer of Dropdown has to use a specific className for their Dropdown and add a css to specify the min-width of the class ` .bx--list-box__menu` in the scope of the Dropdown class , that means having a specific css per Dropdown. ### Desired UX and success metrics <!--alex disable failure--> Describe the full user experience for this feature. Also define the metrics by which we can measure success/failure for the user. <!--alex enable failure--> ### "Must have" functionality Highlight any "must have" needs and functionality for the request. This should not be a full list of functionality; the Carbon team will work with you to define functionality based on the desired UX. ### Specific timeline issues / requests It's a UX requirement for Cognos Analytics release 11.2.1 (scheduled probably for early next year) <!--alex disable period--> Do you want this work within a specific time period? Is it related to an upcoming release? <!--alex enable period--> _NB: The Carbon team will try to work with your timeline, but it's not guaranteed. The earlier you make a request in advance of a desired delivery date, the better!_ ### Available extra resources What resources do you have to assist this effort? _Carbon is a collaborative system. We encourage teams to build components and submit them for integration as either add-ons or core components._
True
Add a new property to component Dropdown to allow specifying the min-width of the` .bx--list-box__menu` class - ### Summary Add a new property to component Dropdown to allow specifying the min-width of the` .bx--list-box__menu` class inside Dropdown so we can see full text (and not truncated text) of all values in the options list. Clarify if you are asking for design, development, or both design and development. ### Justification Without this property, consumer of Dropdown has to use a specific className for their Dropdown and add a css to specify the min-width of the class ` .bx--list-box__menu` in the scope of the Dropdown class , that means having a specific css per Dropdown. ### Desired UX and success metrics <!--alex disable failure--> Describe the full user experience for this feature. Also define the metrics by which we can measure success/failure for the user. <!--alex enable failure--> ### "Must have" functionality Highlight any "must have" needs and functionality for the request. This should not be a full list of functionality; the Carbon team will work with you to define functionality based on the desired UX. ### Specific timeline issues / requests It's a UX requirement for Cognos Analytics release 11.2.1 (scheduled probably for early next year) <!--alex disable period--> Do you want this work within a specific time period? Is it related to an upcoming release? <!--alex enable period--> _NB: The Carbon team will try to work with your timeline, but it's not guaranteed. The earlier you make a request in advance of a desired delivery date, the better!_ ### Available extra resources What resources do you have to assist this effort? _Carbon is a collaborative system. We encourage teams to build components and submit them for integration as either add-ons or core components._
main
add a new property to component dropdown to allow specifying the min width of the bx list box menu class summary add a new property to component dropdown to allow specifying the min width of the bx list box menu class inside dropdown so we can see full text and not truncated text of all values in the options list clarify if you are asking for design development or both design and development justification without this property consumer of dropdown has to use a specific classname for their dropdown and add a css to specify the min width of the class bx list box menu in the scope of the dropdown class that means having a specific css per dropdown desired ux and success metrics describe the full user experience for this feature also define the metrics by which we can measure success failure for the user must have functionality highlight any must have needs and functionality for the request this should not be a full list of functionality the carbon team will work with you to define functionality based on the desired ux specific timeline issues requests it s a ux requirement for cognos analytics release scheduled probably for early next year do you want this work within a specific time period is it related to an upcoming release nb the carbon team will try to work with your timeline but it s not guaranteed the earlier you make a request in advance of a desired delivery date the better available extra resources what resources do you have to assist this effort carbon is a collaborative system we encourage teams to build components and submit them for integration as either add ons or core components
1
1,385
6,011,672,681
IssuesEvent
2017-06-06 15:39:12
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_chocolatey: upgrade feature doesnt seem to work
affects_2.2 bug_report waiting_on_maintainer windows
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME win_chocolatey ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.2.0 (devel 58b9f637a1) last updated 2016/08/24 09:07:32 (GMT +100) lib/ansible/modules/core: (detached HEAD 368ca738fa) last updated 2016/08/23 20:15:23 (GMT +100) lib/ansible/modules/extras: (detached HEAD 0749ce6faa) last updated 2016/08/23 20:15:24 (GMT +100) config file = /home/id/ansible/ansible.cfg configured module search path = ['./library'] ``` ##### OS / ENVIRONMENT win10 pro anniversary update Build 14393 ##### SUMMARY Upgrade feature calls `choco install` instead of `choco upgrade` ##### STEPS TO REPRODUCE ``` - name: Upgrade choco_pkgs win_chocolatey: name: "{{item}}" upgrade: true # <----- no effect state: present allow_empty_checksums: true with_items: "{{ choco_pkgs_auto }}" ignore_errors: yes ``` ##### EXPECTED RESULTS ``` choco.exe upgrade -dv -y --allow-empty-checksums "{{ item }}" ``` ##### ACTUAL RESULTS ``` choco.exe install -dv -y --allow-empty-checksums "{{ item }}" ``` For example as seen here whenever the command errors out: ``` task path: /home/id/ansible/windows/roles/5_pkgs_upgrade/tasks/main.yml:72 changed: [brix-3205] => {"changed": true, "rc": 0, "stderr": "", "stdout": "upgrading\r\npkgs...\r\n", "stdout_lines": ["upgrading", "pkgs..."]} ____________________________________________ < TASK [5_pkgs_upgrade : Upgrade choco_pkgs] > -------------------------------------------- \ ^__^ \ (oo)\_______ (__)\ )\/\ ||----w | || || task path: /home/id/ansible/windows/roles/5_pkgs_upgrade/tasks/main.yml:76 ok: [brix-3205] => (item=windirstat) => {"changed": false, "item": "windirstat"} ok: [brix-3205] => (item=realtemp) => {"changed": false, "item": "realtemp"} ok: [brix-3205] => (item=f.lux) => {"changed": false, "item": "f.lux"} ok: [brix-3205] => (item=procmon) => {"changed": false, "item": "procmon"} ok: [brix-3205] => (item=crystaldiskinfo) => {"changed": false, "item": "crystaldiskinfo"} ok: [brix-3205] => (item=crystaldiskmark) => {"changed": false, "item": "crystaldiskmark"} failed: [brix-3205] (item=easybcd) => {"changed": false, "choco_error_cmd": "choco.exe install -dv -y easybcd --allow-empty-checksums", "choco_error_log": "Chocolatey v0.10.0 Chocolatey is running on Windows v 10.0.14393.0 Attempting to delete file \"C:/ProgramData/chocolatey/choco.exe.old\". Attempting to delete file \"C:\\ProgramData\\chocolatey\\choco.exe.old\". Command line: \"C:\\ProgramData\\chocolatey\\choco.exe\" install -dv -y easybcd --allow-empty-checksums Received arguments: install -dv -y easybcd --allow-empty-checksums RemovePendingPackagesTask is now ready and waiting for PreRunMessage. Sending message 'PreRunMessage' out if there are subscribers... [Pending] Removing all pending packages that should not be considered installed... The source 'https://chocolatey.org/api/v2/' evaluated to a 'normal' source type NOTE: Hiding sensitive configuration data! Please double and triple check to be sure no sensitive data is shown, especially if copying ```
True
win_chocolatey: upgrade feature doesnt seem to work - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME win_chocolatey ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.2.0 (devel 58b9f637a1) last updated 2016/08/24 09:07:32 (GMT +100) lib/ansible/modules/core: (detached HEAD 368ca738fa) last updated 2016/08/23 20:15:23 (GMT +100) lib/ansible/modules/extras: (detached HEAD 0749ce6faa) last updated 2016/08/23 20:15:24 (GMT +100) config file = /home/id/ansible/ansible.cfg configured module search path = ['./library'] ``` ##### OS / ENVIRONMENT win10 pro anniversary update Build 14393 ##### SUMMARY Upgrade feature calls `choco install` instead of `choco upgrade` ##### STEPS TO REPRODUCE ``` - name: Upgrade choco_pkgs win_chocolatey: name: "{{item}}" upgrade: true # <----- no effect state: present allow_empty_checksums: true with_items: "{{ choco_pkgs_auto }}" ignore_errors: yes ``` ##### EXPECTED RESULTS ``` choco.exe upgrade -dv -y --allow-empty-checksums "{{ item }}" ``` ##### ACTUAL RESULTS ``` choco.exe install -dv -y --allow-empty-checksums "{{ item }}" ``` For example as seen here whenever the command errors out: ``` task path: /home/id/ansible/windows/roles/5_pkgs_upgrade/tasks/main.yml:72 changed: [brix-3205] => {"changed": true, "rc": 0, "stderr": "", "stdout": "upgrading\r\npkgs...\r\n", "stdout_lines": ["upgrading", "pkgs..."]} ____________________________________________ < TASK [5_pkgs_upgrade : Upgrade choco_pkgs] > -------------------------------------------- \ ^__^ \ (oo)\_______ (__)\ )\/\ ||----w | || || task path: /home/id/ansible/windows/roles/5_pkgs_upgrade/tasks/main.yml:76 ok: [brix-3205] => (item=windirstat) => {"changed": false, "item": "windirstat"} ok: [brix-3205] => (item=realtemp) => {"changed": false, "item": "realtemp"} ok: [brix-3205] => (item=f.lux) => {"changed": false, "item": "f.lux"} ok: [brix-3205] => (item=procmon) => {"changed": false, "item": "procmon"} ok: [brix-3205] => (item=crystaldiskinfo) => {"changed": false, "item": "crystaldiskinfo"} ok: [brix-3205] => (item=crystaldiskmark) => {"changed": false, "item": "crystaldiskmark"} failed: [brix-3205] (item=easybcd) => {"changed": false, "choco_error_cmd": "choco.exe install -dv -y easybcd --allow-empty-checksums", "choco_error_log": "Chocolatey v0.10.0 Chocolatey is running on Windows v 10.0.14393.0 Attempting to delete file \"C:/ProgramData/chocolatey/choco.exe.old\". Attempting to delete file \"C:\\ProgramData\\chocolatey\\choco.exe.old\". Command line: \"C:\\ProgramData\\chocolatey\\choco.exe\" install -dv -y easybcd --allow-empty-checksums Received arguments: install -dv -y easybcd --allow-empty-checksums RemovePendingPackagesTask is now ready and waiting for PreRunMessage. Sending message 'PreRunMessage' out if there are subscribers... [Pending] Removing all pending packages that should not be considered installed... The source 'https://chocolatey.org/api/v2/' evaluated to a 'normal' source type NOTE: Hiding sensitive configuration data! Please double and triple check to be sure no sensitive data is shown, especially if copying ```
main
win chocolatey upgrade feature doesnt seem to work issue type bug report component name win chocolatey ansible version ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file home id ansible ansible cfg configured module search path os environment pro anniversary update build summary upgrade feature calls choco install instead of choco upgrade steps to reproduce name upgrade choco pkgs win chocolatey name item upgrade true no effect state present allow empty checksums true with items choco pkgs auto ignore errors yes expected results choco exe upgrade dv y allow empty checksums item actual results choco exe install dv y allow empty checksums item for example as seen here whenever the command errors out task path home id ansible windows roles pkgs upgrade tasks main yml changed changed true rc stderr stdout upgrading r npkgs r n stdout lines oo w task path home id ansible windows roles pkgs upgrade tasks main yml ok item windirstat changed false item windirstat ok item realtemp changed false item realtemp ok item f lux changed false item f lux ok item procmon changed false item procmon ok item crystaldiskinfo changed false item crystaldiskinfo ok item crystaldiskmark changed false item crystaldiskmark failed item easybcd changed false choco error cmd choco exe install dv y easybcd allow empty checksums choco error log chocolatey chocolatey is running on windows v attempting to delete file c programdata chocolatey choco exe old attempting to delete file c programdata chocolatey choco exe old command line c programdata chocolatey choco exe install dv y easybcd allow empty checksums received arguments install dv y easybcd allow empty checksums removependingpackagestask is now ready and waiting for prerunmessage sending message prerunmessage out if there are subscribers removing all pending packages that should not be considered installed the source evaluated to a normal source type note hiding sensitive configuration data please double and triple check to be sure no sensitive data is shown especially if copying
1
131,333
12,481,158,896
IssuesEvent
2020-05-29 21:49:45
syl20bnr/spacemacs
https://api.github.com/repos/syl20bnr/spacemacs
closed
Explain the basics of Org-mode for beginners
- Forum - Documentation ✏ Feature request Org stale
All documentation attached with Spacemacs is in org-files. Alas, knowing how to browse and read org files is necessary for a person to get a basic knowledge about Spacemacs. But currently it's doesn't explain the very basics of org-mode. The only reason that I know how to expand the headlines and show their content is that I have read Org-modes' info files in an earlier adventure with Emacs. I propose that the basics of browsing and reading org files should be added besides the Evil and Emacs tutorial in the "Quick Help" box on the start buffer. A relevant headline could be "How to read Spacemacs' documentation in Emacs".
1.0
Explain the basics of Org-mode for beginners - All documentation attached with Spacemacs is in org-files. Alas, knowing how to browse and read org files is necessary for a person to get a basic knowledge about Spacemacs. But currently it's doesn't explain the very basics of org-mode. The only reason that I know how to expand the headlines and show their content is that I have read Org-modes' info files in an earlier adventure with Emacs. I propose that the basics of browsing and reading org files should be added besides the Evil and Emacs tutorial in the "Quick Help" box on the start buffer. A relevant headline could be "How to read Spacemacs' documentation in Emacs".
non_main
explain the basics of org mode for beginners all documentation attached with spacemacs is in org files alas knowing how to browse and read org files is necessary for a person to get a basic knowledge about spacemacs but currently it s doesn t explain the very basics of org mode the only reason that i know how to expand the headlines and show their content is that i have read org modes info files in an earlier adventure with emacs i propose that the basics of browsing and reading org files should be added besides the evil and emacs tutorial in the quick help box on the start buffer a relevant headline could be how to read spacemacs documentation in emacs
0
345,778
10,372,990,992
IssuesEvent
2019-09-09 05:45:59
alexakasanjeev/magento_react_native
https://api.github.com/repos/alexakasanjeev/magento_react_native
opened
Product Detail Page UI Change List
Priority: Medium Status: On Hold Type: Enhancement
**Is your feature request related to a problem? Please describe.** Product Detail page has lots of unimplemented feature, which are essential 1. [ ] `WebView` height displaying product description should be dynamic, currently it is set statically 2. [ ] Use `type_id === 'configurable'` logic to check it has options, everywhere in product detail page 3. [ ] No message is shown when user hit add to cart button, show success or error message accordingly 4. [ ] No check in `configurable` type product to disable certain options which are not available > Example: suppose in size `s` color `red` is not available, so when user select size `s`, `red` option should be disabled. 5. [ ] No check written to check whether product is out of stock or not, if out of stock, disable `add-to-cart` button 6. [ ] Add input box to let user enter quantity of that product for cart currently defaults to 1, and not more then available quantity **Describe the solution you'd like** Implement above list, in Product Detail page
1.0
Product Detail Page UI Change List - **Is your feature request related to a problem? Please describe.** Product Detail page has lots of unimplemented feature, which are essential 1. [ ] `WebView` height displaying product description should be dynamic, currently it is set statically 2. [ ] Use `type_id === 'configurable'` logic to check it has options, everywhere in product detail page 3. [ ] No message is shown when user hit add to cart button, show success or error message accordingly 4. [ ] No check in `configurable` type product to disable certain options which are not available > Example: suppose in size `s` color `red` is not available, so when user select size `s`, `red` option should be disabled. 5. [ ] No check written to check whether product is out of stock or not, if out of stock, disable `add-to-cart` button 6. [ ] Add input box to let user enter quantity of that product for cart currently defaults to 1, and not more then available quantity **Describe the solution you'd like** Implement above list, in Product Detail page
non_main
product detail page ui change list is your feature request related to a problem please describe product detail page has lots of unimplemented feature which are essential webview height displaying product description should be dynamic currently it is set statically use type id configurable logic to check it has options everywhere in product detail page no message is shown when user hit add to cart button show success or error message accordingly no check in configurable type product to disable certain options which are not available example suppose in size s color red is not available so when user select size s red option should be disabled no check written to check whether product is out of stock or not if out of stock disable add to cart button add input box to let user enter quantity of that product for cart currently defaults to and not more then available quantity describe the solution you d like implement above list in product detail page
0
48,632
12,225,260,065
IssuesEvent
2020-05-03 04:01:18
Autodesk/arnold-usd
https://api.github.com/repos/Autodesk/arnold-usd
closed
Update testsuite scripts to match arnold
bug build
**Describe the bug** The arnold-usd testsuite scripts have slightly derived from arnold core ones, which is causing issues when running the testsuite from arnold. The parameter `resave` should be called again `resaved`, and we should check whether it's a string or a boolean. When set to true, we assume the scene has to be resaved to .ass. This way, both test scripts will be similar again
1.0
Update testsuite scripts to match arnold - **Describe the bug** The arnold-usd testsuite scripts have slightly derived from arnold core ones, which is causing issues when running the testsuite from arnold. The parameter `resave` should be called again `resaved`, and we should check whether it's a string or a boolean. When set to true, we assume the scene has to be resaved to .ass. This way, both test scripts will be similar again
non_main
update testsuite scripts to match arnold describe the bug the arnold usd testsuite scripts have slightly derived from arnold core ones which is causing issues when running the testsuite from arnold the parameter resave should be called again resaved and we should check whether it s a string or a boolean when set to true we assume the scene has to be resaved to ass this way both test scripts will be similar again
0
18
2,515,212,688
IssuesEvent
2015-01-15 17:07:08
simplesamlphp/simplesamlphp
https://api.github.com/repos/simplesamlphp/simplesamlphp
opened
Create an AuthMemCookie module
enhancement low maintainability
This should be on its own standalone repository and installable through composer. Since the functionality is already there, it's just about moving `www/authmemcookie.php` and `lib/SimpleSAML/AuthMemCookie.php` out and creating the additional module structure.
True
Create an AuthMemCookie module - This should be on its own standalone repository and installable through composer. Since the functionality is already there, it's just about moving `www/authmemcookie.php` and `lib/SimpleSAML/AuthMemCookie.php` out and creating the additional module structure.
main
create an authmemcookie module this should be on its own standalone repository and installable through composer since the functionality is already there it s just about moving www authmemcookie php and lib simplesaml authmemcookie php out and creating the additional module structure
1
1,584
6,572,353,642
IssuesEvent
2017-09-11 01:39:15
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
[ecs_service_facts] value of details must be one of: true,false, got: True
affects_2.1 aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME `ecs_service_facts` ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = /srv/code/ops/ansible/ansible.cfg configured module search path = ['./library'] ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT OSX ##### SUMMARY Running this playbook (with `details: true`): ``` - hosts: localhost connection: local gather_facts: false tasks: - ecs_service_facts: region: "us-west-1" cluster: "my-cluster" service: "my-service" details: true ``` ##### EXPECTED RESULTS Detailed result as mentioned in [docs](http://docs.ansible.com/ansible/ecs_service_facts_module.html). ##### ACTUAL RESULTS ``` fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"cluster": "mtdev-microservices-cluster", "details": "True", "region": "us-west-2", "service": "mtdev-threat-service", "validate_certs": true}, "module_name": "ecs_service_facts"}, "msg": "value of details must be one of: true,false, got: True"} ```
True
[ecs_service_facts] value of details must be one of: true,false, got: True - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME `ecs_service_facts` ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = /srv/code/ops/ansible/ansible.cfg configured module search path = ['./library'] ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT OSX ##### SUMMARY Running this playbook (with `details: true`): ``` - hosts: localhost connection: local gather_facts: false tasks: - ecs_service_facts: region: "us-west-1" cluster: "my-cluster" service: "my-service" details: true ``` ##### EXPECTED RESULTS Detailed result as mentioned in [docs](http://docs.ansible.com/ansible/ecs_service_facts_module.html). ##### ACTUAL RESULTS ``` fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"cluster": "mtdev-microservices-cluster", "details": "True", "region": "us-west-2", "service": "mtdev-threat-service", "validate_certs": true}, "module_name": "ecs_service_facts"}, "msg": "value of details must be one of: true,false, got: True"} ```
main
value of details must be one of true false got true issue type bug report component name ecs service facts ansible version ansible config file srv code ops ansible ansible cfg configured module search path configuration n a os environment osx summary running this playbook with details true hosts localhost connection local gather facts false tasks ecs service facts region us west cluster my cluster service my service details true expected results detailed result as mentioned in actual results fatal failed changed false failed true invocation module args cluster mtdev microservices cluster details true region us west service mtdev threat service validate certs true module name ecs service facts msg value of details must be one of true false got true
1
88,777
17,663,608,706
IssuesEvent
2021-08-22 02:12:31
adventuregamestudio/ags
https://api.github.com/repos/adventuregamestudio/ags
closed
INFO: a variant of utf8 support for translations
type: information context: unicode
A while ago @mgambrell has pointed me his work on hacking in utf-8 string support for translations. It's based on modifying alfont source to use parts of the **musl** library to convert strings. Here's the commit (and maybe there are more commits around): https://github.com/RatalaikaGames/ags/commit/a00e7d933458d8360c17cf84c8cfbf5125111c3b I am leaving this here only for the reference if someone would like to look into this at some point, because afaik the solution was not applied universally to whole engine, and it's hard to tell which nuances may be met when fully moving to utf8 support in both editor and engine (especially if you care about loading old games into new interpreter).
1.0
INFO: a variant of utf8 support for translations - A while ago @mgambrell has pointed me his work on hacking in utf-8 string support for translations. It's based on modifying alfont source to use parts of the **musl** library to convert strings. Here's the commit (and maybe there are more commits around): https://github.com/RatalaikaGames/ags/commit/a00e7d933458d8360c17cf84c8cfbf5125111c3b I am leaving this here only for the reference if someone would like to look into this at some point, because afaik the solution was not applied universally to whole engine, and it's hard to tell which nuances may be met when fully moving to utf8 support in both editor and engine (especially if you care about loading old games into new interpreter).
non_main
info a variant of support for translations a while ago mgambrell has pointed me his work on hacking in utf string support for translations it s based on modifying alfont source to use parts of the musl library to convert strings here s the commit and maybe there are more commits around i am leaving this here only for the reference if someone would like to look into this at some point because afaik the solution was not applied universally to whole engine and it s hard to tell which nuances may be met when fully moving to support in both editor and engine especially if you care about loading old games into new interpreter
0
4,440
23,067,911,648
IssuesEvent
2022-07-25 15:20:25
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[Bug]: MeterChart generates error in the console
type: bug 🐛 status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬
### Package carbon-components ### Browser _No response_ ### Package version v11.0 ### React version _No response_ ### Description When using a meter chart on a page, the console produces the following error: Error: Invalid value for <circle> attribute cy="calc(1em / 2)" I have to remove the chart from my design, because we have quite a few of these charts on the page. It produces many errors. ### Reproduction/example https://carbondesignsystem.com/data-visualization/simple-charts/#meter ### Steps to reproduce Please open example page. Inspect Switch to the console... and you will see the error message. ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
True
[Bug]: MeterChart generates error in the console - ### Package carbon-components ### Browser _No response_ ### Package version v11.0 ### React version _No response_ ### Description When using a meter chart on a page, the console produces the following error: Error: Invalid value for <circle> attribute cy="calc(1em / 2)" I have to remove the chart from my design, because we have quite a few of these charts on the page. It produces many errors. ### Reproduction/example https://carbondesignsystem.com/data-visualization/simple-charts/#meter ### Steps to reproduce Please open example page. Inspect Switch to the console... and you will see the error message. ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
main
meterchart generates error in the console package carbon components browser no response package version react version no response description when using a meter chart on a page the console produces the following error error invalid value for attribute cy calc i have to remove the chart from my design because we have quite a few of these charts on the page it produces many errors reproduction example steps to reproduce please open example page inspect switch to the console and you will see the error message code of conduct i agree to follow this project s i checked the for duplicate problems
1
460,889
13,219,917,163
IssuesEvent
2020-08-17 11:24:54
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
almudi.org - "Secure connection Failed error message is displayed
browser-focus-geckoview engine-gecko priority-normal severity-critical type-unsupported-tls
<!-- @browser: Firefox Mobile 71.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://almudi.org/calendario-liturgico/meditacion/56/domingo-1-de-adviento-ciclo-a?acm=29397_2119 **Browser / Version**: Firefox Mobile 71.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: conexión segura fallida **Steps to Reproduce**: No se ha podido verificar autenticidad del sitio <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
almudi.org - "Secure connection Failed error message is displayed - <!-- @browser: Firefox Mobile 71.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://almudi.org/calendario-liturgico/meditacion/56/domingo-1-de-adviento-ciclo-a?acm=29397_2119 **Browser / Version**: Firefox Mobile 71.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: conexión segura fallida **Steps to Reproduce**: No se ha podido verificar autenticidad del sitio <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
almudi org secure connection failed error message is displayed url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description conexión segura fallida steps to reproduce no se ha podido verificar autenticidad del sitio browser configuration none from with ❤️
0
3,641
14,730,765,981
IssuesEvent
2021-01-06 13:44:34
AMYMEME/re-cycle-app
https://api.github.com/repos/AMYMEME/re-cycle-app
closed
firebase 통합환경 만들기
maintain
# Database 찾아보니까 Firebase에 DB로 쓸 수 있는게 있는데, 원래 Google Storage를 포함해서 실시간 DB, firestore가 있음 ## Google Storage 구글 스토리지는 90일 300$이용이나 for firebase용도 따로 있는 것 같은데, java지원이 안되고, 우리가 갖고 있는 데이터에 적합하지 않아보임. 이미지같은 큰 바이너리 데이터를 저장할 때 가장 유용할 것 같음 ## realtime DB NoSQL 형태, Android, iOS, 자바스크립트 SDK로 연동할 수 있음 따라서 클라이언트 쪽에서는 잘 모르겠는데, 백엔드 쪽에서는 잘 모르겠음 ## firestore NoSQL 형태. 백엔드 쪽에도 적합해 보임 안드로이드(자바, 코틀린 모두 지원), iOS, 노드JS, 스프링, 파이썬, golang 모두 지원하고 우리가 저장할 데이터가 그렇게 큰 편이 아니어서 적합해보임
True
firebase 통합환경 만들기 - # Database 찾아보니까 Firebase에 DB로 쓸 수 있는게 있는데, 원래 Google Storage를 포함해서 실시간 DB, firestore가 있음 ## Google Storage 구글 스토리지는 90일 300$이용이나 for firebase용도 따로 있는 것 같은데, java지원이 안되고, 우리가 갖고 있는 데이터에 적합하지 않아보임. 이미지같은 큰 바이너리 데이터를 저장할 때 가장 유용할 것 같음 ## realtime DB NoSQL 형태, Android, iOS, 자바스크립트 SDK로 연동할 수 있음 따라서 클라이언트 쪽에서는 잘 모르겠는데, 백엔드 쪽에서는 잘 모르겠음 ## firestore NoSQL 형태. 백엔드 쪽에도 적합해 보임 안드로이드(자바, 코틀린 모두 지원), iOS, 노드JS, 스프링, 파이썬, golang 모두 지원하고 우리가 저장할 데이터가 그렇게 큰 편이 아니어서 적합해보임
main
firebase 통합환경 만들기 database 찾아보니까 firebase에 db로 쓸 수 있는게 있는데 원래 google storage를 포함해서 실시간 db firestore가 있음 google storage 구글 스토리지는 이용이나 for firebase용도 따로 있는 것 같은데 java지원이 안되고 우리가 갖고 있는 데이터에 적합하지 않아보임 이미지같은 큰 바이너리 데이터를 저장할 때 가장 유용할 것 같음 realtime db nosql 형태 android ios 자바스크립트 sdk로 연동할 수 있음 따라서 클라이언트 쪽에서는 잘 모르겠는데 백엔드 쪽에서는 잘 모르겠음 firestore nosql 형태 백엔드 쪽에도 적합해 보임 안드로이드 자바 코틀린 모두 지원 ios 노드js 스프링 파이썬 golang 모두 지원하고 우리가 저장할 데이터가 그렇게 큰 편이 아니어서 적합해보임
1
3,013
11,140,133,870
IssuesEvent
2019-12-21 11:49:39
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
terraform plan_file custom path broken
affects_2.8 bug cloud has_pr module needs_maintainer needs_triage python3 support:community
<!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and devel branch are affected too --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY https://github.com/ansible/ansible/blob/devel/lib/ansible/modules/cloud/misc/terraform.py#L353 expects the plan_file to be a filename not a path to a file.inside of project_path. Works with relative paths but not with absolute paths. Lucky for me I use a path inside project_path but it worked before with paths outside too. ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> terraform ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible 2.8.1 config file = /home/adam/.ansible.cfg configured module search path = ['/home/adam/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /home/adam/.local/lib/python3.6/site-packages/ansible executable location = /home/adam/.local/bin/ansible python version = 3.6.8 (default, Jan 14 2019, 11:02:34) [GCC 8.0.1 20180414 (experimental) [trunk revision 259383]] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ``` ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> <!--- Paste example playbooks or commands between quotes below --> ```yaml use absolute path for plan_file ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> uses the filepath provided ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes --> ```paste below TASK [run terraform apply] *************************************************************************************************************************************************************************************************************************************************************** fatal: [localhost]: FAILED! => {"changed": false, "msg": "Could not find plan_file \"/home/.../plan.tfplan\", check the path and try again."} ```
True
terraform plan_file custom path broken - <!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and devel branch are affected too --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY https://github.com/ansible/ansible/blob/devel/lib/ansible/modules/cloud/misc/terraform.py#L353 expects the plan_file to be a filename not a path to a file.inside of project_path. Works with relative paths but not with absolute paths. Lucky for me I use a path inside project_path but it worked before with paths outside too. ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> terraform ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible 2.8.1 config file = /home/adam/.ansible.cfg configured module search path = ['/home/adam/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /home/adam/.local/lib/python3.6/site-packages/ansible executable location = /home/adam/.local/bin/ansible python version = 3.6.8 (default, Jan 14 2019, 11:02:34) [GCC 8.0.1 20180414 (experimental) [trunk revision 259383]] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ``` ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> <!--- Paste example playbooks or commands between quotes below --> ```yaml use absolute path for plan_file ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> uses the filepath provided ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes --> ```paste below TASK [run terraform apply] *************************************************************************************************************************************************************************************************************************************************************** fatal: [localhost]: FAILED! => {"changed": false, "msg": "Could not find plan_file \"/home/.../plan.tfplan\", check the path and try again."} ```
main
terraform plan file custom path broken summary expects the plan file to be a filename not a path to a file inside of project path works with relative paths but not with absolute paths lucky for me i use a path inside project path but it worked before with paths outside too issue type bug report component name terraform ansible version paste below ansible config file home adam ansible cfg configured module search path ansible python module location home adam local lib site packages ansible executable location home adam local bin ansible python version default jan configuration paste below os environment steps to reproduce yaml use absolute path for plan file expected results uses the filepath provided actual results paste below task fatal failed changed false msg could not find plan file home plan tfplan check the path and try again
1
5,524
27,616,562,954
IssuesEvent
2023-03-09 19:51:52
microsoft/mu_plus
https://api.github.com/repos/microsoft/mu_plus
closed
SetCurrentTextString() in EditBox from SimpleUiToolkit not working bug
type:bug complexity:easy state:needs-maintainer-feedback
This function leaves the box empty ant text is never set. Just add this: StrnCpyS(this->m_EditBoxDisplayText, sizeof(this->m_EditBoxDisplayText) / sizeof(CHAR16), NewTextString, (UIT_EDITBOX_MAX_STRING_LENGTH - 1)); after StrnCpyS (this->m_EditBoxText... on the line 322 in function SetCurrentTextString() in MsGraphics\Library\SimpleUIToolKit\EditBox.c to make it work.
True
SetCurrentTextString() in EditBox from SimpleUiToolkit not working bug - This function leaves the box empty ant text is never set. Just add this: StrnCpyS(this->m_EditBoxDisplayText, sizeof(this->m_EditBoxDisplayText) / sizeof(CHAR16), NewTextString, (UIT_EDITBOX_MAX_STRING_LENGTH - 1)); after StrnCpyS (this->m_EditBoxText... on the line 322 in function SetCurrentTextString() in MsGraphics\Library\SimpleUIToolKit\EditBox.c to make it work.
main
setcurrenttextstring in editbox from simpleuitoolkit not working bug this function leaves the box empty ant text is never set just add this strncpys this m editboxdisplaytext sizeof this m editboxdisplaytext sizeof newtextstring uit editbox max string length after strncpys this m editboxtext on the line in function setcurrenttextstring in msgraphics library simpleuitoolkit editbox c to make it work
1
5,623
28,133,658,972
IssuesEvent
2023-04-01 05:34:58
beefproject/beef
https://api.github.com/repos/beefproject/beef
closed
Install Script: Remove --without test develop from ./install
Maintainability Install
https://github.com/beefproject/beef/blob/1ae320c3bc9db748dff7e53fe919cdef80089062/install#L240 is deprecated and should be removed bundle not requires a local config file to store this information
True
Install Script: Remove --without test develop from ./install - https://github.com/beefproject/beef/blob/1ae320c3bc9db748dff7e53fe919cdef80089062/install#L240 is deprecated and should be removed bundle not requires a local config file to store this information
main
install script remove without test develop from install is deprecated and should be removed bundle not requires a local config file to store this information
1
204,864
15,953,303,083
IssuesEvent
2021-04-15 12:17:01
srijan-sivakumar/redant
https://api.github.com/repos/srijan-sivakumar/redant
opened
Creating a readme for writing test cases
documentation enhancement
Create a readme that can guide anyone to integrate with the ops and write their own tests.
1.0
Creating a readme for writing test cases - Create a readme that can guide anyone to integrate with the ops and write their own tests.
non_main
creating a readme for writing test cases create a readme that can guide anyone to integrate with the ops and write their own tests
0
1,786
6,575,879,768
IssuesEvent
2017-09-11 17:41:12
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Adding Namespace as an editable parameter for the docker_login module.
affects_2.1 cloud docker feature_idea waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME ansible-modules-core/cloud/docker/docker_login.py ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Currently, I'm not using the docker_login module, I am exporting the docker login details via terminal. DOCKERCLOUD_USER=username DOCKERCLOUD_PASS=password DOCKERCLOUD_NAMESPACE=organization --> ##### OS / ENVIRONMENT <!--- Ubuntu 16.04.01 LTS --> ##### SUMMARY I am trying to log into my organization using the docker_login module. Docker Hub and Cloud now both have Organizations which improve your ability to control who can create, edit or delete Docker Hub repositories. ##### STEPS TO REPRODUCE <!--- I would just specify DOCKERCLOUD_NAMESPACE in the config file that docker would have by default. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: Log into DockerHub docker_login: username: docker password: rekcod email: docker@docker.io namespace: docker_organization ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS I would expect that Docker now logs into my personal account but uses the organization that I am linked to and not my own personal account for pushing images or doing any tasks Docker related on the machine I used the docker_login module. This way I can effectively work with multiple teams that have their own repositories and I can effectively deploy Docker Images that are private from different teams. ##### ACTUAL RESULTS This isn't currently possible.
True
Adding Namespace as an editable parameter for the docker_login module. - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME ansible-modules-core/cloud/docker/docker_login.py ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Currently, I'm not using the docker_login module, I am exporting the docker login details via terminal. DOCKERCLOUD_USER=username DOCKERCLOUD_PASS=password DOCKERCLOUD_NAMESPACE=organization --> ##### OS / ENVIRONMENT <!--- Ubuntu 16.04.01 LTS --> ##### SUMMARY I am trying to log into my organization using the docker_login module. Docker Hub and Cloud now both have Organizations which improve your ability to control who can create, edit or delete Docker Hub repositories. ##### STEPS TO REPRODUCE <!--- I would just specify DOCKERCLOUD_NAMESPACE in the config file that docker would have by default. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: Log into DockerHub docker_login: username: docker password: rekcod email: docker@docker.io namespace: docker_organization ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS I would expect that Docker now logs into my personal account but uses the organization that I am linked to and not my own personal account for pushing images or doing any tasks Docker related on the machine I used the docker_login module. This way I can effectively work with multiple teams that have their own repositories and I can effectively deploy Docker Images that are private from different teams. ##### ACTUAL RESULTS This isn't currently possible.
main
adding namespace as an editable parameter for the docker login module issue type feature idea component name ansible modules core cloud docker docker login py ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration currently i m not using the docker login module i am exporting the docker login details via terminal dockercloud user username dockercloud pass password dockercloud namespace organization os environment ubuntu lts summary i am trying to log into my organization using the docker login module docker hub and cloud now both have organizations which improve your ability to control who can create edit or delete docker hub repositories steps to reproduce i would just specify dockercloud namespace in the config file that docker would have by default name log into dockerhub docker login username docker password rekcod email docker docker io namespace docker organization expected results i would expect that docker now logs into my personal account but uses the organization that i am linked to and not my own personal account for pushing images or doing any tasks docker related on the machine i used the docker login module this way i can effectively work with multiple teams that have their own repositories and i can effectively deploy docker images that are private from different teams actual results this isn t currently possible
1
1,185
5,100,927,191
IssuesEvent
2017-01-04 14:01:28
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
opened
minimal install uses scipy
maintainability
Our minimal install now pulls scipy as a dependency. This comes from the griddataformats conda package (which depends on scipy for remapping). To solve this griddataformats would have to be installed from pip for the minimal build. Or we just allow scipy as a dependency.
True
minimal install uses scipy - Our minimal install now pulls scipy as a dependency. This comes from the griddataformats conda package (which depends on scipy for remapping). To solve this griddataformats would have to be installed from pip for the minimal build. Or we just allow scipy as a dependency.
main
minimal install uses scipy our minimal install now pulls scipy as a dependency this comes from the griddataformats conda package which depends on scipy for remapping to solve this griddataformats would have to be installed from pip for the minimal build or we just allow scipy as a dependency
1
121,074
10,149,402,262
IssuesEvent
2019-08-05 15:08:53
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: tpchbench/tpchVec/nodes=3/cpu=4/sf=1 failed
C-test-failure O-roachtest O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/cfdaadc3514e7e8660f6c009ba159fdfd604f0a8 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=tpchbench/tpchVec/nodes=3/cpu=4/sf=1 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1409070&tab=buildLog ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190727-1409070/tpchbench/tpchVec/nodes=3/cpu=4/sf=1/run_1 test_runner.go:706: test timed out (10h0m0s) tpchbench.go:119,cluster.go:2069,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1564208378-06-n4cpu4:4 -- ./workload run querybench --db=tpch --concurrency=1 --query-file=tpchVec --num-runs=3 --max-ops=27 --vectorized=true {pgurl:1-3} --histograms=perf/stats.json --histograms-max-latency=8m20s returned: stderr: stdout: TH AND l_returnflag = 'R' AND c_nationkey = n_nationkey GROUP BY c_custkey, c_name, c_acctbal, c_phone, n_name, c_address, c_comment ORDER BY revenue DESC LIMIT 20 9h57m15s 0 0.0 0.0 0.0 0.0 0.0 0.0 8: SELECT ps_partkey, sum(ps_supplycost * ps_availqty::float) AS value FROM partsupp, supplier, nation WHERE ps_suppkey = s_suppkey AND s_nationkey = n_nationkey AND n_name = 'GERMANY' GROUP BY ps_partkey HAVING sum(ps_supplycost * ps_availqty::float) > ( SELECT sum(ps_supplycost * ps_availqty::float) * 0.0001 FROM partsupp, supplier, nation WHERE ps_suppkey = s_suppkey AND s_nationkey = n_nationkey AND n_name = 'GERMANY') ORDER BY value DESC 9h57m15s 0 0.0 0.0 0.0 0.0 0.0 0.0 9: SELECT sum(l_extendedprice) / 7.0 AS avg_yearly FROM lineitem, part WHERE p_partkey = l_partkey AND p_brand = 'Brand#23' AND p_container = 'MED BOX' AND l_quantity < ( SELECT 0.2 * avg(l_quantity) FROM lineitem WHERE l_partkey = p_partkey) : signal: killed cluster.go:2090,tpchbench.go:123,tpchbench.go:244,test_runner.go:691: Goexit() was called ```
2.0
roachtest: tpchbench/tpchVec/nodes=3/cpu=4/sf=1 failed - SHA: https://github.com/cockroachdb/cockroach/commits/cfdaadc3514e7e8660f6c009ba159fdfd604f0a8 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=tpchbench/tpchVec/nodes=3/cpu=4/sf=1 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1409070&tab=buildLog ``` The test failed on branch=master, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20190727-1409070/tpchbench/tpchVec/nodes=3/cpu=4/sf=1/run_1 test_runner.go:706: test timed out (10h0m0s) tpchbench.go:119,cluster.go:2069,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1564208378-06-n4cpu4:4 -- ./workload run querybench --db=tpch --concurrency=1 --query-file=tpchVec --num-runs=3 --max-ops=27 --vectorized=true {pgurl:1-3} --histograms=perf/stats.json --histograms-max-latency=8m20s returned: stderr: stdout: TH AND l_returnflag = 'R' AND c_nationkey = n_nationkey GROUP BY c_custkey, c_name, c_acctbal, c_phone, n_name, c_address, c_comment ORDER BY revenue DESC LIMIT 20 9h57m15s 0 0.0 0.0 0.0 0.0 0.0 0.0 8: SELECT ps_partkey, sum(ps_supplycost * ps_availqty::float) AS value FROM partsupp, supplier, nation WHERE ps_suppkey = s_suppkey AND s_nationkey = n_nationkey AND n_name = 'GERMANY' GROUP BY ps_partkey HAVING sum(ps_supplycost * ps_availqty::float) > ( SELECT sum(ps_supplycost * ps_availqty::float) * 0.0001 FROM partsupp, supplier, nation WHERE ps_suppkey = s_suppkey AND s_nationkey = n_nationkey AND n_name = 'GERMANY') ORDER BY value DESC 9h57m15s 0 0.0 0.0 0.0 0.0 0.0 0.0 9: SELECT sum(l_extendedprice) / 7.0 AS avg_yearly FROM lineitem, part WHERE p_partkey = l_partkey AND p_brand = 'Brand#23' AND p_container = 'MED BOX' AND l_quantity < ( SELECT 0.2 * avg(l_quantity) FROM lineitem WHERE l_partkey = p_partkey) : signal: killed cluster.go:2090,tpchbench.go:123,tpchbench.go:244,test_runner.go:691: Goexit() was called ```
non_main
roachtest tpchbench tpchvec nodes cpu sf failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests tpchbench tpchvec nodes cpu sf pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch master cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts tpchbench tpchvec nodes cpu sf run test runner go test timed out tpchbench go cluster go errgroup go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run querybench db tpch concurrency query file tpchvec num runs max ops vectorized true pgurl histograms perf stats json histograms max latency returned stderr stdout th and l returnflag r and c nationkey n nationkey group by c custkey c name c acctbal c phone n name c address c comment order by revenue desc limit select ps partkey sum ps supplycost ps availqty float as value from partsupp supplier nation where ps suppkey s suppkey and s nationkey n nationkey and n name germany group by ps partkey having sum ps supplycost ps availqty float select sum ps supplycost ps availqty float from partsupp supplier nation where ps suppkey s suppkey and s nationkey n nationkey and n name germany order by value desc select sum l extendedprice as avg yearly from lineitem part where p partkey l partkey and p brand brand and p container med box and l quantity select avg l quantity from lineitem where l partkey p partkey signal killed cluster go tpchbench go tpchbench go test runner go goexit was called
0
51,390
13,635,116,181
IssuesEvent
2020-09-25 01:56:37
nasifimtiazohi/openmrs-module-reporting-1.20.0
https://api.github.com/repos/nasifimtiazohi/openmrs-module-reporting-1.20.0
opened
CVE-2018-19362 (High) detected in jackson-databind-2.9.0.jar
security vulnerability
## CVE-2018-19362 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: openmrs-module-reporting-1.20.0/api-2.2/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p> <p> Dependency Hierarchy: - openmrs-api-2.2.0.jar (Root Library) - :x: **jackson-databind-2.9.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-module-reporting-1.20.0/commit/43757d56a9ab9f7202e297fea95f1861af41888c">43757d56a9ab9f7202e297fea95f1861af41888c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization. <p>Publish Date: 2019-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19362>CVE-2018-19362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362</a></p> <p>Release Date: 2019-01-02</p> <p>Fix Resolution: 2.9.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-19362 (High) detected in jackson-databind-2.9.0.jar - ## CVE-2018-19362 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: openmrs-module-reporting-1.20.0/api-2.2/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p> <p> Dependency Hierarchy: - openmrs-api-2.2.0.jar (Root Library) - :x: **jackson-databind-2.9.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-module-reporting-1.20.0/commit/43757d56a9ab9f7202e297fea95f1861af41888c">43757d56a9ab9f7202e297fea95f1861af41888c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.8 might allow attackers to have unspecified impact by leveraging failure to block the jboss-common-core class from polymorphic deserialization. <p>Publish Date: 2019-01-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19362>CVE-2018-19362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19362</a></p> <p>Release Date: 2019-01-02</p> <p>Fix Resolution: 2.9.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file openmrs module reporting api pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy openmrs api jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before might allow attackers to have unspecified impact by leveraging failure to block the jboss common core class from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
4,275
21,476,657,365
IssuesEvent
2022-04-26 14:11:16
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
opened
Long-term solution for import sorting
maintainability proposal
## Is your feature request related to a problem? ## As [discussed](https://github.com/MDAnalysis/mdanalysis/pull/3644#issuecomment-1100836585) in #3644 I am always frustrated when I see unsorted imports 🙃. Especially, for a new import, it is not clear where to put it. ## Describe the solution you'd like ## Run `isort` with options (to be discussed) for the complete repo ```toml line_length = 80 indent = 4 multi_line_output = 8 # Vertical Hanging Indent Bracket include_trailing_comma = true lines_after_imports = 2 known_first_party = "MDAnalysis" ``` and add a command like `isort --verbose --check-only --diff` to the CI. ## Describe alternatives you've considered ## Leave everything as it is.
True
Long-term solution for import sorting - ## Is your feature request related to a problem? ## As [discussed](https://github.com/MDAnalysis/mdanalysis/pull/3644#issuecomment-1100836585) in #3644 I am always frustrated when I see unsorted imports 🙃. Especially, for a new import, it is not clear where to put it. ## Describe the solution you'd like ## Run `isort` with options (to be discussed) for the complete repo ```toml line_length = 80 indent = 4 multi_line_output = 8 # Vertical Hanging Indent Bracket include_trailing_comma = true lines_after_imports = 2 known_first_party = "MDAnalysis" ``` and add a command like `isort --verbose --check-only --diff` to the CI. ## Describe alternatives you've considered ## Leave everything as it is.
main
long term solution for import sorting is your feature request related to a problem as in i am always frustrated when i see unsorted imports 🙃 especially for a new import it is not clear where to put it describe the solution you d like run isort with options to be discussed for the complete repo toml line length indent multi line output vertical hanging indent bracket include trailing comma true lines after imports known first party mdanalysis and add a command like isort verbose check only diff to the ci describe alternatives you ve considered leave everything as it is
1
159,690
20,085,892,819
IssuesEvent
2022-02-05 01:08:01
AkshayMukkavilli/Tensorflow
https://api.github.com/repos/AkshayMukkavilli/Tensorflow
opened
CVE-2021-41207 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
security vulnerability
## CVE-2021-41207 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /Tensorflow/src/requirements.txt</p> <p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. In affected versions the implementation of `ParallelConcat` misses some input validation and can produce a division by 0. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range. <p>Publish Date: 2021-11-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41207>CVE-2021-41207</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-7v94-64hj-m82h">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-7v94-64hj-m82h</a></p> <p>Release Date: 2021-11-05</p> <p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-41207 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-41207 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /Tensorflow/src/requirements.txt</p> <p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. In affected versions the implementation of `ParallelConcat` misses some input validation and can produce a division by 0. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range. <p>Publish Date: 2021-11-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41207>CVE-2021-41207</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-7v94-64hj-m82h">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-7v94-64hj-m82h</a></p> <p>Release Date: 2021-11-05</p> <p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file tensorflow src requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source platform for machine learning in affected versions the implementation of parallelconcat misses some input validation and can produce a division by the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
306,675
23,168,976,601
IssuesEvent
2022-07-30 12:02:13
swarm-game/swarm
https://api.github.com/repos/swarm-game/swarm
closed
Documentation for how to author scenarios
Z-Feature C-Low Hanging Fruit S-Critical Z-Documentation
This is potentially related to #346 and #347, but is fundamentally separate. We need some documentation somewhere (e.g. on the [wiki](https://github.com/swarm-game/swarm/wiki)?) explaining the format of the `.yaml` files used to describe scenarios and what all the different options mean.
1.0
Documentation for how to author scenarios - This is potentially related to #346 and #347, but is fundamentally separate. We need some documentation somewhere (e.g. on the [wiki](https://github.com/swarm-game/swarm/wiki)?) explaining the format of the `.yaml` files used to describe scenarios and what all the different options mean.
non_main
documentation for how to author scenarios this is potentially related to and but is fundamentally separate we need some documentation somewhere e g on the explaining the format of the yaml files used to describe scenarios and what all the different options mean
0
808
4,425,771,330
IssuesEvent
2016-08-16 16:20:22
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
AttributeError: 'DockerManager' object has no attribute 'client'
bug_report cloud docker waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME docker module ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### STEPS TO REPRODUCE ``` - name: run the site in a docker container docker: name: app_test env_file: /opt/app/env.conf publish_all_ports: yes cap_add: - "SYS_PTRACE" tty: yes detach: yes volumes: "/usr/share/GeoIP/GeoLiteCity.dat:/usr/share/GeoIP/GeoLiteCity.dat" image: "app_test:{{ RUBY_SEMVER }}" state: started when: RUBY_SEMVER is defined ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 1975, in <module> main() File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 1912, in main manager = DockerManager(module) File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 749, in __init__ self.environment = self.get_environment(env, env_file) File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 895, in get_environment self.ensure_capability('env_file') File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 870, in ensure_capability self._check_capabilities() File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 853, in _check_capabilities api_version = self.client.version()['ApiVersion'] AttributeError: 'DockerManager' object has no attribute 'client' fatal: [srv-1]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "docker"}, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 1975, in <module>\n main()\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 1912, in main\n manager = DockerManager(module)\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 749, in __init__\n self.environment = self.get_environment(env, env_file)\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 895, in get_environment\n self.ensure_capability('env_file')\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 870, in ensure_capability\n self._check_capabilities()\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 853, in _check_capabilities\n api_version = self.client.version()['ApiVersion']\nAttributeError: 'DockerManager' object has no attribute 'client'\n", "module_stdout": "", "msg": "MODULE FAILURE", "parsed": false} ```
True
AttributeError: 'DockerManager' object has no attribute 'client' - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME docker module ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### STEPS TO REPRODUCE ``` - name: run the site in a docker container docker: name: app_test env_file: /opt/app/env.conf publish_all_ports: yes cap_add: - "SYS_PTRACE" tty: yes detach: yes volumes: "/usr/share/GeoIP/GeoLiteCity.dat:/usr/share/GeoIP/GeoLiteCity.dat" image: "app_test:{{ RUBY_SEMVER }}" state: started when: RUBY_SEMVER is defined ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 1975, in <module> main() File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 1912, in main manager = DockerManager(module) File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 749, in __init__ self.environment = self.get_environment(env, env_file) File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 895, in get_environment self.ensure_capability('env_file') File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 870, in ensure_capability self._check_capabilities() File "/tmp/ansible_CBwbjg/ansible_module_docker.py", line 853, in _check_capabilities api_version = self.client.version()['ApiVersion'] AttributeError: 'DockerManager' object has no attribute 'client' fatal: [srv-1]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "docker"}, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 1975, in <module>\n main()\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 1912, in main\n manager = DockerManager(module)\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 749, in __init__\n self.environment = self.get_environment(env, env_file)\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 895, in get_environment\n self.ensure_capability('env_file')\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 870, in ensure_capability\n self._check_capabilities()\n File \"/tmp/ansible_CBwbjg/ansible_module_docker.py\", line 853, in _check_capabilities\n api_version = self.client.version()['ApiVersion']\nAttributeError: 'DockerManager' object has no attribute 'client'\n", "module_stdout": "", "msg": "MODULE FAILURE", "parsed": false} ```
main
attributeerror dockermanager object has no attribute client issue type bug report component name docker module ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables steps to reproduce name run the site in a docker container docker name app test env file opt app env conf publish all ports yes cap add sys ptrace tty yes detach yes volumes usr share geoip geolitecity dat usr share geoip geolitecity dat image app test ruby semver state started when ruby semver is defined actual results an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible cbwbjg ansible module docker py line in main file tmp ansible cbwbjg ansible module docker py line in main manager dockermanager module file tmp ansible cbwbjg ansible module docker py line in init self environment self get environment env env file file tmp ansible cbwbjg ansible module docker py line in get environment self ensure capability env file file tmp ansible cbwbjg ansible module docker py line in ensure capability self check capabilities file tmp ansible cbwbjg ansible module docker py line in check capabilities api version self client version attributeerror dockermanager object has no attribute client fatal failed changed false failed true invocation module name docker module stderr traceback most recent call last n file tmp ansible cbwbjg ansible module docker py line in n main n file tmp ansible cbwbjg ansible module docker py line in main n manager dockermanager module n file tmp ansible cbwbjg ansible module docker py line in init n self environment self get environment env env file n file tmp ansible cbwbjg ansible module docker py line in get environment n self ensure capability env file n file tmp ansible cbwbjg ansible module docker py line in ensure capability n self check capabilities n file tmp ansible cbwbjg ansible module docker py line in check capabilities n api version self client version nattributeerror dockermanager object has no attribute client n module stdout msg module failure parsed false
1
4,711
24,270,835,605
IssuesEvent
2022-09-28 10:07:43
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
SEO | incorrect page found in sitemap.xml 
engineering Maintain
The Same [page identified](https://foundation.mozilla.org/en/blog/why-our-tech-community-is-most-excited-about-mozfest-2022/) with a 5XX code is listed here as well. However, using multiple sitemap validator tools, they were unable to validate the XML sitemap for the domain foundation.mozilla.org - additional attention should be paid to the output for the SITEMAP to ensure it follows the standard with all necessary XML structures.
True
SEO | incorrect page found in sitemap.xml  - The Same [page identified](https://foundation.mozilla.org/en/blog/why-our-tech-community-is-most-excited-about-mozfest-2022/) with a 5XX code is listed here as well. However, using multiple sitemap validator tools, they were unable to validate the XML sitemap for the domain foundation.mozilla.org - additional attention should be paid to the output for the SITEMAP to ensure it follows the standard with all necessary XML structures.
main
seo incorrect page found in sitemap xml  the same with a code is listed here as well however using multiple sitemap validator tools they were unable to validate the xml sitemap for the domain foundation mozilla org additional attention should be paid to the output for the sitemap to ensure it follows the standard with all necessary xml structures
1
4,747
24,489,629,195
IssuesEvent
2022-10-09 22:15:22
rustsec/advisory-db
https://api.github.com/repos/rustsec/advisory-db
closed
`num-format` Status
Unmaintained
_3,232,716 downloads, ~8k a day_ Last release was over three years ago It is using the old version of itoa: https://github.com/rustsec/advisory-db/issues/1404 Ralf was helpful to ping earlier: https://github.com/bcmyers/num-format/issues/29 Maintenance status was asked on 9 Jan 2022 https://github.com/bcmyers/num-format/issues/27 @bcmyers - I wonder if people should be still using this crate today and whether it would be helpful to bump the itoa dependency
True
`num-format` Status - _3,232,716 downloads, ~8k a day_ Last release was over three years ago It is using the old version of itoa: https://github.com/rustsec/advisory-db/issues/1404 Ralf was helpful to ping earlier: https://github.com/bcmyers/num-format/issues/29 Maintenance status was asked on 9 Jan 2022 https://github.com/bcmyers/num-format/issues/27 @bcmyers - I wonder if people should be still using this crate today and whether it would be helpful to bump the itoa dependency
main
num format status downloads a day last release was over three years ago it is using the old version of itoa ralf was helpful to ping earlier maintenance status was asked on jan bcmyers i wonder if people should be still using this crate today and whether it would be helpful to bump the itoa dependency
1
1,052
4,863,765,703
IssuesEvent
2016-11-14 16:14:18
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
template src does not work for roles
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME Module template ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION ``` [defaults] inventory = ./hosts.ini library = ./library forks = 50 gathering = smart roles_path = ./roles vault_password_file = /xxx/ansible_vault_password.txt fact_caching = jsonfile fact_caching_connection = /var/cache/ansible-facts fact_caching_timeout = 86400 var_compression_level = 5 module_compression = 'ZIP_DEFLATED' [privilege_escalation] [paramiko_connection] [ssh_connection] pipelining = True [accelerate] [selinux] [colors] ``` ##### OS / ENVIRONMENT Ubuntu 14.04.5 LTS ##### SUMMARY The module `template` does not search in the `files` directory of a role as it is done by the `copy` module. ##### STEPS TO REPRODUCE I have a file layout as the best practice documentation recommends: site.yml roles/facts/vars/main.yml roles/facts/tasks/main.yml roles/facts/tasks/facts_file.yml roles/facts/files/bash.j2 My `site.yml` includes the role `facts` and `tasks/main.yml` includes `facts_file.yml`. In `facts_file.yml` I use the template module to transfer files to a remote system with the following `src` attribute in a loop: src: "{{item.shell}}.j2" which expands to src: bash.j2 But when I run my `site.yml` I get the error: > IOError: [Errno 2] No such file or directory: u'/home/ziemann/ansible/bash.j2' The file is searched in Ansible's base directory but not in the `files` directory of the role. ##### EXPECTED RESULTS I expect the template module to act on the `src` attribute in the same way as the `copy` module. ##### ACTUAL RESULTS The template module and copy module work in different ways. ##### BTW It seems to me that it is a design error, that there are two different copy modules: one with template expansion and one without. It might be better to merge them together. ##### Workaround? How can I build the path to the file based on the role by myself?
True
template src does not work for roles - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME Module template ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION ``` [defaults] inventory = ./hosts.ini library = ./library forks = 50 gathering = smart roles_path = ./roles vault_password_file = /xxx/ansible_vault_password.txt fact_caching = jsonfile fact_caching_connection = /var/cache/ansible-facts fact_caching_timeout = 86400 var_compression_level = 5 module_compression = 'ZIP_DEFLATED' [privilege_escalation] [paramiko_connection] [ssh_connection] pipelining = True [accelerate] [selinux] [colors] ``` ##### OS / ENVIRONMENT Ubuntu 14.04.5 LTS ##### SUMMARY The module `template` does not search in the `files` directory of a role as it is done by the `copy` module. ##### STEPS TO REPRODUCE I have a file layout as the best practice documentation recommends: site.yml roles/facts/vars/main.yml roles/facts/tasks/main.yml roles/facts/tasks/facts_file.yml roles/facts/files/bash.j2 My `site.yml` includes the role `facts` and `tasks/main.yml` includes `facts_file.yml`. In `facts_file.yml` I use the template module to transfer files to a remote system with the following `src` attribute in a loop: src: "{{item.shell}}.j2" which expands to src: bash.j2 But when I run my `site.yml` I get the error: > IOError: [Errno 2] No such file or directory: u'/home/ziemann/ansible/bash.j2' The file is searched in Ansible's base directory but not in the `files` directory of the role. ##### EXPECTED RESULTS I expect the template module to act on the `src` attribute in the same way as the `copy` module. ##### ACTUAL RESULTS The template module and copy module work in different ways. ##### BTW It seems to me that it is a design error, that there are two different copy modules: one with template expansion and one without. It might be better to merge them together. ##### Workaround? How can I build the path to the file based on the role by myself?
main
template src does not work for roles issue type bug report component name module template ansible version ansible configuration inventory hosts ini library library forks gathering smart roles path roles vault password file xxx ansible vault password txt fact caching jsonfile fact caching connection var cache ansible facts fact caching timeout var compression level module compression zip deflated pipelining true os environment ubuntu lts summary the module template does not search in the files directory of a role as it is done by the copy module steps to reproduce i have a file layout as the best practice documentation recommends site yml roles facts vars main yml roles facts tasks main yml roles facts tasks facts file yml roles facts files bash my site yml includes the role facts and tasks main yml includes facts file yml in facts file yml i use the template module to transfer files to a remote system with the following src attribute in a loop src item shell which expands to src bash but when i run my site yml i get the error ioerror no such file or directory u home ziemann ansible bash the file is searched in ansible s base directory but not in the files directory of the role expected results i expect the template module to act on the src attribute in the same way as the copy module actual results the template module and copy module work in different ways btw it seems to me that it is a design error that there are two different copy modules one with template expansion and one without it might be better to merge them together workaround how can i build the path to the file based on the role by myself
1
3,363
13,033,183,014
IssuesEvent
2020-07-28 06:20:00
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Update Vicino for N-gram clusterer bug fix
bug maintainability
We've got a bug fix for the SIMILE Vicino N-gram clusterer sitting at https://github.com/OpenRefine/simile-vicino/commit/e9e9eda18bf905f5a0ee6c04cc6a1b48d621b8c0 which never got published. We should publish a new version with the bug fix and update OpenRefine to use it. We could use this opportunity to clean up the OpenRefine dependencies a little by: - [x] switching to the official `secondstring` dependency from the original author instead of publishing our own (resolved in https://github.com/OpenRefine/simile-vicino/pull/1/ - [x] switching vicino bzip2 dependency to use Apache commons-compress instead of ant-tools, as we've done for OpenRefine (resolved in https://github.com/OpenRefine/simile-vicino/pull/1/) - [x] updating to arithcode-1.2. This is a very minor release, but it includes tests, which is a plus. (pending in https://github.com/OpenRefine/simile-vicino/pull/2/) - [x] moving the `secondstring` and `arithcode` dependencies from OpenRefine to https://github.com/OpenRefine/simile-vicino
True
Update Vicino for N-gram clusterer bug fix - We've got a bug fix for the SIMILE Vicino N-gram clusterer sitting at https://github.com/OpenRefine/simile-vicino/commit/e9e9eda18bf905f5a0ee6c04cc6a1b48d621b8c0 which never got published. We should publish a new version with the bug fix and update OpenRefine to use it. We could use this opportunity to clean up the OpenRefine dependencies a little by: - [x] switching to the official `secondstring` dependency from the original author instead of publishing our own (resolved in https://github.com/OpenRefine/simile-vicino/pull/1/ - [x] switching vicino bzip2 dependency to use Apache commons-compress instead of ant-tools, as we've done for OpenRefine (resolved in https://github.com/OpenRefine/simile-vicino/pull/1/) - [x] updating to arithcode-1.2. This is a very minor release, but it includes tests, which is a plus. (pending in https://github.com/OpenRefine/simile-vicino/pull/2/) - [x] moving the `secondstring` and `arithcode` dependencies from OpenRefine to https://github.com/OpenRefine/simile-vicino
main
update vicino for n gram clusterer bug fix we ve got a bug fix for the simile vicino n gram clusterer sitting at which never got published we should publish a new version with the bug fix and update openrefine to use it we could use this opportunity to clean up the openrefine dependencies a little by switching to the official secondstring dependency from the original author instead of publishing our own resolved in switching vicino dependency to use apache commons compress instead of ant tools as we ve done for openrefine resolved in updating to arithcode this is a very minor release but it includes tests which is a plus pending in moving the secondstring and arithcode dependencies from openrefine to
1
272
3,040,294,501
IssuesEvent
2015-08-07 14:41:09
simplesamlphp/simplesamlphp
https://api.github.com/repos/simplesamlphp/simplesamlphp
closed
Extract the metaedit module out of the repository
enhancement low maintainability
It should get its own repository and allow installation through composer.
True
Extract the metaedit module out of the repository - It should get its own repository and allow installation through composer.
main
extract the metaedit module out of the repository it should get its own repository and allow installation through composer
1
3,080
2,536,590,014
IssuesEvent
2015-01-26 15:09:28
JMurk/Snowblower_Issues
https://api.github.com/repos/JMurk/Snowblower_Issues
opened
Snow - Color Blind Accessibility
enhancement low priority
**Future Enhancement** Per Amy, there is a color blind user. He will not be able to distinguish anything on the map as-is now. We should develop something that allows him to be able to use the map effectively.
1.0
Snow - Color Blind Accessibility - **Future Enhancement** Per Amy, there is a color blind user. He will not be able to distinguish anything on the map as-is now. We should develop something that allows him to be able to use the map effectively.
non_main
snow color blind accessibility future enhancement per amy there is a color blind user he will not be able to distinguish anything on the map as is now we should develop something that allows him to be able to use the map effectively
0
417,349
28,110,369,032
IssuesEvent
2023-03-31 06:36:33
euph00/ped
https://api.github.com/repos/euph00/ped
opened
command summary table in UG is out of date
type.DocumentationBug severity.Low
delete command suspected to refer to delete_patient find command suspected to refer to find_patient find_details command suspected to refer to find_patient details etc ![bug 8.png](https://raw.githubusercontent.com/euph00/ped/main/files/4476e732-d12d-4109-a7a2-740d27a32e63.png) ![bug 8(1).png](https://raw.githubusercontent.com/euph00/ped/main/files/4e9ae643-3f6f-4a26-a185-fba8c156de6c.png) <!--session: 1680241960933-9b72ee5c-285f-4f6d-9a4f-a1ade050869a--> <!--Version: Web v3.4.7-->
1.0
command summary table in UG is out of date - delete command suspected to refer to delete_patient find command suspected to refer to find_patient find_details command suspected to refer to find_patient details etc ![bug 8.png](https://raw.githubusercontent.com/euph00/ped/main/files/4476e732-d12d-4109-a7a2-740d27a32e63.png) ![bug 8(1).png](https://raw.githubusercontent.com/euph00/ped/main/files/4e9ae643-3f6f-4a26-a185-fba8c156de6c.png) <!--session: 1680241960933-9b72ee5c-285f-4f6d-9a4f-a1ade050869a--> <!--Version: Web v3.4.7-->
non_main
command summary table in ug is out of date delete command suspected to refer to delete patient find command suspected to refer to find patient find details command suspected to refer to find patient details etc
0
646,847
21,076,654,515
IssuesEvent
2022-04-02 08:34:32
SE701-T1/backend
https://api.github.com/repos/SE701-T1/backend
closed
Fix CI action workflow
Priority: Low Status: Review Needed Type: Bug
**Describe the task that needs to be done.** <!-- *(If this issue is about a bug, please describe the problem and steps to reproduce the issue. You can also include screenshots of any stack traces, or any other supporting images).* --> The CI workflow for GitHub actions is not being triggered for execution in all of the desired cases. Considering the jobs in the CI workflow, it could be better to have the CI triggered for execution by all push and pull request actions regardless of branch and to have a manually triggered workflow dispatch in case there is an error that is not related to the source code. **Describe how a solution to your proposed task might look like (and any alternatives considered).** A proposed solution is to remove the specified branch names from the pull_request event trigger, and to add push and workflow_dispatch event triggers for any branch: ``` on: push: pull_request: workflow_dispatch: ``` **Notes**
1.0
Fix CI action workflow - **Describe the task that needs to be done.** <!-- *(If this issue is about a bug, please describe the problem and steps to reproduce the issue. You can also include screenshots of any stack traces, or any other supporting images).* --> The CI workflow for GitHub actions is not being triggered for execution in all of the desired cases. Considering the jobs in the CI workflow, it could be better to have the CI triggered for execution by all push and pull request actions regardless of branch and to have a manually triggered workflow dispatch in case there is an error that is not related to the source code. **Describe how a solution to your proposed task might look like (and any alternatives considered).** A proposed solution is to remove the specified branch names from the pull_request event trigger, and to add push and workflow_dispatch event triggers for any branch: ``` on: push: pull_request: workflow_dispatch: ``` **Notes**
non_main
fix ci action workflow describe the task that needs to be done the ci workflow for github actions is not being triggered for execution in all of the desired cases considering the jobs in the ci workflow it could be better to have the ci triggered for execution by all push and pull request actions regardless of branch and to have a manually triggered workflow dispatch in case there is an error that is not related to the source code describe how a solution to your proposed task might look like and any alternatives considered a proposed solution is to remove the specified branch names from the pull request event trigger and to add push and workflow dispatch event triggers for any branch on push pull request workflow dispatch notes
0
176,171
28,039,053,342
IssuesEvent
2023-03-28 17:03:48
pulumi/pulumi-docker
https://api.github.com/repos/pulumi/pulumi-docker
closed
docker:index:Image could not open dockerfile at relative path Dockerfile
kind/bug resolution/by-design
### What happened? Upgraded from `pulumi-docker` `4.0.0` to `4.1.0`. `pulumi up` unexpectedly errors: ``` Diagnostics: docker:index:Image (my-app:latest): error: could not open dockerfile at relative path Dockerfile. Try setting `dockerfile` to "/Users/shed/Repos/my-repo/projects/my-app/Dockerfile" ``` ### Expected Behavior `pulumi up` to locate `Dockerfile` relative to my project directory. ### Steps to reproduce My pulumi resource looks something like: ```python image = docker.Image( "my_app:latest", build=docker.DockerBuildArgs( context="/Users/shed/Repos/my-repo/projects/my-app/", dockerfile="Dockerfile", args={"BUILDKIT_INLINE_CACHE": "1"}, # may not be necessary? cache_from=docker.CacheFromArgs( images=[ pulumi.Output.concat(repo.repository_url, ":", "latest") ] ), platform="linux/amd64", ), registry=registry, image_name=pulumi.Output.concat(self.repo.repository_url, ":", "latest"), ) ``` There is a file `/Users/shed/Repos/my-repo/projects/my-app/Dockerfile`. `context` is set to the directory and `dockerfile` is explicitly set to `Dockerfile`. ### Output of `pulumi about` I don't want to copy/paste this from a work project. I'll try to provide specific info as requested if relevant. ### Additional context Split from #566 ### Contributing Vote on this issue by adding a 👍 reaction. To contribute a fix for this issue, leave a comment (and link to your pull request, if you've opened one already).
1.0
docker:index:Image could not open dockerfile at relative path Dockerfile - ### What happened? Upgraded from `pulumi-docker` `4.0.0` to `4.1.0`. `pulumi up` unexpectedly errors: ``` Diagnostics: docker:index:Image (my-app:latest): error: could not open dockerfile at relative path Dockerfile. Try setting `dockerfile` to "/Users/shed/Repos/my-repo/projects/my-app/Dockerfile" ``` ### Expected Behavior `pulumi up` to locate `Dockerfile` relative to my project directory. ### Steps to reproduce My pulumi resource looks something like: ```python image = docker.Image( "my_app:latest", build=docker.DockerBuildArgs( context="/Users/shed/Repos/my-repo/projects/my-app/", dockerfile="Dockerfile", args={"BUILDKIT_INLINE_CACHE": "1"}, # may not be necessary? cache_from=docker.CacheFromArgs( images=[ pulumi.Output.concat(repo.repository_url, ":", "latest") ] ), platform="linux/amd64", ), registry=registry, image_name=pulumi.Output.concat(self.repo.repository_url, ":", "latest"), ) ``` There is a file `/Users/shed/Repos/my-repo/projects/my-app/Dockerfile`. `context` is set to the directory and `dockerfile` is explicitly set to `Dockerfile`. ### Output of `pulumi about` I don't want to copy/paste this from a work project. I'll try to provide specific info as requested if relevant. ### Additional context Split from #566 ### Contributing Vote on this issue by adding a 👍 reaction. To contribute a fix for this issue, leave a comment (and link to your pull request, if you've opened one already).
non_main
docker index image could not open dockerfile at relative path dockerfile what happened upgraded from pulumi docker to pulumi up unexpectedly errors diagnostics docker index image my app latest error could not open dockerfile at relative path dockerfile try setting dockerfile to users shed repos my repo projects my app dockerfile expected behavior pulumi up to locate dockerfile relative to my project directory steps to reproduce my pulumi resource looks something like python image docker image my app latest build docker dockerbuildargs context users shed repos my repo projects my app dockerfile dockerfile args buildkit inline cache may not be necessary cache from docker cachefromargs images pulumi output concat repo repository url latest platform linux registry registry image name pulumi output concat self repo repository url latest there is a file users shed repos my repo projects my app dockerfile context is set to the directory and dockerfile is explicitly set to dockerfile output of pulumi about i don t want to copy paste this from a work project i ll try to provide specific info as requested if relevant additional context split from contributing vote on this issue by adding a 👍 reaction to contribute a fix for this issue leave a comment and link to your pull request if you ve opened one already
0
91,265
15,856,386,602
IssuesEvent
2021-04-08 02:13:23
n-devs/NodeJSControUI
https://api.github.com/repos/n-devs/NodeJSControUI
opened
CVE-2021-23337 (High) detected in lodash-4.17.5.tgz
security vulnerability
## CVE-2021-23337 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.5.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz</a></p> <p>Path to dependency file: /NodeJSControUI/package.json</p> <p>Path to vulnerable library: NodeJSControUI/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - jsdom-11.6.2.tgz (Root Library) - request-promise-native-1.0.5.tgz - request-promise-core-1.1.1.tgz - :x: **lodash-4.17.5.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23337 (High) detected in lodash-4.17.5.tgz - ## CVE-2021-23337 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.5.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz</a></p> <p>Path to dependency file: /NodeJSControUI/package.json</p> <p>Path to vulnerable library: NodeJSControUI/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - jsdom-11.6.2.tgz (Root Library) - request-promise-native-1.0.5.tgz - request-promise-core-1.1.1.tgz - :x: **lodash-4.17.5.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file nodejscontroui package json path to vulnerable library nodejscontroui node modules lodash package json dependency hierarchy jsdom tgz root library request promise native tgz request promise core tgz x lodash tgz vulnerable library vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
0
5,707
8,367,726,522
IssuesEvent
2018-10-04 13:05:10
SAEONData/ckanext-metadata
https://api.github.com/repos/SAEONData/ckanext-metadata
closed
Metadata record uniqueness
requirement
An attempt to create a metadata record with the same DOI and download link as an existing one, should be processed as an update to that existing record. A match on DOI but not on download link, or vice versa, should be interpreted as an error.
1.0
Metadata record uniqueness - An attempt to create a metadata record with the same DOI and download link as an existing one, should be processed as an update to that existing record. A match on DOI but not on download link, or vice versa, should be interpreted as an error.
non_main
metadata record uniqueness an attempt to create a metadata record with the same doi and download link as an existing one should be processed as an update to that existing record a match on doi but not on download link or vice versa should be interpreted as an error
0
344,070
10,339,838,277
IssuesEvent
2019-09-03 20:20:30
knative/docs
https://api.github.com/repos/knative/docs
closed
Lacking guidance on nuances of writing controllers
kind/new-docs-needed priority/1
I'd like us to write down or link to some of the accumulated wisdom and best practices learned about how to write Kubernetes controllers. This will make onboarding new contributors easier (see https://github.com/knative/eventing/pull/308#discussion_r207280167) and help keep old contributors honest (and informed) about why we do things a certain way. In particular, I think these questions need well-supported answers: - When should status be updated during reconcile? - What counts as a reconcile error (i.e., what errors should requeue the resource)? - When is it ok to update the spec of a resource? - Can I update a different resource during reconcile? When these answers already exist in https://github.com/kubernetes/community/tree/master/contributors/devel (https://github.com/kubernetes/community/blob/master/contributors/devel/api-conventions.md and https://github.com/kubernetes/community/blob/master/contributors/devel/controllers.md probably have some great info), we should link to them, since those documents are hard to find otherwise. /cc @pmorie @mattmoor @dprotaso @evankanderson @n3wscott
1.0
Lacking guidance on nuances of writing controllers - I'd like us to write down or link to some of the accumulated wisdom and best practices learned about how to write Kubernetes controllers. This will make onboarding new contributors easier (see https://github.com/knative/eventing/pull/308#discussion_r207280167) and help keep old contributors honest (and informed) about why we do things a certain way. In particular, I think these questions need well-supported answers: - When should status be updated during reconcile? - What counts as a reconcile error (i.e., what errors should requeue the resource)? - When is it ok to update the spec of a resource? - Can I update a different resource during reconcile? When these answers already exist in https://github.com/kubernetes/community/tree/master/contributors/devel (https://github.com/kubernetes/community/blob/master/contributors/devel/api-conventions.md and https://github.com/kubernetes/community/blob/master/contributors/devel/controllers.md probably have some great info), we should link to them, since those documents are hard to find otherwise. /cc @pmorie @mattmoor @dprotaso @evankanderson @n3wscott
non_main
lacking guidance on nuances of writing controllers i d like us to write down or link to some of the accumulated wisdom and best practices learned about how to write kubernetes controllers this will make onboarding new contributors easier see and help keep old contributors honest and informed about why we do things a certain way in particular i think these questions need well supported answers when should status be updated during reconcile what counts as a reconcile error i e what errors should requeue the resource when is it ok to update the spec of a resource can i update a different resource during reconcile when these answers already exist in and probably have some great info we should link to them since those documents are hard to find otherwise cc pmorie mattmoor dprotaso evankanderson
0
44,999
18,294,573,972
IssuesEvent
2021-10-05 19:02:43
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Update CTMA GIS Data / Add Automation to Process
Type: Data Service: Geo Workgroup: ATD
CMTA updated their data again in TX Open Data Portal, update all layers in GISDM, them AGOL.
1.0
Update CTMA GIS Data / Add Automation to Process - CMTA updated their data again in TX Open Data Portal, update all layers in GISDM, them AGOL.
non_main
update ctma gis data add automation to process cmta updated their data again in tx open data portal update all layers in gisdm them agol
0
295,400
22,215,131,647
IssuesEvent
2022-06-08 00:17:34
aws/aws-sdk-go-v2
https://api.github.com/repos/aws/aws-sdk-go-v2
opened
README badges are not searchable with ctrl + F
documentation needs-triage
### Describe the issue Reopening #1293. If I do ctrl + F on the landing page and search for "migration" it doesn't find the badge in the results, because the badge is an image not text. This makes it hard to find the migration guide unless you already know it's a badge (so in essence https://github.com/aws/aws-sdk-go-v2/issues/1289 isn't truly resolved). ### Links https://github.com/aws/aws-sdk-go-v2/blob/main/README.md ### AWS Go SDK V2 Module Versions Used n/a
1.0
README badges are not searchable with ctrl + F - ### Describe the issue Reopening #1293. If I do ctrl + F on the landing page and search for "migration" it doesn't find the badge in the results, because the badge is an image not text. This makes it hard to find the migration guide unless you already know it's a badge (so in essence https://github.com/aws/aws-sdk-go-v2/issues/1289 isn't truly resolved). ### Links https://github.com/aws/aws-sdk-go-v2/blob/main/README.md ### AWS Go SDK V2 Module Versions Used n/a
non_main
readme badges are not searchable with ctrl f describe the issue reopening if i do ctrl f on the landing page and search for migration it doesn t find the badge in the results because the badge is an image not text this makes it hard to find the migration guide unless you already know it s a badge so in essence isn t truly resolved links aws go sdk module versions used n a
0
3,913
17,480,659,881
IssuesEvent
2021-08-09 01:13:23
DLR-RM/rl-baselines3-zoo
https://api.github.com/repos/DLR-RM/rl-baselines3-zoo
closed
"MySQLdb._exceptions.OperationalError: (1054, "Unknown column 'infe0' in 'field list'")
Maintainers on vacation
I've recently been using RL Baselines Zoo and I'm getting an error message that I'm not able to comprehend, and is almost certainly a bug. I've attached the giant traceback as a txt file to this. I'd be willing to fix it if I actually understood what was happening because this is holding up moderately important work. After some Googling, the issue appears to be somehow related to the inf showing up in the logs: https://github.com/pandas-dev/pandas/issues/34431 [magical_error.txt](https://github.com/DLR-RM/rl-baselines3-zoo/files/6919284/magical_error.txt)
True
"MySQLdb._exceptions.OperationalError: (1054, "Unknown column 'infe0' in 'field list'") - I've recently been using RL Baselines Zoo and I'm getting an error message that I'm not able to comprehend, and is almost certainly a bug. I've attached the giant traceback as a txt file to this. I'd be willing to fix it if I actually understood what was happening because this is holding up moderately important work. After some Googling, the issue appears to be somehow related to the inf showing up in the logs: https://github.com/pandas-dev/pandas/issues/34431 [magical_error.txt](https://github.com/DLR-RM/rl-baselines3-zoo/files/6919284/magical_error.txt)
main
mysqldb exceptions operationalerror unknown column in field list i ve recently been using rl baselines zoo and i m getting an error message that i m not able to comprehend and is almost certainly a bug i ve attached the giant traceback as a txt file to this i d be willing to fix it if i actually understood what was happening because this is holding up moderately important work after some googling the issue appears to be somehow related to the inf showing up in the logs
1
4,281
21,527,840,627
IssuesEvent
2022-04-28 20:24:07
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Error when trying to load tables
type: enhancement work: frontend status: ready restricted: maintainers
## Reproduce 1. Go to `http://localhost:8000/mathesar_tables/2/?t=W1tdLG51bGxd` to load the "astronomy" schema. 1. Switch to the "public" schema. 1. Observe a 500 response from `/api/db/v0/tables/` with error code 4999 and the following message: > "Got KeyError when attempting to get a value for field `input` on serializer `BooleanDisplayOptionSerializer`.\nThe serializer field might be named incorrectly and not match any attribute or key on the `dict` instance.\nOriginal exception text was: 'input'."
True
Error when trying to load tables - ## Reproduce 1. Go to `http://localhost:8000/mathesar_tables/2/?t=W1tdLG51bGxd` to load the "astronomy" schema. 1. Switch to the "public" schema. 1. Observe a 500 response from `/api/db/v0/tables/` with error code 4999 and the following message: > "Got KeyError when attempting to get a value for field `input` on serializer `BooleanDisplayOptionSerializer`.\nThe serializer field might be named incorrectly and not match any attribute or key on the `dict` instance.\nOriginal exception text was: 'input'."
main
error when trying to load tables reproduce go to to load the astronomy schema switch to the public schema observe a response from api db tables with error code and the following message got keyerror when attempting to get a value for field input on serializer booleandisplayoptionserializer nthe serializer field might be named incorrectly and not match any attribute or key on the dict instance noriginal exception text was input
1
27,514
5,357,755,517
IssuesEvent
2017-02-20 19:38:36
webgme/webgme
https://api.github.com/repos/webgme/webgme
opened
Revise API Documentation
Documentation Minor REST API
- Bearer security scheme should be explained. - Fix wrong/outdated responses - Add security where applicable
1.0
Revise API Documentation - - Bearer security scheme should be explained. - Fix wrong/outdated responses - Add security where applicable
non_main
revise api documentation bearer security scheme should be explained fix wrong outdated responses add security where applicable
0
2,500
8,655,458,140
IssuesEvent
2018-11-27 16:00:19
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
no ID3 tags shown for music
unmaintained
I cant use the album list or anything other than all songs even though id3 tags are verry orderly
True
no ID3 tags shown for music - I cant use the album list or anything other than all songs even though id3 tags are verry orderly
main
no tags shown for music i cant use the album list or anything other than all songs even though tags are verry orderly
1
190,044
15,215,956,697
IssuesEvent
2021-02-17 14:59:16
fduquesne/poker-planning
https://api.github.com/repos/fduquesne/poker-planning
opened
Changer le README.md
documentation
Changer le contenu du README.md pour le spécifier à notre application. Y placer aussi les https://shields.io/
1.0
Changer le README.md - Changer le contenu du README.md pour le spécifier à notre application. Y placer aussi les https://shields.io/
non_main
changer le readme md changer le contenu du readme md pour le spécifier à notre application y placer aussi les
0
100,352
8,736,849,846
IssuesEvent
2018-12-11 20:45:28
Princeton-CDH/ppa-django
https://api.github.com/repos/Princeton-CDH/ppa-django
closed
As an admin, I want to suppress items from the site so that I can pull content that should not be included or was wrongly added as I am going through and assigning collections to archive volumes.
awaiting testing
## Notes for testing - [x] digitized works now have a status field; default is public, you can set manually to suppressed - [x] should see an indicator on the list view if something is public or suppressed - [x] should be able to filter the list view on status - [x] status should be included in CSV export - when you set a record to suppressed the data should be deleted from the hathitrust pairtree data so we don't actually import and index it again (not sure how you can test this; you could ask us to run the hathi import script on the source id?) - [x] if you try to switch a suppressed record back to public, you should get a validation error because it's not yet supported ## Notes for development We don't want to actually delete the record from the database; we'll want to keep a stub at least, to indicate the record was removed and track the history. - [x] add a status field; options public/suppressed, default to public - [x] make editable in admin - [x] display status in the admin list view so removed items are obvious; also configure as a filter. - [x] Include status field in CSV export - [x] when status is changed to suppressed, delete rsync data so it won't be re-added/indexed on a full import - [x] don't allow un-suppressing items (validation? pre-save hook?) ### out of scope - We may eventually want a bulk removal option, but consider that out of scope for now. - Supporting "un-suppress" logic is out of scope for now.
1.0
As an admin, I want to suppress items from the site so that I can pull content that should not be included or was wrongly added as I am going through and assigning collections to archive volumes. - ## Notes for testing - [x] digitized works now have a status field; default is public, you can set manually to suppressed - [x] should see an indicator on the list view if something is public or suppressed - [x] should be able to filter the list view on status - [x] status should be included in CSV export - when you set a record to suppressed the data should be deleted from the hathitrust pairtree data so we don't actually import and index it again (not sure how you can test this; you could ask us to run the hathi import script on the source id?) - [x] if you try to switch a suppressed record back to public, you should get a validation error because it's not yet supported ## Notes for development We don't want to actually delete the record from the database; we'll want to keep a stub at least, to indicate the record was removed and track the history. - [x] add a status field; options public/suppressed, default to public - [x] make editable in admin - [x] display status in the admin list view so removed items are obvious; also configure as a filter. - [x] Include status field in CSV export - [x] when status is changed to suppressed, delete rsync data so it won't be re-added/indexed on a full import - [x] don't allow un-suppressing items (validation? pre-save hook?) ### out of scope - We may eventually want a bulk removal option, but consider that out of scope for now. - Supporting "un-suppress" logic is out of scope for now.
non_main
as an admin i want to suppress items from the site so that i can pull content that should not be included or was wrongly added as i am going through and assigning collections to archive volumes notes for testing digitized works now have a status field default is public you can set manually to suppressed should see an indicator on the list view if something is public or suppressed should be able to filter the list view on status status should be included in csv export when you set a record to suppressed the data should be deleted from the hathitrust pairtree data so we don t actually import and index it again not sure how you can test this you could ask us to run the hathi import script on the source id if you try to switch a suppressed record back to public you should get a validation error because it s not yet supported notes for development we don t want to actually delete the record from the database we ll want to keep a stub at least to indicate the record was removed and track the history add a status field options public suppressed default to public make editable in admin display status in the admin list view so removed items are obvious also configure as a filter include status field in csv export when status is changed to suppressed delete rsync data so it won t be re added indexed on a full import don t allow un suppressing items validation pre save hook out of scope we may eventually want a bulk removal option but consider that out of scope for now supporting un suppress logic is out of scope for now
0
5,452
27,288,408,339
IssuesEvent
2023-02-23 15:01:55
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Handle 404 pages throughout the app
type: bug work: frontend status: ready restricted: maintainers
## Description * Client side routing: A non existent url results in a blank screen, or has obscure errors. * Server side routing: A non existent url entered on the browser results in a 404, which returns a django 404 page. ## Expected behavior * 404s should be clearly presented to the user in all valid cases. * The 404 page should be consistent in appearance in both server side & client side routing.
True
Handle 404 pages throughout the app - ## Description * Client side routing: A non existent url results in a blank screen, or has obscure errors. * Server side routing: A non existent url entered on the browser results in a 404, which returns a django 404 page. ## Expected behavior * 404s should be clearly presented to the user in all valid cases. * The 404 page should be consistent in appearance in both server side & client side routing.
main
handle pages throughout the app description client side routing a non existent url results in a blank screen or has obscure errors server side routing a non existent url entered on the browser results in a which returns a django page expected behavior should be clearly presented to the user in all valid cases the page should be consistent in appearance in both server side client side routing
1
2,534
12,228,232,895
IssuesEvent
2020-05-03 18:31:32
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
ARM template not successfully deploying
Pri2 automation/svc cxp product-question triaged update-management/subsvc
Hi there, I'm having an issue with the ARM file on this page. It fails deployment, it creates the automation account, creates the workspace, links the automation account with the workspace. All successfully. Finally it tries to perform some sort of update on the automation account which appears in the deployment list as "Updates(workspacename)" and returns an error "Not Found" and "Bad Request". It's a bit vague. The powershell error isn't very helpful either: New-AzResourceGroupDeployment : 15:17:41 - Resource Microsoft.OperationsManagement/solutions 'Updates(workspacename)' failed with message '{ "error": { "code": "BadRequest", "message": "" } }' At line:1 char:1 If I can provide any more information please let me know. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: c0123bad-47d3-a6bc-95b4-95aea681ad95 * Version Independent ID: acaf27f7-1d41-954d-f4f2-5065ffd896d6 * Content: [Use Azure Resource Manager templates to onboard Update Management](https://docs.microsoft.com/en-us/azure/automation/automation-update-management-deploy-template#feedback) * Content Source: [articles/automation/automation-update-management-deploy-template.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-update-management-deploy-template.md) * Service: **automation** * Sub-service: **update-management** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
ARM template not successfully deploying - Hi there, I'm having an issue with the ARM file on this page. It fails deployment, it creates the automation account, creates the workspace, links the automation account with the workspace. All successfully. Finally it tries to perform some sort of update on the automation account which appears in the deployment list as "Updates(workspacename)" and returns an error "Not Found" and "Bad Request". It's a bit vague. The powershell error isn't very helpful either: New-AzResourceGroupDeployment : 15:17:41 - Resource Microsoft.OperationsManagement/solutions 'Updates(workspacename)' failed with message '{ "error": { "code": "BadRequest", "message": "" } }' At line:1 char:1 If I can provide any more information please let me know. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: c0123bad-47d3-a6bc-95b4-95aea681ad95 * Version Independent ID: acaf27f7-1d41-954d-f4f2-5065ffd896d6 * Content: [Use Azure Resource Manager templates to onboard Update Management](https://docs.microsoft.com/en-us/azure/automation/automation-update-management-deploy-template#feedback) * Content Source: [articles/automation/automation-update-management-deploy-template.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-update-management-deploy-template.md) * Service: **automation** * Sub-service: **update-management** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
non_main
arm template not successfully deploying hi there i m having an issue with the arm file on this page it fails deployment it creates the automation account creates the workspace links the automation account with the workspace all successfully finally it tries to perform some sort of update on the automation account which appears in the deployment list as updates workspacename and returns an error not found and bad request it s a bit vague the powershell error isn t very helpful either new azresourcegroupdeployment resource microsoft operationsmanagement solutions updates workspacename failed with message error code badrequest message at line char if i can provide any more information please let me know document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service update management github login mgoedtel microsoft alias magoedte
0
3,948
17,873,129,737
IssuesEvent
2021-09-06 19:41:32
antigenomics/vdjdb-db
https://api.github.com/repos/antigenomics/vdjdb-db
opened
TRAJ24*01 vs TRAJ24*02
maintainance formatting&proofreading
Reported by Jamie Heather About 2/3 of the ~400 human TCR entries reportedly making use of TRAJ24*01 are actually probably using TRAJ24*02. There's a two nt mismatch between them that corresponds to a two AA FGXG-proximal mismatch: ``` TTDSWGKFEFGAGTQVVVTP TRAJ24*01 TTDSWGKLQFGAGTQVVVTP TRAJ24*02 ^^ ``` It'd take a lot of deletion to make *02 into *01 it's pretty likely that most of the '-WGKLQF' CDR3s (which is the majority) are actually *02.
True
TRAJ24*01 vs TRAJ24*02 - Reported by Jamie Heather About 2/3 of the ~400 human TCR entries reportedly making use of TRAJ24*01 are actually probably using TRAJ24*02. There's a two nt mismatch between them that corresponds to a two AA FGXG-proximal mismatch: ``` TTDSWGKFEFGAGTQVVVTP TRAJ24*01 TTDSWGKLQFGAGTQVVVTP TRAJ24*02 ^^ ``` It'd take a lot of deletion to make *02 into *01 it's pretty likely that most of the '-WGKLQF' CDR3s (which is the majority) are actually *02.
main
vs reported by jamie heather about of the human tcr entries reportedly making use of are actually probably using there s a two nt mismatch between them that corresponds to a two aa fgxg proximal mismatch ttdswgkfefgagtqvvvtp ttdswgklqfgagtqvvvtp it d take a lot of deletion to make into it s pretty likely that most of the wgklqf which is the majority are actually
1
4,847
24,973,526,603
IssuesEvent
2022-11-02 04:54:57
rthadur/bazel
https://api.github.com/repos/rthadur/bazel
closed
Test 3
awaiting-review awaiting-user-response more data needed awaiting-maintainer
### Description Test 3 ### Issue Type _No response_ ### Operating System _No response_ ### Coral Device _No response_ ### Other Devices _No response_ ### Programming Language _No response_ ### Relevant Log Output _No response_
True
Test 3 - ### Description Test 3 ### Issue Type _No response_ ### Operating System _No response_ ### Coral Device _No response_ ### Other Devices _No response_ ### Programming Language _No response_ ### Relevant Log Output _No response_
main
test description test issue type no response operating system no response coral device no response other devices no response programming language no response relevant log output no response
1
963
4,706,289,311
IssuesEvent
2016-10-13 16:42:58
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ansible-modules-core/network/ - Code review
affects_2.2 bug_report in progress networking P1 waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME eos_facts ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 (devel 70e63ddf6c) last updated 2016/09/15 10:17:19 (GMT +100) lib/ansible/modules/core: (devel 683e5e4d1a) last updated 2016/09/15 10:17:22 (GMT +100) lib/ansible/modules/extras: (devel 170adf16bd) last updated 2016/09/15 10:17:23 (GMT +100) ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY I've raised one issue to track all the issues found rather than having a fairly bitty chain of tickets. If it's easier for you to raise different PRs to address the issues found I'm no issue with that - whatever is easiest for you. I'm wondering if for items we are happy with we should add ignore markers in, as shown here http://stackoverflow.com/questions/28829236/is-it-possible-to-ignore-one-single-specific-line-with-pylint ``` pylint -E network/*/* No config file found, using default configuration ************* Module ansible.modules.core.network.nxos.nxos_hsrp E:402,41: Undefined variable 'module' (undefined-variable) ************* Module ansible.modules.core.network.nxos.nxos_interface E:147,19: Instance of 'list' has no 'split' member (no-member) E:535,56: Undefined variable 'command' (undefined-variable) E:581,13: Undefined variable 'get_module' (undefined-variable) ************* Module ansible.modules.core.network.nxos.nxos_static_route E:158,23: Instance of 'CustomNetworkConfig' has no 'to_lines' member (no-member) E:295,26: Instance of 'list' has no 'split' member (no-member) E:402,66: Using variable 'address' before assignment (used-before-assignment) ************* Module ansible.modules.core.network.nxos.nxos_switchport E:486,56: Undefined variable 'command' (undefined-variable) E:527,13: Undefined variable 'get_module' (undefined-variable) ************* Module ansible.modules.core.network.nxos.nxos_vrf E:501,52: Undefined variable 'cmds' (undefined-variable) ``` ##### STEPS TO REPRODUCE ##### EXPECTED RESULTS ##### ACTUAL RESULTS
True
ansible-modules-core/network/ - Code review - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME eos_facts ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 (devel 70e63ddf6c) last updated 2016/09/15 10:17:19 (GMT +100) lib/ansible/modules/core: (devel 683e5e4d1a) last updated 2016/09/15 10:17:22 (GMT +100) lib/ansible/modules/extras: (devel 170adf16bd) last updated 2016/09/15 10:17:23 (GMT +100) ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY I've raised one issue to track all the issues found rather than having a fairly bitty chain of tickets. If it's easier for you to raise different PRs to address the issues found I'm no issue with that - whatever is easiest for you. I'm wondering if for items we are happy with we should add ignore markers in, as shown here http://stackoverflow.com/questions/28829236/is-it-possible-to-ignore-one-single-specific-line-with-pylint ``` pylint -E network/*/* No config file found, using default configuration ************* Module ansible.modules.core.network.nxos.nxos_hsrp E:402,41: Undefined variable 'module' (undefined-variable) ************* Module ansible.modules.core.network.nxos.nxos_interface E:147,19: Instance of 'list' has no 'split' member (no-member) E:535,56: Undefined variable 'command' (undefined-variable) E:581,13: Undefined variable 'get_module' (undefined-variable) ************* Module ansible.modules.core.network.nxos.nxos_static_route E:158,23: Instance of 'CustomNetworkConfig' has no 'to_lines' member (no-member) E:295,26: Instance of 'list' has no 'split' member (no-member) E:402,66: Using variable 'address' before assignment (used-before-assignment) ************* Module ansible.modules.core.network.nxos.nxos_switchport E:486,56: Undefined variable 'command' (undefined-variable) E:527,13: Undefined variable 'get_module' (undefined-variable) ************* Module ansible.modules.core.network.nxos.nxos_vrf E:501,52: Undefined variable 'cmds' (undefined-variable) ``` ##### STEPS TO REPRODUCE ##### EXPECTED RESULTS ##### ACTUAL RESULTS
main
ansible modules core network code review issue type bug report component name eos facts ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt configuration os environment summary i ve raised one issue to track all the issues found rather than having a fairly bitty chain of tickets if it s easier for you to raise different prs to address the issues found i m no issue with that whatever is easiest for you i m wondering if for items we are happy with we should add ignore markers in as shown here pylint e network no config file found using default configuration module ansible modules core network nxos nxos hsrp e undefined variable module undefined variable module ansible modules core network nxos nxos interface e instance of list has no split member no member e undefined variable command undefined variable e undefined variable get module undefined variable module ansible modules core network nxos nxos static route e instance of customnetworkconfig has no to lines member no member e instance of list has no split member no member e using variable address before assignment used before assignment module ansible modules core network nxos nxos switchport e undefined variable command undefined variable e undefined variable get module undefined variable module ansible modules core network nxos nxos vrf e undefined variable cmds undefined variable steps to reproduce expected results actual results
1
1,915
6,577,706,212
IssuesEvent
2017-09-12 02:44:55
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
AWS Route53 limited in options for setting up routes (weighted latency not available)
affects_2.0 aws cloud feature_idea waiting_on_maintainer
##### Issue Type: - Feature Idea ##### Plugin Name: - cloud/amazon/route53 ##### Ansible Version: ansible 2.0.0.2 config file = /home/db/.ansible.cfg configured module search path = Default w/o overrides ##### Ansible Configuration: None ##### Environment: Ubuntu 12.04 running on a VirtualBox VM. ##### Summary: I have manually built some infrastructure in AWS and I want to automate the build of that with Ansible. This manually built infrastructure uses weighted latency based routing. However, I cannot read back the information fully from the facts nor is there any option for setting this up. Speaking with @defionscode at Ansiblefest 2016 last Thursday, I was asked to raise this issue for a fix. ##### Steps To Reproduce: Not a bug - feature request. ##### Expected Results: Simply trying to get the results back of the current routes to reflect the configuration available through the console. ##### Actual Results: Not applicable.
True
AWS Route53 limited in options for setting up routes (weighted latency not available) - ##### Issue Type: - Feature Idea ##### Plugin Name: - cloud/amazon/route53 ##### Ansible Version: ansible 2.0.0.2 config file = /home/db/.ansible.cfg configured module search path = Default w/o overrides ##### Ansible Configuration: None ##### Environment: Ubuntu 12.04 running on a VirtualBox VM. ##### Summary: I have manually built some infrastructure in AWS and I want to automate the build of that with Ansible. This manually built infrastructure uses weighted latency based routing. However, I cannot read back the information fully from the facts nor is there any option for setting this up. Speaking with @defionscode at Ansiblefest 2016 last Thursday, I was asked to raise this issue for a fix. ##### Steps To Reproduce: Not a bug - feature request. ##### Expected Results: Simply trying to get the results back of the current routes to reflect the configuration available through the console. ##### Actual Results: Not applicable.
main
aws limited in options for setting up routes weighted latency not available issue type feature idea plugin name cloud amazon ansible version ansible config file home db ansible cfg configured module search path default w o overrides ansible configuration none environment ubuntu running on a virtualbox vm summary i have manually built some infrastructure in aws and i want to automate the build of that with ansible this manually built infrastructure uses weighted latency based routing however i cannot read back the information fully from the facts nor is there any option for setting this up speaking with defionscode at ansiblefest last thursday i was asked to raise this issue for a fix steps to reproduce not a bug feature request expected results simply trying to get the results back of the current routes to reflect the configuration available through the console actual results not applicable
1
214,824
24,120,463,020
IssuesEvent
2022-09-20 18:13:17
microsoft/Microsoft365DSC
https://api.github.com/repos/microsoft/Microsoft365DSC
closed
Security and Compliance settings - SCLabelPolicy workload
Enhancement Security & Compliance Center
Wondering if there are any plans to include the "Auto-labelling" policies as part of this tool? ![image](https://user-images.githubusercontent.com/78377887/142685800-a48b15fb-8e7c-4e15-9dac-a802c9c70602.png)
True
Security and Compliance settings - SCLabelPolicy workload - Wondering if there are any plans to include the "Auto-labelling" policies as part of this tool? ![image](https://user-images.githubusercontent.com/78377887/142685800-a48b15fb-8e7c-4e15-9dac-a802c9c70602.png)
non_main
security and compliance settings sclabelpolicy workload wondering if there are any plans to include the auto labelling policies as part of this tool
0
202,058
23,053,921,543
IssuesEvent
2022-07-25 01:17:05
Xi0ngfei/e-mart-backend
https://api.github.com/repos/Xi0ngfei/e-mart-backend
opened
spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar: 1 vulnerabilities (highest severity is: 9.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar</b></p></summary> <p></p> <p>Path to dependency file: /emart-eureka-service/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/5.0.3/woodstox-core-5.0.3.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Xi0ngfei/e-mart-backend/commit/9e2cdf0fabfba0aa30b3a80420cea42d1b714754">9e2cdf0fabfba0aa30b3a80420cea42d1b714754</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [WS-2018-0629](https://github.com/FasterXML/woodstox/commit/7937f97c638ef8afd385ebf4a675a9b096ccdd57) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | woodstox-core-5.0.3.jar | Transitive | 2.2.0.RELEASE | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0629</summary> ### Vulnerable Library - <b>woodstox-core-5.0.3.jar</b></p> <p>Woodstox is a high-performance XML processor that implements Stax (JSR-173), SAX2 and Stax2 APIs</p> <p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p> <p>Path to dependency file: /emart-eureka-service/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/5.0.3/woodstox-core-5.0.3.jar</p> <p> Dependency Hierarchy: - spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar (Root Library) - spring-cloud-netflix-eureka-server-2.1.0.RELEASE.jar - jackson-dataformat-xml-2.9.7.jar - :x: **woodstox-core-5.0.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Xi0ngfei/e-mart-backend/commit/9e2cdf0fabfba0aa30b3a80420cea42d1b714754">9e2cdf0fabfba0aa30b3a80420cea42d1b714754</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The woodstox-core package is vulnerable to improper restriction of XXE reference. <p>Publish Date: 2018-08-23 <p>URL: <a href=https://github.com/FasterXML/woodstox/commit/7937f97c638ef8afd385ebf4a675a9b096ccdd57>WS-2018-0629</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-08-23</p> <p>Fix Resolution (com.fasterxml.woodstox:woodstox-core): 5.2.1</p> <p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-netflix-eureka-server): 2.2.0.RELEASE</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar: 1 vulnerabilities (highest severity is: 9.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar</b></p></summary> <p></p> <p>Path to dependency file: /emart-eureka-service/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/5.0.3/woodstox-core-5.0.3.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Xi0ngfei/e-mart-backend/commit/9e2cdf0fabfba0aa30b3a80420cea42d1b714754">9e2cdf0fabfba0aa30b3a80420cea42d1b714754</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [WS-2018-0629](https://github.com/FasterXML/woodstox/commit/7937f97c638ef8afd385ebf4a675a9b096ccdd57) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | woodstox-core-5.0.3.jar | Transitive | 2.2.0.RELEASE | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0629</summary> ### Vulnerable Library - <b>woodstox-core-5.0.3.jar</b></p> <p>Woodstox is a high-performance XML processor that implements Stax (JSR-173), SAX2 and Stax2 APIs</p> <p>Library home page: <a href="https://github.com/FasterXML/woodstox">https://github.com/FasterXML/woodstox</a></p> <p>Path to dependency file: /emart-eureka-service/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/woodstox/woodstox-core/5.0.3/woodstox-core-5.0.3.jar</p> <p> Dependency Hierarchy: - spring-cloud-starter-netflix-eureka-server-2.1.0.RELEASE.jar (Root Library) - spring-cloud-netflix-eureka-server-2.1.0.RELEASE.jar - jackson-dataformat-xml-2.9.7.jar - :x: **woodstox-core-5.0.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Xi0ngfei/e-mart-backend/commit/9e2cdf0fabfba0aa30b3a80420cea42d1b714754">9e2cdf0fabfba0aa30b3a80420cea42d1b714754</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The woodstox-core package is vulnerable to improper restriction of XXE reference. <p>Publish Date: 2018-08-23 <p>URL: <a href=https://github.com/FasterXML/woodstox/commit/7937f97c638ef8afd385ebf4a675a9b096ccdd57>WS-2018-0629</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-08-23</p> <p>Fix Resolution (com.fasterxml.woodstox:woodstox-core): 5.2.1</p> <p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-netflix-eureka-server): 2.2.0.RELEASE</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_main
spring cloud starter netflix eureka server release jar vulnerabilities highest severity is vulnerable library spring cloud starter netflix eureka server release jar path to dependency file emart eureka service pom xml path to vulnerable library home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high woodstox core jar transitive release details ws vulnerable library woodstox core jar woodstox is a high performance xml processor that implements stax jsr and apis library home page a href path to dependency file emart eureka service pom xml path to vulnerable library home wss scanner repository com fasterxml woodstox woodstox core woodstox core jar dependency hierarchy spring cloud starter netflix eureka server release jar root library spring cloud netflix eureka server release jar jackson dataformat xml jar x woodstox core jar vulnerable library found in head commit a href found in base branch master vulnerability details the woodstox core package is vulnerable to improper restriction of xxe reference publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml woodstox woodstox core direct dependency fix resolution org springframework cloud spring cloud starter netflix eureka server release step up your open source security game with mend
0
176,398
28,087,475,263
IssuesEvent
2023-03-30 10:46:14
audacity/audacity
https://api.github.com/repos/audacity/audacity
reopened
Inconsistency with Linear dB view at default track size
bug Design / UX
### Bug description **Default track size** When using the new **Linear (dB)** view in the Vertical Scale with stereo tracks at **_default_** size, there is an odd **-3dB** shown only in the uppper half of the right channel. The left channel and the lower half of the right channel do not have this **-3dB** labelling. This looks odd and unbalanced: ![image](https://user-images.githubusercontent.com/50205984/225017039-6fe3461b-f1ce-4ca1-afc1-54ea9875f837.png) **Track size reduced slightly from default** If you reduce the height of the stereo track a little then you get a balnced Vertical Scale display: ![image](https://user-images.githubusercontent.com/50205984/225017634-e025223f-5c56-4add-9d33-ea4bf23f7f28.png) **Track size expanded slightly from default** If you expand it slightly more than the defaul stereo track size teh you get a different unbalanced view with the -3db labelling in just the upper half of both tracks: ![image](https://user-images.githubusercontent.com/50205984/225018611-24287627-cedb-49ac-8491-552e2e87cf40.png) **Track size expanded a lot from default** Expanding the track size further (image below is full default window size) gives a nicely balanced view: ![image](https://user-images.githubusercontent.com/50205984/225019340-b0081371-51d8-4219-8d0c-e024ac96ae81.png) ### Steps to reproduce 1. clear audaicty settings folder 2. launch 3.3.0 alpha 3. record a stero track - or import one 4. right-click in the Vertical Scale and select **Linear (dB)** 5. Observe: only the upper half of the right channel has a **-3db** marker ### Expected behavior Consistency. If **-3dB** is to be shown, then it should be shown in both halves of both channels ### Actual behavior See above ### Audacity Version 3.3.0 alpha - audacity-win-3.3.0-alpha-20230310+9b9b935-x64-msvc2022 ### Operating system Windows 11 - but assume all OS ### Additional context This is on my smaller Zurich W11 HP Probook with a 13.5 inch screen. _I can retest this next week back in Manchester on my 17 inch W10 HP Envy - and on my Macbook Pro_ **Importance** Although this looks like a trivial display/cosmetic issue, it is somewhat important as it mars an otherwise useful and nice new feature added by GSoC22 participant @micpap25
1.0
Inconsistency with Linear dB view at default track size - ### Bug description **Default track size** When using the new **Linear (dB)** view in the Vertical Scale with stereo tracks at **_default_** size, there is an odd **-3dB** shown only in the uppper half of the right channel. The left channel and the lower half of the right channel do not have this **-3dB** labelling. This looks odd and unbalanced: ![image](https://user-images.githubusercontent.com/50205984/225017039-6fe3461b-f1ce-4ca1-afc1-54ea9875f837.png) **Track size reduced slightly from default** If you reduce the height of the stereo track a little then you get a balnced Vertical Scale display: ![image](https://user-images.githubusercontent.com/50205984/225017634-e025223f-5c56-4add-9d33-ea4bf23f7f28.png) **Track size expanded slightly from default** If you expand it slightly more than the defaul stereo track size teh you get a different unbalanced view with the -3db labelling in just the upper half of both tracks: ![image](https://user-images.githubusercontent.com/50205984/225018611-24287627-cedb-49ac-8491-552e2e87cf40.png) **Track size expanded a lot from default** Expanding the track size further (image below is full default window size) gives a nicely balanced view: ![image](https://user-images.githubusercontent.com/50205984/225019340-b0081371-51d8-4219-8d0c-e024ac96ae81.png) ### Steps to reproduce 1. clear audaicty settings folder 2. launch 3.3.0 alpha 3. record a stero track - or import one 4. right-click in the Vertical Scale and select **Linear (dB)** 5. Observe: only the upper half of the right channel has a **-3db** marker ### Expected behavior Consistency. If **-3dB** is to be shown, then it should be shown in both halves of both channels ### Actual behavior See above ### Audacity Version 3.3.0 alpha - audacity-win-3.3.0-alpha-20230310+9b9b935-x64-msvc2022 ### Operating system Windows 11 - but assume all OS ### Additional context This is on my smaller Zurich W11 HP Probook with a 13.5 inch screen. _I can retest this next week back in Manchester on my 17 inch W10 HP Envy - and on my Macbook Pro_ **Importance** Although this looks like a trivial display/cosmetic issue, it is somewhat important as it mars an otherwise useful and nice new feature added by GSoC22 participant @micpap25
non_main
inconsistency with linear db view at default track size bug description default track size when using the new linear db view in the vertical scale with stereo tracks at default size there is an odd shown only in the uppper half of the right channel the left channel and the lower half of the right channel do not have this labelling this looks odd and unbalanced track size reduced slightly from default if you reduce the height of the stereo track a little then you get a balnced vertical scale display track size expanded slightly from default if you expand it slightly more than the defaul stereo track size teh you get a different unbalanced view with the labelling in just the upper half of both tracks track size expanded a lot from default expanding the track size further image below is full default window size gives a nicely balanced view steps to reproduce clear audaicty settings folder launch alpha record a stero track or import one right click in the vertical scale and select linear db observe only the upper half of the right channel has a marker expected behavior consistency if is to be shown then it should be shown in both halves of both channels actual behavior see above audacity version alpha audacity win alpha operating system windows but assume all os additional context this is on my smaller zurich hp probook with a inch screen i can retest this next week back in manchester on my inch hp envy and on my macbook pro importance although this looks like a trivial display cosmetic issue it is somewhat important as it mars an otherwise useful and nice new feature added by participant
0
799,290
28,303,979,374
IssuesEvent
2023-04-10 09:06:49
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
chrome.google.com - see bug description
browser-chrome priority-critical
<!-- @browser: Chrome 111.0.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/120632 --> **URL**: https://chrome.google.com/webstore/detail/mozilla-rally/bahhehaddofgkccippmjcecepdakppme/related **Browser / Version**: Chrome 111.0.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Firefox **Problem type**: Something else **Description**: Cant install Rally extension in chrome **Steps to Reproduce**: Cant install Mozilla Rally on Chrome browser even when logged in to google account <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2023/4/64757b61-aed8-4301-a1f2-3949ae03be55.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
chrome.google.com - see bug description - <!-- @browser: Chrome 111.0.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/120632 --> **URL**: https://chrome.google.com/webstore/detail/mozilla-rally/bahhehaddofgkccippmjcecepdakppme/related **Browser / Version**: Chrome 111.0.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Firefox **Problem type**: Something else **Description**: Cant install Rally extension in chrome **Steps to Reproduce**: Cant install Mozilla Rally on Chrome browser even when logged in to google account <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2023/4/64757b61-aed8-4301-a1f2-3949ae03be55.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
chrome google com see bug description url browser version chrome operating system windows tested another browser yes firefox problem type something else description cant install rally extension in chrome steps to reproduce cant install mozilla rally on chrome browser even when logged in to google account view the screenshot img alt screenshot src browser configuration none from with ❤️
0
42,359
6,975,511,753
IssuesEvent
2017-12-12 07:27:06
php-deal/framework
https://api.github.com/repos/php-deal/framework
closed
Library name in Scrutinizer and Packagist
documentation enhancement
To be consistent, scrutinizer-ci.com/g/**lisachenko/php-deal** and packagist.org/packages/**lisachenko/php-deal** should be renamed or a new project should be created :)
1.0
Library name in Scrutinizer and Packagist - To be consistent, scrutinizer-ci.com/g/**lisachenko/php-deal** and packagist.org/packages/**lisachenko/php-deal** should be renamed or a new project should be created :)
non_main
library name in scrutinizer and packagist to be consistent scrutinizer ci com g lisachenko php deal and packagist org packages lisachenko php deal should be renamed or a new project should be created
0
3,418
13,182,092,276
IssuesEvent
2020-08-12 15:15:16
duo-labs/cloudmapper
https://api.github.com/repos/duo-labs/cloudmapper
closed
Network visualization: Filter by subnets
map unmaintained_functionality
It would be great to have a flag that enables to filter the prepare command output w.r.t. to the subnet(s) just like VPC.
True
Network visualization: Filter by subnets - It would be great to have a flag that enables to filter the prepare command output w.r.t. to the subnet(s) just like VPC.
main
network visualization filter by subnets it would be great to have a flag that enables to filter the prepare command output w r t to the subnet s just like vpc
1
95,467
10,880,769,452
IssuesEvent
2019-11-17 13:31:10
DiscipleTools/disciple-tools-theme
https://api.github.com/repos/DiscipleTools/disciple-tools-theme
closed
GULP browser sync optimization Improvement
bug documentation
Developers who desire to improve the DiscipleTools theme, may request for live reloading from a local development standpoint. This needs to be fixed internally, and also documented within the README.md for other developers to use. GULP, and JavaScript coding is required.
1.0
GULP browser sync optimization Improvement - Developers who desire to improve the DiscipleTools theme, may request for live reloading from a local development standpoint. This needs to be fixed internally, and also documented within the README.md for other developers to use. GULP, and JavaScript coding is required.
non_main
gulp browser sync optimization improvement developers who desire to improve the discipletools theme may request for live reloading from a local development standpoint this needs to be fixed internally and also documented within the readme md for other developers to use gulp and javascript coding is required
0
2,328
8,345,162,625
IssuesEvent
2018-10-01 00:10:32
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Airlocks use static types for tool interaction instead of the tool_behaviour system
Consistency Issue Maintainability/Hinders improvements
https://github.com/tgstation/tgstation/blob/master/code/game/machinery/doors/door.dm#L183 Which means they're not weldable with custom welders such as, currently, plasma cutters, and not crowbarable by anything but crowbars and fireaxes.
True
Airlocks use static types for tool interaction instead of the tool_behaviour system - https://github.com/tgstation/tgstation/blob/master/code/game/machinery/doors/door.dm#L183 Which means they're not weldable with custom welders such as, currently, plasma cutters, and not crowbarable by anything but crowbars and fireaxes.
main
airlocks use static types for tool interaction instead of the tool behaviour system which means they re not weldable with custom welders such as currently plasma cutters and not crowbarable by anything but crowbars and fireaxes
1
3,699
15,098,796,194
IssuesEvent
2021-02-08 00:18:49
pypiserver/pypiserver
https://api.github.com/repos/pypiserver/pypiserver
opened
Speed up test-pypy job in GH actions
difficulty.easy good first issue type.Maintainance
The test-pypy job is _much_ slower than the others, largely because its `pip install` on the "Install dependencies" step takes two orders of magnitude longer due to its needing to compile wheels. Investigate whether we can cache the built wheels for pypy to avoid needing to rebuild them every time.
True
Speed up test-pypy job in GH actions - The test-pypy job is _much_ slower than the others, largely because its `pip install` on the "Install dependencies" step takes two orders of magnitude longer due to its needing to compile wheels. Investigate whether we can cache the built wheels for pypy to avoid needing to rebuild them every time.
main
speed up test pypy job in gh actions the test pypy job is much slower than the others largely because its pip install on the install dependencies step takes two orders of magnitude longer due to its needing to compile wheels investigate whether we can cache the built wheels for pypy to avoid needing to rebuild them every time
1
2,111
7,176,917,820
IssuesEvent
2018-01-31 11:44:11
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
opened
Analyzer should report "wrong" kinds of exceptions being thrown
analyzer feature maintainability
The analyzer should report that there are exception thrown (created) which are not acceptable to be created. Such exceptions are: - System.NullReferenceException - System.Exception - System.FatalEngineException
True
Analyzer should report "wrong" kinds of exceptions being thrown - The analyzer should report that there are exception thrown (created) which are not acceptable to be created. Such exceptions are: - System.NullReferenceException - System.Exception - System.FatalEngineException
main
analyzer should report wrong kinds of exceptions being thrown the analyzer should report that there are exception thrown created which are not acceptable to be created such exceptions are system nullreferenceexception system exception system fatalengineexception
1
4,399
22,588,082,467
IssuesEvent
2022-06-28 16:59:35
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[Bug]: Using appendTo with DatePicker causes carriage returns to not work in sibling textarea
type: bug 🐛 status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬
### Package carbon-components-react ### Browser Chrome ### Package version 7.57.1 ### React version 16.14.0 ### Description Some time ago, I reported an issue that when the calendar popup was open, it didn't scroll with the input field. https://github.com/carbon-design-system/carbon/issues/4158. The appendTo prop had been broken, and when it was fixed, I set it to the immediate child of the scroll element per the suggestion given for https://github.com/carbon-design-system/carbon/issues/4818. At some point later, we realized that the TextArea that was in the same "appendTo" as the DatePicker was getting it's carriage returns swallowed by flatpickr. Flatpickr had a PR that at first looked like it might fix it https://github.com/flatpickr/flatpickr/issues/2054 but it has since been merged and it doesn't help. I started to write a defect against Flatpickr but found that I couldn't reproduce the scrolling issue when there's no appendTo, and it doesn't have the carriage return issue when the appendTo is set to a child of the scroll container. Note that my teammate wrote a similar issue here. https://github.com/carbon-design-system/carbon/issues/11640 However, I think this is a more straight-forward use case. ### Reproduction/example https://9qiwkg.csb.app/ ### Steps to reproduce Put your cursor in the text area field and hit Enter. The cursor does not move to the next line. ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
True
[Bug]: Using appendTo with DatePicker causes carriage returns to not work in sibling textarea - ### Package carbon-components-react ### Browser Chrome ### Package version 7.57.1 ### React version 16.14.0 ### Description Some time ago, I reported an issue that when the calendar popup was open, it didn't scroll with the input field. https://github.com/carbon-design-system/carbon/issues/4158. The appendTo prop had been broken, and when it was fixed, I set it to the immediate child of the scroll element per the suggestion given for https://github.com/carbon-design-system/carbon/issues/4818. At some point later, we realized that the TextArea that was in the same "appendTo" as the DatePicker was getting it's carriage returns swallowed by flatpickr. Flatpickr had a PR that at first looked like it might fix it https://github.com/flatpickr/flatpickr/issues/2054 but it has since been merged and it doesn't help. I started to write a defect against Flatpickr but found that I couldn't reproduce the scrolling issue when there's no appendTo, and it doesn't have the carriage return issue when the appendTo is set to a child of the scroll container. Note that my teammate wrote a similar issue here. https://github.com/carbon-design-system/carbon/issues/11640 However, I think this is a more straight-forward use case. ### Reproduction/example https://9qiwkg.csb.app/ ### Steps to reproduce Put your cursor in the text area field and hit Enter. The cursor does not move to the next line. ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
main
using appendto with datepicker causes carriage returns to not work in sibling textarea package carbon components react browser chrome package version react version description some time ago i reported an issue that when the calendar popup was open it didn t scroll with the input field the appendto prop had been broken and when it was fixed i set it to the immediate child of the scroll element per the suggestion given for at some point later we realized that the textarea that was in the same appendto as the datepicker was getting it s carriage returns swallowed by flatpickr flatpickr had a pr that at first looked like it might fix it but it has since been merged and it doesn t help i started to write a defect against flatpickr but found that i couldn t reproduce the scrolling issue when there s no appendto and it doesn t have the carriage return issue when the appendto is set to a child of the scroll container note that my teammate wrote a similar issue here however i think this is a more straight forward use case reproduction example steps to reproduce put your cursor in the text area field and hit enter the cursor does not move to the next line code of conduct i agree to follow this project s i checked the for duplicate problems
1
341,541
30,591,655,455
IssuesEvent
2023-07-21 17:36:17
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix tensor.test_torch_instance_arccos
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix tensor.test_torch_instance_arccos - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5624547597/job/15241590549"><img src=https://img.shields.io/badge/-failure-red></a>
non_main
fix tensor test torch instance arccos tensorflow a href src jax a href src numpy a href src torch a href src paddle a href src
0
3,042
11,277,788,696
IssuesEvent
2020-01-15 04:13:00
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
terraform plan does not provide output for 2.9.0
affects_2.9 bug cloud has_pr module needs_maintainer needs_triage support:community
##### SUMMARY Running terraform plan from ansible 2.9.0 does not provide expected output in stdout. This may be similar to issue #46589. ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME terraform ##### ANSIBLE VERSION ``` # ansible --version ansible 2.9.0 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/dist-packages/ansible executable location = /usr/bin/ansible python version = 2.7.15+ (default, Oct 7 2019, 17:39:04) [GCC 7.4.0] ``` ##### CONFIGURATION ``` # ansible-config dump --only-changed ALLOW_WORLD_READABLE_TMPFILES(/etc/ansible/ansible.cfg) = True ANSIBLE_NOCOWS(/etc/ansible/ansible.cfg) = True DEFAULT_GATHERING(/etc/ansible/ansible.cfg) = explicit DEFAULT_STDOUT_CALLBACK(/etc/ansible/ansible.cfg) = debug HOST_KEY_CHECKING(/etc/ansible/ansible.cfg) = False INTERPRETER_PYTHON(/etc/ansible/ansible.cfg) = auto_silent ``` ##### OS / ENVIRONMENT - Ubuntu 18.04.3 LTS ``` # uname -a Linux 57f42cc031d6 4.15.0-66-generic #75-Ubuntu SMP Tue Oct 1 05:24:09 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux # terraform --version Terraform v0.12.13 # pip freeze ansible==2.9.0 asn1crypto==0.24.0 cryptography==2.1.4 enum34==1.1.6 httplib2==0.9.2 idna==2.6 ipaddress==1.0.17 Jinja2==2.10 keyring==10.6.0 keyrings.alt==3.0 MarkupSafe==1.0 paramiko==2.0.0 pyasn1==0.4.2 pycrypto==2.6.1 pygobject==3.26.1 pyxdg==0.25 PyYAML==3.12 SecretStorage==2.3.1 six==1.11.0 # pip3 freeze asn1crypto==0.24.0 awscli==1.16.272 boto==2.49.0 boto3==1.10.8 botocore==1.13.8 chardet==3.0.4 colorama==0.4.1 cryptography==2.1.4 docutils==0.15.2 idna==2.6 jmespath==0.9.4 keyring==10.6.0 keyrings.alt==3.0 pyasn1==0.4.7 pycrypto==2.6.1 pygobject==3.26.1 python-apt==1.6.4 python-dateutil==2.8.1 python-debian==0.1.32 pyxdg==0.25 PyYAML==5.1.2 rsa==3.4.2 s3transfer==0.2.1 SecretStorage==2.3.1 six==1.11.0 unattended-upgrades==0.1 urllib3==1.25.6 virtualenv==15.1.0 ``` ##### STEPS TO REPRODUCE Run terraform plan from ansible ``` ############ # Run terraform plan - name: Run terraform plan for VPC resources terraform: state: planned project_path: "{{ fileTerraformWorkingPath }}" plan_file: "{{ fileTerraformWorkingPath }}/plan.tfplan" force_init: yes backend_config: region: "{{ nameAWSRegion }}" register: vpc_tf_stack ############ # Print information about the base VPC - name: Display everything with terraform debug: var: vpc_tf_stack - name: Display all terraform output for VPC debug: var: vpc_tf_stack.stdout ``` ##### EXPECTED RESULTS Terraform plan stdout is shown. ##### ACTUAL RESULTS Nothing is shown in stdout. ``` TASK [planTerraform : Display everything with terraform] ************************************************************************************************************************************* ok: [127.0.0.1] => { "vpc_tf_stack": { "changed": false, "command": "/usr/bin/terraform plan -input=false -no-color -detailed-exitcode -out /tmp/terraform-20191104141619238338988/plan.tfplan /tmp/terraform-20191104141619238338988/plan.tfplan", "failed": false, "outputs": { "idAdminHostedZone": { "sensitive": false, "type": "string", "value": "Z23423423423423423423" }, "idExternalSecurityGroup": { "sensitive": false, "type": "string", "value": "sg-12312312312312312" }, "idIGW": { "sensitive": false, "type": "string", "value": "igw-12312312312312312" }, "idLocalHostedZone": { "sensitive": false, "type": "string", "value": "Z12312312312312312312" }, "idPublicRouteTable": { "sensitive": false, "type": "string", "value": "rtb-12312312312312312" }, "idVPC": { "sensitive": false, "type": "string", "value": "vpc-12312312312312312" } }, "state": "planned", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": [], "workspace": "default" } } TASK [planTerraform : Display all terraform output for VPC] ********************************************************************************************************************************** ok: [127.0.0.1] => { "vpc_tf_stack.stdout": "" } ```
True
terraform plan does not provide output for 2.9.0 - ##### SUMMARY Running terraform plan from ansible 2.9.0 does not provide expected output in stdout. This may be similar to issue #46589. ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME terraform ##### ANSIBLE VERSION ``` # ansible --version ansible 2.9.0 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/dist-packages/ansible executable location = /usr/bin/ansible python version = 2.7.15+ (default, Oct 7 2019, 17:39:04) [GCC 7.4.0] ``` ##### CONFIGURATION ``` # ansible-config dump --only-changed ALLOW_WORLD_READABLE_TMPFILES(/etc/ansible/ansible.cfg) = True ANSIBLE_NOCOWS(/etc/ansible/ansible.cfg) = True DEFAULT_GATHERING(/etc/ansible/ansible.cfg) = explicit DEFAULT_STDOUT_CALLBACK(/etc/ansible/ansible.cfg) = debug HOST_KEY_CHECKING(/etc/ansible/ansible.cfg) = False INTERPRETER_PYTHON(/etc/ansible/ansible.cfg) = auto_silent ``` ##### OS / ENVIRONMENT - Ubuntu 18.04.3 LTS ``` # uname -a Linux 57f42cc031d6 4.15.0-66-generic #75-Ubuntu SMP Tue Oct 1 05:24:09 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux # terraform --version Terraform v0.12.13 # pip freeze ansible==2.9.0 asn1crypto==0.24.0 cryptography==2.1.4 enum34==1.1.6 httplib2==0.9.2 idna==2.6 ipaddress==1.0.17 Jinja2==2.10 keyring==10.6.0 keyrings.alt==3.0 MarkupSafe==1.0 paramiko==2.0.0 pyasn1==0.4.2 pycrypto==2.6.1 pygobject==3.26.1 pyxdg==0.25 PyYAML==3.12 SecretStorage==2.3.1 six==1.11.0 # pip3 freeze asn1crypto==0.24.0 awscli==1.16.272 boto==2.49.0 boto3==1.10.8 botocore==1.13.8 chardet==3.0.4 colorama==0.4.1 cryptography==2.1.4 docutils==0.15.2 idna==2.6 jmespath==0.9.4 keyring==10.6.0 keyrings.alt==3.0 pyasn1==0.4.7 pycrypto==2.6.1 pygobject==3.26.1 python-apt==1.6.4 python-dateutil==2.8.1 python-debian==0.1.32 pyxdg==0.25 PyYAML==5.1.2 rsa==3.4.2 s3transfer==0.2.1 SecretStorage==2.3.1 six==1.11.0 unattended-upgrades==0.1 urllib3==1.25.6 virtualenv==15.1.0 ``` ##### STEPS TO REPRODUCE Run terraform plan from ansible ``` ############ # Run terraform plan - name: Run terraform plan for VPC resources terraform: state: planned project_path: "{{ fileTerraformWorkingPath }}" plan_file: "{{ fileTerraformWorkingPath }}/plan.tfplan" force_init: yes backend_config: region: "{{ nameAWSRegion }}" register: vpc_tf_stack ############ # Print information about the base VPC - name: Display everything with terraform debug: var: vpc_tf_stack - name: Display all terraform output for VPC debug: var: vpc_tf_stack.stdout ``` ##### EXPECTED RESULTS Terraform plan stdout is shown. ##### ACTUAL RESULTS Nothing is shown in stdout. ``` TASK [planTerraform : Display everything with terraform] ************************************************************************************************************************************* ok: [127.0.0.1] => { "vpc_tf_stack": { "changed": false, "command": "/usr/bin/terraform plan -input=false -no-color -detailed-exitcode -out /tmp/terraform-20191104141619238338988/plan.tfplan /tmp/terraform-20191104141619238338988/plan.tfplan", "failed": false, "outputs": { "idAdminHostedZone": { "sensitive": false, "type": "string", "value": "Z23423423423423423423" }, "idExternalSecurityGroup": { "sensitive": false, "type": "string", "value": "sg-12312312312312312" }, "idIGW": { "sensitive": false, "type": "string", "value": "igw-12312312312312312" }, "idLocalHostedZone": { "sensitive": false, "type": "string", "value": "Z12312312312312312312" }, "idPublicRouteTable": { "sensitive": false, "type": "string", "value": "rtb-12312312312312312" }, "idVPC": { "sensitive": false, "type": "string", "value": "vpc-12312312312312312" } }, "state": "planned", "stderr": "", "stderr_lines": [], "stdout": "", "stdout_lines": [], "workspace": "default" } } TASK [planTerraform : Display all terraform output for VPC] ********************************************************************************************************************************** ok: [127.0.0.1] => { "vpc_tf_stack.stdout": "" } ```
main
terraform plan does not provide output for summary running terraform plan from ansible does not provide expected output in stdout this may be similar to issue issue type bug report component name terraform ansible version ansible version ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib dist packages ansible executable location usr bin ansible python version default oct configuration ansible config dump only changed allow world readable tmpfiles etc ansible ansible cfg true ansible nocows etc ansible ansible cfg true default gathering etc ansible ansible cfg explicit default stdout callback etc ansible ansible cfg debug host key checking etc ansible ansible cfg false interpreter python etc ansible ansible cfg auto silent os environment ubuntu lts uname a linux generic ubuntu smp tue oct utc gnu linux terraform version terraform pip freeze ansible cryptography idna ipaddress keyring keyrings alt markupsafe paramiko pycrypto pygobject pyxdg pyyaml secretstorage six freeze awscli boto botocore chardet colorama cryptography docutils idna jmespath keyring keyrings alt pycrypto pygobject python apt python dateutil python debian pyxdg pyyaml rsa secretstorage six unattended upgrades virtualenv steps to reproduce run terraform plan from ansible run terraform plan name run terraform plan for vpc resources terraform state planned project path fileterraformworkingpath plan file fileterraformworkingpath plan tfplan force init yes backend config region nameawsregion register vpc tf stack print information about the base vpc name display everything with terraform debug var vpc tf stack name display all terraform output for vpc debug var vpc tf stack stdout expected results terraform plan stdout is shown actual results nothing is shown in stdout task ok vpc tf stack changed false command usr bin terraform plan input false no color detailed exitcode out tmp terraform plan tfplan tmp terraform plan tfplan failed false outputs idadminhostedzone sensitive false type string value idexternalsecuritygroup sensitive false type string value sg idigw sensitive false type string value igw idlocalhostedzone sensitive false type string value idpublicroutetable sensitive false type string value rtb idvpc sensitive false type string value vpc state planned stderr stderr lines stdout stdout lines workspace default task ok vpc tf stack stdout
1
640
4,156,813,608
IssuesEvent
2016-06-16 19:11:51
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Code Search: Change API Endpoint
Maintainer Input Requested
The Spice isn't working because the API endpoint has changed to: `https://searchcode.com/api/jsonp_search_IV/?q=php&callback=myCallback` from `https://searchcode.com/api/jsonp_codesearch_I/?q=php&callback=myCallback` Please double check if the data has changed, i.e., if the Spice is showing up properly or not as well! Thanks! --- Maintainer: @boyter IA Page: https://duck.co/ia/view/code_search
True
Code Search: Change API Endpoint - The Spice isn't working because the API endpoint has changed to: `https://searchcode.com/api/jsonp_search_IV/?q=php&callback=myCallback` from `https://searchcode.com/api/jsonp_codesearch_I/?q=php&callback=myCallback` Please double check if the data has changed, i.e., if the Spice is showing up properly or not as well! Thanks! --- Maintainer: @boyter IA Page: https://duck.co/ia/view/code_search
main
code search change api endpoint the spice isn t working because the api endpoint has changed to from please double check if the data has changed i e if the spice is showing up properly or not as well thanks maintainer boyter ia page
1
2,319
8,303,638,825
IssuesEvent
2018-09-21 18:15:57
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
deprecate start/stop/step in AnalysisBase.__init__()
Component-Analysis deprecation maintainability
In preparation for #1463 where we want analysis classes to use `start`, `stop`, `step` in the `run()` method only, we need to deprecate the use in - [x] `analysis.base.AnalysisBase.__init__`: https://github.com/MDAnalysis/mdanalysis/blob/1a1b2f1442d52d1216b53b32c4c42c3b47368d38/package/MDAnalysis/analysis/base.py#L92 - custom analysis classes (`git grep 'start=.*stop='`) - [x] density.py - [x] diffusionmap.py (clean-up only, already uses AnalysisBase) - [x] hbonds/hbond_analysis.py - [x] hbonds/wbridge_analysis.py - [x] rms.py (removed the *deprecation* for start/stop/step in `run()`!)
True
deprecate start/stop/step in AnalysisBase.__init__() - In preparation for #1463 where we want analysis classes to use `start`, `stop`, `step` in the `run()` method only, we need to deprecate the use in - [x] `analysis.base.AnalysisBase.__init__`: https://github.com/MDAnalysis/mdanalysis/blob/1a1b2f1442d52d1216b53b32c4c42c3b47368d38/package/MDAnalysis/analysis/base.py#L92 - custom analysis classes (`git grep 'start=.*stop='`) - [x] density.py - [x] diffusionmap.py (clean-up only, already uses AnalysisBase) - [x] hbonds/hbond_analysis.py - [x] hbonds/wbridge_analysis.py - [x] rms.py (removed the *deprecation* for start/stop/step in `run()`!)
main
deprecate start stop step in analysisbase init in preparation for where we want analysis classes to use start stop step in the run method only we need to deprecate the use in analysis base analysisbase init custom analysis classes git grep start stop density py diffusionmap py clean up only already uses analysisbase hbonds hbond analysis py hbonds wbridge analysis py rms py removed the deprecation for start stop step in run
1
208,265
16,108,320,435
IssuesEvent
2021-04-27 17:37:09
uber/h3
https://api.github.com/repos/uber/h3
closed
Projected coordinate system
documentation
For those functions related to latitude and longitude, is the projected coordinate system used as WGS84 or something?
1.0
Projected coordinate system - For those functions related to latitude and longitude, is the projected coordinate system used as WGS84 or something?
non_main
projected coordinate system for those functions related to latitude and longitude is the projected coordinate system used as or something
0
2,601
8,837,878,921
IssuesEvent
2019-01-05 10:50:33
OXIDprojects/oxid-module-internals
https://api.github.com/repos/OXIDprojects/oxid-module-internals
closed
duplicate code
bug maintainability pull request
As Maintainer I like to remove the effort in maintaining and to increase the quality by removeing duplicate code. - I found duplicate code in CheckConsistency.php
True
duplicate code - As Maintainer I like to remove the effort in maintaining and to increase the quality by removeing duplicate code. - I found duplicate code in CheckConsistency.php
main
duplicate code as maintainer i like to remove the effort in maintaining and to increase the quality by removeing duplicate code i found duplicate code in checkconsistency php
1
5,054
25,888,706,525
IssuesEvent
2022-12-14 16:17:32
deislabs/spiderlightning
https://api.github.com/repos/deislabs/spiderlightning
closed
`slight` panics when the config path isn't a slightfile.
🐛 bug 🚧 maintainer issue
**Description of the bug** When running slight with a config path that isn't a slightfile, `slight` CLI will panic. I found that there are various causes to the issues 1. path does not exist 2. path isn't a slightfile 3. slightfile is mis-configured. **To Reproduce** ``` > ./target/release/slight -c ./examples/http-demo/random_slightfile.toml run -m ./examples/http-demo/target/wasm32-wasi/release/http-demo.wasm thread 'main' panicked at 'called `Option::unwrap()` on a `None` value', slight/src/commands/run.rs:115:34 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace ``` **Additional context**
True
`slight` panics when the config path isn't a slightfile. - **Description of the bug** When running slight with a config path that isn't a slightfile, `slight` CLI will panic. I found that there are various causes to the issues 1. path does not exist 2. path isn't a slightfile 3. slightfile is mis-configured. **To Reproduce** ``` > ./target/release/slight -c ./examples/http-demo/random_slightfile.toml run -m ./examples/http-demo/target/wasm32-wasi/release/http-demo.wasm thread 'main' panicked at 'called `Option::unwrap()` on a `None` value', slight/src/commands/run.rs:115:34 note: run with `RUST_BACKTRACE=1` environment variable to display a backtrace ``` **Additional context**
main
slight panics when the config path isn t a slightfile description of the bug when running slight with a config path that isn t a slightfile slight cli will panic i found that there are various causes to the issues path does not exist path isn t a slightfile slightfile is mis configured to reproduce target release slight c examples http demo random slightfile toml run m examples http demo target wasi release http demo wasm thread main panicked at called option unwrap on a none value slight src commands run rs note run with rust backtrace environment variable to display a backtrace additional context
1
4,788
24,628,591,726
IssuesEvent
2022-10-16 20:38:37
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Destroy TabularData in RecordSelector, TablePage, and TableWidget to prevent Memory leaks
type: bug work: frontend status: ready restricted: maintainers
## Description * There are a lot of places (eg.,RecordSelector, TablePage, and TableWidget) where TabularData object is created reactively, like this: ``` $: tabularData = new TabularData({ id: table.id, abstractTypesMap, meta, }); ``` This creates a new instance each time any of the arguments get updated, and is never destroyed, leading to memory leaks. Each TabularData instance should be destroyed before the next instance is created. If possible, we should consider the following improvements: * The TabularData instance creation is costly and should be kept to a minimum, if possible it should not be made reactive. * We should update a TabularData's property instead of recreating if it is possible.
True
Destroy TabularData in RecordSelector, TablePage, and TableWidget to prevent Memory leaks - ## Description * There are a lot of places (eg.,RecordSelector, TablePage, and TableWidget) where TabularData object is created reactively, like this: ``` $: tabularData = new TabularData({ id: table.id, abstractTypesMap, meta, }); ``` This creates a new instance each time any of the arguments get updated, and is never destroyed, leading to memory leaks. Each TabularData instance should be destroyed before the next instance is created. If possible, we should consider the following improvements: * The TabularData instance creation is costly and should be kept to a minimum, if possible it should not be made reactive. * We should update a TabularData's property instead of recreating if it is possible.
main
destroy tabulardata in recordselector tablepage and tablewidget to prevent memory leaks description there are a lot of places eg recordselector tablepage and tablewidget where tabulardata object is created reactively like this tabulardata new tabulardata id table id abstracttypesmap meta this creates a new instance each time any of the arguments get updated and is never destroyed leading to memory leaks each tabulardata instance should be destroyed before the next instance is created if possible we should consider the following improvements the tabulardata instance creation is costly and should be kept to a minimum if possible it should not be made reactive we should update a tabulardata s property instead of recreating if it is possible
1
499,632
14,474,846,021
IssuesEvent
2020-12-10 00:12:37
Dynna/PBL
https://api.github.com/repos/Dynna/PBL
closed
Insufficient Logging and Monitoring
Backend/PHP/Laravel Priority: HIGH Security
- [x] When it comes to your application and server, log everything, including failed login attempts and password resets. - [ ] Laravel comes with Monolog out of the box. You can even integrate it with a third party logging service like Papertrail and receive alerts for specific log events.
1.0
Insufficient Logging and Monitoring - - [x] When it comes to your application and server, log everything, including failed login attempts and password resets. - [ ] Laravel comes with Monolog out of the box. You can even integrate it with a third party logging service like Papertrail and receive alerts for specific log events.
non_main
insufficient logging and monitoring when it comes to your application and server log everything including failed login attempts and password resets laravel comes with monolog out of the box you can even integrate it with a third party logging service like papertrail and receive alerts for specific log events
0
332,387
10,092,319,707
IssuesEvent
2019-07-26 16:21:13
cBioPortal/cbioportal
https://api.github.com/repos/cBioPortal/cbioportal
closed
Co-expression service performance
api critical priority
Following request can take 20-60 seconds to complete. It's causing some tests to fail. /api/molecular-profiles/co-expressions/fetch?molecularProfileIdA=coadread_tcga_pub_rna_seq_mrna&molecularProfileIdB=coadread_tcga_pub_rna_seq_mrna&threshold=0 Odd that after our first round of optimization, I don't remember it taking so long. perhaps something else happened since then?
1.0
Co-expression service performance - Following request can take 20-60 seconds to complete. It's causing some tests to fail. /api/molecular-profiles/co-expressions/fetch?molecularProfileIdA=coadread_tcga_pub_rna_seq_mrna&molecularProfileIdB=coadread_tcga_pub_rna_seq_mrna&threshold=0 Odd that after our first round of optimization, I don't remember it taking so long. perhaps something else happened since then?
non_main
co expression service performance following request can take seconds to complete it s causing some tests to fail api molecular profiles co expressions fetch molecularprofileida coadread tcga pub rna seq mrna molecularprofileidb coadread tcga pub rna seq mrna threshold odd that after our first round of optimization i don t remember it taking so long perhaps something else happened since then
0
463,230
13,261,967,788
IssuesEvent
2020-08-20 20:51:48
googleapis/google-api-php-client
https://api.github.com/repos/googleapis/google-api-php-client
closed
Basic Example in README.md does not work
priority: p2 type: bug type: docs
In the current README.md the following basic example is provided, however this is not correct, I suspect it's a bit outdated. ```php // include your composer dependencies require_once 'vendor/autoload.php'; $client = new Google_Client(); $client->setApplicationName("Client_Library_Examples"); $client->setDeveloperKey("YOUR_APP_KEY"); $service = new Google_Service_Books($client); $optParams = array('filter' => 'free-ebooks'); $results = $service->volumes->listVolumes('Henry David Thoreau', $optParams); foreach ($results->getItems() as $item) { echo $item['volumeInfo']['title'], "<br /> \n"; } ``` I changed the following lines to make it work, once I noticed my IDE telling me that listVolumes only expects 1 parameter. ``` $optParams = array( 'filter' => 'free-ebooks', 'q' => 'Henry David Thoreau', ); $results = $service->volumes->listVolumes( $optParams); ```
1.0
Basic Example in README.md does not work - In the current README.md the following basic example is provided, however this is not correct, I suspect it's a bit outdated. ```php // include your composer dependencies require_once 'vendor/autoload.php'; $client = new Google_Client(); $client->setApplicationName("Client_Library_Examples"); $client->setDeveloperKey("YOUR_APP_KEY"); $service = new Google_Service_Books($client); $optParams = array('filter' => 'free-ebooks'); $results = $service->volumes->listVolumes('Henry David Thoreau', $optParams); foreach ($results->getItems() as $item) { echo $item['volumeInfo']['title'], "<br /> \n"; } ``` I changed the following lines to make it work, once I noticed my IDE telling me that listVolumes only expects 1 parameter. ``` $optParams = array( 'filter' => 'free-ebooks', 'q' => 'Henry David Thoreau', ); $results = $service->volumes->listVolumes( $optParams); ```
non_main
basic example in readme md does not work in the current readme md the following basic example is provided however this is not correct i suspect it s a bit outdated php include your composer dependencies require once vendor autoload php client new google client client setapplicationname client library examples client setdeveloperkey your app key service new google service books client optparams array filter free ebooks results service volumes listvolumes henry david thoreau optparams foreach results getitems as item echo item n i changed the following lines to make it work once i noticed my ide telling me that listvolumes only expects parameter optparams array filter free ebooks q henry david thoreau results service volumes listvolumes optparams
0
2,561
8,709,008,829
IssuesEvent
2018-12-06 12:42:59
arcticicestudio/nord-docs
https://api.github.com/repos/arcticicestudio/nord-docs
opened
Core HTML element atoms
context-api context-ui scope-configurability scope-maintainability status-tracking type-epic
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/49325616-f1053900-f545-11e8-9725-0b8f5af8d1e2.png" width="20%" /></p> > Associated epic: #63 To achieve a consistent and uniform style and layout, the basic HTML elements, like e.g. a `<h1>` or `<ul>` should be used through a React component. These components render to the base HTML element they represent, but will apply styles, behavior and layout properties to ensure they match the project's design guidelines instead of using default browser configurations that might be even differ for each user agent. This allows to use base HTML elements with all the advantages of React and JS without worrying about different render output. This collection issue tracks the implementations of all the different individual React components. They are all based on the awesome MDN [HTML elements reference][mdn-html-el-ref] documentation and will use the same categorization. ## [Inline Text Semantics][mdn-html-el-ref-inl-txt-sem] ### `A` > Implementation: #70 Represents the `<a>` HTML element (or anchor element). This is a special dynamic and _failsafe_ component since it'll internally use the [Gatsby Link][gatsby-link] to route within the site (internal links) while also being able to link to external data. This will be handled through a utility function to conditionally render based on the passed target URL (internal & external). [gatsby-link]: https://www.gatsbyjs.org/docs/gatsby-link [mdn-html-el-ref]: https://developer.mozilla.org/en-US/docs/Web/HTML/Element [mdn-html-el-ref-inl-txt-sem]: https://developer.mozilla.org/en-US/docs/Web/HTML/Element#Inline_text_semantics
True
Core HTML element atoms - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/49325616-f1053900-f545-11e8-9725-0b8f5af8d1e2.png" width="20%" /></p> > Associated epic: #63 To achieve a consistent and uniform style and layout, the basic HTML elements, like e.g. a `<h1>` or `<ul>` should be used through a React component. These components render to the base HTML element they represent, but will apply styles, behavior and layout properties to ensure they match the project's design guidelines instead of using default browser configurations that might be even differ for each user agent. This allows to use base HTML elements with all the advantages of React and JS without worrying about different render output. This collection issue tracks the implementations of all the different individual React components. They are all based on the awesome MDN [HTML elements reference][mdn-html-el-ref] documentation and will use the same categorization. ## [Inline Text Semantics][mdn-html-el-ref-inl-txt-sem] ### `A` > Implementation: #70 Represents the `<a>` HTML element (or anchor element). This is a special dynamic and _failsafe_ component since it'll internally use the [Gatsby Link][gatsby-link] to route within the site (internal links) while also being able to link to external data. This will be handled through a utility function to conditionally render based on the passed target URL (internal & external). [gatsby-link]: https://www.gatsbyjs.org/docs/gatsby-link [mdn-html-el-ref]: https://developer.mozilla.org/en-US/docs/Web/HTML/Element [mdn-html-el-ref-inl-txt-sem]: https://developer.mozilla.org/en-US/docs/Web/HTML/Element#Inline_text_semantics
main
core html element atoms associated epic to achieve a consistent and uniform style and layout the basic html elements like e g a or should be used through a react component these components render to the base html element they represent but will apply styles behavior and layout properties to ensure they match the project s design guidelines instead of using default browser configurations that might be even differ for each user agent this allows to use base html elements with all the advantages of react and js without worrying about different render output this collection issue tracks the implementations of all the different individual react components they are all based on the awesome mdn documentation and will use the same categorization a implementation represents the html element or anchor element this is a special dynamic and failsafe component since it ll internally use the to route within the site internal links while also being able to link to external data this will be handled through a utility function to conditionally render based on the passed target url internal external
1
750,362
26,198,979,086
IssuesEvent
2023-01-03 15:48:41
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.amd.com - site is not usable
browser-firefox priority-normal engine-gecko
<!-- @browser: Firefox 108.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:108.0) Gecko/20100101 Firefox/108.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/116181 --> **URL**: https://www.amd.com/en/direct-buy/ **Browser / Version**: Firefox 108.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: I can't seem to add to cart on FF but can do so on Chrome. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.amd.com - site is not usable - <!-- @browser: Firefox 108.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:108.0) Gecko/20100101 Firefox/108.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/116181 --> **URL**: https://www.amd.com/en/direct-buy/ **Browser / Version**: Firefox 108.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: I can't seem to add to cart on FF but can do so on Chrome. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description buttons or links not working steps to reproduce i can t seem to add to cart on ff but can do so on chrome browser configuration none from with ❤️
0
3,812
16,541,533,819
IssuesEvent
2021-05-27 17:26:45
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
reopened
[TableToolbarSearch] icon overlay on text
status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬 type: bug 🐛
<!-- Feel free to remove sections that aren't relevant. ## Title line template: [Title]: Brief description --> ## What package(s) are you using? <!-- Add an x in one of the options below, for example: - [x] package name --> - [x] `carbon-components` - [x] `carbon-components-react` ## Detailed description > Describe in detail the issue you're having. The table search has text overlaying the search icon <img width="854" alt="Screen Shot 2021-05-06 at 12 29 35 PM" src="https://user-images.githubusercontent.com/17915105/117342895-5bcea300-ae69-11eb-815e-530596b30bb1.png"> This bx-input in search.scss set a padding which broke this. <img width="560" alt="Screen Shot 2021-05-06 at 12 55 37 PM" src="https://user-images.githubusercontent.com/17915105/117343914-6c334d80-ae6a-11eb-8820-5cacdfa199b9.png"> > Is this issue related to a specific component? TableToolbarSearch > What did you expect to happen? What happened instead? What would you like to > see changed? The text not overlay on the icon > What browser are you working in? Chrome > What version of the Carbon Design System are you using? we used "carbon-components": "10.33.0", "carbon-components-react": "7.33.0", > What offering/product do you work on? Any pressing ship or release dates we > should be aware of? IBM Cloud ## Steps to reproduce the issue It's complicated to copy the whole table in sandbox. you have to import the search.scss > Please create a reduced test case in CodeSandbox > > - Style and vanilla JS: > https://codesandbox.io/s/github/carbon-design-system/carbon/tree/main/packages/components/examples/codesandbox > - React: > https://codesandbox.io/s/github/carbon-design-system/carbon/tree/main/packages/react/examples/codesandbox https://codesandbox.io/s/pensive-butterfly-miqbx?file=/src/index.js ## Additional information - Screenshots or code - Notes
True
[TableToolbarSearch] icon overlay on text - <!-- Feel free to remove sections that aren't relevant. ## Title line template: [Title]: Brief description --> ## What package(s) are you using? <!-- Add an x in one of the options below, for example: - [x] package name --> - [x] `carbon-components` - [x] `carbon-components-react` ## Detailed description > Describe in detail the issue you're having. The table search has text overlaying the search icon <img width="854" alt="Screen Shot 2021-05-06 at 12 29 35 PM" src="https://user-images.githubusercontent.com/17915105/117342895-5bcea300-ae69-11eb-815e-530596b30bb1.png"> This bx-input in search.scss set a padding which broke this. <img width="560" alt="Screen Shot 2021-05-06 at 12 55 37 PM" src="https://user-images.githubusercontent.com/17915105/117343914-6c334d80-ae6a-11eb-8820-5cacdfa199b9.png"> > Is this issue related to a specific component? TableToolbarSearch > What did you expect to happen? What happened instead? What would you like to > see changed? The text not overlay on the icon > What browser are you working in? Chrome > What version of the Carbon Design System are you using? we used "carbon-components": "10.33.0", "carbon-components-react": "7.33.0", > What offering/product do you work on? Any pressing ship or release dates we > should be aware of? IBM Cloud ## Steps to reproduce the issue It's complicated to copy the whole table in sandbox. you have to import the search.scss > Please create a reduced test case in CodeSandbox > > - Style and vanilla JS: > https://codesandbox.io/s/github/carbon-design-system/carbon/tree/main/packages/components/examples/codesandbox > - React: > https://codesandbox.io/s/github/carbon-design-system/carbon/tree/main/packages/react/examples/codesandbox https://codesandbox.io/s/pensive-butterfly-miqbx?file=/src/index.js ## Additional information - Screenshots or code - Notes
main
icon overlay on text feel free to remove sections that aren t relevant title line template brief description what package s are you using add an x in one of the options below for example package name carbon components carbon components react detailed description describe in detail the issue you re having the table search has text overlaying the search icon img width alt screen shot at pm src this bx input in search scss set a padding which broke this img width alt screen shot at pm src is this issue related to a specific component tabletoolbarsearch what did you expect to happen what happened instead what would you like to see changed the text not overlay on the icon what browser are you working in chrome what version of the carbon design system are you using we used carbon components carbon components react what offering product do you work on any pressing ship or release dates we should be aware of ibm cloud steps to reproduce the issue it s complicated to copy the whole table in sandbox you have to import the search scss please create a reduced test case in codesandbox style and vanilla js react additional information screenshots or code notes
1
2,235
7,875,825,733
IssuesEvent
2018-06-25 21:48:27
react-navigation/react-navigation
https://api.github.com/repos/react-navigation/react-navigation
closed
Shifting and focused boolean with createMaterialBottomTabNavigator
needs response from maintainer
### Current Behavior My aim is to make a MaterialBottomTabNavigator which changes the icon for the selected tab, among 5 tabs, however, the focused boolean appears to be broken when shifting is enabled. Code example: ```js import React from 'react'; import { createMaterialBottomTabNavigator } from 'react-navigation-material-bottom-tabs'; class Screen1 extends React.Component { static navigationOptions = { tabBarIcon: ({ focused }) => { console.log('Screen1', focused); }, activeTintColor: 'blue', }; render() { return null; } } class Screen2 extends React.Component { static navigationOptions = { tabBarIcon: ({ focused }) => { console.log('Screen2', focused); }, activeTintColor: 'red', }; render() { return null; } } const MaterialBottomTabNavigator = createMaterialBottomTabNavigator( { Screen1: Screen1, Screen2: Screen2, }, { shifting: true, } ); export default class App extends React.Component { render() { return <MaterialBottomTabNavigator />; } } ``` Console shows: ![image](https://user-images.githubusercontent.com/39280494/40396042-d195ea2c-5ddf-11e8-91a8-8ef479829d84.png) However, if shifting is changed to false, the focused boolean works as expected, although the tabBarIcon still appears to be referenced twice: ![image](https://user-images.githubusercontent.com/39280494/40396077-fbec057c-5ddf-11e8-9331-2cd5da56d70b.png) ### Expected Behavior - The focused variable should work consistently between shifting enabled or disabled ### How to reproduce - To see the error run the above code and compare console logs between shifting: true, and shifting: false ### Your Environment | software | version | ---------------- | ------- | react-navigation | ^2.0.1 | react-native | 0.55.4 | node | v8.11.2 | yarn | 1.6.0
True
Shifting and focused boolean with createMaterialBottomTabNavigator - ### Current Behavior My aim is to make a MaterialBottomTabNavigator which changes the icon for the selected tab, among 5 tabs, however, the focused boolean appears to be broken when shifting is enabled. Code example: ```js import React from 'react'; import { createMaterialBottomTabNavigator } from 'react-navigation-material-bottom-tabs'; class Screen1 extends React.Component { static navigationOptions = { tabBarIcon: ({ focused }) => { console.log('Screen1', focused); }, activeTintColor: 'blue', }; render() { return null; } } class Screen2 extends React.Component { static navigationOptions = { tabBarIcon: ({ focused }) => { console.log('Screen2', focused); }, activeTintColor: 'red', }; render() { return null; } } const MaterialBottomTabNavigator = createMaterialBottomTabNavigator( { Screen1: Screen1, Screen2: Screen2, }, { shifting: true, } ); export default class App extends React.Component { render() { return <MaterialBottomTabNavigator />; } } ``` Console shows: ![image](https://user-images.githubusercontent.com/39280494/40396042-d195ea2c-5ddf-11e8-91a8-8ef479829d84.png) However, if shifting is changed to false, the focused boolean works as expected, although the tabBarIcon still appears to be referenced twice: ![image](https://user-images.githubusercontent.com/39280494/40396077-fbec057c-5ddf-11e8-9331-2cd5da56d70b.png) ### Expected Behavior - The focused variable should work consistently between shifting enabled or disabled ### How to reproduce - To see the error run the above code and compare console logs between shifting: true, and shifting: false ### Your Environment | software | version | ---------------- | ------- | react-navigation | ^2.0.1 | react-native | 0.55.4 | node | v8.11.2 | yarn | 1.6.0
main
shifting and focused boolean with creatematerialbottomtabnavigator current behavior my aim is to make a materialbottomtabnavigator which changes the icon for the selected tab among tabs however the focused boolean appears to be broken when shifting is enabled code example js import react from react import creatematerialbottomtabnavigator from react navigation material bottom tabs class extends react component static navigationoptions tabbaricon focused console log focused activetintcolor blue render return null class extends react component static navigationoptions tabbaricon focused console log focused activetintcolor red render return null const materialbottomtabnavigator creatematerialbottomtabnavigator shifting true export default class app extends react component render return console shows however if shifting is changed to false the focused boolean works as expected although the tabbaricon still appears to be referenced twice expected behavior the focused variable should work consistently between shifting enabled or disabled how to reproduce to see the error run the above code and compare console logs between shifting true and shifting false your environment software version react navigation react native node yarn
1
331,944
29,174,493,486
IssuesEvent
2023-05-19 06:39:27
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_build_tuple_unpack (__main__.StaticDefaultDynamicShapesMiscTests)
triaged module: flaky-tests skipped module: dynamo
Platforms: linux, mac, macos This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_build_tuple_unpack&suite=StaticDefaultDynamicShapesMiscTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/undefined). Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_build_tuple_unpack` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `dynamo/test_dynamic_shapes.py` or `dynamo/test_dynamic_shapes.py`
1.0
DISABLED test_build_tuple_unpack (__main__.StaticDefaultDynamicShapesMiscTests) - Platforms: linux, mac, macos This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_build_tuple_unpack&suite=StaticDefaultDynamicShapesMiscTests) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/undefined). Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_build_tuple_unpack` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `dynamo/test_dynamic_shapes.py` or `dynamo/test_dynamic_shapes.py`
non_main
disabled test build tuple unpack main staticdefaultdynamicshapesmisctests platforms linux mac macos this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test build tuple unpack there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path dynamo test dynamic shapes py or dynamo test dynamic shapes py
0
609,957
18,891,254,947
IssuesEvent
2021-11-15 13:28:49
ZeligsoftDev/CX4CBDDS
https://api.github.com/repos/ZeligsoftDev/CX4CBDDS
opened
PSDD Example cannot have IDL generated by CX Extensions plugin
bug Priority: Major NGC: Approved To Work
# **Issue and tracking information** ### Developer's time Estimated effort to fix (hours): ### Developer's Actual time spent on fix (hours) # **Issue reporter to provide a detailed description of the issue in the space below** The PSDD example can be validated and generate IDL manually, however, the automatic tooling to do batch validations and generations fail on the model with the following message: `Value 'org.eclipse.uml2.uml.internal.impl.ClassImpl@63aeac6d (name: PSDD_Listen, visibility: <unset>) (isLeaf: false, isAbstract: false, isFinalSpecialization: false) (isActive: false)' is not legal. (file:/home/user/workspace-papyrus/SNA_Core/PSDDPubSub/ModelFiles/PSDDPubSub.uml, -1, -1)`
1.0
PSDD Example cannot have IDL generated by CX Extensions plugin - # **Issue and tracking information** ### Developer's time Estimated effort to fix (hours): ### Developer's Actual time spent on fix (hours) # **Issue reporter to provide a detailed description of the issue in the space below** The PSDD example can be validated and generate IDL manually, however, the automatic tooling to do batch validations and generations fail on the model with the following message: `Value 'org.eclipse.uml2.uml.internal.impl.ClassImpl@63aeac6d (name: PSDD_Listen, visibility: <unset>) (isLeaf: false, isAbstract: false, isFinalSpecialization: false) (isActive: false)' is not legal. (file:/home/user/workspace-papyrus/SNA_Core/PSDDPubSub/ModelFiles/PSDDPubSub.uml, -1, -1)`
non_main
psdd example cannot have idl generated by cx extensions plugin issue and tracking information developer s time estimated effort to fix hours developer s actual time spent on fix hours issue reporter to provide a detailed description of the issue in the space below the psdd example can be validated and generate idl manually however the automatic tooling to do batch validations and generations fail on the model with the following message value org eclipse uml internal impl classimpl name psdd listen visibility isleaf false isabstract false isfinalspecialization false isactive false is not legal file home user workspace papyrus sna core psddpubsub modelfiles psddpubsub uml
0
1,805
6,575,934,355
IssuesEvent
2017-09-11 17:53:35
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Feature Idea - add xattr support in FILE & STAT module. (chattr extended attributes in FS)
affects_2.2 feature_idea waiting_on_maintainer
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME file stat ##### ANSIBLE VERSION ansible 2.2.0 (devel c9a5b1c555) last updated 2016/06/02 15:42:56 (GMT +200) lib/ansible/modules/core: (detached HEAD ca4365b644) last updated 2016/06/02 15:43:14 (GMT +200) lib/ansible/modules/extras: (detached HEAD b0aec50b9a) last updated 2016/06/02 15:43:15 (GMT +200) config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ##### OS / ENVIRONMENT LSB Version: :core-4.1-amd64:core-4.1-noarch Distributor ID: RedHatEnterpriseServer Description: Red Hat Enterprise Linux Server release 7.2 (Maipo) Release: 7.2 Codename: Maipo ##### SUMMARY It would be useful add xattr (extended attributes) native support in file and stat modules. Currently, I have to use shell module (chattr and lsattr) to get xttr setting from files. ##### EXPECTED RESULTS Implementation idea: file module: Could be added an additional parameter named xattr or chattr. stat modue: Could be added an addtional field named stat.xattr
True
Feature Idea - add xattr support in FILE & STAT module. (chattr extended attributes in FS) - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME file stat ##### ANSIBLE VERSION ansible 2.2.0 (devel c9a5b1c555) last updated 2016/06/02 15:42:56 (GMT +200) lib/ansible/modules/core: (detached HEAD ca4365b644) last updated 2016/06/02 15:43:14 (GMT +200) lib/ansible/modules/extras: (detached HEAD b0aec50b9a) last updated 2016/06/02 15:43:15 (GMT +200) config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ##### OS / ENVIRONMENT LSB Version: :core-4.1-amd64:core-4.1-noarch Distributor ID: RedHatEnterpriseServer Description: Red Hat Enterprise Linux Server release 7.2 (Maipo) Release: 7.2 Codename: Maipo ##### SUMMARY It would be useful add xattr (extended attributes) native support in file and stat modules. Currently, I have to use shell module (chattr and lsattr) to get xttr setting from files. ##### EXPECTED RESULTS Implementation idea: file module: Could be added an additional parameter named xattr or chattr. stat modue: Could be added an addtional field named stat.xattr
main
feature idea add xattr support in file stat module chattr extended attributes in fs issue type feature idea component name file stat ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file etc ansible ansible cfg configured module search path default w o overrides os environment lsb version core core noarch distributor id redhatenterpriseserver description red hat enterprise linux server release maipo release codename maipo summary it would be useful add xattr extended attributes native support in file and stat modules currently i have to use shell module chattr and lsattr to get xttr setting from files expected results implementation idea file module could be added an additional parameter named xattr or chattr stat modue could be added an addtional field named stat xattr
1
4,451
23,151,971,751
IssuesEvent
2022-07-29 09:12:27
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Java linting not enforced on source files
bug maintainability java
The linting of our Java source files seem to work only for our test files, not our main source files. For instance, a lot of our source files have mixed tabs and spaces. This seems to be due to the non-standard source directories we are using. We should add those source directories in the POM files to make sure they are linted. This will require fixing the corresponding linting issues, and rebase / merge a lot of pull requests… Sorry about that, I should have checked that when I introduced the linting in the first place!
True
Java linting not enforced on source files - The linting of our Java source files seem to work only for our test files, not our main source files. For instance, a lot of our source files have mixed tabs and spaces. This seems to be due to the non-standard source directories we are using. We should add those source directories in the POM files to make sure they are linted. This will require fixing the corresponding linting issues, and rebase / merge a lot of pull requests… Sorry about that, I should have checked that when I introduced the linting in the first place!
main
java linting not enforced on source files the linting of our java source files seem to work only for our test files not our main source files for instance a lot of our source files have mixed tabs and spaces this seems to be due to the non standard source directories we are using we should add those source directories in the pom files to make sure they are linted this will require fixing the corresponding linting issues and rebase merge a lot of pull requests… sorry about that i should have checked that when i introduced the linting in the first place
1
5,090
26,006,280,933
IssuesEvent
2022-12-20 19:42:58
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Implement custom icons for column extracting and moving
type: enhancement work: frontend status: ready restricted: maintainers
Our Figma design specifies custom icons for these actions, but we don't have them yet. | Figma | App | | -- | -- | | ![image](https://user-images.githubusercontent.com/42411/208752459-4910283f-114e-4a10-9200-496ff0bfd053.png) | ![image](https://user-images.githubusercontent.com/42411/208752517-5632172f-d37d-4660-af32-46766fbafbdb.png) |
True
Implement custom icons for column extracting and moving - Our Figma design specifies custom icons for these actions, but we don't have them yet. | Figma | App | | -- | -- | | ![image](https://user-images.githubusercontent.com/42411/208752459-4910283f-114e-4a10-9200-496ff0bfd053.png) | ![image](https://user-images.githubusercontent.com/42411/208752517-5632172f-d37d-4660-af32-46766fbafbdb.png) |
main
implement custom icons for column extracting and moving our figma design specifies custom icons for these actions but we don t have them yet figma app
1
33,960
7,312,656,995
IssuesEvent
2018-02-28 21:44:11
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
closed
Optibelt - corregir el status de la facturas de este año, con base a la fecha de emisión y condiciones de pago
bug defect
Seguimos atorados con optibelt. Si bien aparentemente se corrigió el tema de la fecha de vencimiento... pero eso va de la mano con el status... si la fecha aún no se alcanza, entonces es vigente... si la fecha ya se alcanzó y no han pagado es vencida... y cuando se paga su status es pagada... ya están muy nerviosos con el tema, termina el 2o mes del año, y aún no puden tener estable la aplicación...
1.0
Optibelt - corregir el status de la facturas de este año, con base a la fecha de emisión y condiciones de pago - Seguimos atorados con optibelt. Si bien aparentemente se corrigió el tema de la fecha de vencimiento... pero eso va de la mano con el status... si la fecha aún no se alcanza, entonces es vigente... si la fecha ya se alcanzó y no han pagado es vencida... y cuando se paga su status es pagada... ya están muy nerviosos con el tema, termina el 2o mes del año, y aún no puden tener estable la aplicación...
non_main
optibelt corregir el status de la facturas de este año con base a la fecha de emisión y condiciones de pago seguimos atorados con optibelt si bien aparentemente se corrigió el tema de la fecha de vencimiento pero eso va de la mano con el status si la fecha aún no se alcanza entonces es vigente si la fecha ya se alcanzó y no han pagado es vencida y cuando se paga su status es pagada ya están muy nerviosos con el tema termina el mes del año y aún no puden tener estable la aplicación
0
3,700
15,099,385,021
IssuesEvent
2021-02-08 02:24:34
pypiserver/pypiserver
https://api.github.com/repos/pypiserver/pypiserver
closed
Update CI to push to Docker Hub directly
type.Maintainance
In order to support multiple arch builds easily (see #364) and to isolate ourselves from the truly awful interface that is Docker Hub, I would like to see if we can update the CI to push there directly, rather than relying on webhooks and tag introspection by Docker Hub. The documentation [here](https://docs.docker.com/ci-cd/github-actions/#push-tagged-versions-to-docker-hub) certainly suggests that it should be possible.
True
Update CI to push to Docker Hub directly - In order to support multiple arch builds easily (see #364) and to isolate ourselves from the truly awful interface that is Docker Hub, I would like to see if we can update the CI to push there directly, rather than relying on webhooks and tag introspection by Docker Hub. The documentation [here](https://docs.docker.com/ci-cd/github-actions/#push-tagged-versions-to-docker-hub) certainly suggests that it should be possible.
main
update ci to push to docker hub directly in order to support multiple arch builds easily see and to isolate ourselves from the truly awful interface that is docker hub i would like to see if we can update the ci to push there directly rather than relying on webhooks and tag introspection by docker hub the documentation certainly suggests that it should be possible
1
1,444
6,265,849,430
IssuesEvent
2017-07-16 20:51:45
enterprisemediawiki/meza
https://api.github.com/repos/enterprisemediawiki/meza
opened
Use role remote-mysqldump for backups and replication
critical: stability difficulty: hard important: maintainability important: performance
Currently roles [dump-db-wikis](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/dump-db-wikis/tasks/main.yml#L20) and [backup-db-wikis](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/backup-db-wikis/tasks/main.yml#L29) dump SQL files on master then move them through the controller to their final location. Likewise, when setting up replication [db-master dumps SQL then the file is moved to controller and then to replica](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/database/tasks/replication.yml#L91). Instead use the role `remote-mysqldump` to dump these directly. Also setup tests of replica server. - [ ] Use remote-mysqldump during backups - [ ] Use remote-mysqldump during replication setup - [ ] Setup tests of replication in Docker (Travis)
True
Use role remote-mysqldump for backups and replication - Currently roles [dump-db-wikis](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/dump-db-wikis/tasks/main.yml#L20) and [backup-db-wikis](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/backup-db-wikis/tasks/main.yml#L29) dump SQL files on master then move them through the controller to their final location. Likewise, when setting up replication [db-master dumps SQL then the file is moved to controller and then to replica](https://github.com/enterprisemediawiki/meza/blob/master/src/roles/database/tasks/replication.yml#L91). Instead use the role `remote-mysqldump` to dump these directly. Also setup tests of replica server. - [ ] Use remote-mysqldump during backups - [ ] Use remote-mysqldump during replication setup - [ ] Setup tests of replication in Docker (Travis)
main
use role remote mysqldump for backups and replication currently roles and dump sql files on master then move them through the controller to their final location likewise when setting up replication instead use the role remote mysqldump to dump these directly also setup tests of replica server use remote mysqldump during backups use remote mysqldump during replication setup setup tests of replication in docker travis
1
66,161
8,883,374,734
IssuesEvent
2019-01-14 15:33:02
LycheeOrg/Lychee
https://api.github.com/repos/LycheeOrg/Lychee
opened
Doc: add Couldron link to wiki
Documentation Low Priority
### Detailed description of the problem from #164 : > We are trying to provide an update for the Lychee Cloudron app. The current version is 3.1.6 and the new version will be 3.2.8-fixed. Since this is for Cloudron, the built-in updater is disabled, since on Cloudron apps run on a read-only filesystem. For reference, our app packaging code can be found at https://git.cloudron.io/cloudron/lychee-app Maybe adding a pointer to this in the Wiki would be a nice idea. :)
1.0
Doc: add Couldron link to wiki - ### Detailed description of the problem from #164 : > We are trying to provide an update for the Lychee Cloudron app. The current version is 3.1.6 and the new version will be 3.2.8-fixed. Since this is for Cloudron, the built-in updater is disabled, since on Cloudron apps run on a read-only filesystem. For reference, our app packaging code can be found at https://git.cloudron.io/cloudron/lychee-app Maybe adding a pointer to this in the Wiki would be a nice idea. :)
non_main
doc add couldron link to wiki detailed description of the problem from we are trying to provide an update for the lychee cloudron app the current version is and the new version will be fixed since this is for cloudron the built in updater is disabled since on cloudron apps run on a read only filesystem for reference our app packaging code can be found at maybe adding a pointer to this in the wiki would be a nice idea
0
1,049
4,861,978,851
IssuesEvent
2016-11-14 10:44:04
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
Blockinfile overwrites symlink with file if changed
affects_2.1 bug_report waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME blockinfile ##### ANSIBLE VERSION ``` ansible 2.1.0.0 config file = /Users/vince/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ansible.cfg mentioned above: ``` [defaults] roles_path = ./roles:/etc/ansible/roles forks=25 retry_files_save_path = ./.retry/ ``` ##### OS / ENVIRONMENT I'm running ansible on OSX El Capitain (10.11.6) and managing localhost I'm also running ansible on debian jessie with ansible 2.2.0 and the same config (as localhost) ##### SUMMARY blockinfile overwrites the file with the new version if the old version does not have the block. The problem arises when the file is a symlink. The symlink is overwritten with the file version of the linked file with the new block added. (happens on both OSs/ansible versions) The links remains in existence if the content does not need to be not changed ##### STEPS TO REPRODUCE To test: put the following in `~/test_file`: ``` this is a regular file ``` then symlink `~/test_file` as `~/test_link` like so: ``` $ ln -s ~/test_file ~/test_link ``` Then run the following playbook ``` yaml - hosts: localhost become: no tasks: - blockinfile: block: | hostsum="$(hostname | cksum | awk '{print $1}')"; export hostnum="$((hostsum%253))"; export hostcolor="$(tput setaf $hostnum)" dest: ~/test_link marker: "# hostcolor {mark} ANSIBLE MANAGED BLOCK" ``` ##### EXPECTED RESULTS note the filetype `l` of the link and the filesize of `228` of the linked file ``` $ ls -laF ~/test_link ~/test_file -rw-r--r-- 1 vince vince 228 Sep 28 10:37 /home/vince/test_file lrwxrwxrwx 1 vince vince 21 Sep 28 10:32 /home/vince/test_link -> /home/vince/test_file ``` I also expect both files to have the same content: ``` $ diff -s ~/test_file ~/test_link Files /home/vince/test_file and /home/vince/test_link are identical ``` ##### ACTUAL RESULTS ``` $ ls -laF ~/test_link ~/test_file -rw-r--r-- 1 vince vince 23 Sep 28 10:39 /home/vince/test_file -rw-r--r-- 1 vince vince 228 Sep 28 10:39 /home/vince/test_link ``` ``` $ diff -s ~/test_file ~/test_link 1a2,6 > # hostcolor BEGIN ANSIBLE MANAGED BLOCK > hostsum="$(hostname | cksum | awk '{print $1}')"; > export hostnum="$((hostsum%253))"; > export hostcolor="$(tput setaf $hostnum)" > # hostcolor END ANSIBLE MANAGED `BLOCK` ``` If I can find the time I'll do some additional research, but this should be enough to get started 😉
True
Blockinfile overwrites symlink with file if changed - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME blockinfile ##### ANSIBLE VERSION ``` ansible 2.1.0.0 config file = /Users/vince/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ansible.cfg mentioned above: ``` [defaults] roles_path = ./roles:/etc/ansible/roles forks=25 retry_files_save_path = ./.retry/ ``` ##### OS / ENVIRONMENT I'm running ansible on OSX El Capitain (10.11.6) and managing localhost I'm also running ansible on debian jessie with ansible 2.2.0 and the same config (as localhost) ##### SUMMARY blockinfile overwrites the file with the new version if the old version does not have the block. The problem arises when the file is a symlink. The symlink is overwritten with the file version of the linked file with the new block added. (happens on both OSs/ansible versions) The links remains in existence if the content does not need to be not changed ##### STEPS TO REPRODUCE To test: put the following in `~/test_file`: ``` this is a regular file ``` then symlink `~/test_file` as `~/test_link` like so: ``` $ ln -s ~/test_file ~/test_link ``` Then run the following playbook ``` yaml - hosts: localhost become: no tasks: - blockinfile: block: | hostsum="$(hostname | cksum | awk '{print $1}')"; export hostnum="$((hostsum%253))"; export hostcolor="$(tput setaf $hostnum)" dest: ~/test_link marker: "# hostcolor {mark} ANSIBLE MANAGED BLOCK" ``` ##### EXPECTED RESULTS note the filetype `l` of the link and the filesize of `228` of the linked file ``` $ ls -laF ~/test_link ~/test_file -rw-r--r-- 1 vince vince 228 Sep 28 10:37 /home/vince/test_file lrwxrwxrwx 1 vince vince 21 Sep 28 10:32 /home/vince/test_link -> /home/vince/test_file ``` I also expect both files to have the same content: ``` $ diff -s ~/test_file ~/test_link Files /home/vince/test_file and /home/vince/test_link are identical ``` ##### ACTUAL RESULTS ``` $ ls -laF ~/test_link ~/test_file -rw-r--r-- 1 vince vince 23 Sep 28 10:39 /home/vince/test_file -rw-r--r-- 1 vince vince 228 Sep 28 10:39 /home/vince/test_link ``` ``` $ diff -s ~/test_file ~/test_link 1a2,6 > # hostcolor BEGIN ANSIBLE MANAGED BLOCK > hostsum="$(hostname | cksum | awk '{print $1}')"; > export hostnum="$((hostsum%253))"; > export hostcolor="$(tput setaf $hostnum)" > # hostcolor END ANSIBLE MANAGED `BLOCK` ``` If I can find the time I'll do some additional research, but this should be enough to get started 😉
main
blockinfile overwrites symlink with file if changed issue type bug report component name blockinfile ansible version ansible config file users vince ansible ansible cfg configured module search path default w o overrides configuration ansible cfg mentioned above roles path roles etc ansible roles forks retry files save path retry os environment i m running ansible on osx el capitain and managing localhost i m also running ansible on debian jessie with ansible and the same config as localhost summary blockinfile overwrites the file with the new version if the old version does not have the block the problem arises when the file is a symlink the symlink is overwritten with the file version of the linked file with the new block added happens on both oss ansible versions the links remains in existence if the content does not need to be not changed steps to reproduce to test put the following in test file this is a regular file then symlink test file as test link like so ln s test file test link then run the following playbook yaml hosts localhost become no tasks blockinfile block hostsum hostname cksum awk print export hostnum hostsum export hostcolor tput setaf hostnum dest test link marker hostcolor mark ansible managed block expected results note the filetype l of the link and the filesize of of the linked file ls laf test link test file rw r r vince vince sep home vince test file lrwxrwxrwx vince vince sep home vince test link home vince test file i also expect both files to have the same content diff s test file test link files home vince test file and home vince test link are identical actual results ls laf test link test file rw r r vince vince sep home vince test file rw r r vince vince sep home vince test link diff s test file test link hostcolor begin ansible managed block hostsum hostname cksum awk print export hostnum hostsum export hostcolor tput setaf hostnum hostcolor end ansible managed block if i can find the time i ll do some additional research but this should be enough to get started 😉
1
358,471
25,193,184,021
IssuesEvent
2022-11-12 06:42:27
crushten/github_workflow_repo
https://api.github.com/repos/crushten/github_workflow_repo
closed
Centralized the workflows
documentation
Would like to centralized all workflows here. Will need to make some sort of structure and then can use that in other projects. https://docs.github.com/en/actions/using-workflows/reusing-workflows https://github.blog/changelog/2020-06-23-github-actions-workflow-templates/ https://github.blog/2022-02-10-using-reusable-workflows-github-actions/ https://github.blog/changelog/2021-08-25-github-actions-reduce-duplication-with-action-composition/ https://docs.github.com/en/actions/creating-actions/creating-a-composite-action https://betterprogramming.pub/how-to-use-github-actions-reusable-workflow-8604e8cbf258 https://dev.to/n3wt0n/avoid-duplication-github-actions-reusable-workflows-3ae8
1.0
Centralized the workflows - Would like to centralized all workflows here. Will need to make some sort of structure and then can use that in other projects. https://docs.github.com/en/actions/using-workflows/reusing-workflows https://github.blog/changelog/2020-06-23-github-actions-workflow-templates/ https://github.blog/2022-02-10-using-reusable-workflows-github-actions/ https://github.blog/changelog/2021-08-25-github-actions-reduce-duplication-with-action-composition/ https://docs.github.com/en/actions/creating-actions/creating-a-composite-action https://betterprogramming.pub/how-to-use-github-actions-reusable-workflow-8604e8cbf258 https://dev.to/n3wt0n/avoid-duplication-github-actions-reusable-workflows-3ae8
non_main
centralized the workflows would like to centralized all workflows here will need to make some sort of structure and then can use that in other projects
0
3,945
17,793,408,091
IssuesEvent
2021-08-31 18:59:14
tgstation/tgstation-server
https://api.github.com/repos/tgstation/tgstation-server
opened
Separate deployment jobs from CI suite and trigger via workflow_dispatch
CI/CD Maintainability Issue
So when we desperately need a build we can force it out IDEALLY THOUGH: we'd fix the spurious failures in regular CI and re-parallelize it
True
Separate deployment jobs from CI suite and trigger via workflow_dispatch - So when we desperately need a build we can force it out IDEALLY THOUGH: we'd fix the spurious failures in regular CI and re-parallelize it
main
separate deployment jobs from ci suite and trigger via workflow dispatch so when we desperately need a build we can force it out ideally though we d fix the spurious failures in regular ci and re parallelize it
1
3,070
11,584,438,428
IssuesEvent
2020-02-22 17:20:43
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
opened
Conditional expressions (? : ) should be short
Area: analyzer Area: maintainability feature
The `Condition`, `WhenTrue` and `WhenFalse` parts of a conditional expression should be short, they should not exceed 20 characters. The reason is: The longer the parts are, the harder the conditional expression is to read.
True
Conditional expressions (? : ) should be short - The `Condition`, `WhenTrue` and `WhenFalse` parts of a conditional expression should be short, they should not exceed 20 characters. The reason is: The longer the parts are, the harder the conditional expression is to read.
main
conditional expressions should be short the condition whentrue and whenfalse parts of a conditional expression should be short they should not exceed characters the reason is the longer the parts are the harder the conditional expression is to read
1