Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
7,214
10,346,996,816
IssuesEvent
2019-09-04 16:22:05
qri-io/desktop
https://api.github.com/repos/qri-io/desktop
closed
ensure auto-update is running
chore main process
qri-io/frontend successfully ran auto-update in the background. We should make sure the same functionality is in place for Desktop
1.0
ensure auto-update is running - qri-io/frontend successfully ran auto-update in the background. We should make sure the same functionality is in place for Desktop
process
ensure auto update is running qri io frontend successfully ran auto update in the background we should make sure the same functionality is in place for desktop
1
21,228
6,132,434,420
IssuesEvent
2017-06-25 02:08:23
ganeti/ganeti
https://api.github.com/repos/ganeti/ganeti
opened
Misnamed option in NEWS file
imported_from_google_code Status:Fixed
Originally reported of Google Code with ID 1177. ``` In the NEWS file: "- RPC security got enhanced by using different client SSL certificates for each node. In this context 'gnt-cluster renew-crypto' got a new option '--renew-node-certificates'" But the option is actually "--new-node-certificates" ``` Originally added on 2016-05-06 14:04:58 +0000 UTC.
1.0
Misnamed option in NEWS file - Originally reported of Google Code with ID 1177. ``` In the NEWS file: "- RPC security got enhanced by using different client SSL certificates for each node. In this context 'gnt-cluster renew-crypto' got a new option '--renew-node-certificates'" But the option is actually "--new-node-certificates" ``` Originally added on 2016-05-06 14:04:58 +0000 UTC.
non_process
misnamed option in news file originally reported of google code with id in the news file rpc security got enhanced by using different client ssl certificates for each node in this context gnt cluster renew crypto got a new option renew node certificates but the option is actually new node certificates originally added on utc
0
4,248
7,187,159,249
IssuesEvent
2018-02-02 03:18:15
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
cacheMan does not correct the file when finding a reversal
monitors-cacheMan status-inprocess type-bug
The cacheMan -f mode does not actually correct the cache. It should, if it finds a reversal, simply truncate the remainder of the cache because once there's a reveral all bets are off. If there's a reversal, there may be very large jumps later in the data (there's a reversal because the data is messed up). Suggestion: the first reversal is teh location of a total truncate at that point.
1.0
cacheMan does not correct the file when finding a reversal - The cacheMan -f mode does not actually correct the cache. It should, if it finds a reversal, simply truncate the remainder of the cache because once there's a reveral all bets are off. If there's a reversal, there may be very large jumps later in the data (there's a reversal because the data is messed up). Suggestion: the first reversal is teh location of a total truncate at that point.
process
cacheman does not correct the file when finding a reversal the cacheman f mode does not actually correct the cache it should if it finds a reversal simply truncate the remainder of the cache because once there s a reveral all bets are off if there s a reversal there may be very large jumps later in the data there s a reversal because the data is messed up suggestion the first reversal is teh location of a total truncate at that point
1
16,095
20,264,412,972
IssuesEvent
2022-02-15 10:39:04
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
jni_md.h not found on linux_mips64 and linux_riscv64
type: support / not a bug (process) untriaged team-OSS
### Description of the problem / feature request: The header `jni_md.h` cannot be found on some architectures when compiling due to these 2 lines. https://github.com/bazelbuild/bazel/blob/eeec121668e6307c21e1a9698a96237988269dba/tools/jdk/BUILD.tools#L109-L110
1.0
jni_md.h not found on linux_mips64 and linux_riscv64 - ### Description of the problem / feature request: The header `jni_md.h` cannot be found on some architectures when compiling due to these 2 lines. https://github.com/bazelbuild/bazel/blob/eeec121668e6307c21e1a9698a96237988269dba/tools/jdk/BUILD.tools#L109-L110
process
jni md h not found on linux and linux description of the problem feature request the header jni md h cannot be found on some architectures when compiling due to these lines
1
10,208
13,067,104,008
IssuesEvent
2020-07-30 23:24:49
googleapis/proto-plus-python
https://api.github.com/repos/googleapis/proto-plus-python
closed
Automated publish to PyPI is broken
type: process
https://app.circleci.com/pipelines/github/googleapis/proto-plus-python/132/workflows/41232c5b-ce36-4a48-9fc5-5eb3943dceb7/jobs/955/steps ``` #!/bin/bash -eo pipefail openssl aes-256-cbc -d \ -in .circleci/.pypirc.enc \ -out ~/.pypirc \ -k "${PYPIRC_ENCRYPTION_KEY}" *** WARNING : deprecated key derivation used. Using -iter or -pbkdf2 would be better. bad decrypt 140293096486016:error:06065064:digital envelope routines:EVP_DecryptFinal_ex:bad decrypt:../crypto/evp/evp_enc.c:570: Exited with code exit status 1 ```
1.0
Automated publish to PyPI is broken - https://app.circleci.com/pipelines/github/googleapis/proto-plus-python/132/workflows/41232c5b-ce36-4a48-9fc5-5eb3943dceb7/jobs/955/steps ``` #!/bin/bash -eo pipefail openssl aes-256-cbc -d \ -in .circleci/.pypirc.enc \ -out ~/.pypirc \ -k "${PYPIRC_ENCRYPTION_KEY}" *** WARNING : deprecated key derivation used. Using -iter or -pbkdf2 would be better. bad decrypt 140293096486016:error:06065064:digital envelope routines:EVP_DecryptFinal_ex:bad decrypt:../crypto/evp/evp_enc.c:570: Exited with code exit status 1 ```
process
automated publish to pypi is broken bin bash eo pipefail openssl aes cbc d in circleci pypirc enc out pypirc k pypirc encryption key warning deprecated key derivation used using iter or would be better bad decrypt error digital envelope routines evp decryptfinal ex bad decrypt crypto evp evp enc c exited with code exit status
1
342,756
24,755,812,731
IssuesEvent
2022-10-21 17:36:43
Py-Contributors/AudioBook
https://api.github.com/repos/Py-Contributors/AudioBook
opened
extends the current documentation for readthedocs
documentation enhancement help wanted good first issue hacktoberfest-accepted
**Is your feature request related to a problem? Please describe.** Current Documentation is not sufficient. we have to extend it. **Plan for Docs** will use Sphinx[¶](https://docs.readthedocs.io/en/stable/intro/getting-started-with-sphinx.html#getting-started-with-sphinx) for documentation and write proper multipage documentation. This issue is open to assigning. Reference:- https://pypdf2.readthedocs.io/en/latest/ https://flask.palletsprojects.com/en/2.2.x/
1.0
extends the current documentation for readthedocs - **Is your feature request related to a problem? Please describe.** Current Documentation is not sufficient. we have to extend it. **Plan for Docs** will use Sphinx[¶](https://docs.readthedocs.io/en/stable/intro/getting-started-with-sphinx.html#getting-started-with-sphinx) for documentation and write proper multipage documentation. This issue is open to assigning. Reference:- https://pypdf2.readthedocs.io/en/latest/ https://flask.palletsprojects.com/en/2.2.x/
non_process
extends the current documentation for readthedocs is your feature request related to a problem please describe current documentation is not sufficient we have to extend it plan for docs will use sphinx for documentation and write proper multipage documentation this issue is open to assigning reference
0
21,582
14,656,962,668
IssuesEvent
2020-12-28 14:34:42
airyhq/airy
https://api.github.com/repos/airyhq/airy
closed
Import core go libraries from anywhere
fix infrastructure
We need to make our go libraries importable from anywhere.
1.0
Import core go libraries from anywhere - We need to make our go libraries importable from anywhere.
non_process
import core go libraries from anywhere we need to make our go libraries importable from anywhere
0
14,220
17,141,349,176
IssuesEvent
2021-07-13 09:57:40
2i2c-org/team-compass
https://api.github.com/repos/2i2c-org/team-compass
opened
Finalize the titles that we use describe roles in a given hub
:label: administration :label: team-process prio: med type: enhancement
# Summary In a few issues now, we've run into ambiguity / multiple kinds of terminology for how we describe roles in a hub. We should just nail this down and agree upon one description so that we can remain consistent. # Proposed titles ## On the community side - **Hub Community**: The group of people that we are (or might be) serving with a hub. **We use this word instead of "client" or "customer"**. - My rationale here is that we are "in between" a SaaS platform and a consultancy, with a strong community+pro-social focus. As such, neither "customer" or "client" feel right to me. This is a stab at something that aligns more with the mission focus of 2i2c. - **Community Representative**: The point of contact with Hub Engineers, and the main interface with a hub's community - **Community Admin Team**: The group of community members with "Administrative privileges" on a hub. They are expected to do most hub administration via the JupyterHub UI. This group must have at least one member (usually the Community Rep). ## On the 2i2c side - **Hub Engineer**: A member of the @2i2c-org/tech-team that performs more complex dev/ops for a hub # Actions - [ ] Decide whether "Hub Community" is an acceptable drop-in for "customer" or "client" - [ ] Agree on the other major roles and what we call them - [ ] Agree that we're not missing any important roles - [ ] Write these up in our Team Compass - [ ] Update respective documents (e.g. the Managed Hub Service doc) # Related issues - Some discussion of this in https://github.com/2i2c-org/team-compass/issues/147
1.0
Finalize the titles that we use describe roles in a given hub - # Summary In a few issues now, we've run into ambiguity / multiple kinds of terminology for how we describe roles in a hub. We should just nail this down and agree upon one description so that we can remain consistent. # Proposed titles ## On the community side - **Hub Community**: The group of people that we are (or might be) serving with a hub. **We use this word instead of "client" or "customer"**. - My rationale here is that we are "in between" a SaaS platform and a consultancy, with a strong community+pro-social focus. As such, neither "customer" or "client" feel right to me. This is a stab at something that aligns more with the mission focus of 2i2c. - **Community Representative**: The point of contact with Hub Engineers, and the main interface with a hub's community - **Community Admin Team**: The group of community members with "Administrative privileges" on a hub. They are expected to do most hub administration via the JupyterHub UI. This group must have at least one member (usually the Community Rep). ## On the 2i2c side - **Hub Engineer**: A member of the @2i2c-org/tech-team that performs more complex dev/ops for a hub # Actions - [ ] Decide whether "Hub Community" is an acceptable drop-in for "customer" or "client" - [ ] Agree on the other major roles and what we call them - [ ] Agree that we're not missing any important roles - [ ] Write these up in our Team Compass - [ ] Update respective documents (e.g. the Managed Hub Service doc) # Related issues - Some discussion of this in https://github.com/2i2c-org/team-compass/issues/147
process
finalize the titles that we use describe roles in a given hub summary in a few issues now we ve run into ambiguity multiple kinds of terminology for how we describe roles in a hub we should just nail this down and agree upon one description so that we can remain consistent proposed titles on the community side hub community the group of people that we are or might be serving with a hub we use this word instead of client or customer my rationale here is that we are in between a saas platform and a consultancy with a strong community pro social focus as such neither customer or client feel right to me this is a stab at something that aligns more with the mission focus of community representative the point of contact with hub engineers and the main interface with a hub s community community admin team the group of community members with administrative privileges on a hub they are expected to do most hub administration via the jupyterhub ui this group must have at least one member usually the community rep on the side hub engineer a member of the org tech team that performs more complex dev ops for a hub actions decide whether hub community is an acceptable drop in for customer or client agree on the other major roles and what we call them agree that we re not missing any important roles write these up in our team compass update respective documents e g the managed hub service doc related issues some discussion of this in
1
802,563
28,967,034,537
IssuesEvent
2023-05-10 08:37:48
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
PHP error after the recent update 1.0.84
bug [Priority: HIGH] Ready for Review
**_Error notice_** ` PHP Notice: Function WP_Scripts::localize was called <strong>incorrectly</strong>. The <code>$l10n</code> parameter must be an array. To pass arbitrary data to scripts, use the <code>wp_add_inline_script()</code> function instead. Please see <a href="https://wordpress.org/documentation/article/debugging-in-wordpress/">Debugging in WordPress</a> for more information. (This message was added in version 5.7.0.) in C:\xampp\htdocs\wordpress\wp-includes\functions.php on line 5865 ` Reference ticket: https://wordpress.org/support/topic/function-wp_scriptslocalize-was-called-incorrectly-the-l10n-parameter-must-b-2/
1.0
PHP error after the recent update 1.0.84 - **_Error notice_** ` PHP Notice: Function WP_Scripts::localize was called <strong>incorrectly</strong>. The <code>$l10n</code> parameter must be an array. To pass arbitrary data to scripts, use the <code>wp_add_inline_script()</code> function instead. Please see <a href="https://wordpress.org/documentation/article/debugging-in-wordpress/">Debugging in WordPress</a> for more information. (This message was added in version 5.7.0.) in C:\xampp\htdocs\wordpress\wp-includes\functions.php on line 5865 ` Reference ticket: https://wordpress.org/support/topic/function-wp_scriptslocalize-was-called-incorrectly-the-l10n-parameter-must-b-2/
non_process
php error after the recent update error notice php notice function wp scripts localize was called incorrectly the parameter must be an array to pass arbitrary data to scripts use the wp add inline script function instead please see for more information this message was added in version in c xampp htdocs wordpress wp includes functions php on line reference ticket
0
15,479
19,688,035,840
IssuesEvent
2022-01-12 01:34:50
dtcenter/MET
https://api.github.com/repos/dtcenter/MET
closed
ioda2nc fails if the same input file is given with -iodafile option
type: bug priority: high alert: NEED MORE DEFINITION component: CI/CD reporting: DTC NCAR Base requestor: METplus Team required: FOR OFFICIAL RELEASE MET: PreProcessing Tools (Point)
*Replace italics below with details for this issue.* ## Describe the Problem ## ioda2nc fails if the same input file is given with -iodafile option ### Expected Behavior ### It should work without errors. ### Environment ### Describe your runtime environment: *1. Machine: (Linux Workstation)* *2. OS: (RedHat Linux)* *3. Software version number(s): 10.1.0 and 11.0 beta* ### To Reproduce ### Describe the steps to reproduce the behavior: *1. Go to seneca* *2. run the following command* /usr/local/met/bin/ioda2nc -v 2 /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda \ /ioda.NC001007.2020031012.nc ioda.NC001007.2020031012.summary.nc \ -iodafile /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda/ioda.NC001007.2020031012.nc *3. See error* terminate called after throwing an instance of 'netCDF::exceptions::NcEdge' what(): NetCDF: Start+count exceeds dimension bound file: ncVar.cpp line:958 Aborted *4. run the following command* export METPLUS_ELEVATION_RANGE_DICT="elevation_range = {beg = -1000;end = 100000;}" export METPLUS_LEVEL_RANGE_DICT="" export METPLUS_MASK_DICT="" export METPLUS_MESSAGE_TYPE="" export METPLUS_MESSAGE_TYPE_GROUP_MAP="" export METPLUS_MESSAGE_TYPE_MAP="" export METPLUS_METADATA_MAP="" export METPLUS_MET_CONFIG_OVERRIDES="" export METPLUS_MISSING_THRESH="" export METPLUS_OBS_NAME_MAP="obs_name_map = [{ key = \"wind_direction\"; val = \"WDIR\"; }, { key = \"wind_speed\"; val = \"WIND\"; }];" export METPLUS_OBS_VAR="" export METPLUS_OBS_WINDOW_DICT="obs_window = {beg = -5400;end = 5400;}" export METPLUS_QUALITY_MARK_THRESH="quality_mark_thresh = 0;" export METPLUS_STATION_ID="" export METPLUS_TIME_SUMMARY_DICT="time_summary = {flag = TRUE;raw_data = TRUE;beg = \"000000\";end = \"235959\";step = 300;width = 600;grib_code = [];obs_var = [\"WIND\"];type = [\"min\", \"max\", \"range\", \"mean\", \"stdev\", \"median\", \"p80\"];vld_freq = 0;vld_thresh = 0.0;}" export MET_TMP_DIR="/d1/personal/mccabe/out2/tmp" /usr/local/met/bin/ioda2nc -v 2 /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda \ /ioda.NC001007.2020031012.nc ioda.NC001007.2020031012.summary.nc \ -config /d1/personal/mccabe/METplus/parm/met_config/IODA2NCConfig_wrapped \ -iodafile /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda/ioda.NC001007.2020031012.nc *5. See error* ERROR : ERROR : get_obs_data_float() -> WDIR@ObsValue does not exist! ERROR : *Post relevant sample data following these instructions:* *https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### 2702691 ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [x] Select **scientist(s)** or **no scientist** required: no scientist ### Labels ### - [x] Select **component(s)** - [x] Select **priority** - [x] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Organization** level **Project** for support of the current coordinated release - [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next bugfix version ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Bugfix Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **main_\<Version>**. Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>` - [ ] Fix the bug and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **main_\<Version>**. Pull request: `bugfix <Issue Number> main_<Version> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Organization** level software support **Project** for the current coordinated release Select: **Milestone** as the next bugfix version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Complete the steps above to fix the bug on the **develop** branch. Branch name: `bugfix_<Issue Number>_develop_<Description>` Pull request: `bugfix <Issue Number> develop <Description>` Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Close this issue.
1.0
ioda2nc fails if the same input file is given with -iodafile option - *Replace italics below with details for this issue.* ## Describe the Problem ## ioda2nc fails if the same input file is given with -iodafile option ### Expected Behavior ### It should work without errors. ### Environment ### Describe your runtime environment: *1. Machine: (Linux Workstation)* *2. OS: (RedHat Linux)* *3. Software version number(s): 10.1.0 and 11.0 beta* ### To Reproduce ### Describe the steps to reproduce the behavior: *1. Go to seneca* *2. run the following command* /usr/local/met/bin/ioda2nc -v 2 /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda \ /ioda.NC001007.2020031012.nc ioda.NC001007.2020031012.summary.nc \ -iodafile /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda/ioda.NC001007.2020031012.nc *3. See error* terminate called after throwing an instance of 'netCDF::exceptions::NcEdge' what(): NetCDF: Start+count exceeds dimension bound file: ncVar.cpp line:958 Aborted *4. run the following command* export METPLUS_ELEVATION_RANGE_DICT="elevation_range = {beg = -1000;end = 100000;}" export METPLUS_LEVEL_RANGE_DICT="" export METPLUS_MASK_DICT="" export METPLUS_MESSAGE_TYPE="" export METPLUS_MESSAGE_TYPE_GROUP_MAP="" export METPLUS_MESSAGE_TYPE_MAP="" export METPLUS_METADATA_MAP="" export METPLUS_MET_CONFIG_OVERRIDES="" export METPLUS_MISSING_THRESH="" export METPLUS_OBS_NAME_MAP="obs_name_map = [{ key = \"wind_direction\"; val = \"WDIR\"; }, { key = \"wind_speed\"; val = \"WIND\"; }];" export METPLUS_OBS_VAR="" export METPLUS_OBS_WINDOW_DICT="obs_window = {beg = -5400;end = 5400;}" export METPLUS_QUALITY_MARK_THRESH="quality_mark_thresh = 0;" export METPLUS_STATION_ID="" export METPLUS_TIME_SUMMARY_DICT="time_summary = {flag = TRUE;raw_data = TRUE;beg = \"000000\";end = \"235959\";step = 300;width = 600;grib_code = [];obs_var = [\"WIND\"];type = [\"min\", \"max\", \"range\", \"mean\", \"stdev\", \"median\", \"p80\"];vld_freq = 0;vld_thresh = 0.0;}" export MET_TMP_DIR="/d1/personal/mccabe/out2/tmp" /usr/local/met/bin/ioda2nc -v 2 /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda \ /ioda.NC001007.2020031012.nc ioda.NC001007.2020031012.summary.nc \ -config /d1/personal/mccabe/METplus/parm/met_config/IODA2NCConfig_wrapped \ -iodafile /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda/ioda.NC001007.2020031012.nc *5. See error* ERROR : ERROR : get_obs_data_float() -> WDIR@ObsValue does not exist! ERROR : *Post relevant sample data following these instructions:* *https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### 2702691 ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [x] Select **scientist(s)** or **no scientist** required: no scientist ### Labels ### - [x] Select **component(s)** - [x] Select **priority** - [x] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Organization** level **Project** for support of the current coordinated release - [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next bugfix version ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Bugfix Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **main_\<Version>**. Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>` - [ ] Fix the bug and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **main_\<Version>**. Pull request: `bugfix <Issue Number> main_<Version> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Organization** level software support **Project** for the current coordinated release Select: **Milestone** as the next bugfix version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Complete the steps above to fix the bug on the **develop** branch. Branch name: `bugfix_<Issue Number>_develop_<Description>` Pull request: `bugfix <Issue Number> develop <Description>` Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Close this issue.
process
fails if the same input file is given with iodafile option replace italics below with details for this issue describe the problem fails if the same input file is given with iodafile option expected behavior it should work without errors environment describe your runtime environment machine linux workstation os redhat linux software version number s and beta to reproduce describe the steps to reproduce the behavior go to seneca run the following command usr local met bin v projects metplus metplus data development feature met test new ioda ioda nc ioda summary nc iodafile projects metplus metplus data development feature met test new ioda ioda nc see error terminate called after throwing an instance of netcdf exceptions ncedge what netcdf start count exceeds dimension bound file ncvar cpp line aborted run the following command export metplus elevation range dict elevation range beg end export metplus level range dict export metplus mask dict export metplus message type export metplus message type group map export metplus message type map export metplus metadata map export metplus met config overrides export metplus missing thresh export metplus obs name map obs name map export metplus obs var export metplus obs window dict obs window beg end export metplus quality mark thresh quality mark thresh export metplus station id export metplus time summary dict time summary flag true raw data true beg end step width grib code obs var type vld freq vld thresh export met tmp dir personal mccabe tmp usr local met bin v projects metplus metplus data development feature met test new ioda ioda nc ioda summary nc config personal mccabe metplus parm met config wrapped iodafile projects metplus metplus data development feature met test new ioda ioda nc see error error error get obs data float wdir obsvalue does not exist error post relevant sample data following these instructions relevant deadlines list relevant project deadlines here or state none funding source define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required no scientist labels select component s select priority select requestor s projects and milestone select organization level project for support of the current coordinated release select repository level project for development toward the next official release or add alert need project assignment label select milestone as the next bugfix version define related issue s consider the impact to the other metplus components bugfix checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of main branch name bugfix main fix the bug and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into main pull request bugfix main define the pull request metadata as permissions allow select reviewer s and linked issues select organization level software support project for the current coordinated release select milestone as the next bugfix version iterate until the reviewer s accept and merge your changes delete your fork or branch complete the steps above to fix the bug on the develop branch branch name bugfix develop pull request bugfix develop select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version close this issue
1
21,910
30,440,085,688
IssuesEvent
2023-07-15 01:22:53
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Medium-Priority Bug Fixes
.Epic .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
```[tasklist] ### Tasks - [x] #29734 - [x] #29702 - [x] #29746 - [x] #29764 - [x] #29902 - [x] #29745 - [x] #29770 - [x] #29988 - [x] #29747 - [x] #29898 - [x] #29964 - [x] #29944 - [x] #29947 - [ ] #30568 - [ ] #29895 - [ ] #29908 - [ ] #30280 - [ ] #29748 - [ ] #29897 - [ ] #29904 - [ ] #29909 - [ ] #29910 - [ ] #29935 - [ ] #29936 - [ ] #29938 - [ ] #29941 - [ ] #29942 - [ ] #29946 - [ ] #29948 - [ ] #29950 - [ ] #29953 - [ ] #29958 - [ ] #30385 - [ ] https://github.com/metabase/metabase/issues/30397 - [ ] #30401 - [ ] https://github.com/metabase/metabase/issues/30857 - [ ] https://github.com/metabase/metabase/issues/30858 - [ ] https://github.com/metabase/metabase/issues/30948 - [ ] https://github.com/metabase/metabase/issues/30949 - [ ] https://github.com/metabase/metabase/issues/30950 - [ ] https://github.com/metabase/metabase/issues/30957 - [ ] #29949 - [ ] https://github.com/metabase/metabase/issues/31053 - [ ] https://github.com/metabase/metabase/issues/31223 - [ ] https://github.com/metabase/metabase/issues/31268 - [ ] https://github.com/metabase/metabase/issues/31365 - [ ] https://github.com/metabase/metabase/issues/31366 - [ ] https://github.com/metabase/metabase/issues/31368 - [ ] #31521 - [ ] https://github.com/metabase/metabase/issues/31624 - [ ] https://github.com/metabase/metabase/issues/31741 - [ ] https://github.com/metabase/metabase/issues/31769 - [ ] https://github.com/metabase/metabase/issues/31775 - [ ] https://github.com/metabase/metabase/issues/31858 - [ ] https://github.com/metabase/metabase/issues/32049 - [ ] https://github.com/metabase/metabase/issues/32063 ```
1.0
[MLv2] Medium-Priority Bug Fixes - ```[tasklist] ### Tasks - [x] #29734 - [x] #29702 - [x] #29746 - [x] #29764 - [x] #29902 - [x] #29745 - [x] #29770 - [x] #29988 - [x] #29747 - [x] #29898 - [x] #29964 - [x] #29944 - [x] #29947 - [ ] #30568 - [ ] #29895 - [ ] #29908 - [ ] #30280 - [ ] #29748 - [ ] #29897 - [ ] #29904 - [ ] #29909 - [ ] #29910 - [ ] #29935 - [ ] #29936 - [ ] #29938 - [ ] #29941 - [ ] #29942 - [ ] #29946 - [ ] #29948 - [ ] #29950 - [ ] #29953 - [ ] #29958 - [ ] #30385 - [ ] https://github.com/metabase/metabase/issues/30397 - [ ] #30401 - [ ] https://github.com/metabase/metabase/issues/30857 - [ ] https://github.com/metabase/metabase/issues/30858 - [ ] https://github.com/metabase/metabase/issues/30948 - [ ] https://github.com/metabase/metabase/issues/30949 - [ ] https://github.com/metabase/metabase/issues/30950 - [ ] https://github.com/metabase/metabase/issues/30957 - [ ] #29949 - [ ] https://github.com/metabase/metabase/issues/31053 - [ ] https://github.com/metabase/metabase/issues/31223 - [ ] https://github.com/metabase/metabase/issues/31268 - [ ] https://github.com/metabase/metabase/issues/31365 - [ ] https://github.com/metabase/metabase/issues/31366 - [ ] https://github.com/metabase/metabase/issues/31368 - [ ] #31521 - [ ] https://github.com/metabase/metabase/issues/31624 - [ ] https://github.com/metabase/metabase/issues/31741 - [ ] https://github.com/metabase/metabase/issues/31769 - [ ] https://github.com/metabase/metabase/issues/31775 - [ ] https://github.com/metabase/metabase/issues/31858 - [ ] https://github.com/metabase/metabase/issues/32049 - [ ] https://github.com/metabase/metabase/issues/32063 ```
process
medium priority bug fixes tasks
1
112,816
17,102,345,872
IssuesEvent
2021-07-09 13:06:09
turkdevops/codecov-action
https://api.github.com/repos/turkdevops/codecov-action
closed
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - autoclosed
security vulnerability
## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: codecov-action/package.json</p> <p>Path to vulnerable library: codecov-action/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - jest-26.6.3.tgz (Root Library) - core-26.6.3.tgz - jest-resolve-26.6.2.tgz - read-pkg-up-7.0.1.tgz - read-pkg-5.2.0.tgz - normalize-package-data-2.5.0.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/codecov-action/commit/55e35bb57ce02b251f09ff66d3ac34c5a60291ea">55e35bb57ce02b251f09ff66d3ac34c5a60291ea</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - autoclosed - ## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: codecov-action/package.json</p> <p>Path to vulnerable library: codecov-action/node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - jest-26.6.3.tgz (Root Library) - core-26.6.3.tgz - jest-resolve-26.6.2.tgz - read-pkg-up-7.0.1.tgz - read-pkg-5.2.0.tgz - normalize-package-data-2.5.0.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/codecov-action/commit/55e35bb57ce02b251f09ff66d3ac34c5a60291ea">55e35bb57ce02b251f09ff66d3ac34c5a60291ea</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in hosted git info tgz autoclosed cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file codecov action package json path to vulnerable library codecov action node modules hosted git info package json dependency hierarchy jest tgz root library core tgz jest resolve tgz read pkg up tgz read pkg tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info step up your open source security game with whitesource
0
16,114
20,376,965,342
IssuesEvent
2022-02-21 16:33:08
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Query Engine tests with sharded MongoDB connection strings
process/candidate topic: internal team/migrations team/client
MongoDB connection strings are not URIs. It would be great to have Rust-level tests to see if sharded strings work correctly with the Query Engine. The URI format is this: ``` mongodb://user:password@srv1.rgyl0.mongodb.net:27017,srv2.rgyl0.mongodb.net:27017,srv3.rgyl0.mongodb.net:27017/database?ssl=true&authSource=admin&retryWrites=true&w=majority ``` So we should have a few replicas, and test with a connection string that lists them like in the example.
1.0
Query Engine tests with sharded MongoDB connection strings - MongoDB connection strings are not URIs. It would be great to have Rust-level tests to see if sharded strings work correctly with the Query Engine. The URI format is this: ``` mongodb://user:password@srv1.rgyl0.mongodb.net:27017,srv2.rgyl0.mongodb.net:27017,srv3.rgyl0.mongodb.net:27017/database?ssl=true&authSource=admin&retryWrites=true&w=majority ``` So we should have a few replicas, and test with a connection string that lists them like in the example.
process
query engine tests with sharded mongodb connection strings mongodb connection strings are not uris it would be great to have rust level tests to see if sharded strings work correctly with the query engine the uri format is this mongodb user password mongodb net mongodb net mongodb net database ssl true authsource admin retrywrites true w majority so we should have a few replicas and test with a connection string that lists them like in the example
1
785,119
27,599,080,620
IssuesEvent
2023-03-09 08:50:04
AY2223S2-CS2113-T15-4/tp
https://api.github.com/repos/AY2223S2-CS2113-T15-4/tp
opened
Delete flashcards
type.Epic priority.High
As a user I can delete any if my cards so that I don’t get asked to review that card later on in case I am confident I have truly memorised the card
1.0
Delete flashcards - As a user I can delete any if my cards so that I don’t get asked to review that card later on in case I am confident I have truly memorised the card
non_process
delete flashcards as a user i can delete any if my cards so that i don’t get asked to review that card later on in case i am confident i have truly memorised the card
0
3,783
6,760,953,854
IssuesEvent
2017-10-24 22:44:58
aspnet/IISIntegration
https://api.github.com/repos/aspnet/IISIntegration
closed
React to ANCM changes for NativeMethods pInvoke layer.
bug in-process
Parallel issue to https://github.com/aspnet/KestrelHttpServer/blob/dev/src/Kestrel.Core/Internal/Http/HttpProtocol.FeatureCollection.cs#L97-L101. With Pan's changes ( to the native ANCM and recognition that we need to expose more pInvoke methods, we will need to react to these appropriately.
1.0
React to ANCM changes for NativeMethods pInvoke layer. - Parallel issue to https://github.com/aspnet/KestrelHttpServer/blob/dev/src/Kestrel.Core/Internal/Http/HttpProtocol.FeatureCollection.cs#L97-L101. With Pan's changes ( to the native ANCM and recognition that we need to expose more pInvoke methods, we will need to react to these appropriately.
process
react to ancm changes for nativemethods pinvoke layer parallel issue to with pan s changes to the native ancm and recognition that we need to expose more pinvoke methods we will need to react to these appropriately
1
186,029
15,044,100,855
IssuesEvent
2021-02-03 02:11:31
walterimaican/nightlight
https://api.github.com/repos/walterimaican/nightlight
opened
Documentation: Stale
documentation
Some of the screenshots used in the README.md are now inconsistent with the current state of the project. These images should be updated when possible.
1.0
Documentation: Stale - Some of the screenshots used in the README.md are now inconsistent with the current state of the project. These images should be updated when possible.
non_process
documentation stale some of the screenshots used in the readme md are now inconsistent with the current state of the project these images should be updated when possible
0
4,208
7,166,652,311
IssuesEvent
2018-01-29 17:54:37
itsyouonline/identityserver
https://api.github.com/repos/itsyouonline/identityserver
reopened
Cloudflare returns error 504 on rare occasions
process_wontfix type_bug
Very rarely, cloudlfare will return a 504 when calling any url, even if others go through just fine. The 504 is not generated on our site as observed from the logs (tracing an `/error504` page load, no earlier requests gave a 504 for that users). This can cause pages to render empty, translations to not load etc...
1.0
Cloudflare returns error 504 on rare occasions - Very rarely, cloudlfare will return a 504 when calling any url, even if others go through just fine. The 504 is not generated on our site as observed from the logs (tracing an `/error504` page load, no earlier requests gave a 504 for that users). This can cause pages to render empty, translations to not load etc...
process
cloudflare returns error on rare occasions very rarely cloudlfare will return a when calling any url even if others go through just fine the is not generated on our site as observed from the logs tracing an page load no earlier requests gave a for that users this can cause pages to render empty translations to not load etc
1
656,733
21,773,565,908
IssuesEvent
2022-05-13 11:35:17
awslabs/aws-lambda-powertools-typescript
https://api.github.com/repos/awslabs/aws-lambda-powertools-typescript
opened
Feature (all): evaluate migration to Middy 3.x
enhancement dependencies utility:all triage priority:low
## Description of the feature request **Problem statement** Middy, the dependency that we use to vend middleware has [recently released](https://github.com/middyjs/middy/releases/tag/3.0.0) a new major version (v 3.x). This version drops support for Node JS version 12.x and introduces some breaking changes. Given that Middy, as a whole, has a large surface and we are only using `@middy/core` we need to investigate on whether there's a path to upgrade to the newer version while still continuing to support Node JS 12.x in Powertools. **Additional context** Running the e2e tests on a branch that upgraded `@middy/core` to 3.0.1 seems to show that the newer release is compatible with Powertools on both versions 12 and 14 of Node ([see run results](https://github.com/awslabs/aws-lambda-powertools-typescript/actions/runs/2318810802)). At the same time, the unit tests for that same branch are failing for all utilities and specifically in the section that relates the middleware implementations. The error message (see detail below) seems to hint at incompatibilities between the new bundling of middy that now supports both CJS and ESM bundling, and our project's configuration since the errors are related to the imports and the tests don't run at all. <details> ```sh FAIL AWS Lambda Powertools utility: TRACER tests/unit/middy.test.ts ● Test suite failed to run Jest encountered an unexpected token Jest failed to parse a file. This happens e.g. when your code or its dependencies use non-standard JavaScript syntax, or when Jest is not configured to support such syntax. Out of the box Jest supports Babel, which will be used to transform your files into valid JS based on your Babel configuration. By default "node_modules" folder is ignored by transformers. Here's what you can do: • If you are trying to use ECMAScript Modules, see https://jestjs.io/docs/ecmascript-modules for how to enable it. • If you are trying to use TypeScript, see https://jestjs.io/docs/getting-started#using-typescript • To have some of your "node_modules" files transformed, you can specify a custom "transformIgnorePatterns" in your config. • If you need a custom transformation specify a "transform" option in your config. • If you simply want to mock your non-JS modules (e.g. binary assets) you can stub them out with the "moduleNameMapper" config option. You'll find more details and examples of these config options in the docs: https://jestjs.io/docs/configuration For information about custom transformations, see: https://jestjs.io/docs/code-transformation Details: /home/ec2-user/aws-lambda-powertools-typescript/node_modules/@middy/core/index.js:1 ({"Object.<anonymous>":function(module,exports,require,__dirname,__filename,jest){import{EventEmitter}from"events";const defaultLambdaHandler=()=>{};const defaultPlugin={timeoutEarlyInMillis:5,timeoutEarlyResponse:()=>{throw new Error("Timeout")}};const middy=(lambdaHandler=defaultLambdaHandler,plugin={})=>{if(typeof lambdaHandler!=="function"){plugin=lambdaHandler;lambdaHandler=defaultLambdaHandler}plugin={...defaultPlugin,...plugin};plugin.timeoutEarly=plugin.timeoutEarlyInMillis>0;plugin.beforePrefetch?.();const beforeMiddlewares=[];const afterMiddlewares=[];const onErrorMiddlewares=[];const middy1=(event={},context={})=>{plugin.requestStart?.();const request={event,context,response:undefined,error:undefined,internal:plugin.internal??{}};return runRequest(request,[...beforeMiddlewares],lambdaHandler,[...afterMiddlewares],[...onErrorMiddlewares],plugin)};middy1.use=middlewares=>{if(!Array.isArray(middlewares)){middlewares=[middlewares]}for(const middleware of middlewares){const{before,after,onError}=middleware;if(!before&&!after&&!onError){throw new Error('Middleware must be an object containing at least one key among "before", "after", "onError"')}if(before)middy1.before(before);if(after)middy1.after(after);if(onError)middy1.onError(onError)}return middy1};middy1.before=beforeMiddleware=>{beforeMiddlewares.push(beforeMiddleware);return middy1};middy1.after=afterMiddleware=>{afterMiddlewares.unshift(afterMiddleware);return middy1};middy1.onError=onErrorMiddleware=>{onErrorMiddlewares.unshift(onErrorMiddleware);return middy1};middy1.handler=replaceLambdaHandler=>{lambdaHandler=replaceLambdaHandler;return middy1};return middy1};const runRequest=async(request,beforeMiddlewares,lambdaHandler,afterMiddlewares,onErrorMiddlewares,plugin)=>{const timeoutEarly=plugin.timeoutEarly&&request.context.getRemainingTimeInMillis;try{await runMiddlewares(request,beforeMiddlewares,plugin);if(request.response===undefined){plugin.beforeHandler?.();const handlerAbort=new AbortController;let timeoutAbort;if(timeoutEarly)timeoutAbort=new AbortController;request.response=await Promise.race([lambdaHandler(request.event,request.context,{signal:handlerAbort.signal}),timeoutEarly?setTimeoutPromise(request.context.getRemainingTimeInMillis()-plugin.timeoutEarlyInMillis,{signal:timeoutAbort.signal}).then(()=>{handlerAbort.abort();return plugin.timeoutEarlyResponse()}):Promise.race([])]);if(timeoutEarly)timeoutAbort.abort();plugin.afterHandler?.();await runMiddlewares(request,afterMiddlewares,plugin)}}catch(e){request.response=undefined;request.error=e;try{await runMiddlewares(request,onErrorMiddlewares,plugin)}catch(e){e.originalError=request.error;request.error=e;throw request.error}if(request.response===undefined)throw request.error}finally{await plugin.requestEnd?.(request)}return request.response};const runMiddlewares=async(request,middlewares,plugin)=>{for(const nextMiddleware of middlewares){plugin.beforeMiddleware?.(nextMiddleware.name);const res=await nextMiddleware(request);plugin.afterMiddleware?.(nextMiddleware.name);if(res!==undefined){request.response=res;return}}};const polyfillAbortController=()=>{if(process.version<"v15.0.0"){let _toStringTag;let _toStringTag1=Symbol.toStringTag;class AbortSignal{toString(){return"[object AbortSignal]"}get[_toStringTag1](){return"AbortSignal"}removeEventListener(name,handler){this.eventEmitter.removeListener(name,handler)}addEventListener(name,handler){this.eventEmitter.on(name,handler)}dispatchEvent(type){const event={type,target:this};const handlerName=`on${type}`;if(typeof this[handlerName]==="function")this[handlerName](event);this.eventEmitter.emit(type,event)}constructor(){this.eventEmitter=new EventEmitter;this.onabort=null;this.aborted=false}}return _toStringTag=Symbol.toStringTag,class AbortController{abort(){if(this.signal.aborted)return;this.signal.aborted=true;this.signal.dispatchEvent("abort")}toString(){return"[object AbortController]"}get[_toStringTag](){return"AbortController"}constructor(){this.signal=new AbortSignal}}}else{return AbortController}};global.AbortController=polyfillAbortController();const polyfillSetTimeoutPromise=()=>{return(ms,{signal})=>{if(signal.aborted){return Promise.reject(new Error("Aborted","AbortError"))}return new Promise((resolve,reject)=>{const abortHandler=()=>{clearTimeout(timeout);reject(new Error("Aborted","AbortError"))};const timeout=setTimeout(()=>{resolve();signal.removeEventListener("abort",abortHandler)},ms);signal.addEventListener("abort",abortHandler)})}};const setTimeoutPromise=polyfillSetTimeoutPromise();export default middy ^^^^^^ SyntaxError: Cannot use import statement outside a module 6 | 7 | import { captureLambdaHandler } from '../../src/middleware/middy'; > 8 | import middy from '@middy/core'; | ^ 9 | import { Tracer } from './../../src'; 10 | import type { Context, Handler } from 'aws-lambda/handler'; 11 | import { Segment, setContextMissingStrategy, Subsegment } from 'aws-xray-sdk-core'; at Runtime.createScriptFromCode (../../node_modules/jest-runtime/build/index.js:1728:14) at Object.<anonymous> (tests/unit/middy.test.ts:8:1) ``` </details> **Code examples** N/A **Benefits for you and the wider AWS community** <!-- What are the benefits your your proposed feature? --> **Describe alternatives you've considered** Not upgrading to 3.x and staying on version 2.5.x **Additional context** <!-- Add any other context or screenshots about the feature request here. --> ### Related issues, RFCs #861
1.0
Feature (all): evaluate migration to Middy 3.x - ## Description of the feature request **Problem statement** Middy, the dependency that we use to vend middleware has [recently released](https://github.com/middyjs/middy/releases/tag/3.0.0) a new major version (v 3.x). This version drops support for Node JS version 12.x and introduces some breaking changes. Given that Middy, as a whole, has a large surface and we are only using `@middy/core` we need to investigate on whether there's a path to upgrade to the newer version while still continuing to support Node JS 12.x in Powertools. **Additional context** Running the e2e tests on a branch that upgraded `@middy/core` to 3.0.1 seems to show that the newer release is compatible with Powertools on both versions 12 and 14 of Node ([see run results](https://github.com/awslabs/aws-lambda-powertools-typescript/actions/runs/2318810802)). At the same time, the unit tests for that same branch are failing for all utilities and specifically in the section that relates the middleware implementations. The error message (see detail below) seems to hint at incompatibilities between the new bundling of middy that now supports both CJS and ESM bundling, and our project's configuration since the errors are related to the imports and the tests don't run at all. <details> ```sh FAIL AWS Lambda Powertools utility: TRACER tests/unit/middy.test.ts ● Test suite failed to run Jest encountered an unexpected token Jest failed to parse a file. This happens e.g. when your code or its dependencies use non-standard JavaScript syntax, or when Jest is not configured to support such syntax. Out of the box Jest supports Babel, which will be used to transform your files into valid JS based on your Babel configuration. By default "node_modules" folder is ignored by transformers. Here's what you can do: • If you are trying to use ECMAScript Modules, see https://jestjs.io/docs/ecmascript-modules for how to enable it. • If you are trying to use TypeScript, see https://jestjs.io/docs/getting-started#using-typescript • To have some of your "node_modules" files transformed, you can specify a custom "transformIgnorePatterns" in your config. • If you need a custom transformation specify a "transform" option in your config. • If you simply want to mock your non-JS modules (e.g. binary assets) you can stub them out with the "moduleNameMapper" config option. You'll find more details and examples of these config options in the docs: https://jestjs.io/docs/configuration For information about custom transformations, see: https://jestjs.io/docs/code-transformation Details: /home/ec2-user/aws-lambda-powertools-typescript/node_modules/@middy/core/index.js:1 ({"Object.<anonymous>":function(module,exports,require,__dirname,__filename,jest){import{EventEmitter}from"events";const defaultLambdaHandler=()=>{};const defaultPlugin={timeoutEarlyInMillis:5,timeoutEarlyResponse:()=>{throw new Error("Timeout")}};const middy=(lambdaHandler=defaultLambdaHandler,plugin={})=>{if(typeof lambdaHandler!=="function"){plugin=lambdaHandler;lambdaHandler=defaultLambdaHandler}plugin={...defaultPlugin,...plugin};plugin.timeoutEarly=plugin.timeoutEarlyInMillis>0;plugin.beforePrefetch?.();const beforeMiddlewares=[];const afterMiddlewares=[];const onErrorMiddlewares=[];const middy1=(event={},context={})=>{plugin.requestStart?.();const request={event,context,response:undefined,error:undefined,internal:plugin.internal??{}};return runRequest(request,[...beforeMiddlewares],lambdaHandler,[...afterMiddlewares],[...onErrorMiddlewares],plugin)};middy1.use=middlewares=>{if(!Array.isArray(middlewares)){middlewares=[middlewares]}for(const middleware of middlewares){const{before,after,onError}=middleware;if(!before&&!after&&!onError){throw new Error('Middleware must be an object containing at least one key among "before", "after", "onError"')}if(before)middy1.before(before);if(after)middy1.after(after);if(onError)middy1.onError(onError)}return middy1};middy1.before=beforeMiddleware=>{beforeMiddlewares.push(beforeMiddleware);return middy1};middy1.after=afterMiddleware=>{afterMiddlewares.unshift(afterMiddleware);return middy1};middy1.onError=onErrorMiddleware=>{onErrorMiddlewares.unshift(onErrorMiddleware);return middy1};middy1.handler=replaceLambdaHandler=>{lambdaHandler=replaceLambdaHandler;return middy1};return middy1};const runRequest=async(request,beforeMiddlewares,lambdaHandler,afterMiddlewares,onErrorMiddlewares,plugin)=>{const timeoutEarly=plugin.timeoutEarly&&request.context.getRemainingTimeInMillis;try{await runMiddlewares(request,beforeMiddlewares,plugin);if(request.response===undefined){plugin.beforeHandler?.();const handlerAbort=new AbortController;let timeoutAbort;if(timeoutEarly)timeoutAbort=new AbortController;request.response=await Promise.race([lambdaHandler(request.event,request.context,{signal:handlerAbort.signal}),timeoutEarly?setTimeoutPromise(request.context.getRemainingTimeInMillis()-plugin.timeoutEarlyInMillis,{signal:timeoutAbort.signal}).then(()=>{handlerAbort.abort();return plugin.timeoutEarlyResponse()}):Promise.race([])]);if(timeoutEarly)timeoutAbort.abort();plugin.afterHandler?.();await runMiddlewares(request,afterMiddlewares,plugin)}}catch(e){request.response=undefined;request.error=e;try{await runMiddlewares(request,onErrorMiddlewares,plugin)}catch(e){e.originalError=request.error;request.error=e;throw request.error}if(request.response===undefined)throw request.error}finally{await plugin.requestEnd?.(request)}return request.response};const runMiddlewares=async(request,middlewares,plugin)=>{for(const nextMiddleware of middlewares){plugin.beforeMiddleware?.(nextMiddleware.name);const res=await nextMiddleware(request);plugin.afterMiddleware?.(nextMiddleware.name);if(res!==undefined){request.response=res;return}}};const polyfillAbortController=()=>{if(process.version<"v15.0.0"){let _toStringTag;let _toStringTag1=Symbol.toStringTag;class AbortSignal{toString(){return"[object AbortSignal]"}get[_toStringTag1](){return"AbortSignal"}removeEventListener(name,handler){this.eventEmitter.removeListener(name,handler)}addEventListener(name,handler){this.eventEmitter.on(name,handler)}dispatchEvent(type){const event={type,target:this};const handlerName=`on${type}`;if(typeof this[handlerName]==="function")this[handlerName](event);this.eventEmitter.emit(type,event)}constructor(){this.eventEmitter=new EventEmitter;this.onabort=null;this.aborted=false}}return _toStringTag=Symbol.toStringTag,class AbortController{abort(){if(this.signal.aborted)return;this.signal.aborted=true;this.signal.dispatchEvent("abort")}toString(){return"[object AbortController]"}get[_toStringTag](){return"AbortController"}constructor(){this.signal=new AbortSignal}}}else{return AbortController}};global.AbortController=polyfillAbortController();const polyfillSetTimeoutPromise=()=>{return(ms,{signal})=>{if(signal.aborted){return Promise.reject(new Error("Aborted","AbortError"))}return new Promise((resolve,reject)=>{const abortHandler=()=>{clearTimeout(timeout);reject(new Error("Aborted","AbortError"))};const timeout=setTimeout(()=>{resolve();signal.removeEventListener("abort",abortHandler)},ms);signal.addEventListener("abort",abortHandler)})}};const setTimeoutPromise=polyfillSetTimeoutPromise();export default middy ^^^^^^ SyntaxError: Cannot use import statement outside a module 6 | 7 | import { captureLambdaHandler } from '../../src/middleware/middy'; > 8 | import middy from '@middy/core'; | ^ 9 | import { Tracer } from './../../src'; 10 | import type { Context, Handler } from 'aws-lambda/handler'; 11 | import { Segment, setContextMissingStrategy, Subsegment } from 'aws-xray-sdk-core'; at Runtime.createScriptFromCode (../../node_modules/jest-runtime/build/index.js:1728:14) at Object.<anonymous> (tests/unit/middy.test.ts:8:1) ``` </details> **Code examples** N/A **Benefits for you and the wider AWS community** <!-- What are the benefits your your proposed feature? --> **Describe alternatives you've considered** Not upgrading to 3.x and staying on version 2.5.x **Additional context** <!-- Add any other context or screenshots about the feature request here. --> ### Related issues, RFCs #861
non_process
feature all evaluate migration to middy x description of the feature request problem statement middy the dependency that we use to vend middleware has a new major version v x this version drops support for node js version x and introduces some breaking changes given that middy as a whole has a large surface and we are only using middy core we need to investigate on whether there s a path to upgrade to the newer version while still continuing to support node js x in powertools additional context running the tests on a branch that upgraded middy core to seems to show that the newer release is compatible with powertools on both versions and of node at the same time the unit tests for that same branch are failing for all utilities and specifically in the section that relates the middleware implementations the error message see detail below seems to hint at incompatibilities between the new bundling of middy that now supports both cjs and esm bundling and our project s configuration since the errors are related to the imports and the tests don t run at all sh fail aws lambda powertools utility tracer tests unit middy test ts ● test suite failed to run jest encountered an unexpected token jest failed to parse a file this happens e g when your code or its dependencies use non standard javascript syntax or when jest is not configured to support such syntax out of the box jest supports babel which will be used to transform your files into valid js based on your babel configuration by default node modules folder is ignored by transformers here s what you can do • if you are trying to use ecmascript modules see for how to enable it • if you are trying to use typescript see • to have some of your node modules files transformed you can specify a custom transformignorepatterns in your config • if you need a custom transformation specify a transform option in your config • if you simply want to mock your non js modules e g binary assets you can stub them out with the modulenamemapper config option you ll find more details and examples of these config options in the docs for information about custom transformations see details home user aws lambda powertools typescript node modules middy core index js object function module exports require dirname filename jest import eventemitter from events const defaultlambdahandler const defaultplugin timeoutearlyinmillis timeoutearlyresponse throw new error timeout const middy lambdahandler defaultlambdahandler plugin if typeof lambdahandler function plugin lambdahandler lambdahandler defaultlambdahandler plugin defaultplugin plugin plugin timeoutearly plugin timeoutearlyinmillis plugin beforeprefetch const beforemiddlewares const aftermiddlewares const onerrormiddlewares const event context plugin requeststart const request event context response undefined error undefined internal plugin internal return runrequest request lambdahandler plugin use middlewares if array isarray middlewares middlewares for const middleware of middlewares const before after onerror middleware if before after onerror throw new error middleware must be an object containing at least one key among before after onerror if before before before if after after after if onerror onerror onerror return before beforemiddleware beforemiddlewares push beforemiddleware return after aftermiddleware aftermiddlewares unshift aftermiddleware return onerror onerrormiddleware onerrormiddlewares unshift onerrormiddleware return handler replacelambdahandler lambdahandler replacelambdahandler return return const runrequest async request beforemiddlewares lambdahandler aftermiddlewares onerrormiddlewares plugin const timeoutearly plugin timeoutearly request context getremainingtimeinmillis try await runmiddlewares request beforemiddlewares plugin if request response undefined plugin beforehandler const handlerabort new abortcontroller let timeoutabort if timeoutearly timeoutabort new abortcontroller request response await promise race if timeoutearly timeoutabort abort plugin afterhandler await runmiddlewares request aftermiddlewares plugin catch e request response undefined request error e try await runmiddlewares request onerrormiddlewares plugin catch e e originalerror request error request error e throw request error if request response undefined throw request error finally await plugin requestend request return request response const runmiddlewares async request middlewares plugin for const nextmiddleware of middlewares plugin beforemiddleware nextmiddleware name const res await nextmiddleware request plugin aftermiddleware nextmiddleware name if res undefined request response res return const polyfillabortcontroller if process version return ms signal if signal aborted return promise reject new error aborted aborterror return new promise resolve reject const aborthandler cleartimeout timeout reject new error aborted aborterror const timeout settimeout resolve signal removeeventlistener abort aborthandler ms signal addeventlistener abort aborthandler const settimeoutpromise polyfillsettimeoutpromise export default middy syntaxerror cannot use import statement outside a module import capturelambdahandler from src middleware middy import middy from middy core import tracer from src import type context handler from aws lambda handler import segment setcontextmissingstrategy subsegment from aws xray sdk core at runtime createscriptfromcode node modules jest runtime build index js at object tests unit middy test ts code examples n a benefits for you and the wider aws community describe alternatives you ve considered not upgrading to x and staying on version x additional context related issues rfcs
0
2,776
5,713,195,463
IssuesEvent
2017-04-19 07:00:09
g8os/grid
https://api.github.com/repos/g8os/grid
closed
AYS service for grid controller
process_wontfix type_feature
relates to : https://github.com/g8os/grid/issues/66 We need a service that deploy and configure the grid controller. actions: - install: - reserve a pair disk of disk, create btrfs raid1 on them for replication. - create container with the [grid flist](https://hub.gig.tech/maxux/grid.flist) - start AYS Server - start Grid API server
1.0
AYS service for grid controller - relates to : https://github.com/g8os/grid/issues/66 We need a service that deploy and configure the grid controller. actions: - install: - reserve a pair disk of disk, create btrfs raid1 on them for replication. - create container with the [grid flist](https://hub.gig.tech/maxux/grid.flist) - start AYS Server - start Grid API server
process
ays service for grid controller relates to we need a service that deploy and configure the grid controller actions install reserve a pair disk of disk create btrfs on them for replication create container with the start ays server start grid api server
1
59,041
14,365,891,805
IssuesEvent
2020-12-01 02:57:02
NixOS/nixpkgs
https://api.github.com/repos/NixOS/nixpkgs
closed
Vulnerability roundup 93: postgresql-9.6.17: 1 advisory [7.3]
1.severity: security
[search](https://search.nix.gsc.io/?q=postgresql&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=postgresql+in%3Apath&type=Code) * [ ] [CVE-2020-10733](https://nvd.nist.gov/vuln/detail/CVE-2020-10733) CVSSv3=7.3 (nixos-20.03) Scanned versions: nixos-20.03: 0d0660fde3b. Cc @danbst Cc @globin Cc @ocharles Cc @thoughtpolice
True
Vulnerability roundup 93: postgresql-9.6.17: 1 advisory [7.3] - [search](https://search.nix.gsc.io/?q=postgresql&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=postgresql+in%3Apath&type=Code) * [ ] [CVE-2020-10733](https://nvd.nist.gov/vuln/detail/CVE-2020-10733) CVSSv3=7.3 (nixos-20.03) Scanned versions: nixos-20.03: 0d0660fde3b. Cc @danbst Cc @globin Cc @ocharles Cc @thoughtpolice
non_process
vulnerability roundup postgresql advisory nixos scanned versions nixos cc danbst cc globin cc ocharles cc thoughtpolice
0
15,317
19,424,845,342
IssuesEvent
2021-12-21 03:08:50
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
GDAL Translate not respecting -a_srs for TFW files
Feedback Processing Bug
### What is the bug or the crash? Choose Translate, set a different SRS as the override AND/OR ``-a_srs EPSG:4326`` in Additional command-line parameters, and the generated TFW will have coordinates in the layer's original SRS. ### Steps to reproduce the issue 1. Run Translate from Processing Toolbox 2. Setup Override CRS as EPSG:4326 3. Add ``-co TFW=YES`` to additional command-line parameters 4. Observe that TFW CRS is input layer's original CRS (in this case, EPSG:32615 5. Add ``-a_srs EPSG:4326`` to additional command-line parameters 6. Observer that the new TFW is still in the wrong SRS QGIS History Item ``` """ Double-click on the history item or paste the command below to re-run the algorithm """ processing.run("gdal:translate", {'INPUT':'C:/Users/USERNAME/AppData/Local/Temp/processing_BxRQJM/6516c5016b0d4816a6bcb0cbdabd618a/OUTPUT.tif','TARGET_CRS':QgsCoordinateReferenceSystem('EPSG:4326'),'NODATA':None,'COPY_SUBDATASETS':False,'OPTIONS':'','EXTRA':'-co TFW=YES -a_srs EPSG:4326','DATA_TYPE':0,'OUTPUT':'TEMPORARY_OUTPUT'}) ``` ### Versions QGIS version 3.22.1-Białowieża QGIS code revision 663dcf8fb9 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.0 PROJ version 8.2.0 EPSG Registry database version v10.038 (2021-10-21) GEOS version 3.10.0-CAPI-1.16.0 SQLite version 3.35.2 PDAL version 2.3.0 PostgreSQL client version 13.0 SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins CalculateGeometry 0.6.4 cartography_tools 1.2.1 ImportPhotos 3.0.3 kmltools 3.1.15 LAStools 1.4 loadthemall 3.3.0 MemoryLayerSaver 4.0.4 pdaltools-master 0.1.5 photo2shape 3.0 processing_fusion 3.0 Projestions 1.0.1 qgis-maptiler-plugin 2.0.0 qgis2web 3.16.0 qgissprp 0.5.0 slyr 3.18.0 zoom_level 0.1 db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context ![image](https://user-images.githubusercontent.com/19295950/145323317-c5d30bdd-35df-4a1c-a4f8-e018181715cf.png) Thread exploring this in usage: https://community.opendronemap.org/t/is-there-a-way-to-produce-world-and-projection-files/9771/14?u=saijin_naib
1.0
GDAL Translate not respecting -a_srs for TFW files - ### What is the bug or the crash? Choose Translate, set a different SRS as the override AND/OR ``-a_srs EPSG:4326`` in Additional command-line parameters, and the generated TFW will have coordinates in the layer's original SRS. ### Steps to reproduce the issue 1. Run Translate from Processing Toolbox 2. Setup Override CRS as EPSG:4326 3. Add ``-co TFW=YES`` to additional command-line parameters 4. Observe that TFW CRS is input layer's original CRS (in this case, EPSG:32615 5. Add ``-a_srs EPSG:4326`` to additional command-line parameters 6. Observer that the new TFW is still in the wrong SRS QGIS History Item ``` """ Double-click on the history item or paste the command below to re-run the algorithm """ processing.run("gdal:translate", {'INPUT':'C:/Users/USERNAME/AppData/Local/Temp/processing_BxRQJM/6516c5016b0d4816a6bcb0cbdabd618a/OUTPUT.tif','TARGET_CRS':QgsCoordinateReferenceSystem('EPSG:4326'),'NODATA':None,'COPY_SUBDATASETS':False,'OPTIONS':'','EXTRA':'-co TFW=YES -a_srs EPSG:4326','DATA_TYPE':0,'OUTPUT':'TEMPORARY_OUTPUT'}) ``` ### Versions QGIS version 3.22.1-Białowieża QGIS code revision 663dcf8fb9 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.0 PROJ version 8.2.0 EPSG Registry database version v10.038 (2021-10-21) GEOS version 3.10.0-CAPI-1.16.0 SQLite version 3.35.2 PDAL version 2.3.0 PostgreSQL client version 13.0 SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins CalculateGeometry 0.6.4 cartography_tools 1.2.1 ImportPhotos 3.0.3 kmltools 3.1.15 LAStools 1.4 loadthemall 3.3.0 MemoryLayerSaver 4.0.4 pdaltools-master 0.1.5 photo2shape 3.0 processing_fusion 3.0 Projestions 1.0.1 qgis-maptiler-plugin 2.0.0 qgis2web 3.16.0 qgissprp 0.5.0 slyr 3.18.0 zoom_level 0.1 db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context ![image](https://user-images.githubusercontent.com/19295950/145323317-c5d30bdd-35df-4a1c-a4f8-e018181715cf.png) Thread exploring this in usage: https://community.opendronemap.org/t/is-there-a-way-to-produce-world-and-projection-files/9771/14?u=saijin_naib
process
gdal translate not respecting a srs for tfw files what is the bug or the crash choose translate set a different srs as the override and or a srs epsg in additional command line parameters and the generated tfw will have coordinates in the layer s original srs steps to reproduce the issue run translate from processing toolbox setup override crs as epsg add co tfw yes to additional command line parameters observe that tfw crs is input layer s original crs in this case epsg add a srs epsg to additional command line parameters observer that the new tfw is still in the wrong srs qgis history item double click on the history item or paste the command below to re run the algorithm processing run gdal translate input c users username appdata local temp processing bxrqjm output tif target crs qgscoordinatereferencesystem epsg nodata none copy subdatasets false options extra co tfw yes a srs epsg data type output temporary output versions qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins calculategeometry cartography tools importphotos kmltools lastools loadthemall memorylayersaver pdaltools master processing fusion projestions qgis maptiler plugin qgissprp slyr zoom level db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context thread exploring this in usage
1
87,753
15,790,301,850
IssuesEvent
2021-04-02 01:04:28
revfactory/task-api
https://api.github.com/repos/revfactory/task-api
opened
CVE-2019-0232 (High) detected in tomcat-embed-core-9.0.16.jar
security vulnerability
## CVE-2019-0232 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.16.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: task-api/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.3.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.3.RELEASE.jar - :x: **tomcat-embed-core-9.0.16.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When running on Windows with enableCmdLineArguments enabled, the CGI Servlet in Apache Tomcat 9.0.0.M1 to 9.0.17, 8.5.0 to 8.5.39 and 7.0.0 to 7.0.93 is vulnerable to Remote Code Execution due to a bug in the way the JRE passes command line arguments to Windows. The CGI Servlet is disabled by default. The CGI option enableCmdLineArguments is disable by default in Tomcat 9.0.x (and will be disabled by default in all versions in response to this vulnerability). For a detailed explanation of the JRE behaviour, see Markus Wulftange's blog (https://codewhitesec.blogspot.com/2016/02/java-and-command-line-injections-in-windows.html) and this archived MSDN blog (https://web.archive.org/web/20161228144344/https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/). <p>Publish Date: 2019-04-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0232>CVE-2019-0232</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232</a></p> <p>Release Date: 2019-04-15</p> <p>Fix Resolution: 9.0.18,8.5.40,7.0.94</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-0232 (High) detected in tomcat-embed-core-9.0.16.jar - ## CVE-2019-0232 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.16.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Path to dependency file: task-api/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.3.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.1.3.RELEASE.jar - :x: **tomcat-embed-core-9.0.16.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> When running on Windows with enableCmdLineArguments enabled, the CGI Servlet in Apache Tomcat 9.0.0.M1 to 9.0.17, 8.5.0 to 8.5.39 and 7.0.0 to 7.0.93 is vulnerable to Remote Code Execution due to a bug in the way the JRE passes command line arguments to Windows. The CGI Servlet is disabled by default. The CGI option enableCmdLineArguments is disable by default in Tomcat 9.0.x (and will be disabled by default in all versions in response to this vulnerability). For a detailed explanation of the JRE behaviour, see Markus Wulftange's blog (https://codewhitesec.blogspot.com/2016/02/java-and-command-line-injections-in-windows.html) and this archived MSDN blog (https://web.archive.org/web/20161228144344/https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/). <p>Publish Date: 2019-04-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0232>CVE-2019-0232</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232</a></p> <p>Release Date: 2019-04-15</p> <p>Fix Resolution: 9.0.18,8.5.40,7.0.94</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation path to dependency file task api build gradle path to vulnerable library root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details when running on windows with enablecmdlinearguments enabled the cgi servlet in apache tomcat to to and to is vulnerable to remote code execution due to a bug in the way the jre passes command line arguments to windows the cgi servlet is disabled by default the cgi option enablecmdlinearguments is disable by default in tomcat x and will be disabled by default in all versions in response to this vulnerability for a detailed explanation of the jre behaviour see markus wulftange s blog and this archived msdn blog publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
149,588
13,283,984,904
IssuesEvent
2020-08-24 05:04:48
BuildForSDGCohort2/team-024-group-a
https://api.github.com/repos/BuildForSDGCohort2/team-024-group-a
closed
Update readme.md
documentation
Readme.md is just the description of our app the backend and the frontend has each readme.md for description of starting the server
1.0
Update readme.md - Readme.md is just the description of our app the backend and the frontend has each readme.md for description of starting the server
non_process
update readme md readme md is just the description of our app the backend and the frontend has each readme md for description of starting the server
0
18,761
24,664,034,442
IssuesEvent
2022-10-18 08:57:07
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Textual definition update: New definition of mitotic spindle assembly checkpoint signaling including START and END
cell cycle and DNA processes textual definition
GO:0007094 mitotic spindle assembly checkpoint signaling Existing A signal transduction process that contributes to a mitotic cell cycle spindle assembly checkpoint, that delays the metaphase/anaphase transition of a mitotic nuclear division until the spindle is correctly assembled and chromosomes are attached to the spindle. Suggested A mitotic cell cycle checkpoint that delays mitotic sister chromatid separation and consequently the mitotic metaphase/anaphase transition until the spindle is correctly assembled and chromosomes are attached to the spindle. BEGINNING AND END Spindle assembly checkpoint signalling begins with the activated Mph family kinase, and results in the inhibition of the Anaphase Promoting Complex and its activator Sleepy/Cdc20 by the mitotic checkpoint complex (MCC). (This works for all species) pers comm VW and Silke Hauf
1.0
Textual definition update: New definition of mitotic spindle assembly checkpoint signaling including START and END - GO:0007094 mitotic spindle assembly checkpoint signaling Existing A signal transduction process that contributes to a mitotic cell cycle spindle assembly checkpoint, that delays the metaphase/anaphase transition of a mitotic nuclear division until the spindle is correctly assembled and chromosomes are attached to the spindle. Suggested A mitotic cell cycle checkpoint that delays mitotic sister chromatid separation and consequently the mitotic metaphase/anaphase transition until the spindle is correctly assembled and chromosomes are attached to the spindle. BEGINNING AND END Spindle assembly checkpoint signalling begins with the activated Mph family kinase, and results in the inhibition of the Anaphase Promoting Complex and its activator Sleepy/Cdc20 by the mitotic checkpoint complex (MCC). (This works for all species) pers comm VW and Silke Hauf
process
textual definition update new definition of mitotic spindle assembly checkpoint signaling including start and end go mitotic spindle assembly checkpoint signaling existing a signal transduction process that contributes to a mitotic cell cycle spindle assembly checkpoint that delays the metaphase anaphase transition of a mitotic nuclear division until the spindle is correctly assembled and chromosomes are attached to the spindle suggested a mitotic cell cycle checkpoint that delays mitotic sister chromatid separation and consequently the mitotic metaphase anaphase transition until the spindle is correctly assembled and chromosomes are attached to the spindle beginning and end spindle assembly checkpoint signalling begins with the activated mph family kinase and results in the inhibition of the anaphase promoting complex and its activator sleepy by the mitotic checkpoint complex mcc this works for all species pers comm vw and silke hauf
1
5,331
8,147,954,812
IssuesEvent
2018-08-22 02:46:49
zotero/zotero
https://api.github.com/repos/zotero/zotero
closed
Style/translator initialization failure shouldn't freeze the doc plugins
Word Processor Integration
See https://forums.zotero.org/discussion/73274/word-2010-plugin-only-gives-progress-bar-when-trying-to-add-citation-or-bibliography I suppose subsequent invocations of `Zotero.Styles.init()` should throw an error?
1.0
Style/translator initialization failure shouldn't freeze the doc plugins - See https://forums.zotero.org/discussion/73274/word-2010-plugin-only-gives-progress-bar-when-trying-to-add-citation-or-bibliography I suppose subsequent invocations of `Zotero.Styles.init()` should throw an error?
process
style translator initialization failure shouldn t freeze the doc plugins see i suppose subsequent invocations of zotero styles init should throw an error
1
185,933
6,732,055,655
IssuesEvent
2017-10-18 09:57:14
ballerinalang/composer
https://api.github.com/repos/ballerinalang/composer
closed
Cannot delete the try-catch block when finally is added
0.94-pre-release Priority/High Severity/Major Type/Bug
1. Add finally block from the source view 2. Try to delete the entire try-catch Cannot delete the try-catch block when finally is added
1.0
Cannot delete the try-catch block when finally is added - 1. Add finally block from the source view 2. Try to delete the entire try-catch Cannot delete the try-catch block when finally is added
non_process
cannot delete the try catch block when finally is added add finally block from the source view try to delete the entire try catch cannot delete the try catch block when finally is added
0
262,311
19,782,695,307
IssuesEvent
2022-01-18 00:02:23
elastic/eui
https://api.github.com/repos/elastic/eui
closed
[EuiSearchBar] Document defaultFields option on toESQuery
documentation stale-issue stale-issue-closed
It seems that toESQuery options accepts a `defaultFields` array, but the docs to not indicate that, as well as a few other options available.
1.0
[EuiSearchBar] Document defaultFields option on toESQuery - It seems that toESQuery options accepts a `defaultFields` array, but the docs to not indicate that, as well as a few other options available.
non_process
document defaultfields option on toesquery it seems that toesquery options accepts a defaultfields array but the docs to not indicate that as well as a few other options available
0
407,226
11,908,259,639
IssuesEvent
2020-03-31 00:26:05
kubeflow/examples
https://api.github.com/repos/kubeflow/examples
closed
Create an mnist E2E notebook for Azure
area/examples kind/feature platform/azure priority/p1
For 1.0 the core CUJ is build, train, deploy. In kubeflow/examples#723 we've created a notebook that highlights this core CUJ. We'd like to have variations of this notebook for all the different platforms/clouds on which Kubeflow is supported. So we'd like to have a variant of this notebook that is suited for Azure. That primarily means updating the notebook to use Azure blob store. Alternatively you could use minio running on the cluster but using Azure blob store would probably be better. /cc @aronchick
1.0
Create an mnist E2E notebook for Azure - For 1.0 the core CUJ is build, train, deploy. In kubeflow/examples#723 we've created a notebook that highlights this core CUJ. We'd like to have variations of this notebook for all the different platforms/clouds on which Kubeflow is supported. So we'd like to have a variant of this notebook that is suited for Azure. That primarily means updating the notebook to use Azure blob store. Alternatively you could use minio running on the cluster but using Azure blob store would probably be better. /cc @aronchick
non_process
create an mnist notebook for azure for the core cuj is build train deploy in kubeflow examples we ve created a notebook that highlights this core cuj we d like to have variations of this notebook for all the different platforms clouds on which kubeflow is supported so we d like to have a variant of this notebook that is suited for azure that primarily means updating the notebook to use azure blob store alternatively you could use minio running on the cluster but using azure blob store would probably be better cc aronchick
0
5,127
7,894,683,629
IssuesEvent
2018-06-28 22:32:39
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
vSphere post-processor mark as template
enhancement post-processor/vsphere
Is it possible to have the vSphere provisioner automatically mark the uploaded VM as a template?
1.0
vSphere post-processor mark as template - Is it possible to have the vSphere provisioner automatically mark the uploaded VM as a template?
process
vsphere post processor mark as template is it possible to have the vsphere provisioner automatically mark the uploaded vm as a template
1
4,729
7,571,883,366
IssuesEvent
2018-04-23 13:34:24
zotero/zotero
https://api.github.com/repos/zotero/zotero
closed
Add uninstallation options for word processor plugins to Cite pane
Word Processor Integration
Shouldn't have to tell people to find their Word startup folder https://forums.zotero.org/discussion/69411/how-remove-zotero-plugin-in-word
1.0
Add uninstallation options for word processor plugins to Cite pane - Shouldn't have to tell people to find their Word startup folder https://forums.zotero.org/discussion/69411/how-remove-zotero-plugin-in-word
process
add uninstallation options for word processor plugins to cite pane shouldn t have to tell people to find their word startup folder
1
160,730
12,519,732,887
IssuesEvent
2020-06-03 14:51:24
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
closed
RVD#1979: Using Fault to parse untrusted XML data is known to be vulnerable to XML..., /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52
bandit bug components software duplicate invalid robot component: ROS static analysis testing triage version: noetic
```yaml { "id": 1979, "title": "RVD#1979: Using Fault to parse untrusted XML data is known to be vulnerable to XML..., /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52", "type": "bug", "description": "HIGH confidence of HIGH severity bug. Using Fault to parse untrusted XML data is known to be vulnerable to XML attacks. Use defused.xmlrpc.monkey_patch() function to monkey-patch xmlrpclib and mitigate XML vulnerabilities. at /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52 See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": [ "https://github.com/aliasrobotics/RVD/issues/1979", "https://bandit.readthedocs.io/en/latest/blacklists/blacklist_imports.html#b411-import-xmlrpclib" ], "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-29 (08:41)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-29 (08:41)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "https://github.com/aliasrobotics/RVD/issues/1979", "reproducibility": "always", "trace": "/opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
RVD#1979: Using Fault to parse untrusted XML data is known to be vulnerable to XML..., /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52 - ```yaml { "id": 1979, "title": "RVD#1979: Using Fault to parse untrusted XML data is known to be vulnerable to XML..., /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52", "type": "bug", "description": "HIGH confidence of HIGH severity bug. Using Fault to parse untrusted XML data is known to be vulnerable to XML attacks. Use defused.xmlrpc.monkey_patch() function to monkey-patch xmlrpclib and mitigate XML vulnerabilities. at /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52 See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": [ "https://github.com/aliasrobotics/RVD/issues/1979", "https://bandit.readthedocs.io/en/latest/blacklists/blacklist_imports.html#b411-import-xmlrpclib" ], "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-29 (08:41)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-29 (08:41)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "https://github.com/aliasrobotics/RVD/issues/1979", "reproducibility": "always", "trace": "/opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
non_process
rvd using fault to parse untrusted xml data is known to be vulnerable to xml opt ros noetic ws src ros comm rostopic src rostopic init py yaml id title rvd using fault to parse untrusted xml data is known to be vulnerable to xml opt ros noetic ws src ros comm rostopic src rostopic init py type bug description high confidence of high severity bug using fault to parse untrusted xml data is known to be vulnerable to xml attacks use defused xmlrpc monkey patch function to monkey patch xmlrpclib and mitigate xml vulnerabilities at opt ros noetic ws src ros comm rostopic src rostopic init py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace opt ros noetic ws src ros comm rostopic src rostopic init py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
0
4,644
7,488,632,794
IssuesEvent
2018-04-06 02:40:15
UnbFeelings/unb-feelings-GQA
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
closed
Aplicar atualizações de FeedBack no Processo
help wanted process
Atualizações: - [x] Renomear o artefato Resultado Padronizado para Resultado Consolidado; - [x] Reorganizar o subprocesso que lida com as não conformidades; - [x] Criar atividade de **Definir Critérios de Avaliação** (Após definir objetos de Avaliação); - Gerar artefato de Critérios de Avaliação; - [x] Arrumar no último sub-Processo a forma como serão tratadas as inconformidades do ciclo (Atividade ou Sub-Processo menor); - [x] Criar atividade para Definir ferramentas; - Gerar artefato de ferramentas; - [x] Tirar ferramenta de dentro do Sub-Processo Detalhar; - [x] Analisar Processo, Critérios de Avaliação, Definir Ferramentas; - [x] Eliminar tarefas desnecessárias do Processo; OBS: MUDANÇAS NO PROCESSO SERÃO PROVIDAS POR UM PROCESSO DE APOIA QUE FOGE O ESCOPO DA EQUIPE DE GQA; **Comentário com as palavras da professora:** Na atividade 3.3 do processo, chamada Registrar resultados, existe uma saída chamada "Resultado Padronizado", o que não deixa muito claro o que é, sugeriu-se trocar para "Resultado Consolidado" ** Deve ser atualizado tanto no diagrama quanto na descrição da atividade do processo - Explicitar no processo que a não conformidade deve ser acompanhada por um prazo diferente das auditorias. Por exemplo, o prazo para resolver o problema X é dois dias, e não até a próxima auditoria - Colocar na descrição da atividade do processo: ao registrar a não conformidade, é necessário informar a complexidade dela
1.0
Aplicar atualizações de FeedBack no Processo - Atualizações: - [x] Renomear o artefato Resultado Padronizado para Resultado Consolidado; - [x] Reorganizar o subprocesso que lida com as não conformidades; - [x] Criar atividade de **Definir Critérios de Avaliação** (Após definir objetos de Avaliação); - Gerar artefato de Critérios de Avaliação; - [x] Arrumar no último sub-Processo a forma como serão tratadas as inconformidades do ciclo (Atividade ou Sub-Processo menor); - [x] Criar atividade para Definir ferramentas; - Gerar artefato de ferramentas; - [x] Tirar ferramenta de dentro do Sub-Processo Detalhar; - [x] Analisar Processo, Critérios de Avaliação, Definir Ferramentas; - [x] Eliminar tarefas desnecessárias do Processo; OBS: MUDANÇAS NO PROCESSO SERÃO PROVIDAS POR UM PROCESSO DE APOIA QUE FOGE O ESCOPO DA EQUIPE DE GQA; **Comentário com as palavras da professora:** Na atividade 3.3 do processo, chamada Registrar resultados, existe uma saída chamada "Resultado Padronizado", o que não deixa muito claro o que é, sugeriu-se trocar para "Resultado Consolidado" ** Deve ser atualizado tanto no diagrama quanto na descrição da atividade do processo - Explicitar no processo que a não conformidade deve ser acompanhada por um prazo diferente das auditorias. Por exemplo, o prazo para resolver o problema X é dois dias, e não até a próxima auditoria - Colocar na descrição da atividade do processo: ao registrar a não conformidade, é necessário informar a complexidade dela
process
aplicar atualizações de feedback no processo atualizações renomear o artefato resultado padronizado para resultado consolidado reorganizar o subprocesso que lida com as não conformidades criar atividade de definir critérios de avaliação após definir objetos de avaliação gerar artefato de critérios de avaliação arrumar no último sub processo a forma como serão tratadas as inconformidades do ciclo atividade ou sub processo menor criar atividade para definir ferramentas gerar artefato de ferramentas tirar ferramenta de dentro do sub processo detalhar analisar processo critérios de avaliação definir ferramentas eliminar tarefas desnecessárias do processo obs mudanças no processo serão providas por um processo de apoia que foge o escopo da equipe de gqa comentário com as palavras da professora na atividade do processo chamada registrar resultados existe uma saída chamada resultado padronizado o que não deixa muito claro o que é sugeriu se trocar para resultado consolidado deve ser atualizado tanto no diagrama quanto na descrição da atividade do processo explicitar no processo que a não conformidade deve ser acompanhada por um prazo diferente das auditorias por exemplo o prazo para resolver o problema x é dois dias e não até a próxima auditoria colocar na descrição da atividade do processo ao registrar a não conformidade é necessário informar a complexidade dela
1
12,562
14,981,350,171
IssuesEvent
2021-01-28 14:44:36
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
reopened
Add new user page > Role > 'Super Admin' label should be changed to 'Make this user a superadmin'
Bug P2 Participant manager Process: Dev Process: Fixed Process: Tested QA Process: Tested dev
AR : Role is displaying as 'Super Admin' ER : Add new user page > Role > 'Super Admin' label should be changed to 'Make this user a superadmin' ![role](https://user-images.githubusercontent.com/71445210/100261797-4d529600-2f71-11eb-83ff-5a6aa5ec70aa.png)
4.0
Add new user page > Role > 'Super Admin' label should be changed to 'Make this user a superadmin' - AR : Role is displaying as 'Super Admin' ER : Add new user page > Role > 'Super Admin' label should be changed to 'Make this user a superadmin' ![role](https://user-images.githubusercontent.com/71445210/100261797-4d529600-2f71-11eb-83ff-5a6aa5ec70aa.png)
process
add new user page role super admin label should be changed to make this user a superadmin ar role is displaying as super admin er add new user page role super admin label should be changed to make this user a superadmin
1
2,283
5,108,317,835
IssuesEvent
2017-01-05 17:19:55
DynareTeam/dynare
https://api.github.com/repos/DynareTeam/dynare
closed
Depth issue
bug preprocessor
Looking into #1175 , by reverting commit 3c7e60b744567f6f39a9c611bce6dcaadcd52bc6, I obtained the following error from matlab when trying to run Christiano-Motto-Rostagno model (the one in subfolder `figure4` of the archive available [here](http://faculty.wcas.northwestern.edu/~lchrist/research/ECB/risk_shocks/20100922_data.zip) ``` Error: File: cmr_static.m Line: 1292 Column: 16331 Nesting of {, [, and ( cannot exceed a depth of 32. ``` @MichelJuillard May this be a consequence of your patch about on auxiliary variables in steady state and static files (see #1133)?
1.0
Depth issue - Looking into #1175 , by reverting commit 3c7e60b744567f6f39a9c611bce6dcaadcd52bc6, I obtained the following error from matlab when trying to run Christiano-Motto-Rostagno model (the one in subfolder `figure4` of the archive available [here](http://faculty.wcas.northwestern.edu/~lchrist/research/ECB/risk_shocks/20100922_data.zip) ``` Error: File: cmr_static.m Line: 1292 Column: 16331 Nesting of {, [, and ( cannot exceed a depth of 32. ``` @MichelJuillard May this be a consequence of your patch about on auxiliary variables in steady state and static files (see #1133)?
process
depth issue looking into by reverting commit i obtained the following error from matlab when trying to run christiano motto rostagno model the one in subfolder of the archive available error file cmr static m line column nesting of and cannot exceed a depth of micheljuillard may this be a consequence of your patch about on auxiliary variables in steady state and static files see
1
276,043
23,961,956,210
IssuesEvent
2022-09-12 19:59:18
Sheldonfrith/sst-scheduled-api-test-testee-repo
https://api.github.com/repos/Sheldonfrith/sst-scheduled-api-test-testee-repo
closed
Auto-Generated: g/json - get - 200 - Test: deliberateBug - Response body is not an empty object empty
sst-scheduled-api-testing
{"userDefinedErrorMessage":"Test: deliberateBug - Response body is not an empty object empty","endpointUrl":"https://api.sunrise-sunset.org/json","response":{"status":200,"headers":{"server":"nginx","date":"Tue, 30 Aug 2022 00:54:02 GMT","content-type":"application/json","transfer-encoding":"chunked","connection":"close","vary":"Accept-Encoding","access-control-allow-origin":"*"},"body":"{\"results\":{\"sunrise\":\"10:16:46 AM\",\"sunset\":\"11:28:22 PM\",\"solar_noon\":\"4:52:34 PM\",\"day_length\":\"13:11:36\",\"civil_twilight_begin\":\"9:50:18 AM\",\"civil_twilight_end\":\"11:54:50 PM\",\"nautical_twilight_begin\":\"9:17:01 AM\",\"nautical_twilight_end\":\"12:28:07 AM\",\"astronomical_twilight_begin\":\"8:42:14 AM\",\"astronomical_twilight_end\":\"1:02:54 AM\"},\"status\":\"OK\"}"},"request":{"headers":{"Accept":"application/json, text/plain, */*","User-Agent":"axios/0.27.2"}}}
1.0
Auto-Generated: g/json - get - 200 - Test: deliberateBug - Response body is not an empty object empty - {"userDefinedErrorMessage":"Test: deliberateBug - Response body is not an empty object empty","endpointUrl":"https://api.sunrise-sunset.org/json","response":{"status":200,"headers":{"server":"nginx","date":"Tue, 30 Aug 2022 00:54:02 GMT","content-type":"application/json","transfer-encoding":"chunked","connection":"close","vary":"Accept-Encoding","access-control-allow-origin":"*"},"body":"{\"results\":{\"sunrise\":\"10:16:46 AM\",\"sunset\":\"11:28:22 PM\",\"solar_noon\":\"4:52:34 PM\",\"day_length\":\"13:11:36\",\"civil_twilight_begin\":\"9:50:18 AM\",\"civil_twilight_end\":\"11:54:50 PM\",\"nautical_twilight_begin\":\"9:17:01 AM\",\"nautical_twilight_end\":\"12:28:07 AM\",\"astronomical_twilight_begin\":\"8:42:14 AM\",\"astronomical_twilight_end\":\"1:02:54 AM\"},\"status\":\"OK\"}"},"request":{"headers":{"Accept":"application/json, text/plain, */*","User-Agent":"axios/0.27.2"}}}
non_process
auto generated g json get test deliberatebug response body is not an empty object empty userdefinederrormessage test deliberatebug response body is not an empty object empty endpointurl aug gmt content type application json transfer encoding chunked connection close vary accept encoding access control allow origin body results sunrise am sunset pm solar noon pm day length civil twilight begin am civil twilight end pm nautical twilight begin am nautical twilight end am astronomical twilight begin am astronomical twilight end am status ok request headers accept application json text plain user agent axios
0
15,727
19,899,566,280
IssuesEvent
2022-01-25 05:46:29
SigNoz/signoz
https://api.github.com/repos/SigNoz/signoz
closed
Add Events data from OpenTelemetry spans
processors query-service
While flattening the data, we should also store `Events` data from spans to Apache Druid. This contains `exception` data apart from reasons for erroneous spans like `Redis timeout`
1.0
Add Events data from OpenTelemetry spans - While flattening the data, we should also store `Events` data from spans to Apache Druid. This contains `exception` data apart from reasons for erroneous spans like `Redis timeout`
process
add events data from opentelemetry spans while flattening the data we should also store events data from spans to apache druid this contains exception data apart from reasons for erroneous spans like redis timeout
1
368,348
25,789,442,856
IssuesEvent
2022-12-10 01:10:24
Navnedia/Link-Shortener
https://api.github.com/repos/Navnedia/Link-Shortener
opened
Add documentation for the project
documentation
Write an organized README and potentially other documents detailing information about the project: - [ ] What is the project? - [ ] What's the purpose of the project? Why was this project developed? - [ ] What tools and concepts were used to develop this project? - [ ] Skills and tools learned and developed along the way - [ ] Document the components of the project - [ ] The API endpoints - [ ] Describe how to install and use the application
1.0
Add documentation for the project - Write an organized README and potentially other documents detailing information about the project: - [ ] What is the project? - [ ] What's the purpose of the project? Why was this project developed? - [ ] What tools and concepts were used to develop this project? - [ ] Skills and tools learned and developed along the way - [ ] Document the components of the project - [ ] The API endpoints - [ ] Describe how to install and use the application
non_process
add documentation for the project write an organized readme and potentially other documents detailing information about the project what is the project what s the purpose of the project why was this project developed what tools and concepts were used to develop this project skills and tools learned and developed along the way document the components of the project the api endpoints describe how to install and use the application
0
11,420
14,246,229,201
IssuesEvent
2020-11-19 09:47:20
pystatgen/sgkit
https://api.github.com/repos/pystatgen/sgkit
closed
Release 0.1.0a1
process + tools
This is an umbrella issue for our first release. This would be an alpha [pre-release](https://www.python.org/dev/peps/pep-0440/#pre-releases) (suggested version number: 0.1.0a1), which means that users would have to opt in to use it by specifying the `--pre` flag: `pip install --pre sgkit`. Here are some of the tasks we need to do - please suggest more (or say if they are not needed). - [ ] cyvcf2 wheel (should we wait for this?) - [ ] Build release artifacts locally - [ ] Add release number to documentation URL - [ ] Use TestPyPI to test release process - [ ] Update installation instructions for users - [ ] Document release process Note that these are all about what we need to do to make a release; since it's a pre-release, the exact set of features doesn't matter so much.
1.0
Release 0.1.0a1 - This is an umbrella issue for our first release. This would be an alpha [pre-release](https://www.python.org/dev/peps/pep-0440/#pre-releases) (suggested version number: 0.1.0a1), which means that users would have to opt in to use it by specifying the `--pre` flag: `pip install --pre sgkit`. Here are some of the tasks we need to do - please suggest more (or say if they are not needed). - [ ] cyvcf2 wheel (should we wait for this?) - [ ] Build release artifacts locally - [ ] Add release number to documentation URL - [ ] Use TestPyPI to test release process - [ ] Update installation instructions for users - [ ] Document release process Note that these are all about what we need to do to make a release; since it's a pre-release, the exact set of features doesn't matter so much.
process
release this is an umbrella issue for our first release this would be an alpha suggested version number which means that users would have to opt in to use it by specifying the pre flag pip install pre sgkit here are some of the tasks we need to do please suggest more or say if they are not needed wheel should we wait for this build release artifacts locally add release number to documentation url use testpypi to test release process update installation instructions for users document release process note that these are all about what we need to do to make a release since it s a pre release the exact set of features doesn t matter so much
1
6,450
9,546,286,441
IssuesEvent
2019-05-01 19:29:41
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Bug: Lose education data when education added
Apply Process Bug State Dept.
Environment: Test Browser: Chrome Steps to reproduce: 1) Answer the 3 education questions and GPA 2) Select Add Education (when you don't already have education) 3) Enter education and save 4) Questions and GPA are lost - Once saved, I can edit the questions and follow the same path and the edits will save. Appears to be an issue when blank and the page hasn't been saved.
1.0
Bug: Lose education data when education added - Environment: Test Browser: Chrome Steps to reproduce: 1) Answer the 3 education questions and GPA 2) Select Add Education (when you don't already have education) 3) Enter education and save 4) Questions and GPA are lost - Once saved, I can edit the questions and follow the same path and the edits will save. Appears to be an issue when blank and the page hasn't been saved.
process
bug lose education data when education added environment test browser chrome steps to reproduce answer the education questions and gpa select add education when you don t already have education enter education and save questions and gpa are lost once saved i can edit the questions and follow the same path and the edits will save appears to be an issue when blank and the page hasn t been saved
1
11,786
14,617,358,501
IssuesEvent
2020-12-22 14:40:17
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
opened
[PM] Participant details > Consent document is not shown for closed study having Eligibility test
Bug P0 Participant manager Process: Dev
**Steps:** 1. Publish a study having token validation and Eligibility test 2. PM admin invites a user 3. Mobile participant enrolls into the study using valid token and passes the eligibility test 4. User enrolls successfully 5. Observe the participant details --> Consent History A/R: Consent document is not shown for closed study having Eligibility test E/R: Consent document should be shown for all type of studies ![Screenshot_7](https://user-images.githubusercontent.com/60386291/102899937-a854ab80-4491-11eb-9f82-55489b1c8b98.png)
1.0
[PM] Participant details > Consent document is not shown for closed study having Eligibility test - **Steps:** 1. Publish a study having token validation and Eligibility test 2. PM admin invites a user 3. Mobile participant enrolls into the study using valid token and passes the eligibility test 4. User enrolls successfully 5. Observe the participant details --> Consent History A/R: Consent document is not shown for closed study having Eligibility test E/R: Consent document should be shown for all type of studies ![Screenshot_7](https://user-images.githubusercontent.com/60386291/102899937-a854ab80-4491-11eb-9f82-55489b1c8b98.png)
process
participant details consent document is not shown for closed study having eligibility test steps publish a study having token validation and eligibility test pm admin invites a user mobile participant enrolls into the study using valid token and passes the eligibility test user enrolls successfully observe the participant details consent history a r consent document is not shown for closed study having eligibility test e r consent document should be shown for all type of studies
1
46,942
7,296,112,410
IssuesEvent
2018-02-26 09:41:22
alphagov/govuk-frontend
https://api.github.com/repos/alphagov/govuk-frontend
opened
Explain why we use the current CSS methodologies
documentation
We have detail on what we're doing, but not why here https://github.com/alphagov/govuk-frontend/blob/master/docs/coding-standards/css.md This has come up already and it'd be good have something comprehensive to link people to. Might be good as a blog post.
1.0
Explain why we use the current CSS methodologies - We have detail on what we're doing, but not why here https://github.com/alphagov/govuk-frontend/blob/master/docs/coding-standards/css.md This has come up already and it'd be good have something comprehensive to link people to. Might be good as a blog post.
non_process
explain why we use the current css methodologies we have detail on what we re doing but not why here this has come up already and it d be good have something comprehensive to link people to might be good as a blog post
0
186,046
14,394,554,906
IssuesEvent
2020-12-03 01:34:08
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
chengfangang/fabric-ca-gm-test: cmd/fabric-ca-client/main_test.go; 3 LoC
fresh test tiny
Found a possible issue in [chengfangang/fabric-ca-gm-test](https://www.github.com/chengfangang/fabric-ca-gm-test) at [cmd/fabric-ca-client/main_test.go](https://github.com/chengfangang/fabric-ca-gm-test/blob/6aca151d16f6978ce651134c28cb34f1a6c29c85/cmd/fabric-ca-client/main_test.go#L132-L134) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to e at line 133 may start a goroutine [Click here to see the code in its original context.](https://github.com/chengfangang/fabric-ca-gm-test/blob/6aca151d16f6978ce651134c28cb34f1a6c29c85/cmd/fabric-ca-client/main_test.go#L132-L134) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, e := range errCases { extraArgErrorTest(&e, t) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 6aca151d16f6978ce651134c28cb34f1a6c29c85
1.0
chengfangang/fabric-ca-gm-test: cmd/fabric-ca-client/main_test.go; 3 LoC - Found a possible issue in [chengfangang/fabric-ca-gm-test](https://www.github.com/chengfangang/fabric-ca-gm-test) at [cmd/fabric-ca-client/main_test.go](https://github.com/chengfangang/fabric-ca-gm-test/blob/6aca151d16f6978ce651134c28cb34f1a6c29c85/cmd/fabric-ca-client/main_test.go#L132-L134) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to e at line 133 may start a goroutine [Click here to see the code in its original context.](https://github.com/chengfangang/fabric-ca-gm-test/blob/6aca151d16f6978ce651134c28cb34f1a6c29c85/cmd/fabric-ca-client/main_test.go#L132-L134) <details> <summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary> ```go for _, e := range errCases { extraArgErrorTest(&e, t) } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 6aca151d16f6978ce651134c28cb34f1a6c29c85
non_process
chengfangang fabric ca gm test cmd fabric ca client main test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to e at line may start a goroutine click here to show the line s of go which triggered the analyzer go for e range errcases extraargerrortest e t leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
14,230
4,856,810,590
IssuesEvent
2016-11-12 08:08:56
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
new menu item check
No Code Attached Yet
### Steps to reproduce the issue It is not allowed that a menu item alias is the same as a sub-folder in your Joomla installation. The message will be: "Save failed with the following error: A first level menu item alias can't be 'tmp' because 'tmp' is a sub-folder of your joomla installation folder." My problem is that when you enter a new menu item this check is not done, before creating the menu item. So you can create a menu item with the name of a sub-folder. The moment the item is creacted you can't save it again. ### Expected result Check before the item is created. ### Actual result No check and an error on the frontend of the site because your not allowed to browse a sub-folder. ### System information (as much as possible) ### Additional comments
1.0
new menu item check - ### Steps to reproduce the issue It is not allowed that a menu item alias is the same as a sub-folder in your Joomla installation. The message will be: "Save failed with the following error: A first level menu item alias can't be 'tmp' because 'tmp' is a sub-folder of your joomla installation folder." My problem is that when you enter a new menu item this check is not done, before creating the menu item. So you can create a menu item with the name of a sub-folder. The moment the item is creacted you can't save it again. ### Expected result Check before the item is created. ### Actual result No check and an error on the frontend of the site because your not allowed to browse a sub-folder. ### System information (as much as possible) ### Additional comments
non_process
new menu item check steps to reproduce the issue it is not allowed that a menu item alias is the same as a sub folder in your joomla installation the message will be save failed with the following error a first level menu item alias can t be tmp because tmp is a sub folder of your joomla installation folder my problem is that when you enter a new menu item this check is not done before creating the menu item so you can create a menu item with the name of a sub folder the moment the item is creacted you can t save it again expected result check before the item is created actual result no check and an error on the frontend of the site because your not allowed to browse a sub folder system information as much as possible additional comments
0
2,646
5,425,882,156
IssuesEvent
2017-03-03 08:13:46
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
closed
Can't initiate Parse in Office XP
bug critical parse-tree-processing
The Parse button is enabled and clickable in Excel, Word and Access, but the Parse doesn't appear to commence, and so never completes. In PowerPoint, the button is clickable, and the spinning circles appear in the Code Explorer and Code Inspections windows, but there is no other visible sign of progress, and parsing never completes.
1.0
Can't initiate Parse in Office XP - The Parse button is enabled and clickable in Excel, Word and Access, but the Parse doesn't appear to commence, and so never completes. In PowerPoint, the button is clickable, and the spinning circles appear in the Code Explorer and Code Inspections windows, but there is no other visible sign of progress, and parsing never completes.
process
can t initiate parse in office xp the parse button is enabled and clickable in excel word and access but the parse doesn t appear to commence and so never completes in powerpoint the button is clickable and the spinning circles appear in the code explorer and code inspections windows but there is no other visible sign of progress and parsing never completes
1
21,944
30,446,799,875
IssuesEvent
2023-07-15 19:28:47
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1b7 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b7", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:204", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpylvasmk3/pyutils" } }```
1.0
pyutils 0.0.1b7 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b7", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils/exec_utils.py/pyutils/exec_utils.py:204", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpylvasmk3/pyutils" } }```
process
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt python utils pytils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pyutils
1
16,350
21,009,456,998
IssuesEvent
2022-03-30 04:08:16
bitPogo/kmock
https://api.github.com/repos/bitPogo/kmock
opened
Add Argument types to invoke
enhancement kmock-processor
## Description <!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug --> Currently invoke accepts any value, since there is no generic value given.
1.0
Add Argument types to invoke - ## Description <!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug --> Currently invoke accepts any value, since there is no generic value given.
process
add argument types to invoke description currently invoke accepts any value since there is no generic value given
1
15,262
19,211,018,701
IssuesEvent
2021-12-07 01:56:31
googleapis/python-spanner
https://api.github.com/repos/googleapis/python-spanner
closed
tests.system.test_system_dbapi.TestTransactionsManagement: test_commit failed
api: spanner type: process flakybot: issue flakybot: flaky
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 2487800e31842a44dcc37937c325e130c8c926b0 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/306a2e02-87cb-4be9-be31-37456ec7a8a2), [Sponge](http://sponge2/306a2e02-87cb-4be9-be31-37456ec7a8a2) status: failed <details><summary>Test output</summary><br><pre>args = (parent: "projects/precise-truck-742" instance_id: "google-cloud-1627550679627" instance { name: "projects/precise-t...1627550946" } labels { key: "python-spanner-dbapi-systests" value: "true" } processing_units: 1000 } ,) kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/precise-truck-742/instances/google-cloud-1627550679627'), ('x...ms', 'parent=projects/precise-truck-742'), ('x-goog-api-client', 'gl-python/3.8.6 grpc/1.39.0 gax/1.31.1 gccl/3.6.0')]} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:67: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7fe7942014f0> request = parent: "projects/precise-truck-742" instance_id: "google-cloud-1627550679627" instance { name: "projects/precise-tr... "1627550946" } labels { key: "python-spanner-dbapi-systests" value: "true" } processing_units: 1000 } timeout = None metadata = [('google-cloud-resource-prefix', 'projects/precise-truck-742/instances/google-cloud-1627550679627'), ('x-goog-request...ams', 'parent=projects/precise-truck-742'), ('x-goog-api-client', 'gl-python/3.8.6 grpc/1.39.0 gax/1.31.1 gccl/3.6.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:946: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7fe78e5b66d0> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7fe79445bc00> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _MultiThreadedRendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _InactiveRpcError(state) E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: E status = StatusCode.ALREADY_EXISTS E details = "Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627" E debug_error_string = "{"created":"@1627550946.587652143","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1069,"grpc_message":"Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627","grpc_status":6}" E > .nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:849: _InactiveRpcError The above exception was the direct cause of the following exception: def setUpModule(): if USE_EMULATOR: from google.auth.credentials import AnonymousCredentials emulator_project = os.getenv("GCLOUD_PROJECT", "emulator-test-project") Config.CLIENT = Client( project=emulator_project, credentials=AnonymousCredentials() ) else: Config.CLIENT = Client() retry = RetryErrors(exceptions.ServiceUnavailable) configs = list(retry(Config.CLIENT.list_instance_configs)()) instances = retry(_list_instances)() EXISTING_INSTANCES[:] = instances # Delete test instances that are older than an hour. cutoff = int(time.time()) - 1 * 60 * 60 for instance_pb in Config.CLIENT.list_instances( "labels.python-spanner-dbapi-systests:true" ): instance = Instance.from_pb(instance_pb, Config.CLIENT) if "created" not in instance.labels: continue create_time = int(instance.labels["created"]) if create_time > cutoff: continue # Instance cannot be deleted while backups exist. for backup_pb in instance.list_backups(): backup = Backup.from_pb(backup_pb, instance) backup.delete() instance.delete() if CREATE_INSTANCE: if not USE_EMULATOR: # Defend against back-end returning configs for regions we aren't # actually allowed to use. configs = [config for config in configs if "-us-" in config.name] if not configs: raise ValueError("List instance configs failed in module set up.") Config.INSTANCE_CONFIG = configs[0] config_name = configs[0].name create_time = str(int(time.time())) labels = {"python-spanner-dbapi-systests": "true", "created": create_time} Config.INSTANCE = Config.CLIENT.instance( INSTANCE_ID, config_name, labels=labels ) > created_op = Config.INSTANCE.create() tests/system/test_system_dbapi.py:98: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/spanner_v1/instance.py:318: in create future = api.create_instance( google/cloud/spanner_admin_instance_v1/services/instance_admin/client.py:829: in create_instance response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_InactiveRpcError of RPC that terminated with: status = StatusCode.ALREADY_EXISTS details = "Instance already exist...message":"Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627","grpc_status":6}" > > ??? E google.api_core.exceptions.AlreadyExists: 409 Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627 <string>:3: AlreadyExists</pre></details>
1.0
tests.system.test_system_dbapi.TestTransactionsManagement: test_commit failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 2487800e31842a44dcc37937c325e130c8c926b0 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/306a2e02-87cb-4be9-be31-37456ec7a8a2), [Sponge](http://sponge2/306a2e02-87cb-4be9-be31-37456ec7a8a2) status: failed <details><summary>Test output</summary><br><pre>args = (parent: "projects/precise-truck-742" instance_id: "google-cloud-1627550679627" instance { name: "projects/precise-t...1627550946" } labels { key: "python-spanner-dbapi-systests" value: "true" } processing_units: 1000 } ,) kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/precise-truck-742/instances/google-cloud-1627550679627'), ('x...ms', 'parent=projects/precise-truck-742'), ('x-goog-api-client', 'gl-python/3.8.6 grpc/1.39.0 gax/1.31.1 gccl/3.6.0')]} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:67: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7fe7942014f0> request = parent: "projects/precise-truck-742" instance_id: "google-cloud-1627550679627" instance { name: "projects/precise-tr... "1627550946" } labels { key: "python-spanner-dbapi-systests" value: "true" } processing_units: 1000 } timeout = None metadata = [('google-cloud-resource-prefix', 'projects/precise-truck-742/instances/google-cloud-1627550679627'), ('x-goog-request...ams', 'parent=projects/precise-truck-742'), ('x-goog-api-client', 'gl-python/3.8.6 grpc/1.39.0 gax/1.31.1 gccl/3.6.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:946: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7fe78e5b66d0> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7fe79445bc00> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _MultiThreadedRendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _InactiveRpcError(state) E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: E status = StatusCode.ALREADY_EXISTS E details = "Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627" E debug_error_string = "{"created":"@1627550946.587652143","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1069,"grpc_message":"Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627","grpc_status":6}" E > .nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:849: _InactiveRpcError The above exception was the direct cause of the following exception: def setUpModule(): if USE_EMULATOR: from google.auth.credentials import AnonymousCredentials emulator_project = os.getenv("GCLOUD_PROJECT", "emulator-test-project") Config.CLIENT = Client( project=emulator_project, credentials=AnonymousCredentials() ) else: Config.CLIENT = Client() retry = RetryErrors(exceptions.ServiceUnavailable) configs = list(retry(Config.CLIENT.list_instance_configs)()) instances = retry(_list_instances)() EXISTING_INSTANCES[:] = instances # Delete test instances that are older than an hour. cutoff = int(time.time()) - 1 * 60 * 60 for instance_pb in Config.CLIENT.list_instances( "labels.python-spanner-dbapi-systests:true" ): instance = Instance.from_pb(instance_pb, Config.CLIENT) if "created" not in instance.labels: continue create_time = int(instance.labels["created"]) if create_time > cutoff: continue # Instance cannot be deleted while backups exist. for backup_pb in instance.list_backups(): backup = Backup.from_pb(backup_pb, instance) backup.delete() instance.delete() if CREATE_INSTANCE: if not USE_EMULATOR: # Defend against back-end returning configs for regions we aren't # actually allowed to use. configs = [config for config in configs if "-us-" in config.name] if not configs: raise ValueError("List instance configs failed in module set up.") Config.INSTANCE_CONFIG = configs[0] config_name = configs[0].name create_time = str(int(time.time())) labels = {"python-spanner-dbapi-systests": "true", "created": create_time} Config.INSTANCE = Config.CLIENT.instance( INSTANCE_ID, config_name, labels=labels ) > created_op = Config.INSTANCE.create() tests/system/test_system_dbapi.py:98: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/spanner_v1/instance.py:318: in create future = api.create_instance( google/cloud/spanner_admin_instance_v1/services/instance_admin/client.py:829: in create_instance response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_InactiveRpcError of RPC that terminated with: status = StatusCode.ALREADY_EXISTS details = "Instance already exist...message":"Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627","grpc_status":6}" > > ??? E google.api_core.exceptions.AlreadyExists: 409 Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627 <string>:3: AlreadyExists</pre></details>
process
tests system test system dbapi testtransactionsmanagement test commit failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output args parent projects precise truck instance id google cloud instance name projects precise t labels key python spanner dbapi systests value true processing units kwargs metadata six wraps callable def error remapped callable args kwargs try return callable args kwargs nox system lib site packages google api core grpc helpers py self request parent projects precise truck instance id google cloud instance name projects precise tr labels key python spanner dbapi systests value true processing units timeout none metadata credentials none wait for ready none compression none def call self request timeout none metadata none credentials none wait for ready none compression none state call self blocking request timeout metadata credentials wait for ready compression return end unary response blocking state call false none nox system lib site packages grpc channel py state call with call false deadline none def end unary response blocking state call with call deadline if state code is grpc statuscode ok if with call rendezvous multithreadedrendezvous state call none deadline return state response rendezvous else return state response else raise inactiverpcerror state e grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with e status statuscode already exists e details instance already exists projects precise truck instances google cloud e debug error string created description error received from peer file src core lib surface call cc file line grpc message instance already exists projects precise truck instances google cloud grpc status e nox system lib site packages grpc channel py inactiverpcerror the above exception was the direct cause of the following exception def setupmodule if use emulator from google auth credentials import anonymouscredentials emulator project os getenv gcloud project emulator test project config client client project emulator project credentials anonymouscredentials else config client client retry retryerrors exceptions serviceunavailable configs list retry config client list instance configs instances retry list instances existing instances instances delete test instances that are older than an hour cutoff int time time for instance pb in config client list instances labels python spanner dbapi systests true instance instance from pb instance pb config client if created not in instance labels continue create time int instance labels if create time cutoff continue instance cannot be deleted while backups exist for backup pb in instance list backups backup backup from pb backup pb instance backup delete instance delete if create instance if not use emulator defend against back end returning configs for regions we aren t actually allowed to use configs if not configs raise valueerror list instance configs failed in module set up config instance config configs config name configs name create time str int time time labels python spanner dbapi systests true created create time config instance config client instance instance id config name labels labels created op config instance create tests system test system dbapi py google cloud spanner instance py in create future api create instance google cloud spanner admin instance services instance admin client py in create instance response rpc request retry retry timeout timeout metadata metadata nox system lib site packages google api core gapic method py in call return wrapped func args kwargs nox system lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value inactiverpcerror of rpc that terminated with status statuscode already exists details instance already exist message instance already exists projects precise truck instances google cloud grpc status e google api core exceptions alreadyexists instance already exists projects precise truck instances google cloud alreadyexists
1
13,049
15,387,938,537
IssuesEvent
2021-03-03 10:08:57
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Introspection: Secret flag to disable guardrails
kind/feature process/candidate team/migrations tech/engines tech/typescript topic: introspection
Right now we have a bunch of guardrails for Introspection in place that make sure, we do not introspect things into our schema that do not pass validation or would not work in Prisma Client or Query Engine. We should have a "secret" flag on `introspect` that let's you disable those guardrails, so we can observe and test the behavior if those guardrails are disabled. That way we could from time to time run Introspection CI with this flag enabled, and see what still breaks and what the effect of removing one of those guardrails would be.
1.0
Introspection: Secret flag to disable guardrails - Right now we have a bunch of guardrails for Introspection in place that make sure, we do not introspect things into our schema that do not pass validation or would not work in Prisma Client or Query Engine. We should have a "secret" flag on `introspect` that let's you disable those guardrails, so we can observe and test the behavior if those guardrails are disabled. That way we could from time to time run Introspection CI with this flag enabled, and see what still breaks and what the effect of removing one of those guardrails would be.
process
introspection secret flag to disable guardrails right now we have a bunch of guardrails for introspection in place that make sure we do not introspect things into our schema that do not pass validation or would not work in prisma client or query engine we should have a secret flag on introspect that let s you disable those guardrails so we can observe and test the behavior if those guardrails are disabled that way we could from time to time run introspection ci with this flag enabled and see what still breaks and what the effect of removing one of those guardrails would be
1
11,611
14,478,980,146
IssuesEvent
2020-12-10 09:11:47
decidim/decidim
https://api.github.com/repos/decidim/decidim
closed
Statistics content block for Process Groups
contract: process-groups
Ref. PG02-4 **Is your feature request related to a problem? Please describe.** As a visitor I want a quick way to see the activity of all the processes inside the PG. **Describe the solution you'd like** To have implemented Statistics for all the PG, with the sum of every one of the Resources/Components in every Participatory Process. **Describe alternatives you've considered** To use Metrics as they give more data and are more visual (with graphics), but the Metrics features not work well (they have issues with showing the last dates instead of what makes sense on a Participatory Process, such as taking in account the beginning and finishing of the PP). To have the unique sum in the case of Participants, but that would be difficult to understand. **Additional context** Statistics on a Participatory Process: ![](https://i.imgur.com/Ksh4Ckn.png) **Does this issue could impact on users private data?** No **Acceptance criteria** - [x] As a visitor I can see the Statistics for a PG in its homepage - [x] As an administrator I want to be able to decide whether or not to display the statistics on this page.
1.0
Statistics content block for Process Groups - Ref. PG02-4 **Is your feature request related to a problem? Please describe.** As a visitor I want a quick way to see the activity of all the processes inside the PG. **Describe the solution you'd like** To have implemented Statistics for all the PG, with the sum of every one of the Resources/Components in every Participatory Process. **Describe alternatives you've considered** To use Metrics as they give more data and are more visual (with graphics), but the Metrics features not work well (they have issues with showing the last dates instead of what makes sense on a Participatory Process, such as taking in account the beginning and finishing of the PP). To have the unique sum in the case of Participants, but that would be difficult to understand. **Additional context** Statistics on a Participatory Process: ![](https://i.imgur.com/Ksh4Ckn.png) **Does this issue could impact on users private data?** No **Acceptance criteria** - [x] As a visitor I can see the Statistics for a PG in its homepage - [x] As an administrator I want to be able to decide whether or not to display the statistics on this page.
process
statistics content block for process groups ref is your feature request related to a problem please describe as a visitor i want a quick way to see the activity of all the processes inside the pg describe the solution you d like to have implemented statistics for all the pg with the sum of every one of the resources components in every participatory process describe alternatives you ve considered to use metrics as they give more data and are more visual with graphics but the metrics features not work well they have issues with showing the last dates instead of what makes sense on a participatory process such as taking in account the beginning and finishing of the pp to have the unique sum in the case of participants but that would be difficult to understand additional context statistics on a participatory process does this issue could impact on users private data no acceptance criteria as a visitor i can see the statistics for a pg in its homepage as an administrator i want to be able to decide whether or not to display the statistics on this page
1
6,834
9,977,591,930
IssuesEvent
2019-07-09 17:40:41
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
MP modulation by symbiont of host defense-related PCD
multi-species process
GO:0034053 modulation by symbiont of host defense-related programmed cell death should be is_a regulation of GO:0097300 programmed necrotic cell death @CuzickA
1.0
MP modulation by symbiont of host defense-related PCD - GO:0034053 modulation by symbiont of host defense-related programmed cell death should be is_a regulation of GO:0097300 programmed necrotic cell death @CuzickA
process
mp modulation by symbiont of host defense related pcd go modulation by symbiont of host defense related programmed cell death should be is a regulation of go programmed necrotic cell death cuzicka
1
8,544
11,717,252,876
IssuesEvent
2020-03-09 16:56:57
MHRA/products
https://api.github.com/repos/MHRA/products
opened
AUTO BATCH - Support XML Requests
EPIC - Auto Batch Process :oncoming_automobile: HIGH PRIORITY :arrow_double_up:
Accenture require an XML interface to the `doc-index-updater` API. **Acceptance Criteria** - The API should check the `Content-Type` or `Accept` headers. If the header is `application/xml`, then we should deserialize body & serialize response assuming XML rather than JSON. **Stages** - [ ] Backlog - [ ] Discovery - [ ] Dev - [ ] Review - [ ] QA - [ ] UAT
1.0
AUTO BATCH - Support XML Requests - Accenture require an XML interface to the `doc-index-updater` API. **Acceptance Criteria** - The API should check the `Content-Type` or `Accept` headers. If the header is `application/xml`, then we should deserialize body & serialize response assuming XML rather than JSON. **Stages** - [ ] Backlog - [ ] Discovery - [ ] Dev - [ ] Review - [ ] QA - [ ] UAT
process
auto batch support xml requests accenture require an xml interface to the doc index updater api acceptance criteria the api should check the content type or accept headers if the header is application xml then we should deserialize body serialize response assuming xml rather than json stages backlog discovery dev review qa uat
1
10,020
3,082,097,318
IssuesEvent
2015-08-23 12:07:06
tntim96/JSCover
https://api.github.com/repos/tntim96/JSCover
closed
Syntax error bug in generated JavaScript
bug Fix applied - please re-test
When trying to run JSCover on a large JavaScript file I get an error in the browser developer console: ``` Uncaught SyntaxError: Unexpected token - ``` Here is the relevant block of code that JSCover generated, note the offending line that tries to create a JavaScript function with the (invalid) name `visit8876_-1_1` ``` javascript function visit8878_59_15(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][15].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][14].init(25, 21, '0 < (f = this.data[n] >> u)'); function visit8877_59_14(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][14].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['-1'][1].init(23, 36, 'visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f))'); function visit8876_-1_1(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['-1'][1].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][13].init(14, 8, 'u < this.A'); function visit8875_59_13(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][13].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][12].init(218, 46, 'visit8870_59_8(u < this.A) && visit8871_59_9(visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f)))'); function visit8874_59_12(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][12].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][11].init(204, 1, 'visit8869_59_7(visit8870_59_8(u < this.A) && visit8871_59_9(visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f))))'); function visit8873_59_11(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][11].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][10].init(228, 21, '0 < (f = this.data[n] >> u)'); ``` If you would like to see this error for yourself, here is a link to the HTML page containing the JavaScript that triggers this bug: https://bitc.github.io/tojs-benchmark/particles-ghcjs/index.min.html?barriers=l5&particles=500
1.0
Syntax error bug in generated JavaScript - When trying to run JSCover on a large JavaScript file I get an error in the browser developer console: ``` Uncaught SyntaxError: Unexpected token - ``` Here is the relevant block of code that JSCover generated, note the offending line that tries to create a JavaScript function with the (invalid) name `visit8876_-1_1` ``` javascript function visit8878_59_15(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][15].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][14].init(25, 21, '0 < (f = this.data[n] >> u)'); function visit8877_59_14(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][14].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['-1'][1].init(23, 36, 'visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f))'); function visit8876_-1_1(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['-1'][1].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][13].init(14, 8, 'u < this.A'); function visit8875_59_13(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][13].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][12].init(218, 46, 'visit8870_59_8(u < this.A) && visit8871_59_9(visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f)))'); function visit8874_59_12(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][12].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][11].init(204, 1, 'visit8869_59_7(visit8870_59_8(u < this.A) && visit8871_59_9(visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f))))'); function visit8873_59_11(result) { _$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][11].ranCondition(result); return result; }_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][10].init(228, 21, '0 < (f = this.data[n] >> u)'); ``` If you would like to see this error for yourself, here is a link to the HTML page containing the JavaScript that triggers this bug: https://bitc.github.io/tojs-benchmark/particles-ghcjs/index.min.html?barriers=l5&particles=500
non_process
syntax error bug in generated javascript when trying to run jscover on a large javascript file i get an error in the browser developer console uncaught syntaxerror unexpected token here is the relevant block of code that jscover generated note the offending line that tries to create a javascript function with the invalid name javascript function result jscoverage branchdata rancondition result return result jscoverage branchdata init u function result jscoverage branchdata rancondition result return result jscoverage branchdata init u k l m f function result jscoverage branchdata rancondition result return result jscoverage branchdata init u this a function result jscoverage branchdata rancondition result return result jscoverage branchdata init u u k l m f function result jscoverage branchdata rancondition result return result jscoverage branchdata init u u k l m f function result jscoverage branchdata rancondition result return result jscoverage branchdata init u if you would like to see this error for yourself here is a link to the html page containing the javascript that triggers this bug
0
4,317
2,849,431,231
IssuesEvent
2015-05-30 17:46:13
SoftverInzenjeringETFSA/SI2014Tim7
https://api.github.com/repos/SoftverInzenjeringETFSA/SI2014Tim7
closed
Nije prikazana klasa Ocitanja u Class diagramu
documentation wontfix
Komponenta: JKP Vodovod i Kanalizacija desktop Commit ID: bb0e5cac45f01fadbfae2afe4829398aa4b231ae Opis problema: Class diagram iz dokumenta: https://github.com/SoftverInzenjeringETFSA/SI2014Tim7/blob/master/Dizajn%20Sistema.pdf ne sadrzi klasu Ocitanja kao sto je to uradjeno u sistemu: http://imgur.com/J02RFWu
1.0
Nije prikazana klasa Ocitanja u Class diagramu - Komponenta: JKP Vodovod i Kanalizacija desktop Commit ID: bb0e5cac45f01fadbfae2afe4829398aa4b231ae Opis problema: Class diagram iz dokumenta: https://github.com/SoftverInzenjeringETFSA/SI2014Tim7/blob/master/Dizajn%20Sistema.pdf ne sadrzi klasu Ocitanja kao sto je to uradjeno u sistemu: http://imgur.com/J02RFWu
non_process
nije prikazana klasa ocitanja u class diagramu komponenta jkp vodovod i kanalizacija desktop commit id opis problema class diagram iz dokumenta ne sadrzi klasu ocitanja kao sto je to uradjeno u sistemu
0
237,253
18,154,763,019
IssuesEvent
2021-09-26 21:52:00
secheaper/slash
https://api.github.com/repos/secheaper/slash
opened
Add section links to README
documentation enhancement good first issue
Add links that scrolls you to particular sections in the README files. Something like a legend
1.0
Add section links to README - Add links that scrolls you to particular sections in the README files. Something like a legend
non_process
add section links to readme add links that scrolls you to particular sections in the readme files something like a legend
0
254,988
27,484,688,374
IssuesEvent
2023-03-04 01:08:20
panasalap/linux-4.1.15
https://api.github.com/repos/panasalap/linux-4.1.15
opened
CVE-2018-1120 (Medium) detected in linux-yocto-devv4.2.8
security vulnerability
## CVE-2018-1120 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary> <p> <p>Linux Embedded Kernel - tracks the next mainline release</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/gup.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/gup.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found affecting the Linux kernel before version 4.17. By mmap()ing a FUSE-backed file onto a process's memory containing command line arguments (or environment strings), an attacker can cause utilities from psutils or procps (such as ps, w) or any other program which makes a read() call to the /proc/<pid>/cmdline (or /proc/<pid>/environ) files to block indefinitely (denial of service) or for some controlled time (as a synchronization primitive for other attacks). <p>Publish Date: 2018-06-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-1120>CVE-2018-1120</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-1120">https://nvd.nist.gov/vuln/detail/CVE-2018-1120</a></p> <p>Release Date: 2018-06-20</p> <p>Fix Resolution: 4.17</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-1120 (Medium) detected in linux-yocto-devv4.2.8 - ## CVE-2018-1120 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary> <p> <p>Linux Embedded Kernel - tracks the next mainline release</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/gup.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/gup.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found affecting the Linux kernel before version 4.17. By mmap()ing a FUSE-backed file onto a process's memory containing command line arguments (or environment strings), an attacker can cause utilities from psutils or procps (such as ps, w) or any other program which makes a read() call to the /proc/<pid>/cmdline (or /proc/<pid>/environ) files to block indefinitely (denial of service) or for some controlled time (as a synchronization primitive for other attacks). <p>Publish Date: 2018-06-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-1120>CVE-2018-1120</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-1120">https://nvd.nist.gov/vuln/detail/CVE-2018-1120</a></p> <p>Release Date: 2018-06-20</p> <p>Fix Resolution: 4.17</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux yocto cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel tracks the next mainline release library home page a href found in base branch master vulnerable source files mm gup c mm gup c vulnerability details a flaw was found affecting the linux kernel before version by mmap ing a fuse backed file onto a process s memory containing command line arguments or environment strings an attacker can cause utilities from psutils or procps such as ps w or any other program which makes a read call to the proc cmdline or proc environ files to block indefinitely denial of service or for some controlled time as a synchronization primitive for other attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
231,591
17,694,531,343
IssuesEvent
2021-08-24 13:59:06
MeisterOfPizza/file-variants
https://api.github.com/repos/MeisterOfPizza/file-variants
opened
Multiple file-variants of the same variant inside input
documentation enhancement
Output will be however many file-variants of variant found. file-variant naming: <variant>(|.<part>)*.<extension> outputName: "{name}.{part0}.{part1}" (extension is appended automatically)
1.0
Multiple file-variants of the same variant inside input - Output will be however many file-variants of variant found. file-variant naming: <variant>(|.<part>)*.<extension> outputName: "{name}.{part0}.{part1}" (extension is appended automatically)
non_process
multiple file variants of the same variant inside input output will be however many file variants of variant found file variant naming outputname name extension is appended automatically
0
227,599
18,073,015,186
IssuesEvent
2021-09-21 06:29:02
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: sqlsmith/setup=empty/setting=no-mutations failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.sqlsmith/setup=empty/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3473201&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3473201&tab=artifacts#/sqlsmith/setup=empty/setting=no-mutations) on master @ [676dc076dc6a137d88eac9217764bbb63beae86f](https://github.com/cockroachdb/cockroach/commits/676dc076dc6a137d88eac9217764bbb63beae86f): ``` ('17:28:14.764522+14:43:00':::TIMETZ), ('07:23:03.812281+14:50:00':::TIMETZ), ( parse_timetz(COALESCE(st_astext('01030000C0010000000600000070EBF2E3C6B0FCC12418761831E2FEC1BEF7A174620F014290AE7E7800F10042A0E5BFE035CDF8C1FBF9AEBA467F00C2955B7F1FADF900C22E37DC6732D7F9C1F8291D6E8EC7F941342DD7B74BECF6C12CDCD92FA44FFE415C8813B08A060142CA9A4CD716F2FAC160827E0A3AD0E441C2A16B9537A6F441E8E10023E840FF41BC4F7B62083AFBC1009D25D3324EECC1647CD166902AFB418C99946456D0014270EBF2E3C6B0FCC12418761831E2FEC1BEF7A174620F014290AE7E7800F10042':::GEOMETRY::GEOMETRY)::STRING, e'\x1c':::STRING)::STRING)::TIMETZ ), ('04:50:55.334281-00:35:00':::TIMETZ) ) AS tab_97115 (col_167897) LEFT JOIN ( VALUES ('ffc2e94c-474a-45b0-933c-cf76cc75bce3':::UUID), ('2c6a5785-6910-4180-a14e-71319f6862a5':::UUID), (uuid_generate_v4()::UUID), ('203ac89f-0554-4aed-8f55-e9259898b6a9':::UUID), (NULL) ) AS tab_97116 (col_167898) ON false, ( VALUES ('ffff:ffff:ffff:ffff:ffff:ffff:ffff:ffff':::INET), ('7a5f:be04:ed4a:c9ca:1e60:169f:8a5f:2ab/108':::INET) ) AS tab_97117 (col_167899) ORDER BY tab_97117.col_167899, tab_97117.col_167899, tab_97117.col_167899 LIMIT 1:::INT8 ) ), ('0104000060E610000000000000':::GEOGRAPHY, NULL) ) AS tab_97118 (col_167901, col_167902) ) SELECT tab_97119.col_167904 AS col_167905 FROM ( VALUES (NULL), ('20 years 181 days 06:29:21.419479':::INTERVAL), ('59 years 524 days 02:49:03.527908':::INTERVAL), ('-51 years -5 mons -748 days -06:39:54.748217':::INTERVAL), (parse_interval(e',1-\x04Bh``':::STRING::STRING, e'\x05\x04%R!)EbC':::STRING::STRING)::INTERVAL), ('-46 years -2 mons -829 days -13:15:27.684704':::INTERVAL) ) AS tab_97119 (col_167904) WHERE true LIMIT 87:::INT8; ``` <details><summary>Reproduce</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=empty/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: sqlsmith/setup=empty/setting=no-mutations failed - roachtest.sqlsmith/setup=empty/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3473201&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3473201&tab=artifacts#/sqlsmith/setup=empty/setting=no-mutations) on master @ [676dc076dc6a137d88eac9217764bbb63beae86f](https://github.com/cockroachdb/cockroach/commits/676dc076dc6a137d88eac9217764bbb63beae86f): ``` ('17:28:14.764522+14:43:00':::TIMETZ), ('07:23:03.812281+14:50:00':::TIMETZ), ( parse_timetz(COALESCE(st_astext('01030000C0010000000600000070EBF2E3C6B0FCC12418761831E2FEC1BEF7A174620F014290AE7E7800F10042A0E5BFE035CDF8C1FBF9AEBA467F00C2955B7F1FADF900C22E37DC6732D7F9C1F8291D6E8EC7F941342DD7B74BECF6C12CDCD92FA44FFE415C8813B08A060142CA9A4CD716F2FAC160827E0A3AD0E441C2A16B9537A6F441E8E10023E840FF41BC4F7B62083AFBC1009D25D3324EECC1647CD166902AFB418C99946456D0014270EBF2E3C6B0FCC12418761831E2FEC1BEF7A174620F014290AE7E7800F10042':::GEOMETRY::GEOMETRY)::STRING, e'\x1c':::STRING)::STRING)::TIMETZ ), ('04:50:55.334281-00:35:00':::TIMETZ) ) AS tab_97115 (col_167897) LEFT JOIN ( VALUES ('ffc2e94c-474a-45b0-933c-cf76cc75bce3':::UUID), ('2c6a5785-6910-4180-a14e-71319f6862a5':::UUID), (uuid_generate_v4()::UUID), ('203ac89f-0554-4aed-8f55-e9259898b6a9':::UUID), (NULL) ) AS tab_97116 (col_167898) ON false, ( VALUES ('ffff:ffff:ffff:ffff:ffff:ffff:ffff:ffff':::INET), ('7a5f:be04:ed4a:c9ca:1e60:169f:8a5f:2ab/108':::INET) ) AS tab_97117 (col_167899) ORDER BY tab_97117.col_167899, tab_97117.col_167899, tab_97117.col_167899 LIMIT 1:::INT8 ) ), ('0104000060E610000000000000':::GEOGRAPHY, NULL) ) AS tab_97118 (col_167901, col_167902) ) SELECT tab_97119.col_167904 AS col_167905 FROM ( VALUES (NULL), ('20 years 181 days 06:29:21.419479':::INTERVAL), ('59 years 524 days 02:49:03.527908':::INTERVAL), ('-51 years -5 mons -748 days -06:39:54.748217':::INTERVAL), (parse_interval(e',1-\x04Bh``':::STRING::STRING, e'\x05\x04%R!)EbC':::STRING::STRING)::INTERVAL), ('-46 years -2 mons -829 days -13:15:27.684704':::INTERVAL) ) AS tab_97119 (col_167904) WHERE true LIMIT 87:::INT8; ``` <details><summary>Reproduce</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=empty/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_process
roachtest sqlsmith setup empty setting no mutations failed roachtest sqlsmith setup empty setting no mutations with on master timetz timetz parse timetz coalesce st astext geometry geometry string e string string timetz timetz as tab col left join values uuid uuid uuid generate uuid uuid null as tab col on false values ffff ffff ffff ffff ffff ffff ffff ffff inet inet as tab col order by tab col tab col tab col limit geography null as tab col col select tab col as col from values null years days interval years days interval years mons days interval parse interval e string string e r ebc string string interval years mons days interval as tab col where true limit reproduce see cc cockroachdb sql queries
0
157,455
12,374,004,927
IssuesEvent
2020-05-19 00:12:02
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
Create buttons are missing for workload
[zube]: To Test kind/bug
Version: master-head(1f4a2c388) Steps: 1. Navigate to the workload section from the dashboard UI "Create from Yaml" and "Create" buttons are missing ![Screen Shot 2020-05-18 at 1 45 17 PM](https://user-images.githubusercontent.com/18536626/82258024-ebeca300-990d-11ea-9ff9-2af19154b67d.png)
1.0
Create buttons are missing for workload - Version: master-head(1f4a2c388) Steps: 1. Navigate to the workload section from the dashboard UI "Create from Yaml" and "Create" buttons are missing ![Screen Shot 2020-05-18 at 1 45 17 PM](https://user-images.githubusercontent.com/18536626/82258024-ebeca300-990d-11ea-9ff9-2af19154b67d.png)
non_process
create buttons are missing for workload version master head steps navigate to the workload section from the dashboard ui create from yaml and create buttons are missing
0
4,596
7,434,937,657
IssuesEvent
2018-03-26 12:48:31
decidim/decidim
https://api.github.com/repos/decidim/decidim
closed
Title and subtitle of a participatory process are required at model level but not at form level
space: processes type: bug
<!-- 1. Please check if an issue already exists so there are no duplicates 2. Fill out the whole template so we have a good overview on the issue 3. Do not remove any section of the template. If something is not applicable leave it empty but leave it in the Issue 4. Please follow the template, otherwise we'll have to ask you to update it --> # This is a Bug Report #### :tophat: Description For bug reports: * What went wrong? Just a confusing behaviour when you create a participatory process (and it's maybe the case for other elements), the form doesn't have a 'required' asterisk for the title and subtitle fields. But when you try to save the element without the subtitle, you get an error message. * What did you expect should have happened? Add an asterisk on all the required fields * What was the config you used? Standard * What stacktrace or error message from your provider did you see? can't be blank For feature proposals: * What is the use case that should be solved. The more detail you describe this in the easier it is to understand for us. * If there is additional config how would it look #### :pushpin: Related issues #3054 #3056 #### :clipboard: Additional Data * ***Decidim deployment where you found the issue***: * ***Browser & version***: Chrome latest * ***Screenshot***: ![capture d ecran 2018-03-26 a 10 35 15](https://user-images.githubusercontent.com/7223028/37898022-f90a3d12-30e7-11e8-94a9-75ff3ceba34d.png) * ***Error messages***: can't be blank * ***URL to reproduce the error***: localhost:3000/admin/participatory_processes/new
1.0
Title and subtitle of a participatory process are required at model level but not at form level - <!-- 1. Please check if an issue already exists so there are no duplicates 2. Fill out the whole template so we have a good overview on the issue 3. Do not remove any section of the template. If something is not applicable leave it empty but leave it in the Issue 4. Please follow the template, otherwise we'll have to ask you to update it --> # This is a Bug Report #### :tophat: Description For bug reports: * What went wrong? Just a confusing behaviour when you create a participatory process (and it's maybe the case for other elements), the form doesn't have a 'required' asterisk for the title and subtitle fields. But when you try to save the element without the subtitle, you get an error message. * What did you expect should have happened? Add an asterisk on all the required fields * What was the config you used? Standard * What stacktrace or error message from your provider did you see? can't be blank For feature proposals: * What is the use case that should be solved. The more detail you describe this in the easier it is to understand for us. * If there is additional config how would it look #### :pushpin: Related issues #3054 #3056 #### :clipboard: Additional Data * ***Decidim deployment where you found the issue***: * ***Browser & version***: Chrome latest * ***Screenshot***: ![capture d ecran 2018-03-26 a 10 35 15](https://user-images.githubusercontent.com/7223028/37898022-f90a3d12-30e7-11e8-94a9-75ff3ceba34d.png) * ***Error messages***: can't be blank * ***URL to reproduce the error***: localhost:3000/admin/participatory_processes/new
process
title and subtitle of a participatory process are required at model level but not at form level please check if an issue already exists so there are no duplicates fill out the whole template so we have a good overview on the issue do not remove any section of the template if something is not applicable leave it empty but leave it in the issue please follow the template otherwise we ll have to ask you to update it this is a bug report tophat description for bug reports what went wrong just a confusing behaviour when you create a participatory process and it s maybe the case for other elements the form doesn t have a required asterisk for the title and subtitle fields but when you try to save the element without the subtitle you get an error message what did you expect should have happened add an asterisk on all the required fields what was the config you used standard what stacktrace or error message from your provider did you see can t be blank for feature proposals what is the use case that should be solved the more detail you describe this in the easier it is to understand for us if there is additional config how would it look pushpin related issues clipboard additional data decidim deployment where you found the issue browser version chrome latest screenshot error messages can t be blank url to reproduce the error localhost admin participatory processes new
1
255,981
19,349,336,417
IssuesEvent
2021-12-15 14:12:48
NOAA-EMC/NCEPLIBS-grib_util
https://api.github.com/repos/NOAA-EMC/NCEPLIBS-grib_util
closed
update README to include authors of each utility
documentation
update README to include authors of each utility
1.0
update README to include authors of each utility - update README to include authors of each utility
non_process
update readme to include authors of each utility update readme to include authors of each utility
0
163,172
6,192,404,266
IssuesEvent
2017-07-05 01:33:44
ykoblick/Shabbat-Issue-Tracker
https://api.github.com/repos/ykoblick/Shabbat-Issue-Tracker
closed
About me being erased
app High Priority website
The "about me" message in my profile was deleted. I asked R. Klatzko about it and he said he doesn't know how it happened
1.0
About me being erased - The "about me" message in my profile was deleted. I asked R. Klatzko about it and he said he doesn't know how it happened
non_process
about me being erased the about me message in my profile was deleted i asked r klatzko about it and he said he doesn t know how it happened
0
38,867
10,257,960,274
IssuesEvent
2019-08-21 21:26:04
tmrowco/electricitymap-contrib
https://api.github.com/repos/tmrowco/electricitymap-contrib
opened
USA hourly generation by fuel type is now available
parser parser buildable!
The EIA have just released a new visualization for the USA Electric Operating System. One of the graphs is hourly generation by fuel type for the lower 48 states! From a quick glance it appears to lag behind by about a day but it's still hugely useful! :tada: https://www.eia.gov/beta/electricity/gridmonitor/dashboard/electric_overview/US48/US48 It also seems like you can drill down into individual balancing authorities and get the exact same data (hourly by type D-1), for instance checkout [Tennessee Valley Authority](https://www.eia.gov/beta/electricity/gridmonitor/dashboard/electric_overview/balancing_authority/TVA)
1.0
USA hourly generation by fuel type is now available - The EIA have just released a new visualization for the USA Electric Operating System. One of the graphs is hourly generation by fuel type for the lower 48 states! From a quick glance it appears to lag behind by about a day but it's still hugely useful! :tada: https://www.eia.gov/beta/electricity/gridmonitor/dashboard/electric_overview/US48/US48 It also seems like you can drill down into individual balancing authorities and get the exact same data (hourly by type D-1), for instance checkout [Tennessee Valley Authority](https://www.eia.gov/beta/electricity/gridmonitor/dashboard/electric_overview/balancing_authority/TVA)
non_process
usa hourly generation by fuel type is now available the eia have just released a new visualization for the usa electric operating system one of the graphs is hourly generation by fuel type for the lower states from a quick glance it appears to lag behind by about a day but it s still hugely useful tada it also seems like you can drill down into individual balancing authorities and get the exact same data hourly by type d for instance checkout
0
4,157
7,103,958,236
IssuesEvent
2018-01-16 08:09:19
Incubaid/crm
https://api.github.com/repos/Incubaid/crm
closed
FR: show node status of member
process_wontfix
At the moment there's no way to see the node status of a members node. Need to have this before PTO.
1.0
FR: show node status of member - At the moment there's no way to see the node status of a members node. Need to have this before PTO.
process
fr show node status of member at the moment there s no way to see the node status of a members node need to have this before pto
1
323,046
9,842,352,132
IssuesEvent
2019-06-18 09:05:57
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.twitch.tv - see bug description
browser-firefox engine-gecko priority-critical type-webrender-enabled
<!-- @browser: Firefox 69.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 --> <!-- @reported_with: desktop-reporter --> <!-- @extra_labels: type-webrender-enabled --> **URL**: https://www.twitch.tv/riotgames **Browser / Version**: Firefox 69.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: stream is stuttery even in safe mode **Steps to Reproduce**: 1. click on a stream to watch it [![Screenshot Description](https://webcompat.com/uploads/2019/6/8b3acc0c-5043-4c46-a0e1-674d45f58917-thumb.jpeg)](https://webcompat.com/uploads/2019/6/8b3acc0c-5043-4c46-a0e1-674d45f58917.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190614100022</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: true</li><li>gfx.webrender.all: true</li><li>channel: nightly</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Loading failed for the <script> with source https://c.amazon-adsystem.com/aax2/apstag.js." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://c.amazon-adsystem.com/aax2/apstag.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://www.gstatic.com/cv/js/sender/v1/cast_sender.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://cdn-gl.imrworldwide.com/conf/P34D56F9D-5684-4C83-8EE1-5EA7DE9CF45D.js#name=1a0UrMElc2nTFhA32wPD2WhXYc6RFiFi&ns=NOLBUNDLE." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://client-event-reporter.twitch.tv/v1/stats. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Request to access cookie or storage on https://cdn-gl.imrworldwide.com/conf/P34D56F9D-5684-4C83-8EE1-5EA7DE9CF45D.js#name=1a0UrMElc2nTFhA32wPD2WhXYc6RFiFi&ns=NOLBUNDLE was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://imasdk.googleapis.com/js/sdkloader/ima3.js." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted." {file: "https://static.twitchcdn.net/assets/player-core-base-e939ccadccca15e51973.js" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://s.amazon-adsystem.com/iui3?d=3p-hbg&ex-src=twitch.tv&ex-hargs=v%3D1.0%3Bc%3D8858214122683%3Bp%3De75425fb-5407-7bd5-fd20-f462e98a8777 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://cdn.krxd.net/controltag?confid=JYk8LCsC was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://ssl.google-analytics.com/ga.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/beacon.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted." {file: "https://static.twitchcdn.net/assets/player-core-base-e939ccadccca15e51973.js" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/p?c1=2&c2=6745306&ns_type=hidden&ns_st_sv=5.1.3.160420&ns_st_smv=5.1&ns_st_it=r&ns_st_id=1560539414288&ns_st_ec=1&ns_st_sp=1&ns_st_sc=1&ns_st_sq=1&ns_st_ppc=1&ns_st_apc=1&ns_st_spc=1&ns_st_cn=1&ns_st_ev=play&ns_st_po=0&ns_st_cl=0&ns_st_pb=1&ns_st_mp=js_api&ns_st_mv=5.1.3.160420&ns_st_pn=1&ns_st_tp=0&ns_st_ci=34526695408&ns_st_pt=0&ns_st_dpt=0&ns_st_ipt=0&ns_st_et=0&ns_st_det=0&ns_st_upc=0&ns_st_dupc=0&ns_st_iupc=0&ns_st_upa=0&ns_st_dupa=0&ns_st_iupa=0&ns_st_lpc=0&ns_st_dlpc=0&ns_st_lpa=0&ns_st_dlpa=0&ns_st_pa=0&ns_ts=1560539416234&ns_st_bc=0&ns_st_dbc=0&ns_st_bt=0&ns_st_dbt=0&ns_st_bp=0&ns_st_lt=1946&ns_st_skc=0&ns_st_dskc=0&ns_st_ska=0&ns_st_dska=0&ns_st_skd=0&ns_st_skt=0&ns_st_dskt=0&ns_st_pc=0&ns_st_dpc=0&ns_st_pp=0&ns_st_br=0&ns_st_ub=0&ns_st_ki=1200000&ns_st_pr=*null&ns_st_sn=*null&ns_st_en=*null&ns_st_ep=Week%202%20Day%201%20%7C%20LEC%20Summer%20Split%20(2019)%20-%20Splyce%20vs.%20Schalke%2004&ns_st_ct=vc13&ns_st_ge=Gaming&ns_st_st=Riot%20Games&ns_st_ce=1&ns_st_ia=0&ns_st_ddt=2019-06-14&ns_st_tdt=*null&ns_st_pu=*null&ns_st_ti=*null&c3=TWITCH&c4=twitch.tv&c6=*null&c7=https%3A%2F%2Fwww.twitch.tv%2Friotgames&c8=Riot%20Games%20-%20Twitch&c9=&cs_ucfr=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/p?c1=2&c2=6745306&ns_type=hidden&ns_st_sv=5.1.3.160420&ns_st_smv=5.1&ns_st_it=r&ns_st_id=1560539414288&ns_st_ec=2&ns_st_sp=1&ns_st_sc=1&ns_st_sq=1&ns_st_ppc=1&ns_st_apc=1&ns_st_spc=1&ns_st_cn=1&ns_st_ev=hb&ns_st_po=10003&ns_st_cl=0&ns_st_hc=1&ns_st_mp=js_api&ns_st_mv=5.1.3.160420&ns_st_pn=1&ns_st_tp=0&ns_st_ci=34526695408&ns_st_pt=10003&ns_st_dpt=10003&ns_st_ipt=10003&ns_st_et=10003&ns_st_det=10003&ns_st_upc=10003&ns_st_dupc=10003&ns_st_iupc=10003&ns_st_upa=10003&ns_st_dupa=10003&ns_st_iupa=10003&ns_st_lpc=10003&ns_st_dlpc=10003&ns_st_lpa=10003&ns_st_dlpa=10003&ns_st_pa=10003&ns_ts=1560539426237&ns_st_bc=0&ns_st_dbc=0&ns_st_bt=0&ns_st_dbt=0&ns_st_bp=0&ns_st_skc=0&ns_st_dskc=0&ns_st_ska=0&ns_st_dska=0&ns_st_skd=0&ns_st_skt=0&ns_st_dskt=0&ns_st_pc=0&ns_st_dpc=0&ns_st_pp=0&ns_st_br=0&ns_st_ub=0&ns_st_ki=1200000&ns_st_pr=*null&ns_st_sn=*null&ns_st_en=*null&ns_st_ep=Week%202%20Day%201%20%7C%20LEC%20Summer%20Split%20(2019)%20-%20Splyce%20vs.%20Schalke%2004&ns_st_ct=vc13&ns_st_ge=Gaming&ns_st_st=Riot%20Games&ns_st_ce=1&ns_st_ia=0&ns_st_ddt=2019-06-14&ns_st_tdt=*null&ns_st_pu=*null&ns_st_ti=*null&c3=TWITCH&c4=twitch.tv&c6=*null&c7=https%3A%2F%2Fwww.twitch.tv%2Friotgames&c8=Riot%20Games%20-%20Twitch&c9=&cs_ucfr=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.twitch.tv - see bug description - <!-- @browser: Firefox 69.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 --> <!-- @reported_with: desktop-reporter --> <!-- @extra_labels: type-webrender-enabled --> **URL**: https://www.twitch.tv/riotgames **Browser / Version**: Firefox 69.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: stream is stuttery even in safe mode **Steps to Reproduce**: 1. click on a stream to watch it [![Screenshot Description](https://webcompat.com/uploads/2019/6/8b3acc0c-5043-4c46-a0e1-674d45f58917-thumb.jpeg)](https://webcompat.com/uploads/2019/6/8b3acc0c-5043-4c46-a0e1-674d45f58917.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190614100022</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: true</li><li>gfx.webrender.all: true</li><li>channel: nightly</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Loading failed for the <script> with source https://c.amazon-adsystem.com/aax2/apstag.js." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://c.amazon-adsystem.com/aax2/apstag.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://www.gstatic.com/cv/js/sender/v1/cast_sender.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://cdn-gl.imrworldwide.com/conf/P34D56F9D-5684-4C83-8EE1-5EA7DE9CF45D.js#name=1a0UrMElc2nTFhA32wPD2WhXYc6RFiFi&ns=NOLBUNDLE." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://client-event-reporter.twitch.tv/v1/stats. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Request to access cookie or storage on https://cdn-gl.imrworldwide.com/conf/P34D56F9D-5684-4C83-8EE1-5EA7DE9CF45D.js#name=1a0UrMElc2nTFhA32wPD2WhXYc6RFiFi&ns=NOLBUNDLE was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://imasdk.googleapis.com/js/sdkloader/ima3.js." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted." {file: "https://static.twitchcdn.net/assets/player-core-base-e939ccadccca15e51973.js" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://s.amazon-adsystem.com/iui3?d=3p-hbg&ex-src=twitch.tv&ex-hargs=v%3D1.0%3Bc%3D8858214122683%3Bp%3De75425fb-5407-7bd5-fd20-f462e98a8777 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://cdn.krxd.net/controltag?confid=JYk8LCsC was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://ssl.google-analytics.com/ga.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/beacon.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted." {file: "https://static.twitchcdn.net/assets/player-core-base-e939ccadccca15e51973.js" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/p?c1=2&c2=6745306&ns_type=hidden&ns_st_sv=5.1.3.160420&ns_st_smv=5.1&ns_st_it=r&ns_st_id=1560539414288&ns_st_ec=1&ns_st_sp=1&ns_st_sc=1&ns_st_sq=1&ns_st_ppc=1&ns_st_apc=1&ns_st_spc=1&ns_st_cn=1&ns_st_ev=play&ns_st_po=0&ns_st_cl=0&ns_st_pb=1&ns_st_mp=js_api&ns_st_mv=5.1.3.160420&ns_st_pn=1&ns_st_tp=0&ns_st_ci=34526695408&ns_st_pt=0&ns_st_dpt=0&ns_st_ipt=0&ns_st_et=0&ns_st_det=0&ns_st_upc=0&ns_st_dupc=0&ns_st_iupc=0&ns_st_upa=0&ns_st_dupa=0&ns_st_iupa=0&ns_st_lpc=0&ns_st_dlpc=0&ns_st_lpa=0&ns_st_dlpa=0&ns_st_pa=0&ns_ts=1560539416234&ns_st_bc=0&ns_st_dbc=0&ns_st_bt=0&ns_st_dbt=0&ns_st_bp=0&ns_st_lt=1946&ns_st_skc=0&ns_st_dskc=0&ns_st_ska=0&ns_st_dska=0&ns_st_skd=0&ns_st_skt=0&ns_st_dskt=0&ns_st_pc=0&ns_st_dpc=0&ns_st_pp=0&ns_st_br=0&ns_st_ub=0&ns_st_ki=1200000&ns_st_pr=*null&ns_st_sn=*null&ns_st_en=*null&ns_st_ep=Week%202%20Day%201%20%7C%20LEC%20Summer%20Split%20(2019)%20-%20Splyce%20vs.%20Schalke%2004&ns_st_ct=vc13&ns_st_ge=Gaming&ns_st_st=Riot%20Games&ns_st_ce=1&ns_st_ia=0&ns_st_ddt=2019-06-14&ns_st_tdt=*null&ns_st_pu=*null&ns_st_ti=*null&c3=TWITCH&c4=twitch.tv&c6=*null&c7=https%3A%2F%2Fwww.twitch.tv%2Friotgames&c8=Riot%20Games%20-%20Twitch&c9=&cs_ucfr=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/p?c1=2&c2=6745306&ns_type=hidden&ns_st_sv=5.1.3.160420&ns_st_smv=5.1&ns_st_it=r&ns_st_id=1560539414288&ns_st_ec=2&ns_st_sp=1&ns_st_sc=1&ns_st_sq=1&ns_st_ppc=1&ns_st_apc=1&ns_st_spc=1&ns_st_cn=1&ns_st_ev=hb&ns_st_po=10003&ns_st_cl=0&ns_st_hc=1&ns_st_mp=js_api&ns_st_mv=5.1.3.160420&ns_st_pn=1&ns_st_tp=0&ns_st_ci=34526695408&ns_st_pt=10003&ns_st_dpt=10003&ns_st_ipt=10003&ns_st_et=10003&ns_st_det=10003&ns_st_upc=10003&ns_st_dupc=10003&ns_st_iupc=10003&ns_st_upa=10003&ns_st_dupa=10003&ns_st_iupa=10003&ns_st_lpc=10003&ns_st_dlpc=10003&ns_st_lpa=10003&ns_st_dlpa=10003&ns_st_pa=10003&ns_ts=1560539426237&ns_st_bc=0&ns_st_dbc=0&ns_st_bt=0&ns_st_dbt=0&ns_st_bp=0&ns_st_skc=0&ns_st_dskc=0&ns_st_ska=0&ns_st_dska=0&ns_st_skd=0&ns_st_skt=0&ns_st_dskt=0&ns_st_pc=0&ns_st_dpc=0&ns_st_pp=0&ns_st_br=0&ns_st_ub=0&ns_st_ki=1200000&ns_st_pr=*null&ns_st_sn=*null&ns_st_en=*null&ns_st_ep=Week%202%20Day%201%20%7C%20LEC%20Summer%20Split%20(2019)%20-%20Splyce%20vs.%20Schalke%2004&ns_st_ct=vc13&ns_st_ge=Gaming&ns_st_st=Riot%20Games&ns_st_ce=1&ns_st_ia=0&ns_st_ddt=2019-06-14&ns_st_tdt=*null&ns_st_pu=*null&ns_st_ti=*null&c3=TWITCH&c4=twitch.tv&c6=*null&c7=https%3A%2F%2Fwww.twitch.tv%2Friotgames&c8=Riot%20Games%20-%20Twitch&c9=&cs_ucfr=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
see bug description url browser version firefox operating system windows tested another browser yes problem type something else description stream is stuttery even in safe mode steps to reproduce click on a stream to watch it browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled true gfx webrender all true channel nightly console messages u u u u u u u u u u u u u u from with ❤️
0
6,794
9,923,696,062
IssuesEvent
2019-07-01 07:55:26
pmem/pmemkv
https://api.github.com/repos/pmem/pmemkv
closed
Continuous Integration for bindings
Type: Process
We need to include checks in CI for our bindings, so we know if we didn't break anything by changes in the main pmemkv repo
1.0
Continuous Integration for bindings - We need to include checks in CI for our bindings, so we know if we didn't break anything by changes in the main pmemkv repo
process
continuous integration for bindings we need to include checks in ci for our bindings so we know if we didn t break anything by changes in the main pmemkv repo
1
55,052
13,507,515,196
IssuesEvent
2020-09-14 06:07:50
rsx-labs/aide-frontend
https://api.github.com/repos/rsx-labs/aide-frontend
closed
[Daily Workplace Audit] Application error on first access
Bug Fixed - ready for build For Next Build High Priority
**Describe the bug** Application error on first access [4:54 PM] Trilles, Marvin 2020-09-03 16:46:04.7696::ERROR::UI_AIDE_CommCellServices.DailyAuditPage::System.NullReferenceException: Object reference not set to an instance of an object. at UI_AIDE_CommCellServices.DailyAuditPage.GenerateQuestions() **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Version (please complete the following information):** - Version 3.40 **Additional context** Add any other context about the problem here.
2.0
[Daily Workplace Audit] Application error on first access - **Describe the bug** Application error on first access [4:54 PM] Trilles, Marvin 2020-09-03 16:46:04.7696::ERROR::UI_AIDE_CommCellServices.DailyAuditPage::System.NullReferenceException: Object reference not set to an instance of an object. at UI_AIDE_CommCellServices.DailyAuditPage.GenerateQuestions() **To Reproduce** Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error **Expected behavior** A clear and concise description of what you expected to happen. **Screenshots** If applicable, add screenshots to help explain your problem. **Version (please complete the following information):** - Version 3.40 **Additional context** Add any other context about the problem here.
non_process
application error on first access describe the bug application error on first access trilles marvin error ui aide commcellservices dailyauditpage system nullreferenceexception object reference not set to an instance of an object at ui aide commcellservices dailyauditpage generatequestions to reproduce steps to reproduce the behavior go to click on scroll down to see error expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem version please complete the following information version additional context add any other context about the problem here
0
177,776
13,748,232,269
IssuesEvent
2020-10-06 08:47:10
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Metal false walls instantly disappear when clicked, and cannot be closed unless you alt click
Bug Tested/Reproduced
## Reproduction 1. Build a false wall. 2. Open it. ![2020-09-30T12-40-02](https://user-images.githubusercontent.com/35135081/94732230-d5724180-031a-11eb-89fc-18dba8da8cf5.gif) This only seems to happen with metal. ![2020-09-30T12-42-57](https://user-images.githubusercontent.com/35135081/94732296-ea4ed500-031a-11eb-9604-1952fe94acee.gif)
1.0
Metal false walls instantly disappear when clicked, and cannot be closed unless you alt click - ## Reproduction 1. Build a false wall. 2. Open it. ![2020-09-30T12-40-02](https://user-images.githubusercontent.com/35135081/94732230-d5724180-031a-11eb-89fc-18dba8da8cf5.gif) This only seems to happen with metal. ![2020-09-30T12-42-57](https://user-images.githubusercontent.com/35135081/94732296-ea4ed500-031a-11eb-9604-1952fe94acee.gif)
non_process
metal false walls instantly disappear when clicked and cannot be closed unless you alt click reproduction build a false wall open it this only seems to happen with metal
0
5,149
7,689,143,767
IssuesEvent
2018-05-17 11:47:52
unb-cic-esw/youtube-data-monitor
https://api.github.com/repos/unb-cic-esw/youtube-data-monitor
opened
Informar inexistência de certos dados
functional requirements
EU : desenvolvedor(a) QUERO : Informar a inexistência de determinados dados: * Criar um arquivo .txt informando os canais que não possuem vídeos criados a partir de 01.01.2018. * Nos arquivos youtube.csv informar a inexistência dos dados. PARA : Melhorar a visibilidade das informações obtidas da API. Testes de aceitação: * Criar um arquivo .txt informando os canais que não possuem vídeos criados a partir de 01.01.2018, bem como a criação de testes para validar essa informação; * Nos arquivos youtube.csv informar a inexistência dos dados, bem como a criação de testes para validar essa informação.
1.0
Informar inexistência de certos dados - EU : desenvolvedor(a) QUERO : Informar a inexistência de determinados dados: * Criar um arquivo .txt informando os canais que não possuem vídeos criados a partir de 01.01.2018. * Nos arquivos youtube.csv informar a inexistência dos dados. PARA : Melhorar a visibilidade das informações obtidas da API. Testes de aceitação: * Criar um arquivo .txt informando os canais que não possuem vídeos criados a partir de 01.01.2018, bem como a criação de testes para validar essa informação; * Nos arquivos youtube.csv informar a inexistência dos dados, bem como a criação de testes para validar essa informação.
non_process
informar inexistência de certos dados eu desenvolvedor a quero informar a inexistência de determinados dados criar um arquivo txt informando os canais que não possuem vídeos criados a partir de nos arquivos youtube csv informar a inexistência dos dados para melhorar a visibilidade das informações obtidas da api testes de aceitação criar um arquivo txt informando os canais que não possuem vídeos criados a partir de bem como a criação de testes para validar essa informação nos arquivos youtube csv informar a inexistência dos dados bem como a criação de testes para validar essa informação
0
15,297
26,890,456,833
IssuesEvent
2023-02-06 08:29:56
bytedance/bitsail
https://api.github.com/repos/bytedance/bitsail
closed
[BitSail][Connector] Support local file source connector
help wanted connector requirement difficulty-easy
## Is your feature request related to a problem? Please describe We need to read local file as source. CSV file format could be a good start. ## Describe the solution you'd like Use V1 source connector interface, read CSV file as source and able to write to `PrintSink` ## Additional context
1.0
[BitSail][Connector] Support local file source connector - ## Is your feature request related to a problem? Please describe We need to read local file as source. CSV file format could be a good start. ## Describe the solution you'd like Use V1 source connector interface, read CSV file as source and able to write to `PrintSink` ## Additional context
non_process
support local file source connector is your feature request related to a problem please describe we need to read local file as source csv file format could be a good start describe the solution you d like use source connector interface read csv file as source and able to write to printsink additional context
0
524,067
15,195,377,986
IssuesEvent
2021-02-16 06:12:37
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.1.8] No fertilisers appear in the EcoPedia
Category: Usability Priority: Low Status: Fixed
Seems the various fertilisers are missing entirely from EcoPedia: ![image](https://user-images.githubusercontent.com/964559/102599318-28fe5b00-40d2-11eb-81e3-564f8af058ba.png)
1.0
[0.9.1.8] No fertilisers appear in the EcoPedia - Seems the various fertilisers are missing entirely from EcoPedia: ![image](https://user-images.githubusercontent.com/964559/102599318-28fe5b00-40d2-11eb-81e3-564f8af058ba.png)
non_process
no fertilisers appear in the ecopedia seems the various fertilisers are missing entirely from ecopedia
0
8,407
11,572,751,108
IssuesEvent
2020-02-21 01:06:03
kubeflow/testing
https://api.github.com/repos/kubeflow/testing
closed
Auto-deploy V2
area/engprod kind/feature kind/process priority/p1
This is the uber bug for updating the auto-deployment infrastructure. The key feature that we'd like to achieve with V2 is to auto-deploy as soon as a PR is merged. The current implementation doesn't facilitate that because we have a simple cron job which launches a new deployment every X hours. If we decrease X in order to deploy sooner we will end up accumulating multiple instances of the auto-deployed infra and using up all our quota. We want more of a rate limiting queue type pattern similar to what K8s controllers and reconcilers do. This is critical to being able to rapidly verify changes to the deployments.
1.0
Auto-deploy V2 - This is the uber bug for updating the auto-deployment infrastructure. The key feature that we'd like to achieve with V2 is to auto-deploy as soon as a PR is merged. The current implementation doesn't facilitate that because we have a simple cron job which launches a new deployment every X hours. If we decrease X in order to deploy sooner we will end up accumulating multiple instances of the auto-deployed infra and using up all our quota. We want more of a rate limiting queue type pattern similar to what K8s controllers and reconcilers do. This is critical to being able to rapidly verify changes to the deployments.
process
auto deploy this is the uber bug for updating the auto deployment infrastructure the key feature that we d like to achieve with is to auto deploy as soon as a pr is merged the current implementation doesn t facilitate that because we have a simple cron job which launches a new deployment every x hours if we decrease x in order to deploy sooner we will end up accumulating multiple instances of the auto deployed infra and using up all our quota we want more of a rate limiting queue type pattern similar to what controllers and reconcilers do this is critical to being able to rapidly verify changes to the deployments
1
389,045
11,496,609,682
IssuesEvent
2020-02-12 08:22:51
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
nrf52 telnet_shell panic. Mutex using in ISR.
bug has-pr priority: medium
Stack is ``` #0 assert_post_action (file=file@entry=0x4ab31 "ZEPHYR_BASE/kernel/include/ksched.h", line=line@entry=269) at ZEPHYR_BASE/lib/os/assert.c:45 #1 0x0003086a in z_sched_lock () at ZEPHYR_BASE/kernel/include/ksched.h:269 #2 z_impl_k_mutex_unlock (mutex=mutex@entry=0x200023d8 <contexts+248>) at ZEPHYR_BASE/kernel/mutex.c:232 #3 0x0003c556 in k_mutex_unlock (mutex=mutex@entry=0x200023d8 <contexts+248>) at zephyr/include/generated/syscalls/kernel.h:644 #4 0x0003c692 in net_context_send (context=0x200023d0 <contexts+240>, buf=0x20000938 <shell_transport_telnet_shell_telnet+8>, len=13, cb=cb@entry=0x5399 <telnet_sent_cb>, timeout=timeout@entry=-1, user_data=user_data@entry=0x0 <sys_pm_notify_power_state_entry>) at ZEPHYR_BASE/subsys/net/ip/net_context.c:1668 #5 0x0000525e in telnet_send () at ZEPHYR_BASE/subsys/shell/shell_telnet.c:138 #6 0x00032eea in z_timer_expiration_handler (t=0x200009a4 <shell_transport_telnet_shell_telnet+116>) at ZEPHYR_BASE/kernel/timer.c:65 #7 0x00032dc4 in z_clock_announce (ticks=<optimized out>) at ZEPHYR_BASE/kernel/timeout.c:204 #8 0x00007dee in _isr_wrapper () at ZEPHYR_BASE/arch/arm/core/aarch32/isr_wrapper.S:163 #9 <signal handler called> ``` Since timer handlers in nrf52 is in ISR and network uses mutexes we got panic. Maybe moving `telnet_send` in separate thread will resolve this issue.
1.0
nrf52 telnet_shell panic. Mutex using in ISR. - Stack is ``` #0 assert_post_action (file=file@entry=0x4ab31 "ZEPHYR_BASE/kernel/include/ksched.h", line=line@entry=269) at ZEPHYR_BASE/lib/os/assert.c:45 #1 0x0003086a in z_sched_lock () at ZEPHYR_BASE/kernel/include/ksched.h:269 #2 z_impl_k_mutex_unlock (mutex=mutex@entry=0x200023d8 <contexts+248>) at ZEPHYR_BASE/kernel/mutex.c:232 #3 0x0003c556 in k_mutex_unlock (mutex=mutex@entry=0x200023d8 <contexts+248>) at zephyr/include/generated/syscalls/kernel.h:644 #4 0x0003c692 in net_context_send (context=0x200023d0 <contexts+240>, buf=0x20000938 <shell_transport_telnet_shell_telnet+8>, len=13, cb=cb@entry=0x5399 <telnet_sent_cb>, timeout=timeout@entry=-1, user_data=user_data@entry=0x0 <sys_pm_notify_power_state_entry>) at ZEPHYR_BASE/subsys/net/ip/net_context.c:1668 #5 0x0000525e in telnet_send () at ZEPHYR_BASE/subsys/shell/shell_telnet.c:138 #6 0x00032eea in z_timer_expiration_handler (t=0x200009a4 <shell_transport_telnet_shell_telnet+116>) at ZEPHYR_BASE/kernel/timer.c:65 #7 0x00032dc4 in z_clock_announce (ticks=<optimized out>) at ZEPHYR_BASE/kernel/timeout.c:204 #8 0x00007dee in _isr_wrapper () at ZEPHYR_BASE/arch/arm/core/aarch32/isr_wrapper.S:163 #9 <signal handler called> ``` Since timer handlers in nrf52 is in ISR and network uses mutexes we got panic. Maybe moving `telnet_send` in separate thread will resolve this issue.
non_process
telnet shell panic mutex using in isr stack is assert post action file file entry zephyr base kernel include ksched h line line entry at zephyr base lib os assert c in z sched lock at zephyr base kernel include ksched h z impl k mutex unlock mutex mutex entry at zephyr base kernel mutex c in k mutex unlock mutex mutex entry at zephyr include generated syscalls kernel h in net context send context buf len cb cb entry timeout timeout entry user data user data entry at zephyr base subsys net ip net context c in telnet send at zephyr base subsys shell shell telnet c in z timer expiration handler t at zephyr base kernel timer c in z clock announce ticks at zephyr base kernel timeout c in isr wrapper at zephyr base arch arm core isr wrapper s since timer handlers in is in isr and network uses mutexes we got panic maybe moving telnet send in separate thread will resolve this issue
0
280
2,718,446,390
IssuesEvent
2015-04-12 09:45:57
tomatocart/TomatoCart-v1
https://api.github.com/repos/tomatocart/TomatoCart-v1
opened
Replace mootools popup cart with (jquery, zepto) popup cart in bootstrap template
Priority: High Status: In process
In order to support mobile device, we have to remove mootools framework from bootstrap template. Only one framework should be loaded. Jquery for pc and Zepto for mobile. In this way, the bootstrap store front will work smoothly in mobile devices. This is the first step - Replace mootools popup cart with (jquery, zepto) popup cart in bootstrap template
1.0
Replace mootools popup cart with (jquery, zepto) popup cart in bootstrap template - In order to support mobile device, we have to remove mootools framework from bootstrap template. Only one framework should be loaded. Jquery for pc and Zepto for mobile. In this way, the bootstrap store front will work smoothly in mobile devices. This is the first step - Replace mootools popup cart with (jquery, zepto) popup cart in bootstrap template
process
replace mootools popup cart with jquery zepto popup cart in bootstrap template in order to support mobile device we have to remove mootools framework from bootstrap template only one framework should be loaded jquery for pc and zepto for mobile in this way the bootstrap store front will work smoothly in mobile devices this is the first step replace mootools popup cart with jquery zepto popup cart in bootstrap template
1
514,020
14,931,620,596
IssuesEvent
2021-01-25 06:11:09
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
logging: missing assignment in function "toHTTPRequest"
api: logging priority: p2 type: question
In this [file](https://github.com/googleapis/google-cloud-go/blob/master/logging/logadmin/logadmin.go#L164) of line 164, we can see the assignment of logging.HTTPRequest's field "RemoteIP", and maybe the field "LocalIP" should be assigned also? here shows the fixed: ![image](https://user-images.githubusercontent.com/17964471/105351642-c85dc300-5c27-11eb-89c4-ca7579cacbd7.png)
1.0
logging: missing assignment in function "toHTTPRequest" - In this [file](https://github.com/googleapis/google-cloud-go/blob/master/logging/logadmin/logadmin.go#L164) of line 164, we can see the assignment of logging.HTTPRequest's field "RemoteIP", and maybe the field "LocalIP" should be assigned also? here shows the fixed: ![image](https://user-images.githubusercontent.com/17964471/105351642-c85dc300-5c27-11eb-89c4-ca7579cacbd7.png)
non_process
logging missing assignment in function tohttprequest in this of line we can see the assignment of logging httprequest s field remoteip and maybe the field localip should be assigned also here shows the fixed
0
111,549
24,146,142,141
IssuesEvent
2022-09-21 18:57:22
rkd103/Group-11
https://api.github.com/repos/rkd103/Group-11
closed
Implement User Story B
enhancement code
Taking the User Stories assigned in previous issues, begin brainstorming how to implement the requirements in code. Namely, perform independent research regarding what packages, techniques, etc. are needed to satisfy the stakeholder's image. User Story B: "compose, edit, and post statuses on personal timeline" Original Assignee: HeFeng (ho170)
1.0
Implement User Story B - Taking the User Stories assigned in previous issues, begin brainstorming how to implement the requirements in code. Namely, perform independent research regarding what packages, techniques, etc. are needed to satisfy the stakeholder's image. User Story B: "compose, edit, and post statuses on personal timeline" Original Assignee: HeFeng (ho170)
non_process
implement user story b taking the user stories assigned in previous issues begin brainstorming how to implement the requirements in code namely perform independent research regarding what packages techniques etc are needed to satisfy the stakeholder s image user story b compose edit and post statuses on personal timeline original assignee hefeng
0
17,442
23,265,968,540
IssuesEvent
2022-08-04 17:24:44
MPMG-DCC-UFMG/C01
https://api.github.com/repos/MPMG-DCC-UFMG/C01
opened
Transparência - Detalhes do coletor/Opções avançadas (extração de links e download de arquivos)
[1] Requisito [0] Desenvolvimento [2] Média Prioridade [3] Processamento Dinâmico
## Comportamento Esperado Espera-se que as configurações de extração e download das Opções avançadas se apliquem também às coletas que usam processamento dinâmico. ## Comportamento Atual Ao configurar um coletor dinâmico com essa ferramenta, as opções avançadas disponíveis na tela de detalhes do coletor não se aplicam da mesma maneira que funcionam na extração de links "estáticos" pelo Scrapy. Para usar essas ferramentas dinamicamente, apesar de haver opções extras para filtrar xpaths em alguns passos do mecanismo de processamento dinâmico, não há uma opção nos detalhes do coletor para aplicar as opções avançadas desejadas. Isso pode ser pouco intuitivo para o usuário. ## Passos para reproduzir o erro Não se aplica. ## Sistema - MP ou local: ambos - Branch específica: master - Sistema diferente: não ## Screenshots Não se aplica.
1.0
Transparência - Detalhes do coletor/Opções avançadas (extração de links e download de arquivos) - ## Comportamento Esperado Espera-se que as configurações de extração e download das Opções avançadas se apliquem também às coletas que usam processamento dinâmico. ## Comportamento Atual Ao configurar um coletor dinâmico com essa ferramenta, as opções avançadas disponíveis na tela de detalhes do coletor não se aplicam da mesma maneira que funcionam na extração de links "estáticos" pelo Scrapy. Para usar essas ferramentas dinamicamente, apesar de haver opções extras para filtrar xpaths em alguns passos do mecanismo de processamento dinâmico, não há uma opção nos detalhes do coletor para aplicar as opções avançadas desejadas. Isso pode ser pouco intuitivo para o usuário. ## Passos para reproduzir o erro Não se aplica. ## Sistema - MP ou local: ambos - Branch específica: master - Sistema diferente: não ## Screenshots Não se aplica.
process
transparência detalhes do coletor opções avançadas extração de links e download de arquivos comportamento esperado espera se que as configurações de extração e download das opções avançadas se apliquem também às coletas que usam processamento dinâmico comportamento atual ao configurar um coletor dinâmico com essa ferramenta as opções avançadas disponíveis na tela de detalhes do coletor não se aplicam da mesma maneira que funcionam na extração de links estáticos pelo scrapy para usar essas ferramentas dinamicamente apesar de haver opções extras para filtrar xpaths em alguns passos do mecanismo de processamento dinâmico não há uma opção nos detalhes do coletor para aplicar as opções avançadas desejadas isso pode ser pouco intuitivo para o usuário passos para reproduzir o erro não se aplica sistema mp ou local ambos branch específica master sistema diferente não screenshots não se aplica
1
37,563
10,029,748,102
IssuesEvent
2019-07-17 14:30:57
akka/akka-http
https://api.github.com/repos/akka/akka-http
closed
Release Akka HTTP 10.1.9
3 - in progress t:build
<!-- (Liberally copied and adopted from Scala itself https://github.com/scala/scala-dev/blob/b11cd2e4a4431de7867db6b39362bea8fa6650e7/notes/releases/template.md) For every Akka HTTP release, make a copy of this file named after the release, and expand the variables. Ideally replacing variables could become a script you can run on your local machine. Variables to be expanded in this template: - VERSION=??? Key links: - akka/akka-http milestone: https://github.com/akka/akka-http/milestone/? --> ### ~ 1 week before the release - [x] Check that open PRs and issues assigned to the milestone are reasonable - [x] Triage tickets that should be ready for this release, add "pick next" label and release milestone - [x] Triage open PRs and apply "pick next" label and maybe add to release milestone. Some PRs might be explicitly scheduled for this release, others might be ready enough to bring them over the finish line Wind down PR queue. There has to be enough time after the last (non-trivial) PR is merged and the next phase. The core of the eco-system needs time to prepare for the final! - [x] Decide on planned release date - [x] Notify depending projects (notably Play + cinnamon) about upcoming release ### 1 day before the release - [x] Make sure all important / big PRs have been merged by now - [x] Check that latest nightly (once we have that) / master still works with depending projects (notably Play + cinnamon) - [x] Communicate that a release is about to be released in [Gitter Akka Dev Channel](https://gitter.im/akka/dev), so that no new Pull Requests are merged ### Preparing release notes in the documentation / announcement - [x] If this is a new minor (not patch) release, rename the 'akka-http-x.x-stable' reporting project in [WhiteSource](https://saas.whitesourcesoftware.com/) accordingly - [x] Add a release notes entry in docs/src/main/paradox/release-notes/. As a helper run `scripts/commits-for-release-notes.sh <last-version-tag>` which will output a list of commits grouped by submodule, and the closed issues for this milestone. - [x] Create a PR on https://github.com/akka/akka.github.com with a news item draft, using the milestones and `scripts/authors.scala previousVersion thisVersion` - [x] Create a PR on https://github.com/akka/akka.github.com updating Akka HTTP dependency for akka.io website: https://github.com/akka/akka.github.com/blob/master/_config.yml - [x] Release notes PR has been merged - [x] Create a new milestone for the next version at https://github.com/akka/akka-http/milestones - [x] Move all unclosed issues to the newly created milestone (or remove milestone) and close the version you're releasing ### Cutting the release - [x] Make sure there are no stray staging repos on sonatype - [x] Wait until [master build finished](https://travis-ci.org/akka/akka-http/builds/) after merging the release notes - [x] Create a [new release](https://github.com/akka/akka-http/releases/new) with the next tag version (e.g. `v13.3.7`), title and release description linking to announcement, release notes and milestone. - [x] Check that the Travis CI release build has executed successfully - [x] Go to https://bintray.com/akka/maven/com.typesafe.akka:akka-http_2.11 and select the just released version - [x] Go to the Maven Central tab and sync with Sonatype - [x] Log in to Sonatype to Close the staging repository (optional, should happen automatically if selected in Bintray) - [x] Notify Telemetry / Play team to check against staged artifacts - [x] Run a test against the staging repository to make sure the release went well, for example by using https://github.com/akka/akka-http-scala-seed.g8 and adding the sonatype staging repo with `resolvers += "Staging Repo" at "https://oss.sonatype.org/content/repositories/comtypesafe-xxx"` - [x] Release the staging repository to Maven Central. - [x] Checkout the newly created tag and run `sbt -Dakka.genjavadoc.enabled=true ++2.12.8 "deployRsync akkarepo@gustav.akka.io"` to deploy API and reference documentation. ### Check availability - [x] Check release on sonatype: https://oss.sonatype.org/content/repositories/releases/com/typesafe/akka/akka-http-core_2.11/ - [x] Check the release on maven central: http://central.maven.org/maven2/com/typesafe/akka/akka-http-core_2.11/ ### When everything is on maven central - [x] Log into `gustav.akka.io` as `akkarepo` - [ ] update the `10.1` and `current` links on `repo.akka.io` to point to the latest version with (**replace the minor appropriately**) ``` ln -nsf 10.1.9 www/docs/akka-http/10.1 ln -nsf 10.1.9 www/api/akka-http/10.1 ln -nsf 10.1.9 www/japi/akka-http/10.1 ln -nsf 10.1.9 www/docs/akka-http/current ln -nsf 10.1.9 www/api/akka-http/current ln -nsf 10.1.9 www/jap/akka-http/current ``` - [x] check changes and commit the new version to the local git repository ``` cd ~/www git add docs/akka-http/ api/akka-http/ japi/akka-http/ git commit -m "Akka HTTP 10.1.9" ``` ### Announcements - [x] Merge draft news item at https://github.com/akka/akka.github.com - [x] Send a release notification to https://discuss.akka.io - [x] Tweet using the akka account (or ask someone to) about the new release - [x] Announce on Gitter at https://gitter.im/akka/akka ### Afterwards - [x] Add the released version to `project/MiMa.scala` to the `mimaPreviousArtifacts` key *of all current compatible branches*. - [ ] Update Akka HTTP dependency in [akka-management](https://github.com/akka/akka-management/edit/master/project/Dependencies.scala) - [x] Update Akka HTTP reference in [reactive-platform-docs](https://github.com/typesafehub/reactive-platform-docs/edit/master/build.sbt#L29) - Close this issue
1.0
Release Akka HTTP 10.1.9 - <!-- (Liberally copied and adopted from Scala itself https://github.com/scala/scala-dev/blob/b11cd2e4a4431de7867db6b39362bea8fa6650e7/notes/releases/template.md) For every Akka HTTP release, make a copy of this file named after the release, and expand the variables. Ideally replacing variables could become a script you can run on your local machine. Variables to be expanded in this template: - VERSION=??? Key links: - akka/akka-http milestone: https://github.com/akka/akka-http/milestone/? --> ### ~ 1 week before the release - [x] Check that open PRs and issues assigned to the milestone are reasonable - [x] Triage tickets that should be ready for this release, add "pick next" label and release milestone - [x] Triage open PRs and apply "pick next" label and maybe add to release milestone. Some PRs might be explicitly scheduled for this release, others might be ready enough to bring them over the finish line Wind down PR queue. There has to be enough time after the last (non-trivial) PR is merged and the next phase. The core of the eco-system needs time to prepare for the final! - [x] Decide on planned release date - [x] Notify depending projects (notably Play + cinnamon) about upcoming release ### 1 day before the release - [x] Make sure all important / big PRs have been merged by now - [x] Check that latest nightly (once we have that) / master still works with depending projects (notably Play + cinnamon) - [x] Communicate that a release is about to be released in [Gitter Akka Dev Channel](https://gitter.im/akka/dev), so that no new Pull Requests are merged ### Preparing release notes in the documentation / announcement - [x] If this is a new minor (not patch) release, rename the 'akka-http-x.x-stable' reporting project in [WhiteSource](https://saas.whitesourcesoftware.com/) accordingly - [x] Add a release notes entry in docs/src/main/paradox/release-notes/. As a helper run `scripts/commits-for-release-notes.sh <last-version-tag>` which will output a list of commits grouped by submodule, and the closed issues for this milestone. - [x] Create a PR on https://github.com/akka/akka.github.com with a news item draft, using the milestones and `scripts/authors.scala previousVersion thisVersion` - [x] Create a PR on https://github.com/akka/akka.github.com updating Akka HTTP dependency for akka.io website: https://github.com/akka/akka.github.com/blob/master/_config.yml - [x] Release notes PR has been merged - [x] Create a new milestone for the next version at https://github.com/akka/akka-http/milestones - [x] Move all unclosed issues to the newly created milestone (or remove milestone) and close the version you're releasing ### Cutting the release - [x] Make sure there are no stray staging repos on sonatype - [x] Wait until [master build finished](https://travis-ci.org/akka/akka-http/builds/) after merging the release notes - [x] Create a [new release](https://github.com/akka/akka-http/releases/new) with the next tag version (e.g. `v13.3.7`), title and release description linking to announcement, release notes and milestone. - [x] Check that the Travis CI release build has executed successfully - [x] Go to https://bintray.com/akka/maven/com.typesafe.akka:akka-http_2.11 and select the just released version - [x] Go to the Maven Central tab and sync with Sonatype - [x] Log in to Sonatype to Close the staging repository (optional, should happen automatically if selected in Bintray) - [x] Notify Telemetry / Play team to check against staged artifacts - [x] Run a test against the staging repository to make sure the release went well, for example by using https://github.com/akka/akka-http-scala-seed.g8 and adding the sonatype staging repo with `resolvers += "Staging Repo" at "https://oss.sonatype.org/content/repositories/comtypesafe-xxx"` - [x] Release the staging repository to Maven Central. - [x] Checkout the newly created tag and run `sbt -Dakka.genjavadoc.enabled=true ++2.12.8 "deployRsync akkarepo@gustav.akka.io"` to deploy API and reference documentation. ### Check availability - [x] Check release on sonatype: https://oss.sonatype.org/content/repositories/releases/com/typesafe/akka/akka-http-core_2.11/ - [x] Check the release on maven central: http://central.maven.org/maven2/com/typesafe/akka/akka-http-core_2.11/ ### When everything is on maven central - [x] Log into `gustav.akka.io` as `akkarepo` - [ ] update the `10.1` and `current` links on `repo.akka.io` to point to the latest version with (**replace the minor appropriately**) ``` ln -nsf 10.1.9 www/docs/akka-http/10.1 ln -nsf 10.1.9 www/api/akka-http/10.1 ln -nsf 10.1.9 www/japi/akka-http/10.1 ln -nsf 10.1.9 www/docs/akka-http/current ln -nsf 10.1.9 www/api/akka-http/current ln -nsf 10.1.9 www/jap/akka-http/current ``` - [x] check changes and commit the new version to the local git repository ``` cd ~/www git add docs/akka-http/ api/akka-http/ japi/akka-http/ git commit -m "Akka HTTP 10.1.9" ``` ### Announcements - [x] Merge draft news item at https://github.com/akka/akka.github.com - [x] Send a release notification to https://discuss.akka.io - [x] Tweet using the akka account (or ask someone to) about the new release - [x] Announce on Gitter at https://gitter.im/akka/akka ### Afterwards - [x] Add the released version to `project/MiMa.scala` to the `mimaPreviousArtifacts` key *of all current compatible branches*. - [ ] Update Akka HTTP dependency in [akka-management](https://github.com/akka/akka-management/edit/master/project/Dependencies.scala) - [x] Update Akka HTTP reference in [reactive-platform-docs](https://github.com/typesafehub/reactive-platform-docs/edit/master/build.sbt#L29) - Close this issue
non_process
release akka http liberally copied and adopted from scala itself for every akka http release make a copy of this file named after the release and expand the variables ideally replacing variables could become a script you can run on your local machine variables to be expanded in this template version key links akka akka http milestone week before the release check that open prs and issues assigned to the milestone are reasonable triage tickets that should be ready for this release add pick next label and release milestone triage open prs and apply pick next label and maybe add to release milestone some prs might be explicitly scheduled for this release others might be ready enough to bring them over the finish line wind down pr queue there has to be enough time after the last non trivial pr is merged and the next phase the core of the eco system needs time to prepare for the final decide on planned release date notify depending projects notably play cinnamon about upcoming release day before the release make sure all important big prs have been merged by now check that latest nightly once we have that master still works with depending projects notably play cinnamon communicate that a release is about to be released in so that no new pull requests are merged preparing release notes in the documentation announcement if this is a new minor not patch release rename the akka http x x stable reporting project in accordingly add a release notes entry in docs src main paradox release notes as a helper run scripts commits for release notes sh which will output a list of commits grouped by submodule and the closed issues for this milestone create a pr on with a news item draft using the milestones and scripts authors scala previousversion thisversion create a pr on updating akka http dependency for akka io website release notes pr has been merged create a new milestone for the next version at move all unclosed issues to the newly created milestone or remove milestone and close the version you re releasing cutting the release make sure there are no stray staging repos on sonatype wait until after merging the release notes create a with the next tag version e g title and release description linking to announcement release notes and milestone check that the travis ci release build has executed successfully go to and select the just released version go to the maven central tab and sync with sonatype log in to sonatype to close the staging repository optional should happen automatically if selected in bintray notify telemetry play team to check against staged artifacts run a test against the staging repository to make sure the release went well for example by using and adding the sonatype staging repo with resolvers staging repo at release the staging repository to maven central checkout the newly created tag and run sbt dakka genjavadoc enabled true deployrsync akkarepo gustav akka io to deploy api and reference documentation check availability check release on sonatype check the release on maven central when everything is on maven central log into gustav akka io as akkarepo update the and current links on repo akka io to point to the latest version with replace the minor appropriately ln nsf www docs akka http ln nsf www api akka http ln nsf www japi akka http ln nsf www docs akka http current ln nsf www api akka http current ln nsf www jap akka http current check changes and commit the new version to the local git repository cd www git add docs akka http api akka http japi akka http git commit m akka http announcements merge draft news item at send a release notification to tweet using the akka account or ask someone to about the new release announce on gitter at afterwards add the released version to project mima scala to the mimapreviousartifacts key of all current compatible branches update akka http dependency in update akka http reference in close this issue
0
11,247
14,015,567,798
IssuesEvent
2020-10-29 13:29:43
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - dcterms:modified
Class - Record-level Process - implement Term - change
## Change term * Submitter: John Wieczorek * Justification (why is this change necessary?): The recommended specification for dates has been deprecated. * Proponents (who needs this change): Everyone Proposed new attributes of the term: * Usage comments (recommendations regarding content, etc.): "Recommended best practice is to use a date that conforms to ISO 8601-1:2019."
1.0
Change term - dcterms:modified - ## Change term * Submitter: John Wieczorek * Justification (why is this change necessary?): The recommended specification for dates has been deprecated. * Proponents (who needs this change): Everyone Proposed new attributes of the term: * Usage comments (recommendations regarding content, etc.): "Recommended best practice is to use a date that conforms to ISO 8601-1:2019."
process
change term dcterms modified change term submitter john wieczorek justification why is this change necessary the recommended specification for dates has been deprecated proponents who needs this change everyone proposed new attributes of the term usage comments recommendations regarding content etc recommended best practice is to use a date that conforms to iso
1
22,234
3,777,304,118
IssuesEvent
2016-03-17 19:34:15
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Elimination of the use of semicolon at the end of every statement in C# 7
Area-Language Design
**Version Used**: C# 6 **Steps to Reproduce**: We all have to put the ; at the end of every statement. The compiler complain if we don't. Could it be possible to eliminate the need of adding the semicolon? It is a dramatic request, but I think at this point we can get rid of it. The compiler knows when we missed it, it give us an error. Since it knows, could the IDE analyzer add it automatically or better just not expect us to type it anymore. F# do not have it, Swift do not have it. C# can get rid of it :) 1. 2. 3. **Expected Behavior**: Love to see this: Console.Write("No more semicolons, yeah!!") **Actual Behavior**: Console.Write("Hating the semicolon at the end, yeah!!");
1.0
Elimination of the use of semicolon at the end of every statement in C# 7 - **Version Used**: C# 6 **Steps to Reproduce**: We all have to put the ; at the end of every statement. The compiler complain if we don't. Could it be possible to eliminate the need of adding the semicolon? It is a dramatic request, but I think at this point we can get rid of it. The compiler knows when we missed it, it give us an error. Since it knows, could the IDE analyzer add it automatically or better just not expect us to type it anymore. F# do not have it, Swift do not have it. C# can get rid of it :) 1. 2. 3. **Expected Behavior**: Love to see this: Console.Write("No more semicolons, yeah!!") **Actual Behavior**: Console.Write("Hating the semicolon at the end, yeah!!");
non_process
elimination of the use of semicolon at the end of every statement in c version used c steps to reproduce we all have to put the at the end of every statement the compiler complain if we don t could it be possible to eliminate the need of adding the semicolon it is a dramatic request but i think at this point we can get rid of it the compiler knows when we missed it it give us an error since it knows could the ide analyzer add it automatically or better just not expect us to type it anymore f do not have it swift do not have it c can get rid of it expected behavior love to see this console write no more semicolons yeah actual behavior console write hating the semicolon at the end yeah
0
13,190
15,613,698,160
IssuesEvent
2021-03-19 16:47:31
Kreusada/Kreusada-Cogs
https://api.github.com/repos/Kreusada/Kreusada-Cogs
opened
[Dehoister] Some builtins do not have a `__file__` attr
Cog: Dehoister Status: Processing Type: Bug
**What cog are you reporting for?** Dehoister **What command(s) inside the cog are you reporting for?** `[p]vinfo mod` **Description of the issue** Some builtin (or external) modules and/or dependencies do not have `__file__` attributes. This means that the module cannot be checked against `pypath` (`str(distutils.sysconfig.get_python_lib(standard_lib=True))`) because they do not have those attributes. Suggested fixes would probably be to check via `dep.__spec__.origin`. This seems like a more reliable way, although I noticed that one module `sys` returns None (perhaps because it is quite literally the python core, idk). Examples of modules that threw this attribute error were math, and array. I'm sure there will probably be more. ![2021-03-19](https://user-images.githubusercontent.com/67752638/111814627-99d03180-88d2-11eb-8f1c-79a2314781f1.png) I am making this issue so that I can get round to resolving it as soon as possible. Please do not open any subsequent issues for this bug.
1.0
[Dehoister] Some builtins do not have a `__file__` attr - **What cog are you reporting for?** Dehoister **What command(s) inside the cog are you reporting for?** `[p]vinfo mod` **Description of the issue** Some builtin (or external) modules and/or dependencies do not have `__file__` attributes. This means that the module cannot be checked against `pypath` (`str(distutils.sysconfig.get_python_lib(standard_lib=True))`) because they do not have those attributes. Suggested fixes would probably be to check via `dep.__spec__.origin`. This seems like a more reliable way, although I noticed that one module `sys` returns None (perhaps because it is quite literally the python core, idk). Examples of modules that threw this attribute error were math, and array. I'm sure there will probably be more. ![2021-03-19](https://user-images.githubusercontent.com/67752638/111814627-99d03180-88d2-11eb-8f1c-79a2314781f1.png) I am making this issue so that I can get round to resolving it as soon as possible. Please do not open any subsequent issues for this bug.
process
some builtins do not have a file attr what cog are you reporting for dehoister what command s inside the cog are you reporting for vinfo mod description of the issue some builtin or external modules and or dependencies do not have file attributes this means that the module cannot be checked against pypath str distutils sysconfig get python lib standard lib true because they do not have those attributes suggested fixes would probably be to check via dep spec origin this seems like a more reliable way although i noticed that one module sys returns none perhaps because it is quite literally the python core idk examples of modules that threw this attribute error were math and array i m sure there will probably be more i am making this issue so that i can get round to resolving it as soon as possible please do not open any subsequent issues for this bug
1
290,397
32,076,510,509
IssuesEvent
2023-09-25 11:22:49
exasol/exasol-testcontainers
https://api.github.com/repos/exasol/exasol-testcontainers
closed
Fix vulnerability in dependency commons-compress
security
``` Error: Failed to execute goal org.sonatype.ossindex.maven:ossindex-maven-plugin:3.2.0:audit (default-cli) on project exasol-testcontainers: Detected 1 vulnerable components: Error: org.apache.commons:commons-compress:jar:1.23.0:compile; https://ossindex.sonatype.org/component/pkg:maven/org.apache.commons/commons-compress@1.23.0?utm_source=ossindex-client&utm_medium=integration&utm_content=1.8.1 Error: * [CVE-2023-42503] CWE-20: Improper Input Validation (5.5); https://ossindex.sonatype.org/vulnerability/CVE-2023-42503?component-type=maven&component-name=org.apache.commons%2Fcommons-compress&utm_source=ossindex-client&utm_medium=integration&utm_content=1.8.1 ```
True
Fix vulnerability in dependency commons-compress - ``` Error: Failed to execute goal org.sonatype.ossindex.maven:ossindex-maven-plugin:3.2.0:audit (default-cli) on project exasol-testcontainers: Detected 1 vulnerable components: Error: org.apache.commons:commons-compress:jar:1.23.0:compile; https://ossindex.sonatype.org/component/pkg:maven/org.apache.commons/commons-compress@1.23.0?utm_source=ossindex-client&utm_medium=integration&utm_content=1.8.1 Error: * [CVE-2023-42503] CWE-20: Improper Input Validation (5.5); https://ossindex.sonatype.org/vulnerability/CVE-2023-42503?component-type=maven&component-name=org.apache.commons%2Fcommons-compress&utm_source=ossindex-client&utm_medium=integration&utm_content=1.8.1 ```
non_process
fix vulnerability in dependency commons compress error failed to execute goal org sonatype ossindex maven ossindex maven plugin audit default cli on project exasol testcontainers detected vulnerable components error org apache commons commons compress jar compile error cwe improper input validation
0
95,486
10,881,096,189
IssuesEvent
2019-11-17 15:35:15
hsloot/rmo
https://api.github.com/repos/hsloot/rmo
closed
[DOCUMENTATION]: Package documentation in manual
documentation enhancement good first issue help wanted
## Summary The documentation of the package-manual is missing. ## Proposal Create a file`R/rmo.R` With a short description of the package and its main applications. ## Additional context See [Object documentation - R packages](http://r-pkgs.had.co.nz/man.html#man-packages).
1.0
[DOCUMENTATION]: Package documentation in manual - ## Summary The documentation of the package-manual is missing. ## Proposal Create a file`R/rmo.R` With a short description of the package and its main applications. ## Additional context See [Object documentation - R packages](http://r-pkgs.had.co.nz/man.html#man-packages).
non_process
package documentation in manual summary the documentation of the package manual is missing proposal create a file r rmo r with a short description of the package and its main applications additional context see
0
184,721
6,715,683,664
IssuesEvent
2017-10-13 22:29:15
FStarLang/FStar
https://api.github.com/repos/FStarLang/FStar
closed
Type annotations in match (and destructuring let) get completely ignored
area/syntax area/usability component/typechecker kind/bug priority/high
The following should in my eyes be refused (or at least trigger a warning). ``` module Test type test : Type = | T : x:bool -> test val f : test -> bool let f h = match h with T (t : int) -> t ```
1.0
Type annotations in match (and destructuring let) get completely ignored - The following should in my eyes be refused (or at least trigger a warning). ``` module Test type test : Type = | T : x:bool -> test val f : test -> bool let f h = match h with T (t : int) -> t ```
non_process
type annotations in match and destructuring let get completely ignored the following should in my eyes be refused or at least trigger a warning module test type test type t x bool test val f test bool let f h match h with t t int t
0
217,957
24,351,699,753
IssuesEvent
2022-10-03 01:11:03
jyothsna/votingapplication
https://api.github.com/repos/jyothsna/votingapplication
opened
CVE-2022-42004 (Medium) detected in jackson-databind-2.9.6.jar, jackson-databind-2.9.0.jar
security vulnerability
## CVE-2022-42004 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.0.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.3.RELEASE.jar (Root Library) - spring-boot-starter-json-2.0.3.RELEASE.jar - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /epository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.0.jar** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization. <p>Publish Date: 2022-10-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-42004 (Medium) detected in jackson-databind-2.9.6.jar, jackson-databind-2.9.0.jar - ## CVE-2022-42004 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.0.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.0.3.RELEASE.jar (Root Library) - spring-boot-starter-json-2.0.3.RELEASE.jar - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /epository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.0.jar** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization. <p>Publish Date: 2022-10-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jackson databind jar jackson databind jar cve medium severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library epository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details in fasterxml jackson databind before resource exhaustion can occur because of a lack of a check in beandeserializer deserializefromarray to prevent use of deeply nested arrays an application is vulnerable only with certain customized choices for deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with mend
0
31,906
6,017,146,993
IssuesEvent
2017-06-07 08:57:57
mysociety/alaveteli
https://api.github.com/repos/mysociety/alaveteli
closed
Create a diff of the English alaveteli documentation since we deployed the Spanish version
documentation re-user experience
Give it to @JenMysoc to create/update the translations.
1.0
Create a diff of the English alaveteli documentation since we deployed the Spanish version - Give it to @JenMysoc to create/update the translations.
non_process
create a diff of the english alaveteli documentation since we deployed the spanish version give it to jenmysoc to create update the translations
0
68,887
8,358,420,429
IssuesEvent
2018-10-03 02:44:18
ByteClubGames/YumiAndTheYokai
https://api.github.com/repos/ByteClubGames/YumiAndTheYokai
closed
Wall Clinging
Bug HIGH PRIORITY Level Design
They player can currently stick to walls by holding the direction of the wall. -This could be an issue with either the tiles themselves -The could also have to do with the shape of the player model's collider -If the other approaches don't work, it could probably be fixed with a small script.
1.0
Wall Clinging - They player can currently stick to walls by holding the direction of the wall. -This could be an issue with either the tiles themselves -The could also have to do with the shape of the player model's collider -If the other approaches don't work, it could probably be fixed with a small script.
non_process
wall clinging they player can currently stick to walls by holding the direction of the wall this could be an issue with either the tiles themselves the could also have to do with the shape of the player model s collider if the other approaches don t work it could probably be fixed with a small script
0
250,396
18,888,278,219
IssuesEvent
2021-11-15 10:20:47
do-mpc/do-mpc
https://api.github.com/repos/do-mpc/do-mpc
opened
Minor error in the MHE Formulation
Documentation
Regarding the documentation on our [do-mpc](do-mpc.com) homepage. **Link to page / section of the issue** https://www.do-mpc.com/en/latest/mhe_example.html **What is wrong / missing?** Some minor errors in the MHE formuation. **How could it be improved / what should be added?** 1- The sum of the measurement noise term v_k should be from 0 to N 2- The model ode constraint should be from k = 0 to k = N-1 3- I am a bit skiptical about using the control input as a decision variable. Thanks for offering such a useful toolbox!
1.0
Minor error in the MHE Formulation - Regarding the documentation on our [do-mpc](do-mpc.com) homepage. **Link to page / section of the issue** https://www.do-mpc.com/en/latest/mhe_example.html **What is wrong / missing?** Some minor errors in the MHE formuation. **How could it be improved / what should be added?** 1- The sum of the measurement noise term v_k should be from 0 to N 2- The model ode constraint should be from k = 0 to k = N-1 3- I am a bit skiptical about using the control input as a decision variable. Thanks for offering such a useful toolbox!
non_process
minor error in the mhe formulation regarding the documentation on our do mpc com homepage link to page section of the issue what is wrong missing some minor errors in the mhe formuation how could it be improved what should be added the sum of the measurement noise term v k should be from to n the model ode constraint should be from k to k n i am a bit skiptical about using the control input as a decision variable thanks for offering such a useful toolbox
0
137,724
20,203,861,273
IssuesEvent
2022-02-11 17:57:13
CDCgov/prime-reportstream
https://api.github.com/repos/CDCgov/prime-reportstream
closed
Submission History Design
design Epic experience
As the Experience Team, we need some design for the Submission History & Lineage feature as outlined in this Epic [https://app.zenhub.com/workspaces/experience-607d9d5e68b95200150fec37/issues/cdcgov/prime-reportstream/1652](url).
1.0
Submission History Design - As the Experience Team, we need some design for the Submission History & Lineage feature as outlined in this Epic [https://app.zenhub.com/workspaces/experience-607d9d5e68b95200150fec37/issues/cdcgov/prime-reportstream/1652](url).
non_process
submission history design as the experience team we need some design for the submission history lineage feature as outlined in this epic url
0
201
2,610,024,982
IssuesEvent
2015-02-26 18:05:00
numenta/htm.java
https://api.github.com/repos/numenta/htm.java
closed
Sync HTM.java Releases w/NuPIC
administrative help wanted P3 process wiki
The release numbering (in the grade.build and pom.xml files) and the Roadmap need to be updated with compatible numbering and sync'd up with the new release and Roadmap present in NuPIC. @rhyolight
1.0
Sync HTM.java Releases w/NuPIC - The release numbering (in the grade.build and pom.xml files) and the Roadmap need to be updated with compatible numbering and sync'd up with the new release and Roadmap present in NuPIC. @rhyolight
process
sync htm java releases w nupic the release numbering in the grade build and pom xml files and the roadmap need to be updated with compatible numbering and sync d up with the new release and roadmap present in nupic rhyolight
1
158,306
12,412,255,011
IssuesEvent
2020-05-22 10:11:10
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
716]
bug cppcheck static analysis testing triage
```yaml { "id": 1, "title": "716]", "type": "bug", "description": "[src/opencv3/3rdparty/libjpeg/jdarith.c:718] -> [src/opencv3/3rdparty/libjpeg/jdarith.c:716]: (warning) Either the condition 'tbl>=16' is redundant or the array 'entropy->dc_stats[16]' is accessed at index 16, which is out of bounds.", "cwe": "None", "cve": "None", "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "system": "src/opencv3/3rdparty/libjpeg/jdarith.c", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "N/A", "architectural-location": "N/A", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-22 (10:10)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-22 (10:10)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "", "reproduction": "See artifacts below (if available)", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
716] - ```yaml { "id": 1, "title": "716]", "type": "bug", "description": "[src/opencv3/3rdparty/libjpeg/jdarith.c:718] -> [src/opencv3/3rdparty/libjpeg/jdarith.c:716]: (warning) Either the condition 'tbl>=16' is redundant or the array 'entropy->dc_stats[16]' is accessed at index 16, which is out of bounds.", "cwe": "None", "cve": "None", "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "system": "src/opencv3/3rdparty/libjpeg/jdarith.c", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "N/A", "architectural-location": "N/A", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-22 (10:10)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-22 (10:10)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "", "reproduction": "See artifacts below (if available)", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
non_process
yaml id title type bug description warning either the condition tbl is redundant or the array entropy dc stats is accessed at index which is out of bounds cwe none cve none keywords cppcheck static analysis testing triage bug system src libjpeg jdarith c vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity n a architectural location n a application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros kinetic jobs artifacts download exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
0
173,146
21,103,687,134
IssuesEvent
2022-04-04 16:36:07
opensearch-project/data-prepper
https://api.github.com/repos/opensearch-project/data-prepper
closed
CVE-2019-10782 (Medium) detected in checkstyle-8.27.jar - autoclosed
security vulnerability
## CVE-2019-10782 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>checkstyle-8.27.jar</b></p></summary> <p>Checkstyle is a development tool to help programmers write Java code that adheres to a coding standard</p> <p>Library home page: <a href="https://checkstyle.org/">https://checkstyle.org/</a></p> <p> Dependency Hierarchy: <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/data-prepper/commit/022b333dc9be3548b8eb8bb73d0337fd26425056">022b333dc9be3548b8eb8bb73d0337fd26425056</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of com.puppycrawl.tools:checkstyle before 8.29 are vulnerable to XML External Entity (XXE) Injection due to an incomplete fix for CVE-2019-9658. <p>Publish Date: 2020-01-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10782>CVE-2019-10782</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10782">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10782</a></p> <p>Release Date: 2020-02-10</p> <p>Fix Resolution: com.puppycrawl.tools:checkstyle:8.29</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.puppycrawl.tools","packageName":"checkstyle","packageVersion":"8.27","packageFilePaths":[],"isTransitiveDependency":false,"isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.puppycrawl.tools:checkstyle:8.29","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-10782","vulnerabilityDetails":"All versions of com.puppycrawl.tools:checkstyle before 8.29 are vulnerable to XML External Entity (XXE) Injection due to an incomplete fix for CVE-2019-9658.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10782","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-10782 (Medium) detected in checkstyle-8.27.jar - autoclosed - ## CVE-2019-10782 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>checkstyle-8.27.jar</b></p></summary> <p>Checkstyle is a development tool to help programmers write Java code that adheres to a coding standard</p> <p>Library home page: <a href="https://checkstyle.org/">https://checkstyle.org/</a></p> <p> Dependency Hierarchy: <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/data-prepper/commit/022b333dc9be3548b8eb8bb73d0337fd26425056">022b333dc9be3548b8eb8bb73d0337fd26425056</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of com.puppycrawl.tools:checkstyle before 8.29 are vulnerable to XML External Entity (XXE) Injection due to an incomplete fix for CVE-2019-9658. <p>Publish Date: 2020-01-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10782>CVE-2019-10782</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10782">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10782</a></p> <p>Release Date: 2020-02-10</p> <p>Fix Resolution: com.puppycrawl.tools:checkstyle:8.29</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.puppycrawl.tools","packageName":"checkstyle","packageVersion":"8.27","packageFilePaths":[],"isTransitiveDependency":false,"isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.puppycrawl.tools:checkstyle:8.29","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-10782","vulnerabilityDetails":"All versions of com.puppycrawl.tools:checkstyle before 8.29 are vulnerable to XML External Entity (XXE) Injection due to an incomplete fix for CVE-2019-9658.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10782","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in checkstyle jar autoclosed cve medium severity vulnerability vulnerable library checkstyle jar checkstyle is a development tool to help programmers write java code that adheres to a coding standard library home page a href dependency hierarchy found in head commit a href found in base branch main vulnerability details all versions of com puppycrawl tools checkstyle before are vulnerable to xml external entity xxe injection due to an incomplete fix for cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com puppycrawl tools checkstyle isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false isminimumfixversionavailable true minimumfixversion com puppycrawl tools checkstyle isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails all versions of com puppycrawl tools checkstyle before are vulnerable to xml external entity xxe injection due to an incomplete fix for cve vulnerabilityurl
0
18,909
3,098,087,211
IssuesEvent
2015-08-28 08:40:22
mbj4668/pyang
https://api.github.com/repos/mbj4668/pyang
closed
Tree plugin: missing i_config attribute
bug Priority-High Type-Defect
Tree plugin gives the following stack trace for ietf-routing. The statement on which it fails is `net-hop-address` leaf. ``` $ git clone git@github.com:netmod-wg/routing-cfg.git ... $ cd routing-cfg/ $ make yang $ make hello.xml $ pyang -f tree -o model.tree -L hello.xml Traceback (most recent call last): File "/Users/lhotka/bin/pyang", line 403, in <module> run() File "/Users/lhotka/bin/pyang", line 377, in run emit_obj.emit(ctx, modules, fd) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 54, in emit emit_tree(ctx, modules, fd, ctx.opts.tree_depth, path) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 117, in emit_tree print_children(chs, module, fd, ' ', path, 'data', depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 259, in print_node print_children(chs, module, fd, prefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 212, in print_node flags = get_flags_str(s, mode) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 279, in get_flags_str elif s.i_config == True: AttributeError: 'Statement' object has no attribute 'i_config' ```
1.0
Tree plugin: missing i_config attribute - Tree plugin gives the following stack trace for ietf-routing. The statement on which it fails is `net-hop-address` leaf. ``` $ git clone git@github.com:netmod-wg/routing-cfg.git ... $ cd routing-cfg/ $ make yang $ make hello.xml $ pyang -f tree -o model.tree -L hello.xml Traceback (most recent call last): File "/Users/lhotka/bin/pyang", line 403, in <module> run() File "/Users/lhotka/bin/pyang", line 377, in run emit_obj.emit(ctx, modules, fd) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 54, in emit emit_tree(ctx, modules, fd, ctx.opts.tree_depth, path) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 117, in emit_tree print_children(chs, module, fd, ' ', path, 'data', depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node print_children(chs, module, fd, prefix, path, mode, depth) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 259, in print_node print_children(chs, module, fd, prefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children print_node(ch, module, fd, newprefix, path, mode, depth, width) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 212, in print_node flags = get_flags_str(s, mode) File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 279, in get_flags_str elif s.i_config == True: AttributeError: 'Statement' object has no attribute 'i_config' ```
non_process
tree plugin missing i config attribute tree plugin gives the following stack trace for ietf routing the statement on which it fails is net hop address leaf git clone git github com netmod wg routing cfg git cd routing cfg make yang make hello xml pyang f tree o model tree l hello xml traceback most recent call last file users lhotka bin pyang line in run file users lhotka bin pyang line in run emit obj emit ctx modules fd file users lhotka lib python pyang plugins tree py line in emit emit tree ctx modules fd ctx opts tree depth path file users lhotka lib python pyang plugins tree py line in emit tree print children chs module fd path data depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth width file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node flags get flags str s mode file users lhotka lib python pyang plugins tree py line in get flags str elif s i config true attributeerror statement object has no attribute i config
0
759,543
26,600,514,829
IssuesEvent
2023-01-23 15:27:34
Assignaly/assignaly-api
https://api.github.com/repos/Assignaly/assignaly-api
closed
Create router middleware
api feature update core mvp priority:high
As a developer I want to be able to execute code before reaching the callback method given to a route. This could be something like user authentication or turning empty interfaces into strings. - [ ] Create middleware interface and necessary methods and variables
1.0
Create router middleware - As a developer I want to be able to execute code before reaching the callback method given to a route. This could be something like user authentication or turning empty interfaces into strings. - [ ] Create middleware interface and necessary methods and variables
non_process
create router middleware as a developer i want to be able to execute code before reaching the callback method given to a route this could be something like user authentication or turning empty interfaces into strings create middleware interface and necessary methods and variables
0
98,934
12,379,149,314
IssuesEvent
2020-05-19 11:59:49
raiden-network/webui
https://api.github.com/repos/raiden-network/webui
closed
Add complete transfer history in subpage
Alderaan Redesign
### Description Add a functionality / screen to see the complete transfer history. In #349 we decided to have more pages. This will be in a separate page. ### Tasks - [x] Add complete transfer history - [x] Add pagination for transfer history
1.0
Add complete transfer history in subpage - ### Description Add a functionality / screen to see the complete transfer history. In #349 we decided to have more pages. This will be in a separate page. ### Tasks - [x] Add complete transfer history - [x] Add pagination for transfer history
non_process
add complete transfer history in subpage description add a functionality screen to see the complete transfer history in we decided to have more pages this will be in a separate page tasks add complete transfer history add pagination for transfer history
0
150,686
19,610,137,307
IssuesEvent
2022-01-06 14:30:23
ghc-dev/Michelle-Jones
https://api.github.com/repos/ghc-dev/Michelle-Jones
opened
CVE-2020-9488 (Low) detected in log4j-core-2.8.2.jar
security vulnerability
## CVE-2020-9488 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-core-2.8.2.jar</b></p></summary> <p>The Apache Log4j Implementation</p> <p>Library home page: <a href="https://logging.apache.org/log4j/2.x/log4j-core/">https://logging.apache.org/log4j/2.x/log4j-core/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /sitory/org/apache/logging/log4j/log4j-core/2.8.2/log4j-core-2.8.2.jar</p> <p> Dependency Hierarchy: - :x: **log4j-core-2.8.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Michelle-Jones/commit/48bc7ec513d624a25e5109795f26c020f7b0241a">48bc7ec513d624a25e5109795f26c020f7b0241a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender. <p>Publish Date: 2020-04-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p> <p>Release Date: 2020-04-27</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.8.2","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.13.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-9488","vulnerabilityDetails":"Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-9488 (Low) detected in log4j-core-2.8.2.jar - ## CVE-2020-9488 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-core-2.8.2.jar</b></p></summary> <p>The Apache Log4j Implementation</p> <p>Library home page: <a href="https://logging.apache.org/log4j/2.x/log4j-core/">https://logging.apache.org/log4j/2.x/log4j-core/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /sitory/org/apache/logging/log4j/log4j-core/2.8.2/log4j-core-2.8.2.jar</p> <p> Dependency Hierarchy: - :x: **log4j-core-2.8.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Michelle-Jones/commit/48bc7ec513d624a25e5109795f26c020f7b0241a">48bc7ec513d624a25e5109795f26c020f7b0241a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender. <p>Publish Date: 2020-04-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p> <p>Release Date: 2020-04-27</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.8.2","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.13.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-9488","vulnerabilityDetails":"Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve low detected in core jar cve low severity vulnerability vulnerable library core jar the apache implementation library home page a href path to dependency file pom xml path to vulnerable library sitory org apache logging core core jar dependency hierarchy x core jar vulnerable library found in head commit a href found in base branch master vulnerability details improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache logging core isminimumfixversionavailable true minimumfixversion org apache logging core isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender vulnerabilityurl
0
11,649
14,503,020,924
IssuesEvent
2020-12-11 21:58:07
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
[automl] no information about `AUTOML_PROJECT_ID` and `OBJECT_DETECTION_MODEL_ID`
api: automl priority: p2 type: process
I tried to run `automl/cloud-client/vision_object_detection_predict_test.py` against my project then it errored out because I don't set `AUTOML_PROJECT_ID` and `OBJECT_DETECTION_MODEL_ID`. It's great if we have those environment variables in [`testing/test-env.tmpl.sh`](https://github.com/GoogleCloudPlatform/python-docs-samples/blob/master/testing/test-env.tmpl.sh) with some explanations for why you need a different cloud project and how to prepare the model.
1.0
[automl] no information about `AUTOML_PROJECT_ID` and `OBJECT_DETECTION_MODEL_ID` - I tried to run `automl/cloud-client/vision_object_detection_predict_test.py` against my project then it errored out because I don't set `AUTOML_PROJECT_ID` and `OBJECT_DETECTION_MODEL_ID`. It's great if we have those environment variables in [`testing/test-env.tmpl.sh`](https://github.com/GoogleCloudPlatform/python-docs-samples/blob/master/testing/test-env.tmpl.sh) with some explanations for why you need a different cloud project and how to prepare the model.
process
no information about automl project id and object detection model id i tried to run automl cloud client vision object detection predict test py against my project then it errored out because i don t set automl project id and object detection model id it s great if we have those environment variables in with some explanations for why you need a different cloud project and how to prepare the model
1
289,034
8,854,304,687
IssuesEvent
2019-01-09 00:43:47
jstanden/cerb
https://api.github.com/repos/jstanden/cerb
closed
Bulk delete
bug priority-support
In the latest 9.1 release, bulk delete _tickets_ or attachments does not appear to work. It lets you submit the task but the worklist just refreshes without any action taken. If you delete the _tickets_ individually on the worklist, this works.
1.0
Bulk delete - In the latest 9.1 release, bulk delete _tickets_ or attachments does not appear to work. It lets you submit the task but the worklist just refreshes without any action taken. If you delete the _tickets_ individually on the worklist, this works.
non_process
bulk delete in the latest release bulk delete tickets or attachments does not appear to work it lets you submit the task but the worklist just refreshes without any action taken if you delete the tickets individually on the worklist this works
0
453,645
13,086,232,602
IssuesEvent
2020-08-02 05:14:13
Journaly/journaly
https://api.github.com/repos/Journaly/journaly
closed
Visual Bugs & Improvements
bug high priority visual
#### Visual Bugs To Fix - [x] My Feed search box isn't the right width [#181] - [x] Chevron/down arrow is in wrong position on My Feed filters [#181] - [x] Chevron/down arrow is in wrong position on Settings Add Languages - [x] Chevron/down arrow is in wrong position on New Post drop downs - [x] Settings > languages drop downs are off on mobile view
1.0
Visual Bugs & Improvements - #### Visual Bugs To Fix - [x] My Feed search box isn't the right width [#181] - [x] Chevron/down arrow is in wrong position on My Feed filters [#181] - [x] Chevron/down arrow is in wrong position on Settings Add Languages - [x] Chevron/down arrow is in wrong position on New Post drop downs - [x] Settings > languages drop downs are off on mobile view
non_process
visual bugs improvements visual bugs to fix my feed search box isn t the right width chevron down arrow is in wrong position on my feed filters chevron down arrow is in wrong position on settings add languages chevron down arrow is in wrong position on new post drop downs settings languages drop downs are off on mobile view
0
17,972
23,984,015,533
IssuesEvent
2022-09-13 17:22:19
Open-Data-Product-Initiative/open-data-product-spec
https://api.github.com/repos/Open-Data-Product-Initiative/open-data-product-spec
closed
New elements to dataPipeline: platform, technology, storageType
enhancement processed
`"platform": "GCP"` Options: Azure, AWS, GCP, or probably just Terraform. `"technology": "Amazon S3" ` Describes the internal storage area technology, such as: Amazon S3, Google Cloud Storage, Azure Blob Storage. `"storageType": "files"` Describes the internal storage type: files, sql, events, MQTT, etc. All types are string. The purpose is to model the principles of data mesh
1.0
New elements to dataPipeline: platform, technology, storageType - `"platform": "GCP"` Options: Azure, AWS, GCP, or probably just Terraform. `"technology": "Amazon S3" ` Describes the internal storage area technology, such as: Amazon S3, Google Cloud Storage, Azure Blob Storage. `"storageType": "files"` Describes the internal storage type: files, sql, events, MQTT, etc. All types are string. The purpose is to model the principles of data mesh
process
new elements to datapipeline platform technology storagetype platform gcp options azure aws gcp or probably just terraform technology amazon describes the internal storage area technology such as amazon google cloud storage azure blob storage storagetype files describes the internal storage type files sql events mqtt etc all types are string the purpose is to model the principles of data mesh
1
6,127
8,996,635,507
IssuesEvent
2019-02-02 03:04:04
bow-simulation/virtualbow
https://api.github.com/repos/bow-simulation/virtualbow
closed
Use Gitlab-CI to build and run tests automatically
area: software process prio: normal type: improvement
In GitLab by **spfeifer** on Dec 7, 2018, 11:52 This is a first step towards #33 Optional Bonus: Provide development build artifacts (AppImage, deb, rpm)
1.0
Use Gitlab-CI to build and run tests automatically - In GitLab by **spfeifer** on Dec 7, 2018, 11:52 This is a first step towards #33 Optional Bonus: Provide development build artifacts (AppImage, deb, rpm)
process
use gitlab ci to build and run tests automatically in gitlab by spfeifer on dec this is a first step towards optional bonus provide development build artifacts appimage deb rpm
1