Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7,214
| 10,346,996,816
|
IssuesEvent
|
2019-09-04 16:22:05
|
qri-io/desktop
|
https://api.github.com/repos/qri-io/desktop
|
closed
|
ensure auto-update is running
|
chore main process
|
qri-io/frontend successfully ran auto-update in the background. We should make sure the same functionality is in place for Desktop
|
1.0
|
ensure auto-update is running - qri-io/frontend successfully ran auto-update in the background. We should make sure the same functionality is in place for Desktop
|
process
|
ensure auto update is running qri io frontend successfully ran auto update in the background we should make sure the same functionality is in place for desktop
| 1
|
21,228
| 6,132,434,420
|
IssuesEvent
|
2017-06-25 02:08:23
|
ganeti/ganeti
|
https://api.github.com/repos/ganeti/ganeti
|
opened
|
Misnamed option in NEWS file
|
imported_from_google_code Status:Fixed
|
Originally reported of Google Code with ID 1177.
```
In the NEWS file:
"- RPC security got enhanced by using different client SSL certificates
for each node. In this context 'gnt-cluster renew-crypto' got a new
option '--renew-node-certificates'"
But the option is actually "--new-node-certificates"
```
Originally added on 2016-05-06 14:04:58 +0000 UTC.
|
1.0
|
Misnamed option in NEWS file - Originally reported of Google Code with ID 1177.
```
In the NEWS file:
"- RPC security got enhanced by using different client SSL certificates
for each node. In this context 'gnt-cluster renew-crypto' got a new
option '--renew-node-certificates'"
But the option is actually "--new-node-certificates"
```
Originally added on 2016-05-06 14:04:58 +0000 UTC.
|
non_process
|
misnamed option in news file originally reported of google code with id in the news file rpc security got enhanced by using different client ssl certificates for each node in this context gnt cluster renew crypto got a new option renew node certificates but the option is actually new node certificates originally added on utc
| 0
|
4,248
| 7,187,159,249
|
IssuesEvent
|
2018-02-02 03:18:15
|
Great-Hill-Corporation/quickBlocks
|
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
|
closed
|
cacheMan does not correct the file when finding a reversal
|
monitors-cacheMan status-inprocess type-bug
|
The cacheMan -f mode does not actually correct the cache. It should, if it finds a reversal, simply truncate the remainder of the cache because once there's a reveral all bets are off.
If there's a reversal, there may be very large jumps later in the data (there's a reversal because the data is messed up).
Suggestion: the first reversal is teh location of a total truncate at that point.
|
1.0
|
cacheMan does not correct the file when finding a reversal - The cacheMan -f mode does not actually correct the cache. It should, if it finds a reversal, simply truncate the remainder of the cache because once there's a reveral all bets are off.
If there's a reversal, there may be very large jumps later in the data (there's a reversal because the data is messed up).
Suggestion: the first reversal is teh location of a total truncate at that point.
|
process
|
cacheman does not correct the file when finding a reversal the cacheman f mode does not actually correct the cache it should if it finds a reversal simply truncate the remainder of the cache because once there s a reveral all bets are off if there s a reversal there may be very large jumps later in the data there s a reversal because the data is messed up suggestion the first reversal is teh location of a total truncate at that point
| 1
|
16,095
| 20,264,412,972
|
IssuesEvent
|
2022-02-15 10:39:04
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
jni_md.h not found on linux_mips64 and linux_riscv64
|
type: support / not a bug (process) untriaged team-OSS
|
### Description of the problem / feature request:
The header `jni_md.h` cannot be found on some architectures when compiling due to these 2 lines.
https://github.com/bazelbuild/bazel/blob/eeec121668e6307c21e1a9698a96237988269dba/tools/jdk/BUILD.tools#L109-L110
|
1.0
|
jni_md.h not found on linux_mips64 and linux_riscv64 - ### Description of the problem / feature request:
The header `jni_md.h` cannot be found on some architectures when compiling due to these 2 lines.
https://github.com/bazelbuild/bazel/blob/eeec121668e6307c21e1a9698a96237988269dba/tools/jdk/BUILD.tools#L109-L110
|
process
|
jni md h not found on linux and linux description of the problem feature request the header jni md h cannot be found on some architectures when compiling due to these lines
| 1
|
10,208
| 13,067,104,008
|
IssuesEvent
|
2020-07-30 23:24:49
|
googleapis/proto-plus-python
|
https://api.github.com/repos/googleapis/proto-plus-python
|
closed
|
Automated publish to PyPI is broken
|
type: process
|
https://app.circleci.com/pipelines/github/googleapis/proto-plus-python/132/workflows/41232c5b-ce36-4a48-9fc5-5eb3943dceb7/jobs/955/steps
```
#!/bin/bash -eo pipefail
openssl aes-256-cbc -d \
-in .circleci/.pypirc.enc \
-out ~/.pypirc \
-k "${PYPIRC_ENCRYPTION_KEY}"
*** WARNING : deprecated key derivation used.
Using -iter or -pbkdf2 would be better.
bad decrypt
140293096486016:error:06065064:digital envelope routines:EVP_DecryptFinal_ex:bad decrypt:../crypto/evp/evp_enc.c:570:
Exited with code exit status 1
```
|
1.0
|
Automated publish to PyPI is broken - https://app.circleci.com/pipelines/github/googleapis/proto-plus-python/132/workflows/41232c5b-ce36-4a48-9fc5-5eb3943dceb7/jobs/955/steps
```
#!/bin/bash -eo pipefail
openssl aes-256-cbc -d \
-in .circleci/.pypirc.enc \
-out ~/.pypirc \
-k "${PYPIRC_ENCRYPTION_KEY}"
*** WARNING : deprecated key derivation used.
Using -iter or -pbkdf2 would be better.
bad decrypt
140293096486016:error:06065064:digital envelope routines:EVP_DecryptFinal_ex:bad decrypt:../crypto/evp/evp_enc.c:570:
Exited with code exit status 1
```
|
process
|
automated publish to pypi is broken bin bash eo pipefail openssl aes cbc d in circleci pypirc enc out pypirc k pypirc encryption key warning deprecated key derivation used using iter or would be better bad decrypt error digital envelope routines evp decryptfinal ex bad decrypt crypto evp evp enc c exited with code exit status
| 1
|
342,756
| 24,755,812,731
|
IssuesEvent
|
2022-10-21 17:36:43
|
Py-Contributors/AudioBook
|
https://api.github.com/repos/Py-Contributors/AudioBook
|
opened
|
extends the current documentation for readthedocs
|
documentation enhancement help wanted good first issue hacktoberfest-accepted
|
**Is your feature request related to a problem? Please describe.**
Current Documentation is not sufficient. we have to extend it.
**Plan for Docs**
will use Sphinx[¶](https://docs.readthedocs.io/en/stable/intro/getting-started-with-sphinx.html#getting-started-with-sphinx) for documentation and write proper multipage documentation.
This issue is open to assigning.
Reference:-
https://pypdf2.readthedocs.io/en/latest/
https://flask.palletsprojects.com/en/2.2.x/
|
1.0
|
extends the current documentation for readthedocs - **Is your feature request related to a problem? Please describe.**
Current Documentation is not sufficient. we have to extend it.
**Plan for Docs**
will use Sphinx[¶](https://docs.readthedocs.io/en/stable/intro/getting-started-with-sphinx.html#getting-started-with-sphinx) for documentation and write proper multipage documentation.
This issue is open to assigning.
Reference:-
https://pypdf2.readthedocs.io/en/latest/
https://flask.palletsprojects.com/en/2.2.x/
|
non_process
|
extends the current documentation for readthedocs is your feature request related to a problem please describe current documentation is not sufficient we have to extend it plan for docs will use sphinx for documentation and write proper multipage documentation this issue is open to assigning reference
| 0
|
21,582
| 14,656,962,668
|
IssuesEvent
|
2020-12-28 14:34:42
|
airyhq/airy
|
https://api.github.com/repos/airyhq/airy
|
closed
|
Import core go libraries from anywhere
|
fix infrastructure
|
We need to make our go libraries importable from anywhere.
|
1.0
|
Import core go libraries from anywhere - We need to make our go libraries importable from anywhere.
|
non_process
|
import core go libraries from anywhere we need to make our go libraries importable from anywhere
| 0
|
14,220
| 17,141,349,176
|
IssuesEvent
|
2021-07-13 09:57:40
|
2i2c-org/team-compass
|
https://api.github.com/repos/2i2c-org/team-compass
|
opened
|
Finalize the titles that we use describe roles in a given hub
|
:label: administration :label: team-process prio: med type: enhancement
|
# Summary
In a few issues now, we've run into ambiguity / multiple kinds of terminology for how we describe roles in a hub. We should just nail this down and agree upon one description so that we can remain consistent.
# Proposed titles
## On the community side
- **Hub Community**: The group of people that we are (or might be) serving with a hub. **We use this word instead of "client" or "customer"**.
- My rationale here is that we are "in between" a SaaS platform and a consultancy, with a strong community+pro-social focus. As such, neither "customer" or "client" feel right to me. This is a stab at something that aligns more with the mission focus of 2i2c.
- **Community Representative**: The point of contact with Hub Engineers, and the main interface with a hub's community
- **Community Admin Team**: The group of community members with "Administrative privileges" on a hub. They are expected to do most hub administration via the JupyterHub UI. This group must have at least one member (usually the Community Rep).
## On the 2i2c side
- **Hub Engineer**: A member of the @2i2c-org/tech-team that performs more complex dev/ops for a hub
# Actions
- [ ] Decide whether "Hub Community" is an acceptable drop-in for "customer" or "client"
- [ ] Agree on the other major roles and what we call them
- [ ] Agree that we're not missing any important roles
- [ ] Write these up in our Team Compass
- [ ] Update respective documents (e.g. the Managed Hub Service doc)
# Related issues
- Some discussion of this in https://github.com/2i2c-org/team-compass/issues/147
|
1.0
|
Finalize the titles that we use describe roles in a given hub - # Summary
In a few issues now, we've run into ambiguity / multiple kinds of terminology for how we describe roles in a hub. We should just nail this down and agree upon one description so that we can remain consistent.
# Proposed titles
## On the community side
- **Hub Community**: The group of people that we are (or might be) serving with a hub. **We use this word instead of "client" or "customer"**.
- My rationale here is that we are "in between" a SaaS platform and a consultancy, with a strong community+pro-social focus. As such, neither "customer" or "client" feel right to me. This is a stab at something that aligns more with the mission focus of 2i2c.
- **Community Representative**: The point of contact with Hub Engineers, and the main interface with a hub's community
- **Community Admin Team**: The group of community members with "Administrative privileges" on a hub. They are expected to do most hub administration via the JupyterHub UI. This group must have at least one member (usually the Community Rep).
## On the 2i2c side
- **Hub Engineer**: A member of the @2i2c-org/tech-team that performs more complex dev/ops for a hub
# Actions
- [ ] Decide whether "Hub Community" is an acceptable drop-in for "customer" or "client"
- [ ] Agree on the other major roles and what we call them
- [ ] Agree that we're not missing any important roles
- [ ] Write these up in our Team Compass
- [ ] Update respective documents (e.g. the Managed Hub Service doc)
# Related issues
- Some discussion of this in https://github.com/2i2c-org/team-compass/issues/147
|
process
|
finalize the titles that we use describe roles in a given hub summary in a few issues now we ve run into ambiguity multiple kinds of terminology for how we describe roles in a hub we should just nail this down and agree upon one description so that we can remain consistent proposed titles on the community side hub community the group of people that we are or might be serving with a hub we use this word instead of client or customer my rationale here is that we are in between a saas platform and a consultancy with a strong community pro social focus as such neither customer or client feel right to me this is a stab at something that aligns more with the mission focus of community representative the point of contact with hub engineers and the main interface with a hub s community community admin team the group of community members with administrative privileges on a hub they are expected to do most hub administration via the jupyterhub ui this group must have at least one member usually the community rep on the side hub engineer a member of the org tech team that performs more complex dev ops for a hub actions decide whether hub community is an acceptable drop in for customer or client agree on the other major roles and what we call them agree that we re not missing any important roles write these up in our team compass update respective documents e g the managed hub service doc related issues some discussion of this in
| 1
|
802,563
| 28,967,034,537
|
IssuesEvent
|
2023-05-10 08:37:48
|
ahmedkaludi/accelerated-mobile-pages
|
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
|
closed
|
PHP error after the recent update 1.0.84
|
bug [Priority: HIGH] Ready for Review
|
**_Error notice_**
` PHP Notice: Function WP_Scripts::localize was called <strong>incorrectly</strong>. The <code>$l10n</code> parameter must be an array. To pass arbitrary data to scripts, use the <code>wp_add_inline_script()</code> function instead. Please see <a href="https://wordpress.org/documentation/article/debugging-in-wordpress/">Debugging in WordPress</a> for more information. (This message was added in version 5.7.0.) in C:\xampp\htdocs\wordpress\wp-includes\functions.php on line 5865
`
Reference ticket: https://wordpress.org/support/topic/function-wp_scriptslocalize-was-called-incorrectly-the-l10n-parameter-must-b-2/
|
1.0
|
PHP error after the recent update 1.0.84 - **_Error notice_**
` PHP Notice: Function WP_Scripts::localize was called <strong>incorrectly</strong>. The <code>$l10n</code> parameter must be an array. To pass arbitrary data to scripts, use the <code>wp_add_inline_script()</code> function instead. Please see <a href="https://wordpress.org/documentation/article/debugging-in-wordpress/">Debugging in WordPress</a> for more information. (This message was added in version 5.7.0.) in C:\xampp\htdocs\wordpress\wp-includes\functions.php on line 5865
`
Reference ticket: https://wordpress.org/support/topic/function-wp_scriptslocalize-was-called-incorrectly-the-l10n-parameter-must-b-2/
|
non_process
|
php error after the recent update error notice php notice function wp scripts localize was called incorrectly the parameter must be an array to pass arbitrary data to scripts use the wp add inline script function instead please see for more information this message was added in version in c xampp htdocs wordpress wp includes functions php on line reference ticket
| 0
|
15,479
| 19,688,035,840
|
IssuesEvent
|
2022-01-12 01:34:50
|
dtcenter/MET
|
https://api.github.com/repos/dtcenter/MET
|
closed
|
ioda2nc fails if the same input file is given with -iodafile option
|
type: bug priority: high alert: NEED MORE DEFINITION component: CI/CD reporting: DTC NCAR Base requestor: METplus Team required: FOR OFFICIAL RELEASE MET: PreProcessing Tools (Point)
|
*Replace italics below with details for this issue.*
## Describe the Problem ##
ioda2nc fails if the same input file is given with -iodafile option
### Expected Behavior ###
It should work without errors.
### Environment ###
Describe your runtime environment:
*1. Machine: (Linux Workstation)*
*2. OS: (RedHat Linux)*
*3. Software version number(s): 10.1.0 and 11.0 beta*
### To Reproduce ###
Describe the steps to reproduce the behavior:
*1. Go to seneca*
*2. run the following command*
/usr/local/met/bin/ioda2nc -v 2 /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda \
/ioda.NC001007.2020031012.nc ioda.NC001007.2020031012.summary.nc \
-iodafile /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda/ioda.NC001007.2020031012.nc
*3. See error*
terminate called after throwing an instance of 'netCDF::exceptions::NcEdge'
what(): NetCDF: Start+count exceeds dimension bound
file: ncVar.cpp line:958
Aborted
*4. run the following command*
export METPLUS_ELEVATION_RANGE_DICT="elevation_range = {beg = -1000;end = 100000;}"
export METPLUS_LEVEL_RANGE_DICT=""
export METPLUS_MASK_DICT=""
export METPLUS_MESSAGE_TYPE=""
export METPLUS_MESSAGE_TYPE_GROUP_MAP=""
export METPLUS_MESSAGE_TYPE_MAP=""
export METPLUS_METADATA_MAP=""
export METPLUS_MET_CONFIG_OVERRIDES=""
export METPLUS_MISSING_THRESH=""
export METPLUS_OBS_NAME_MAP="obs_name_map = [{ key = \"wind_direction\"; val = \"WDIR\"; }, { key = \"wind_speed\"; val = \"WIND\"; }];"
export METPLUS_OBS_VAR=""
export METPLUS_OBS_WINDOW_DICT="obs_window = {beg = -5400;end = 5400;}"
export METPLUS_QUALITY_MARK_THRESH="quality_mark_thresh = 0;"
export METPLUS_STATION_ID=""
export METPLUS_TIME_SUMMARY_DICT="time_summary = {flag = TRUE;raw_data = TRUE;beg = \"000000\";end = \"235959\";step = 300;width = 600;grib_code = [];obs_var = [\"WIND\"];type = [\"min\", \"max\", \"range\", \"mean\", \"stdev\", \"median\", \"p80\"];vld_freq = 0;vld_thresh = 0.0;}"
export MET_TMP_DIR="/d1/personal/mccabe/out2/tmp"
/usr/local/met/bin/ioda2nc -v 2 /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda \
/ioda.NC001007.2020031012.nc ioda.NC001007.2020031012.summary.nc \
-config /d1/personal/mccabe/METplus/parm/met_config/IODA2NCConfig_wrapped \
-iodafile /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda/ioda.NC001007.2020031012.nc
*5. See error*
ERROR :
ERROR : get_obs_data_float() -> WDIR@ObsValue does not exist!
ERROR :
*Post relevant sample data following these instructions:*
*https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
2702691
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [x] Select **scientist(s)** or **no scientist** required: no scientist
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Organization** level **Project** for support of the current coordinated release
- [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next bugfix version
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Bugfix Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **main_\<Version>**.
Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>`
- [ ] Fix the bug and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **main_\<Version>**.
Pull request: `bugfix <Issue Number> main_<Version> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Organization** level software support **Project** for the current coordinated release
Select: **Milestone** as the next bugfix version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Complete the steps above to fix the bug on the **develop** branch.
Branch name: `bugfix_<Issue Number>_develop_<Description>`
Pull request: `bugfix <Issue Number> develop <Description>`
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Close this issue.
|
1.0
|
ioda2nc fails if the same input file is given with -iodafile option - *Replace italics below with details for this issue.*
## Describe the Problem ##
ioda2nc fails if the same input file is given with -iodafile option
### Expected Behavior ###
It should work without errors.
### Environment ###
Describe your runtime environment:
*1. Machine: (Linux Workstation)*
*2. OS: (RedHat Linux)*
*3. Software version number(s): 10.1.0 and 11.0 beta*
### To Reproduce ###
Describe the steps to reproduce the behavior:
*1. Go to seneca*
*2. run the following command*
/usr/local/met/bin/ioda2nc -v 2 /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda \
/ioda.NC001007.2020031012.nc ioda.NC001007.2020031012.summary.nc \
-iodafile /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda/ioda.NC001007.2020031012.nc
*3. See error*
terminate called after throwing an instance of 'netCDF::exceptions::NcEdge'
what(): NetCDF: Start+count exceeds dimension bound
file: ncVar.cpp line:958
Aborted
*4. run the following command*
export METPLUS_ELEVATION_RANGE_DICT="elevation_range = {beg = -1000;end = 100000;}"
export METPLUS_LEVEL_RANGE_DICT=""
export METPLUS_MASK_DICT=""
export METPLUS_MESSAGE_TYPE=""
export METPLUS_MESSAGE_TYPE_GROUP_MAP=""
export METPLUS_MESSAGE_TYPE_MAP=""
export METPLUS_METADATA_MAP=""
export METPLUS_MET_CONFIG_OVERRIDES=""
export METPLUS_MISSING_THRESH=""
export METPLUS_OBS_NAME_MAP="obs_name_map = [{ key = \"wind_direction\"; val = \"WDIR\"; }, { key = \"wind_speed\"; val = \"WIND\"; }];"
export METPLUS_OBS_VAR=""
export METPLUS_OBS_WINDOW_DICT="obs_window = {beg = -5400;end = 5400;}"
export METPLUS_QUALITY_MARK_THRESH="quality_mark_thresh = 0;"
export METPLUS_STATION_ID=""
export METPLUS_TIME_SUMMARY_DICT="time_summary = {flag = TRUE;raw_data = TRUE;beg = \"000000\";end = \"235959\";step = 300;width = 600;grib_code = [];obs_var = [\"WIND\"];type = [\"min\", \"max\", \"range\", \"mean\", \"stdev\", \"median\", \"p80\"];vld_freq = 0;vld_thresh = 0.0;}"
export MET_TMP_DIR="/d1/personal/mccabe/out2/tmp"
/usr/local/met/bin/ioda2nc -v 2 /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda \
/ioda.NC001007.2020031012.nc ioda.NC001007.2020031012.summary.nc \
-config /d1/personal/mccabe/METplus/parm/met_config/IODA2NCConfig_wrapped \
-iodafile /d1/projects/METplus/METplus_Data/development/feature_1203_ioda2nc/met_test/new/ioda/ioda.NC001007.2020031012.nc
*5. See error*
ERROR :
ERROR : get_obs_data_float() -> WDIR@ObsValue does not exist!
ERROR :
*Post relevant sample data following these instructions:*
*https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
2702691
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [x] Select **scientist(s)** or **no scientist** required: no scientist
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Organization** level **Project** for support of the current coordinated release
- [ ] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next bugfix version
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Bugfix Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **main_\<Version>**.
Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>`
- [ ] Fix the bug and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **main_\<Version>**.
Pull request: `bugfix <Issue Number> main_<Version> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Organization** level software support **Project** for the current coordinated release
Select: **Milestone** as the next bugfix version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Complete the steps above to fix the bug on the **develop** branch.
Branch name: `bugfix_<Issue Number>_develop_<Description>`
Pull request: `bugfix <Issue Number> develop <Description>`
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Close this issue.
|
process
|
fails if the same input file is given with iodafile option replace italics below with details for this issue describe the problem fails if the same input file is given with iodafile option expected behavior it should work without errors environment describe your runtime environment machine linux workstation os redhat linux software version number s and beta to reproduce describe the steps to reproduce the behavior go to seneca run the following command usr local met bin v projects metplus metplus data development feature met test new ioda ioda nc ioda summary nc iodafile projects metplus metplus data development feature met test new ioda ioda nc see error terminate called after throwing an instance of netcdf exceptions ncedge what netcdf start count exceeds dimension bound file ncvar cpp line aborted run the following command export metplus elevation range dict elevation range beg end export metplus level range dict export metplus mask dict export metplus message type export metplus message type group map export metplus message type map export metplus metadata map export metplus met config overrides export metplus missing thresh export metplus obs name map obs name map export metplus obs var export metplus obs window dict obs window beg end export metplus quality mark thresh quality mark thresh export metplus station id export metplus time summary dict time summary flag true raw data true beg end step width grib code obs var type vld freq vld thresh export met tmp dir personal mccabe tmp usr local met bin v projects metplus metplus data development feature met test new ioda ioda nc ioda summary nc config personal mccabe metplus parm met config wrapped iodafile projects metplus metplus data development feature met test new ioda ioda nc see error error error get obs data float wdir obsvalue does not exist error post relevant sample data following these instructions relevant deadlines list relevant project deadlines here or state none funding source define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required no scientist labels select component s select priority select requestor s projects and milestone select organization level project for support of the current coordinated release select repository level project for development toward the next official release or add alert need project assignment label select milestone as the next bugfix version define related issue s consider the impact to the other metplus components bugfix checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of main branch name bugfix main fix the bug and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into main pull request bugfix main define the pull request metadata as permissions allow select reviewer s and linked issues select organization level software support project for the current coordinated release select milestone as the next bugfix version iterate until the reviewer s accept and merge your changes delete your fork or branch complete the steps above to fix the bug on the develop branch branch name bugfix develop pull request bugfix develop select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version close this issue
| 1
|
21,910
| 30,440,085,688
|
IssuesEvent
|
2023-07-15 01:22:53
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] Medium-Priority Bug Fixes
|
.Epic .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
```[tasklist]
### Tasks
- [x] #29734
- [x] #29702
- [x] #29746
- [x] #29764
- [x] #29902
- [x] #29745
- [x] #29770
- [x] #29988
- [x] #29747
- [x] #29898
- [x] #29964
- [x] #29944
- [x] #29947
- [ ] #30568
- [ ] #29895
- [ ] #29908
- [ ] #30280
- [ ] #29748
- [ ] #29897
- [ ] #29904
- [ ] #29909
- [ ] #29910
- [ ] #29935
- [ ] #29936
- [ ] #29938
- [ ] #29941
- [ ] #29942
- [ ] #29946
- [ ] #29948
- [ ] #29950
- [ ] #29953
- [ ] #29958
- [ ] #30385
- [ ] https://github.com/metabase/metabase/issues/30397
- [ ] #30401
- [ ] https://github.com/metabase/metabase/issues/30857
- [ ] https://github.com/metabase/metabase/issues/30858
- [ ] https://github.com/metabase/metabase/issues/30948
- [ ] https://github.com/metabase/metabase/issues/30949
- [ ] https://github.com/metabase/metabase/issues/30950
- [ ] https://github.com/metabase/metabase/issues/30957
- [ ] #29949
- [ ] https://github.com/metabase/metabase/issues/31053
- [ ] https://github.com/metabase/metabase/issues/31223
- [ ] https://github.com/metabase/metabase/issues/31268
- [ ] https://github.com/metabase/metabase/issues/31365
- [ ] https://github.com/metabase/metabase/issues/31366
- [ ] https://github.com/metabase/metabase/issues/31368
- [ ] #31521
- [ ] https://github.com/metabase/metabase/issues/31624
- [ ] https://github.com/metabase/metabase/issues/31741
- [ ] https://github.com/metabase/metabase/issues/31769
- [ ] https://github.com/metabase/metabase/issues/31775
- [ ] https://github.com/metabase/metabase/issues/31858
- [ ] https://github.com/metabase/metabase/issues/32049
- [ ] https://github.com/metabase/metabase/issues/32063
```
|
1.0
|
[MLv2] Medium-Priority Bug Fixes - ```[tasklist]
### Tasks
- [x] #29734
- [x] #29702
- [x] #29746
- [x] #29764
- [x] #29902
- [x] #29745
- [x] #29770
- [x] #29988
- [x] #29747
- [x] #29898
- [x] #29964
- [x] #29944
- [x] #29947
- [ ] #30568
- [ ] #29895
- [ ] #29908
- [ ] #30280
- [ ] #29748
- [ ] #29897
- [ ] #29904
- [ ] #29909
- [ ] #29910
- [ ] #29935
- [ ] #29936
- [ ] #29938
- [ ] #29941
- [ ] #29942
- [ ] #29946
- [ ] #29948
- [ ] #29950
- [ ] #29953
- [ ] #29958
- [ ] #30385
- [ ] https://github.com/metabase/metabase/issues/30397
- [ ] #30401
- [ ] https://github.com/metabase/metabase/issues/30857
- [ ] https://github.com/metabase/metabase/issues/30858
- [ ] https://github.com/metabase/metabase/issues/30948
- [ ] https://github.com/metabase/metabase/issues/30949
- [ ] https://github.com/metabase/metabase/issues/30950
- [ ] https://github.com/metabase/metabase/issues/30957
- [ ] #29949
- [ ] https://github.com/metabase/metabase/issues/31053
- [ ] https://github.com/metabase/metabase/issues/31223
- [ ] https://github.com/metabase/metabase/issues/31268
- [ ] https://github.com/metabase/metabase/issues/31365
- [ ] https://github.com/metabase/metabase/issues/31366
- [ ] https://github.com/metabase/metabase/issues/31368
- [ ] #31521
- [ ] https://github.com/metabase/metabase/issues/31624
- [ ] https://github.com/metabase/metabase/issues/31741
- [ ] https://github.com/metabase/metabase/issues/31769
- [ ] https://github.com/metabase/metabase/issues/31775
- [ ] https://github.com/metabase/metabase/issues/31858
- [ ] https://github.com/metabase/metabase/issues/32049
- [ ] https://github.com/metabase/metabase/issues/32063
```
|
process
|
medium priority bug fixes tasks
| 1
|
112,816
| 17,102,345,872
|
IssuesEvent
|
2021-07-09 13:06:09
|
turkdevops/codecov-action
|
https://api.github.com/repos/turkdevops/codecov-action
|
closed
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: codecov-action/package.json</p>
<p>Path to vulnerable library: codecov-action/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- jest-26.6.3.tgz (Root Library)
- core-26.6.3.tgz
- jest-resolve-26.6.2.tgz
- read-pkg-up-7.0.1.tgz
- read-pkg-5.2.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/codecov-action/commit/55e35bb57ce02b251f09ff66d3ac34c5a60291ea">55e35bb57ce02b251f09ff66d3ac34c5a60291ea</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - autoclosed - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: codecov-action/package.json</p>
<p>Path to vulnerable library: codecov-action/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- jest-26.6.3.tgz (Root Library)
- core-26.6.3.tgz
- jest-resolve-26.6.2.tgz
- read-pkg-up-7.0.1.tgz
- read-pkg-5.2.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/codecov-action/commit/55e35bb57ce02b251f09ff66d3ac34c5a60291ea">55e35bb57ce02b251f09ff66d3ac34c5a60291ea</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in hosted git info tgz autoclosed cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file codecov action package json path to vulnerable library codecov action node modules hosted git info package json dependency hierarchy jest tgz root library core tgz jest resolve tgz read pkg up tgz read pkg tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info step up your open source security game with whitesource
| 0
|
16,114
| 20,376,965,342
|
IssuesEvent
|
2022-02-21 16:33:08
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Query Engine tests with sharded MongoDB connection strings
|
process/candidate topic: internal team/migrations team/client
|
MongoDB connection strings are not URIs. It would be great to have Rust-level tests to see if sharded strings work correctly with the Query Engine. The URI format is this:
```
mongodb://user:password@srv1.rgyl0.mongodb.net:27017,srv2.rgyl0.mongodb.net:27017,srv3.rgyl0.mongodb.net:27017/database?ssl=true&authSource=admin&retryWrites=true&w=majority
```
So we should have a few replicas, and test with a connection string that lists them like in the example.
|
1.0
|
Query Engine tests with sharded MongoDB connection strings - MongoDB connection strings are not URIs. It would be great to have Rust-level tests to see if sharded strings work correctly with the Query Engine. The URI format is this:
```
mongodb://user:password@srv1.rgyl0.mongodb.net:27017,srv2.rgyl0.mongodb.net:27017,srv3.rgyl0.mongodb.net:27017/database?ssl=true&authSource=admin&retryWrites=true&w=majority
```
So we should have a few replicas, and test with a connection string that lists them like in the example.
|
process
|
query engine tests with sharded mongodb connection strings mongodb connection strings are not uris it would be great to have rust level tests to see if sharded strings work correctly with the query engine the uri format is this mongodb user password mongodb net mongodb net mongodb net database ssl true authsource admin retrywrites true w majority so we should have a few replicas and test with a connection string that lists them like in the example
| 1
|
785,119
| 27,599,080,620
|
IssuesEvent
|
2023-03-09 08:50:04
|
AY2223S2-CS2113-T15-4/tp
|
https://api.github.com/repos/AY2223S2-CS2113-T15-4/tp
|
opened
|
Delete flashcards
|
type.Epic priority.High
|
As a user I can delete any if my cards so that I don’t get asked to review that card later on in case I am confident I have truly memorised the card
|
1.0
|
Delete flashcards - As a user I can delete any if my cards so that I don’t get asked to review that card later on in case I am confident I have truly memorised the card
|
non_process
|
delete flashcards as a user i can delete any if my cards so that i don’t get asked to review that card later on in case i am confident i have truly memorised the card
| 0
|
3,783
| 6,760,953,854
|
IssuesEvent
|
2017-10-24 22:44:58
|
aspnet/IISIntegration
|
https://api.github.com/repos/aspnet/IISIntegration
|
closed
|
React to ANCM changes for NativeMethods pInvoke layer.
|
bug in-process
|
Parallel issue to https://github.com/aspnet/KestrelHttpServer/blob/dev/src/Kestrel.Core/Internal/Http/HttpProtocol.FeatureCollection.cs#L97-L101. With Pan's changes ( to the native ANCM and recognition that we need to expose more pInvoke methods, we will need to react to these appropriately.
|
1.0
|
React to ANCM changes for NativeMethods pInvoke layer. - Parallel issue to https://github.com/aspnet/KestrelHttpServer/blob/dev/src/Kestrel.Core/Internal/Http/HttpProtocol.FeatureCollection.cs#L97-L101. With Pan's changes ( to the native ANCM and recognition that we need to expose more pInvoke methods, we will need to react to these appropriately.
|
process
|
react to ancm changes for nativemethods pinvoke layer parallel issue to with pan s changes to the native ancm and recognition that we need to expose more pinvoke methods we will need to react to these appropriately
| 1
|
186,029
| 15,044,100,855
|
IssuesEvent
|
2021-02-03 02:11:31
|
walterimaican/nightlight
|
https://api.github.com/repos/walterimaican/nightlight
|
opened
|
Documentation: Stale
|
documentation
|
Some of the screenshots used in the README.md are now inconsistent with the current state of the project. These images should be updated when possible.
|
1.0
|
Documentation: Stale - Some of the screenshots used in the README.md are now inconsistent with the current state of the project. These images should be updated when possible.
|
non_process
|
documentation stale some of the screenshots used in the readme md are now inconsistent with the current state of the project these images should be updated when possible
| 0
|
4,208
| 7,166,652,311
|
IssuesEvent
|
2018-01-29 17:54:37
|
itsyouonline/identityserver
|
https://api.github.com/repos/itsyouonline/identityserver
|
reopened
|
Cloudflare returns error 504 on rare occasions
|
process_wontfix type_bug
|
Very rarely, cloudlfare will return a 504 when calling any url, even if others go through just fine. The 504 is not generated on our site as observed from the logs (tracing an `/error504` page load, no earlier requests gave a 504 for that users). This can cause pages to render empty, translations to not load etc...
|
1.0
|
Cloudflare returns error 504 on rare occasions - Very rarely, cloudlfare will return a 504 when calling any url, even if others go through just fine. The 504 is not generated on our site as observed from the logs (tracing an `/error504` page load, no earlier requests gave a 504 for that users). This can cause pages to render empty, translations to not load etc...
|
process
|
cloudflare returns error on rare occasions very rarely cloudlfare will return a when calling any url even if others go through just fine the is not generated on our site as observed from the logs tracing an page load no earlier requests gave a for that users this can cause pages to render empty translations to not load etc
| 1
|
656,733
| 21,773,565,908
|
IssuesEvent
|
2022-05-13 11:35:17
|
awslabs/aws-lambda-powertools-typescript
|
https://api.github.com/repos/awslabs/aws-lambda-powertools-typescript
|
opened
|
Feature (all): evaluate migration to Middy 3.x
|
enhancement dependencies utility:all triage priority:low
|
## Description of the feature request
**Problem statement**
Middy, the dependency that we use to vend middleware has [recently released](https://github.com/middyjs/middy/releases/tag/3.0.0) a new major version (v 3.x). This version drops support for Node JS version 12.x and introduces some breaking changes.
Given that Middy, as a whole, has a large surface and we are only using `@middy/core` we need to investigate on whether there's a path to upgrade to the newer version while still continuing to support Node JS 12.x in Powertools.
**Additional context**
Running the e2e tests on a branch that upgraded `@middy/core` to 3.0.1 seems to show that the newer release is compatible with Powertools on both versions 12 and 14 of Node ([see run results](https://github.com/awslabs/aws-lambda-powertools-typescript/actions/runs/2318810802)).
At the same time, the unit tests for that same branch are failing for all utilities and specifically in the section that relates the middleware implementations. The error message (see detail below) seems to hint at incompatibilities between the new bundling of middy that now supports both CJS and ESM bundling, and our project's configuration since the errors are related to the imports and the tests don't run at all.
<details>
```sh
FAIL AWS Lambda Powertools utility: TRACER tests/unit/middy.test.ts
● Test suite failed to run
Jest encountered an unexpected token
Jest failed to parse a file. This happens e.g. when your code or its dependencies use non-standard JavaScript syntax, or when Jest is not configured to support such syntax.
Out of the box Jest supports Babel, which will be used to transform your files into valid JS based on your Babel configuration.
By default "node_modules" folder is ignored by transformers.
Here's what you can do:
• If you are trying to use ECMAScript Modules, see https://jestjs.io/docs/ecmascript-modules for how to enable it.
• If you are trying to use TypeScript, see https://jestjs.io/docs/getting-started#using-typescript
• To have some of your "node_modules" files transformed, you can specify a custom "transformIgnorePatterns" in your config.
• If you need a custom transformation specify a "transform" option in your config.
• If you simply want to mock your non-JS modules (e.g. binary assets) you can stub them out with the "moduleNameMapper" config option.
You'll find more details and examples of these config options in the docs:
https://jestjs.io/docs/configuration
For information about custom transformations, see:
https://jestjs.io/docs/code-transformation
Details:
/home/ec2-user/aws-lambda-powertools-typescript/node_modules/@middy/core/index.js:1
({"Object.<anonymous>":function(module,exports,require,__dirname,__filename,jest){import{EventEmitter}from"events";const defaultLambdaHandler=()=>{};const defaultPlugin={timeoutEarlyInMillis:5,timeoutEarlyResponse:()=>{throw new Error("Timeout")}};const middy=(lambdaHandler=defaultLambdaHandler,plugin={})=>{if(typeof lambdaHandler!=="function"){plugin=lambdaHandler;lambdaHandler=defaultLambdaHandler}plugin={...defaultPlugin,...plugin};plugin.timeoutEarly=plugin.timeoutEarlyInMillis>0;plugin.beforePrefetch?.();const beforeMiddlewares=[];const afterMiddlewares=[];const onErrorMiddlewares=[];const middy1=(event={},context={})=>{plugin.requestStart?.();const request={event,context,response:undefined,error:undefined,internal:plugin.internal??{}};return runRequest(request,[...beforeMiddlewares],lambdaHandler,[...afterMiddlewares],[...onErrorMiddlewares],plugin)};middy1.use=middlewares=>{if(!Array.isArray(middlewares)){middlewares=[middlewares]}for(const middleware of middlewares){const{before,after,onError}=middleware;if(!before&&!after&&!onError){throw new Error('Middleware must be an object containing at least one key among "before", "after", "onError"')}if(before)middy1.before(before);if(after)middy1.after(after);if(onError)middy1.onError(onError)}return middy1};middy1.before=beforeMiddleware=>{beforeMiddlewares.push(beforeMiddleware);return middy1};middy1.after=afterMiddleware=>{afterMiddlewares.unshift(afterMiddleware);return middy1};middy1.onError=onErrorMiddleware=>{onErrorMiddlewares.unshift(onErrorMiddleware);return middy1};middy1.handler=replaceLambdaHandler=>{lambdaHandler=replaceLambdaHandler;return middy1};return middy1};const runRequest=async(request,beforeMiddlewares,lambdaHandler,afterMiddlewares,onErrorMiddlewares,plugin)=>{const timeoutEarly=plugin.timeoutEarly&&request.context.getRemainingTimeInMillis;try{await runMiddlewares(request,beforeMiddlewares,plugin);if(request.response===undefined){plugin.beforeHandler?.();const handlerAbort=new AbortController;let timeoutAbort;if(timeoutEarly)timeoutAbort=new AbortController;request.response=await Promise.race([lambdaHandler(request.event,request.context,{signal:handlerAbort.signal}),timeoutEarly?setTimeoutPromise(request.context.getRemainingTimeInMillis()-plugin.timeoutEarlyInMillis,{signal:timeoutAbort.signal}).then(()=>{handlerAbort.abort();return plugin.timeoutEarlyResponse()}):Promise.race([])]);if(timeoutEarly)timeoutAbort.abort();plugin.afterHandler?.();await runMiddlewares(request,afterMiddlewares,plugin)}}catch(e){request.response=undefined;request.error=e;try{await runMiddlewares(request,onErrorMiddlewares,plugin)}catch(e){e.originalError=request.error;request.error=e;throw request.error}if(request.response===undefined)throw request.error}finally{await plugin.requestEnd?.(request)}return request.response};const runMiddlewares=async(request,middlewares,plugin)=>{for(const nextMiddleware of middlewares){plugin.beforeMiddleware?.(nextMiddleware.name);const res=await nextMiddleware(request);plugin.afterMiddleware?.(nextMiddleware.name);if(res!==undefined){request.response=res;return}}};const polyfillAbortController=()=>{if(process.version<"v15.0.0"){let _toStringTag;let _toStringTag1=Symbol.toStringTag;class AbortSignal{toString(){return"[object AbortSignal]"}get[_toStringTag1](){return"AbortSignal"}removeEventListener(name,handler){this.eventEmitter.removeListener(name,handler)}addEventListener(name,handler){this.eventEmitter.on(name,handler)}dispatchEvent(type){const event={type,target:this};const handlerName=`on${type}`;if(typeof this[handlerName]==="function")this[handlerName](event);this.eventEmitter.emit(type,event)}constructor(){this.eventEmitter=new EventEmitter;this.onabort=null;this.aborted=false}}return _toStringTag=Symbol.toStringTag,class AbortController{abort(){if(this.signal.aborted)return;this.signal.aborted=true;this.signal.dispatchEvent("abort")}toString(){return"[object AbortController]"}get[_toStringTag](){return"AbortController"}constructor(){this.signal=new AbortSignal}}}else{return AbortController}};global.AbortController=polyfillAbortController();const polyfillSetTimeoutPromise=()=>{return(ms,{signal})=>{if(signal.aborted){return Promise.reject(new Error("Aborted","AbortError"))}return new Promise((resolve,reject)=>{const abortHandler=()=>{clearTimeout(timeout);reject(new Error("Aborted","AbortError"))};const timeout=setTimeout(()=>{resolve();signal.removeEventListener("abort",abortHandler)},ms);signal.addEventListener("abort",abortHandler)})}};const setTimeoutPromise=polyfillSetTimeoutPromise();export default middy
^^^^^^
SyntaxError: Cannot use import statement outside a module
6 |
7 | import { captureLambdaHandler } from '../../src/middleware/middy';
> 8 | import middy from '@middy/core';
| ^
9 | import { Tracer } from './../../src';
10 | import type { Context, Handler } from 'aws-lambda/handler';
11 | import { Segment, setContextMissingStrategy, Subsegment } from 'aws-xray-sdk-core';
at Runtime.createScriptFromCode (../../node_modules/jest-runtime/build/index.js:1728:14)
at Object.<anonymous> (tests/unit/middy.test.ts:8:1)
```
</details>
**Code examples**
N/A
**Benefits for you and the wider AWS community**
<!-- What are the benefits your your proposed feature? -->
**Describe alternatives you've considered**
Not upgrading to 3.x and staying on version 2.5.x
**Additional context**
<!-- Add any other context or screenshots about the feature request here. -->
### Related issues, RFCs
#861
|
1.0
|
Feature (all): evaluate migration to Middy 3.x - ## Description of the feature request
**Problem statement**
Middy, the dependency that we use to vend middleware has [recently released](https://github.com/middyjs/middy/releases/tag/3.0.0) a new major version (v 3.x). This version drops support for Node JS version 12.x and introduces some breaking changes.
Given that Middy, as a whole, has a large surface and we are only using `@middy/core` we need to investigate on whether there's a path to upgrade to the newer version while still continuing to support Node JS 12.x in Powertools.
**Additional context**
Running the e2e tests on a branch that upgraded `@middy/core` to 3.0.1 seems to show that the newer release is compatible with Powertools on both versions 12 and 14 of Node ([see run results](https://github.com/awslabs/aws-lambda-powertools-typescript/actions/runs/2318810802)).
At the same time, the unit tests for that same branch are failing for all utilities and specifically in the section that relates the middleware implementations. The error message (see detail below) seems to hint at incompatibilities between the new bundling of middy that now supports both CJS and ESM bundling, and our project's configuration since the errors are related to the imports and the tests don't run at all.
<details>
```sh
FAIL AWS Lambda Powertools utility: TRACER tests/unit/middy.test.ts
● Test suite failed to run
Jest encountered an unexpected token
Jest failed to parse a file. This happens e.g. when your code or its dependencies use non-standard JavaScript syntax, or when Jest is not configured to support such syntax.
Out of the box Jest supports Babel, which will be used to transform your files into valid JS based on your Babel configuration.
By default "node_modules" folder is ignored by transformers.
Here's what you can do:
• If you are trying to use ECMAScript Modules, see https://jestjs.io/docs/ecmascript-modules for how to enable it.
• If you are trying to use TypeScript, see https://jestjs.io/docs/getting-started#using-typescript
• To have some of your "node_modules" files transformed, you can specify a custom "transformIgnorePatterns" in your config.
• If you need a custom transformation specify a "transform" option in your config.
• If you simply want to mock your non-JS modules (e.g. binary assets) you can stub them out with the "moduleNameMapper" config option.
You'll find more details and examples of these config options in the docs:
https://jestjs.io/docs/configuration
For information about custom transformations, see:
https://jestjs.io/docs/code-transformation
Details:
/home/ec2-user/aws-lambda-powertools-typescript/node_modules/@middy/core/index.js:1
({"Object.<anonymous>":function(module,exports,require,__dirname,__filename,jest){import{EventEmitter}from"events";const defaultLambdaHandler=()=>{};const defaultPlugin={timeoutEarlyInMillis:5,timeoutEarlyResponse:()=>{throw new Error("Timeout")}};const middy=(lambdaHandler=defaultLambdaHandler,plugin={})=>{if(typeof lambdaHandler!=="function"){plugin=lambdaHandler;lambdaHandler=defaultLambdaHandler}plugin={...defaultPlugin,...plugin};plugin.timeoutEarly=plugin.timeoutEarlyInMillis>0;plugin.beforePrefetch?.();const beforeMiddlewares=[];const afterMiddlewares=[];const onErrorMiddlewares=[];const middy1=(event={},context={})=>{plugin.requestStart?.();const request={event,context,response:undefined,error:undefined,internal:plugin.internal??{}};return runRequest(request,[...beforeMiddlewares],lambdaHandler,[...afterMiddlewares],[...onErrorMiddlewares],plugin)};middy1.use=middlewares=>{if(!Array.isArray(middlewares)){middlewares=[middlewares]}for(const middleware of middlewares){const{before,after,onError}=middleware;if(!before&&!after&&!onError){throw new Error('Middleware must be an object containing at least one key among "before", "after", "onError"')}if(before)middy1.before(before);if(after)middy1.after(after);if(onError)middy1.onError(onError)}return middy1};middy1.before=beforeMiddleware=>{beforeMiddlewares.push(beforeMiddleware);return middy1};middy1.after=afterMiddleware=>{afterMiddlewares.unshift(afterMiddleware);return middy1};middy1.onError=onErrorMiddleware=>{onErrorMiddlewares.unshift(onErrorMiddleware);return middy1};middy1.handler=replaceLambdaHandler=>{lambdaHandler=replaceLambdaHandler;return middy1};return middy1};const runRequest=async(request,beforeMiddlewares,lambdaHandler,afterMiddlewares,onErrorMiddlewares,plugin)=>{const timeoutEarly=plugin.timeoutEarly&&request.context.getRemainingTimeInMillis;try{await runMiddlewares(request,beforeMiddlewares,plugin);if(request.response===undefined){plugin.beforeHandler?.();const handlerAbort=new AbortController;let timeoutAbort;if(timeoutEarly)timeoutAbort=new AbortController;request.response=await Promise.race([lambdaHandler(request.event,request.context,{signal:handlerAbort.signal}),timeoutEarly?setTimeoutPromise(request.context.getRemainingTimeInMillis()-plugin.timeoutEarlyInMillis,{signal:timeoutAbort.signal}).then(()=>{handlerAbort.abort();return plugin.timeoutEarlyResponse()}):Promise.race([])]);if(timeoutEarly)timeoutAbort.abort();plugin.afterHandler?.();await runMiddlewares(request,afterMiddlewares,plugin)}}catch(e){request.response=undefined;request.error=e;try{await runMiddlewares(request,onErrorMiddlewares,plugin)}catch(e){e.originalError=request.error;request.error=e;throw request.error}if(request.response===undefined)throw request.error}finally{await plugin.requestEnd?.(request)}return request.response};const runMiddlewares=async(request,middlewares,plugin)=>{for(const nextMiddleware of middlewares){plugin.beforeMiddleware?.(nextMiddleware.name);const res=await nextMiddleware(request);plugin.afterMiddleware?.(nextMiddleware.name);if(res!==undefined){request.response=res;return}}};const polyfillAbortController=()=>{if(process.version<"v15.0.0"){let _toStringTag;let _toStringTag1=Symbol.toStringTag;class AbortSignal{toString(){return"[object AbortSignal]"}get[_toStringTag1](){return"AbortSignal"}removeEventListener(name,handler){this.eventEmitter.removeListener(name,handler)}addEventListener(name,handler){this.eventEmitter.on(name,handler)}dispatchEvent(type){const event={type,target:this};const handlerName=`on${type}`;if(typeof this[handlerName]==="function")this[handlerName](event);this.eventEmitter.emit(type,event)}constructor(){this.eventEmitter=new EventEmitter;this.onabort=null;this.aborted=false}}return _toStringTag=Symbol.toStringTag,class AbortController{abort(){if(this.signal.aborted)return;this.signal.aborted=true;this.signal.dispatchEvent("abort")}toString(){return"[object AbortController]"}get[_toStringTag](){return"AbortController"}constructor(){this.signal=new AbortSignal}}}else{return AbortController}};global.AbortController=polyfillAbortController();const polyfillSetTimeoutPromise=()=>{return(ms,{signal})=>{if(signal.aborted){return Promise.reject(new Error("Aborted","AbortError"))}return new Promise((resolve,reject)=>{const abortHandler=()=>{clearTimeout(timeout);reject(new Error("Aborted","AbortError"))};const timeout=setTimeout(()=>{resolve();signal.removeEventListener("abort",abortHandler)},ms);signal.addEventListener("abort",abortHandler)})}};const setTimeoutPromise=polyfillSetTimeoutPromise();export default middy
^^^^^^
SyntaxError: Cannot use import statement outside a module
6 |
7 | import { captureLambdaHandler } from '../../src/middleware/middy';
> 8 | import middy from '@middy/core';
| ^
9 | import { Tracer } from './../../src';
10 | import type { Context, Handler } from 'aws-lambda/handler';
11 | import { Segment, setContextMissingStrategy, Subsegment } from 'aws-xray-sdk-core';
at Runtime.createScriptFromCode (../../node_modules/jest-runtime/build/index.js:1728:14)
at Object.<anonymous> (tests/unit/middy.test.ts:8:1)
```
</details>
**Code examples**
N/A
**Benefits for you and the wider AWS community**
<!-- What are the benefits your your proposed feature? -->
**Describe alternatives you've considered**
Not upgrading to 3.x and staying on version 2.5.x
**Additional context**
<!-- Add any other context or screenshots about the feature request here. -->
### Related issues, RFCs
#861
|
non_process
|
feature all evaluate migration to middy x description of the feature request problem statement middy the dependency that we use to vend middleware has a new major version v x this version drops support for node js version x and introduces some breaking changes given that middy as a whole has a large surface and we are only using middy core we need to investigate on whether there s a path to upgrade to the newer version while still continuing to support node js x in powertools additional context running the tests on a branch that upgraded middy core to seems to show that the newer release is compatible with powertools on both versions and of node at the same time the unit tests for that same branch are failing for all utilities and specifically in the section that relates the middleware implementations the error message see detail below seems to hint at incompatibilities between the new bundling of middy that now supports both cjs and esm bundling and our project s configuration since the errors are related to the imports and the tests don t run at all sh fail aws lambda powertools utility tracer tests unit middy test ts ● test suite failed to run jest encountered an unexpected token jest failed to parse a file this happens e g when your code or its dependencies use non standard javascript syntax or when jest is not configured to support such syntax out of the box jest supports babel which will be used to transform your files into valid js based on your babel configuration by default node modules folder is ignored by transformers here s what you can do • if you are trying to use ecmascript modules see for how to enable it • if you are trying to use typescript see • to have some of your node modules files transformed you can specify a custom transformignorepatterns in your config • if you need a custom transformation specify a transform option in your config • if you simply want to mock your non js modules e g binary assets you can stub them out with the modulenamemapper config option you ll find more details and examples of these config options in the docs for information about custom transformations see details home user aws lambda powertools typescript node modules middy core index js object function module exports require dirname filename jest import eventemitter from events const defaultlambdahandler const defaultplugin timeoutearlyinmillis timeoutearlyresponse throw new error timeout const middy lambdahandler defaultlambdahandler plugin if typeof lambdahandler function plugin lambdahandler lambdahandler defaultlambdahandler plugin defaultplugin plugin plugin timeoutearly plugin timeoutearlyinmillis plugin beforeprefetch const beforemiddlewares const aftermiddlewares const onerrormiddlewares const event context plugin requeststart const request event context response undefined error undefined internal plugin internal return runrequest request lambdahandler plugin use middlewares if array isarray middlewares middlewares for const middleware of middlewares const before after onerror middleware if before after onerror throw new error middleware must be an object containing at least one key among before after onerror if before before before if after after after if onerror onerror onerror return before beforemiddleware beforemiddlewares push beforemiddleware return after aftermiddleware aftermiddlewares unshift aftermiddleware return onerror onerrormiddleware onerrormiddlewares unshift onerrormiddleware return handler replacelambdahandler lambdahandler replacelambdahandler return return const runrequest async request beforemiddlewares lambdahandler aftermiddlewares onerrormiddlewares plugin const timeoutearly plugin timeoutearly request context getremainingtimeinmillis try await runmiddlewares request beforemiddlewares plugin if request response undefined plugin beforehandler const handlerabort new abortcontroller let timeoutabort if timeoutearly timeoutabort new abortcontroller request response await promise race if timeoutearly timeoutabort abort plugin afterhandler await runmiddlewares request aftermiddlewares plugin catch e request response undefined request error e try await runmiddlewares request onerrormiddlewares plugin catch e e originalerror request error request error e throw request error if request response undefined throw request error finally await plugin requestend request return request response const runmiddlewares async request middlewares plugin for const nextmiddleware of middlewares plugin beforemiddleware nextmiddleware name const res await nextmiddleware request plugin aftermiddleware nextmiddleware name if res undefined request response res return const polyfillabortcontroller if process version return ms signal if signal aborted return promise reject new error aborted aborterror return new promise resolve reject const aborthandler cleartimeout timeout reject new error aborted aborterror const timeout settimeout resolve signal removeeventlistener abort aborthandler ms signal addeventlistener abort aborthandler const settimeoutpromise polyfillsettimeoutpromise export default middy syntaxerror cannot use import statement outside a module import capturelambdahandler from src middleware middy import middy from middy core import tracer from src import type context handler from aws lambda handler import segment setcontextmissingstrategy subsegment from aws xray sdk core at runtime createscriptfromcode node modules jest runtime build index js at object tests unit middy test ts code examples n a benefits for you and the wider aws community describe alternatives you ve considered not upgrading to x and staying on version x additional context related issues rfcs
| 0
|
2,776
| 5,713,195,463
|
IssuesEvent
|
2017-04-19 07:00:09
|
g8os/grid
|
https://api.github.com/repos/g8os/grid
|
closed
|
AYS service for grid controller
|
process_wontfix type_feature
|
relates to : https://github.com/g8os/grid/issues/66
We need a service that deploy and configure the grid controller.
actions:
- install:
- reserve a pair disk of disk, create btrfs raid1 on them for replication.
- create container with the [grid flist](https://hub.gig.tech/maxux/grid.flist)
- start AYS Server
- start Grid API server
|
1.0
|
AYS service for grid controller - relates to : https://github.com/g8os/grid/issues/66
We need a service that deploy and configure the grid controller.
actions:
- install:
- reserve a pair disk of disk, create btrfs raid1 on them for replication.
- create container with the [grid flist](https://hub.gig.tech/maxux/grid.flist)
- start AYS Server
- start Grid API server
|
process
|
ays service for grid controller relates to we need a service that deploy and configure the grid controller actions install reserve a pair disk of disk create btrfs on them for replication create container with the start ays server start grid api server
| 1
|
59,041
| 14,365,891,805
|
IssuesEvent
|
2020-12-01 02:57:02
|
NixOS/nixpkgs
|
https://api.github.com/repos/NixOS/nixpkgs
|
closed
|
Vulnerability roundup 93: postgresql-9.6.17: 1 advisory [7.3]
|
1.severity: security
|
[search](https://search.nix.gsc.io/?q=postgresql&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=postgresql+in%3Apath&type=Code)
* [ ] [CVE-2020-10733](https://nvd.nist.gov/vuln/detail/CVE-2020-10733) CVSSv3=7.3 (nixos-20.03)
Scanned versions: nixos-20.03: 0d0660fde3b.
Cc @danbst
Cc @globin
Cc @ocharles
Cc @thoughtpolice
|
True
|
Vulnerability roundup 93: postgresql-9.6.17: 1 advisory [7.3] - [search](https://search.nix.gsc.io/?q=postgresql&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=postgresql+in%3Apath&type=Code)
* [ ] [CVE-2020-10733](https://nvd.nist.gov/vuln/detail/CVE-2020-10733) CVSSv3=7.3 (nixos-20.03)
Scanned versions: nixos-20.03: 0d0660fde3b.
Cc @danbst
Cc @globin
Cc @ocharles
Cc @thoughtpolice
|
non_process
|
vulnerability roundup postgresql advisory nixos scanned versions nixos cc danbst cc globin cc ocharles cc thoughtpolice
| 0
|
15,317
| 19,424,845,342
|
IssuesEvent
|
2021-12-21 03:08:50
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
GDAL Translate not respecting -a_srs for TFW files
|
Feedback Processing Bug
|
### What is the bug or the crash?
Choose Translate, set a different SRS as the override AND/OR ``-a_srs EPSG:4326`` in Additional command-line parameters, and the generated TFW will have coordinates in the layer's original SRS.
### Steps to reproduce the issue
1. Run Translate from Processing Toolbox
2. Setup Override CRS as EPSG:4326
3. Add ``-co TFW=YES`` to additional command-line parameters
4. Observe that TFW CRS is input layer's original CRS (in this case, EPSG:32615
5. Add ``-a_srs EPSG:4326`` to additional command-line parameters
6. Observer that the new TFW is still in the wrong SRS
QGIS History Item
```
"""
Double-click on the history item or paste the command below to re-run the algorithm
"""
processing.run("gdal:translate", {'INPUT':'C:/Users/USERNAME/AppData/Local/Temp/processing_BxRQJM/6516c5016b0d4816a6bcb0cbdabd618a/OUTPUT.tif','TARGET_CRS':QgsCoordinateReferenceSystem('EPSG:4326'),'NODATA':None,'COPY_SUBDATASETS':False,'OPTIONS':'','EXTRA':'-co TFW=YES -a_srs EPSG:4326','DATA_TYPE':0,'OUTPUT':'TEMPORARY_OUTPUT'})
```
### Versions
QGIS version
3.22.1-Białowieża
QGIS code revision
663dcf8fb9
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.0
PROJ version
8.2.0
EPSG Registry database version
v10.038 (2021-10-21)
GEOS version
3.10.0-CAPI-1.16.0
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
13.0
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
CalculateGeometry
0.6.4
cartography_tools
1.2.1
ImportPhotos
3.0.3
kmltools
3.1.15
LAStools
1.4
loadthemall
3.3.0
MemoryLayerSaver
4.0.4
pdaltools-master
0.1.5
photo2shape
3.0
processing_fusion
3.0
Projestions
1.0.1
qgis-maptiler-plugin
2.0.0
qgis2web
3.16.0
qgissprp
0.5.0
slyr
3.18.0
zoom_level
0.1
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context

Thread exploring this in usage:
https://community.opendronemap.org/t/is-there-a-way-to-produce-world-and-projection-files/9771/14?u=saijin_naib
|
1.0
|
GDAL Translate not respecting -a_srs for TFW files - ### What is the bug or the crash?
Choose Translate, set a different SRS as the override AND/OR ``-a_srs EPSG:4326`` in Additional command-line parameters, and the generated TFW will have coordinates in the layer's original SRS.
### Steps to reproduce the issue
1. Run Translate from Processing Toolbox
2. Setup Override CRS as EPSG:4326
3. Add ``-co TFW=YES`` to additional command-line parameters
4. Observe that TFW CRS is input layer's original CRS (in this case, EPSG:32615
5. Add ``-a_srs EPSG:4326`` to additional command-line parameters
6. Observer that the new TFW is still in the wrong SRS
QGIS History Item
```
"""
Double-click on the history item or paste the command below to re-run the algorithm
"""
processing.run("gdal:translate", {'INPUT':'C:/Users/USERNAME/AppData/Local/Temp/processing_BxRQJM/6516c5016b0d4816a6bcb0cbdabd618a/OUTPUT.tif','TARGET_CRS':QgsCoordinateReferenceSystem('EPSG:4326'),'NODATA':None,'COPY_SUBDATASETS':False,'OPTIONS':'','EXTRA':'-co TFW=YES -a_srs EPSG:4326','DATA_TYPE':0,'OUTPUT':'TEMPORARY_OUTPUT'})
```
### Versions
QGIS version
3.22.1-Białowieża
QGIS code revision
663dcf8fb9
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.0
PROJ version
8.2.0
EPSG Registry database version
v10.038 (2021-10-21)
GEOS version
3.10.0-CAPI-1.16.0
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
13.0
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
CalculateGeometry
0.6.4
cartography_tools
1.2.1
ImportPhotos
3.0.3
kmltools
3.1.15
LAStools
1.4
loadthemall
3.3.0
MemoryLayerSaver
4.0.4
pdaltools-master
0.1.5
photo2shape
3.0
processing_fusion
3.0
Projestions
1.0.1
qgis-maptiler-plugin
2.0.0
qgis2web
3.16.0
qgissprp
0.5.0
slyr
3.18.0
zoom_level
0.1
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context

Thread exploring this in usage:
https://community.opendronemap.org/t/is-there-a-way-to-produce-world-and-projection-files/9771/14?u=saijin_naib
|
process
|
gdal translate not respecting a srs for tfw files what is the bug or the crash choose translate set a different srs as the override and or a srs epsg in additional command line parameters and the generated tfw will have coordinates in the layer s original srs steps to reproduce the issue run translate from processing toolbox setup override crs as epsg add co tfw yes to additional command line parameters observe that tfw crs is input layer s original crs in this case epsg add a srs epsg to additional command line parameters observer that the new tfw is still in the wrong srs qgis history item double click on the history item or paste the command below to re run the algorithm processing run gdal translate input c users username appdata local temp processing bxrqjm output tif target crs qgscoordinatereferencesystem epsg nodata none copy subdatasets false options extra co tfw yes a srs epsg data type output temporary output versions qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins calculategeometry cartography tools importphotos kmltools lastools loadthemall memorylayersaver pdaltools master processing fusion projestions qgis maptiler plugin qgissprp slyr zoom level db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context thread exploring this in usage
| 1
|
87,753
| 15,790,301,850
|
IssuesEvent
|
2021-04-02 01:04:28
|
revfactory/task-api
|
https://api.github.com/repos/revfactory/task-api
|
opened
|
CVE-2019-0232 (High) detected in tomcat-embed-core-9.0.16.jar
|
security vulnerability
|
## CVE-2019-0232 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.16.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Path to dependency file: task-api/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.3.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.1.3.RELEASE.jar
- :x: **tomcat-embed-core-9.0.16.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When running on Windows with enableCmdLineArguments enabled, the CGI Servlet in Apache Tomcat 9.0.0.M1 to 9.0.17, 8.5.0 to 8.5.39 and 7.0.0 to 7.0.93 is vulnerable to Remote Code Execution due to a bug in the way the JRE passes command line arguments to Windows. The CGI Servlet is disabled by default. The CGI option enableCmdLineArguments is disable by default in Tomcat 9.0.x (and will be disabled by default in all versions in response to this vulnerability). For a detailed explanation of the JRE behaviour, see Markus Wulftange's blog (https://codewhitesec.blogspot.com/2016/02/java-and-command-line-injections-in-windows.html) and this archived MSDN blog (https://web.archive.org/web/20161228144344/https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/).
<p>Publish Date: 2019-04-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0232>CVE-2019-0232</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232</a></p>
<p>Release Date: 2019-04-15</p>
<p>Fix Resolution: 9.0.18,8.5.40,7.0.94</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-0232 (High) detected in tomcat-embed-core-9.0.16.jar - ## CVE-2019-0232 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.16.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>Path to dependency file: task-api/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.16/d7069e3d0f760035b26b68b7b6af5eaa0c1862f/tomcat-embed-core-9.0.16.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.3.RELEASE.jar (Root Library)
- spring-boot-starter-tomcat-2.1.3.RELEASE.jar
- :x: **tomcat-embed-core-9.0.16.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
When running on Windows with enableCmdLineArguments enabled, the CGI Servlet in Apache Tomcat 9.0.0.M1 to 9.0.17, 8.5.0 to 8.5.39 and 7.0.0 to 7.0.93 is vulnerable to Remote Code Execution due to a bug in the way the JRE passes command line arguments to Windows. The CGI Servlet is disabled by default. The CGI option enableCmdLineArguments is disable by default in Tomcat 9.0.x (and will be disabled by default in all versions in response to this vulnerability). For a detailed explanation of the JRE behaviour, see Markus Wulftange's blog (https://codewhitesec.blogspot.com/2016/02/java-and-command-line-injections-in-windows.html) and this archived MSDN blog (https://web.archive.org/web/20161228144344/https://blogs.msdn.microsoft.com/twistylittlepassagesallalike/2011/04/23/everyone-quotes-command-line-arguments-the-wrong-way/).
<p>Publish Date: 2019-04-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0232>CVE-2019-0232</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-0232</a></p>
<p>Release Date: 2019-04-15</p>
<p>Fix Resolution: 9.0.18,8.5.40,7.0.94</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in tomcat embed core jar cve high severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation path to dependency file task api build gradle path to vulnerable library root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar root gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library vulnerability details when running on windows with enablecmdlinearguments enabled the cgi servlet in apache tomcat to to and to is vulnerable to remote code execution due to a bug in the way the jre passes command line arguments to windows the cgi servlet is disabled by default the cgi option enablecmdlinearguments is disable by default in tomcat x and will be disabled by default in all versions in response to this vulnerability for a detailed explanation of the jre behaviour see markus wulftange s blog and this archived msdn blog publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
149,588
| 13,283,984,904
|
IssuesEvent
|
2020-08-24 05:04:48
|
BuildForSDGCohort2/team-024-group-a
|
https://api.github.com/repos/BuildForSDGCohort2/team-024-group-a
|
closed
|
Update readme.md
|
documentation
|
Readme.md is just the description of our app
the backend and the frontend has each readme.md for description of starting the server
|
1.0
|
Update readme.md - Readme.md is just the description of our app
the backend and the frontend has each readme.md for description of starting the server
|
non_process
|
update readme md readme md is just the description of our app the backend and the frontend has each readme md for description of starting the server
| 0
|
18,761
| 24,664,034,442
|
IssuesEvent
|
2022-10-18 08:57:07
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Textual definition update: New definition of mitotic spindle assembly checkpoint signaling including START and END
|
cell cycle and DNA processes textual definition
|
GO:0007094
mitotic spindle assembly checkpoint signaling
Existing
A signal transduction process that contributes to a mitotic cell cycle spindle assembly checkpoint, that delays the metaphase/anaphase transition of a mitotic nuclear division until the spindle is correctly assembled and chromosomes are attached to the spindle.
Suggested
A mitotic cell cycle checkpoint that delays mitotic sister chromatid separation and consequently the mitotic metaphase/anaphase transition until the spindle is correctly assembled and chromosomes are attached to the spindle.
BEGINNING AND END
Spindle assembly checkpoint signalling begins with the activated Mph family kinase, and results in the inhibition of the Anaphase Promoting Complex and its activator Sleepy/Cdc20 by the mitotic checkpoint complex (MCC).
(This works for all species)
pers comm VW and Silke Hauf
|
1.0
|
Textual definition update: New definition of mitotic spindle assembly checkpoint signaling including START and END - GO:0007094
mitotic spindle assembly checkpoint signaling
Existing
A signal transduction process that contributes to a mitotic cell cycle spindle assembly checkpoint, that delays the metaphase/anaphase transition of a mitotic nuclear division until the spindle is correctly assembled and chromosomes are attached to the spindle.
Suggested
A mitotic cell cycle checkpoint that delays mitotic sister chromatid separation and consequently the mitotic metaphase/anaphase transition until the spindle is correctly assembled and chromosomes are attached to the spindle.
BEGINNING AND END
Spindle assembly checkpoint signalling begins with the activated Mph family kinase, and results in the inhibition of the Anaphase Promoting Complex and its activator Sleepy/Cdc20 by the mitotic checkpoint complex (MCC).
(This works for all species)
pers comm VW and Silke Hauf
|
process
|
textual definition update new definition of mitotic spindle assembly checkpoint signaling including start and end go mitotic spindle assembly checkpoint signaling existing a signal transduction process that contributes to a mitotic cell cycle spindle assembly checkpoint that delays the metaphase anaphase transition of a mitotic nuclear division until the spindle is correctly assembled and chromosomes are attached to the spindle suggested a mitotic cell cycle checkpoint that delays mitotic sister chromatid separation and consequently the mitotic metaphase anaphase transition until the spindle is correctly assembled and chromosomes are attached to the spindle beginning and end spindle assembly checkpoint signalling begins with the activated mph family kinase and results in the inhibition of the anaphase promoting complex and its activator sleepy by the mitotic checkpoint complex mcc this works for all species pers comm vw and silke hauf
| 1
|
5,331
| 8,147,954,812
|
IssuesEvent
|
2018-08-22 02:46:49
|
zotero/zotero
|
https://api.github.com/repos/zotero/zotero
|
closed
|
Style/translator initialization failure shouldn't freeze the doc plugins
|
Word Processor Integration
|
See https://forums.zotero.org/discussion/73274/word-2010-plugin-only-gives-progress-bar-when-trying-to-add-citation-or-bibliography
I suppose subsequent invocations of `Zotero.Styles.init()` should throw an error?
|
1.0
|
Style/translator initialization failure shouldn't freeze the doc plugins - See https://forums.zotero.org/discussion/73274/word-2010-plugin-only-gives-progress-bar-when-trying-to-add-citation-or-bibliography
I suppose subsequent invocations of `Zotero.Styles.init()` should throw an error?
|
process
|
style translator initialization failure shouldn t freeze the doc plugins see i suppose subsequent invocations of zotero styles init should throw an error
| 1
|
185,933
| 6,732,055,655
|
IssuesEvent
|
2017-10-18 09:57:14
|
ballerinalang/composer
|
https://api.github.com/repos/ballerinalang/composer
|
closed
|
Cannot delete the try-catch block when finally is added
|
0.94-pre-release Priority/High Severity/Major Type/Bug
|
1. Add finally block from the source view
2. Try to delete the entire try-catch
Cannot delete the try-catch block when finally is added
|
1.0
|
Cannot delete the try-catch block when finally is added - 1. Add finally block from the source view
2. Try to delete the entire try-catch
Cannot delete the try-catch block when finally is added
|
non_process
|
cannot delete the try catch block when finally is added add finally block from the source view try to delete the entire try catch cannot delete the try catch block when finally is added
| 0
|
262,311
| 19,782,695,307
|
IssuesEvent
|
2022-01-18 00:02:23
|
elastic/eui
|
https://api.github.com/repos/elastic/eui
|
closed
|
[EuiSearchBar] Document defaultFields option on toESQuery
|
documentation stale-issue stale-issue-closed
|
It seems that toESQuery options accepts a `defaultFields` array, but the docs to not indicate that, as well as a few other options available.
|
1.0
|
[EuiSearchBar] Document defaultFields option on toESQuery - It seems that toESQuery options accepts a `defaultFields` array, but the docs to not indicate that, as well as a few other options available.
|
non_process
|
document defaultfields option on toesquery it seems that toesquery options accepts a defaultfields array but the docs to not indicate that as well as a few other options available
| 0
|
407,226
| 11,908,259,639
|
IssuesEvent
|
2020-03-31 00:26:05
|
kubeflow/examples
|
https://api.github.com/repos/kubeflow/examples
|
closed
|
Create an mnist E2E notebook for Azure
|
area/examples kind/feature platform/azure priority/p1
|
For 1.0 the core CUJ is build, train, deploy.
In kubeflow/examples#723
we've created a notebook that highlights this core CUJ.
We'd like to have variations of this notebook for all the different platforms/clouds on which Kubeflow is supported.
So we'd like to have a variant of this notebook that is suited for Azure.
That primarily means updating the notebook to use Azure blob store. Alternatively you could use minio running on the cluster but using Azure blob store would probably be better.
/cc @aronchick
|
1.0
|
Create an mnist E2E notebook for Azure - For 1.0 the core CUJ is build, train, deploy.
In kubeflow/examples#723
we've created a notebook that highlights this core CUJ.
We'd like to have variations of this notebook for all the different platforms/clouds on which Kubeflow is supported.
So we'd like to have a variant of this notebook that is suited for Azure.
That primarily means updating the notebook to use Azure blob store. Alternatively you could use minio running on the cluster but using Azure blob store would probably be better.
/cc @aronchick
|
non_process
|
create an mnist notebook for azure for the core cuj is build train deploy in kubeflow examples we ve created a notebook that highlights this core cuj we d like to have variations of this notebook for all the different platforms clouds on which kubeflow is supported so we d like to have a variant of this notebook that is suited for azure that primarily means updating the notebook to use azure blob store alternatively you could use minio running on the cluster but using azure blob store would probably be better cc aronchick
| 0
|
5,127
| 7,894,683,629
|
IssuesEvent
|
2018-06-28 22:32:39
|
hashicorp/packer
|
https://api.github.com/repos/hashicorp/packer
|
closed
|
vSphere post-processor mark as template
|
enhancement post-processor/vsphere
|
Is it possible to have the vSphere provisioner automatically mark the uploaded VM as a template?
|
1.0
|
vSphere post-processor mark as template - Is it possible to have the vSphere provisioner automatically mark the uploaded VM as a template?
|
process
|
vsphere post processor mark as template is it possible to have the vsphere provisioner automatically mark the uploaded vm as a template
| 1
|
4,729
| 7,571,883,366
|
IssuesEvent
|
2018-04-23 13:34:24
|
zotero/zotero
|
https://api.github.com/repos/zotero/zotero
|
closed
|
Add uninstallation options for word processor plugins to Cite pane
|
Word Processor Integration
|
Shouldn't have to tell people to find their Word startup folder
https://forums.zotero.org/discussion/69411/how-remove-zotero-plugin-in-word
|
1.0
|
Add uninstallation options for word processor plugins to Cite pane - Shouldn't have to tell people to find their Word startup folder
https://forums.zotero.org/discussion/69411/how-remove-zotero-plugin-in-word
|
process
|
add uninstallation options for word processor plugins to cite pane shouldn t have to tell people to find their word startup folder
| 1
|
160,730
| 12,519,732,887
|
IssuesEvent
|
2020-06-03 14:51:24
|
aliasrobotics/RVD
|
https://api.github.com/repos/aliasrobotics/RVD
|
closed
|
RVD#1979: Using Fault to parse untrusted XML data is known to be vulnerable to XML..., /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52
|
bandit bug components software duplicate invalid robot component: ROS static analysis testing triage version: noetic
|
```yaml
{
"id": 1979,
"title": "RVD#1979: Using Fault to parse untrusted XML data is known to be vulnerable to XML..., /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52",
"type": "bug",
"description": "HIGH confidence of HIGH severity bug. Using Fault to parse untrusted XML data is known to be vulnerable to XML attacks. Use defused.xmlrpc.monkey_patch() function to monkey-patch xmlrpclib and mitigate XML vulnerabilities. at /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52 See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": [
"https://github.com/aliasrobotics/RVD/issues/1979",
"https://bandit.readthedocs.io/en/latest/blacklists/blacklist_imports.html#b411-import-xmlrpclib"
],
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-29 (08:41)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-29 (08:41)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "https://github.com/aliasrobotics/RVD/issues/1979",
"reproducibility": "always",
"trace": "/opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
```
|
1.0
|
RVD#1979: Using Fault to parse untrusted XML data is known to be vulnerable to XML..., /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52 - ```yaml
{
"id": 1979,
"title": "RVD#1979: Using Fault to parse untrusted XML data is known to be vulnerable to XML..., /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52",
"type": "bug",
"description": "HIGH confidence of HIGH severity bug. Using Fault to parse untrusted XML data is known to be vulnerable to XML attacks. Use defused.xmlrpc.monkey_patch() function to monkey-patch xmlrpclib and mitigate XML vulnerabilities. at /opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52 See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": [
"https://github.com/aliasrobotics/RVD/issues/1979",
"https://bandit.readthedocs.io/en/latest/blacklists/blacklist_imports.html#b411-import-xmlrpclib"
],
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-29 (08:41)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-29 (08:41)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "https://github.com/aliasrobotics/RVD/issues/1979",
"reproducibility": "always",
"trace": "/opt/ros_noetic_ws/src/ros_comm/rostopic/src/rostopic/__init__.py:52",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
```
|
non_process
|
rvd using fault to parse untrusted xml data is known to be vulnerable to xml opt ros noetic ws src ros comm rostopic src rostopic init py yaml id title rvd using fault to parse untrusted xml data is known to be vulnerable to xml opt ros noetic ws src ros comm rostopic src rostopic init py type bug description high confidence of high severity bug using fault to parse untrusted xml data is known to be vulnerable to xml attacks use defused xmlrpc monkey patch function to monkey patch xmlrpclib and mitigate xml vulnerabilities at opt ros noetic ws src ros comm rostopic src rostopic init py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace opt ros noetic ws src ros comm rostopic src rostopic init py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
| 0
|
4,644
| 7,488,632,794
|
IssuesEvent
|
2018-04-06 02:40:15
|
UnbFeelings/unb-feelings-GQA
|
https://api.github.com/repos/UnbFeelings/unb-feelings-GQA
|
closed
|
Aplicar atualizações de FeedBack no Processo
|
help wanted process
|
Atualizações:
- [x] Renomear o artefato Resultado Padronizado para Resultado Consolidado;
- [x] Reorganizar o subprocesso que lida com as não conformidades;
- [x] Criar atividade de **Definir Critérios de Avaliação** (Após definir objetos de Avaliação);
- Gerar artefato de Critérios de Avaliação;
- [x] Arrumar no último sub-Processo a forma como serão tratadas as inconformidades do ciclo (Atividade ou Sub-Processo menor);
- [x] Criar atividade para Definir ferramentas;
- Gerar artefato de ferramentas;
- [x] Tirar ferramenta de dentro do Sub-Processo Detalhar;
- [x] Analisar Processo, Critérios de Avaliação, Definir Ferramentas;
- [x] Eliminar tarefas desnecessárias do Processo;
OBS: MUDANÇAS NO PROCESSO SERÃO PROVIDAS POR UM PROCESSO DE APOIA QUE FOGE O ESCOPO DA EQUIPE DE GQA;
**Comentário com as palavras da professora:**
Na atividade 3.3 do processo, chamada Registrar resultados, existe uma saída chamada "Resultado Padronizado", o que não deixa muito claro o que é, sugeriu-se trocar para "Resultado Consolidado"
** Deve ser atualizado tanto no diagrama quanto na descrição da atividade do processo
- Explicitar no processo que a não conformidade deve ser acompanhada por um prazo diferente das auditorias. Por exemplo, o prazo para resolver o problema X é dois dias, e não até a próxima auditoria
- Colocar na descrição da atividade do processo: ao registrar a não conformidade, é necessário informar a complexidade dela
|
1.0
|
Aplicar atualizações de FeedBack no Processo - Atualizações:
- [x] Renomear o artefato Resultado Padronizado para Resultado Consolidado;
- [x] Reorganizar o subprocesso que lida com as não conformidades;
- [x] Criar atividade de **Definir Critérios de Avaliação** (Após definir objetos de Avaliação);
- Gerar artefato de Critérios de Avaliação;
- [x] Arrumar no último sub-Processo a forma como serão tratadas as inconformidades do ciclo (Atividade ou Sub-Processo menor);
- [x] Criar atividade para Definir ferramentas;
- Gerar artefato de ferramentas;
- [x] Tirar ferramenta de dentro do Sub-Processo Detalhar;
- [x] Analisar Processo, Critérios de Avaliação, Definir Ferramentas;
- [x] Eliminar tarefas desnecessárias do Processo;
OBS: MUDANÇAS NO PROCESSO SERÃO PROVIDAS POR UM PROCESSO DE APOIA QUE FOGE O ESCOPO DA EQUIPE DE GQA;
**Comentário com as palavras da professora:**
Na atividade 3.3 do processo, chamada Registrar resultados, existe uma saída chamada "Resultado Padronizado", o que não deixa muito claro o que é, sugeriu-se trocar para "Resultado Consolidado"
** Deve ser atualizado tanto no diagrama quanto na descrição da atividade do processo
- Explicitar no processo que a não conformidade deve ser acompanhada por um prazo diferente das auditorias. Por exemplo, o prazo para resolver o problema X é dois dias, e não até a próxima auditoria
- Colocar na descrição da atividade do processo: ao registrar a não conformidade, é necessário informar a complexidade dela
|
process
|
aplicar atualizações de feedback no processo atualizações renomear o artefato resultado padronizado para resultado consolidado reorganizar o subprocesso que lida com as não conformidades criar atividade de definir critérios de avaliação após definir objetos de avaliação gerar artefato de critérios de avaliação arrumar no último sub processo a forma como serão tratadas as inconformidades do ciclo atividade ou sub processo menor criar atividade para definir ferramentas gerar artefato de ferramentas tirar ferramenta de dentro do sub processo detalhar analisar processo critérios de avaliação definir ferramentas eliminar tarefas desnecessárias do processo obs mudanças no processo serão providas por um processo de apoia que foge o escopo da equipe de gqa comentário com as palavras da professora na atividade do processo chamada registrar resultados existe uma saída chamada resultado padronizado o que não deixa muito claro o que é sugeriu se trocar para resultado consolidado deve ser atualizado tanto no diagrama quanto na descrição da atividade do processo explicitar no processo que a não conformidade deve ser acompanhada por um prazo diferente das auditorias por exemplo o prazo para resolver o problema x é dois dias e não até a próxima auditoria colocar na descrição da atividade do processo ao registrar a não conformidade é necessário informar a complexidade dela
| 1
|
12,562
| 14,981,350,171
|
IssuesEvent
|
2021-01-28 14:44:36
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
reopened
|
Add new user page > Role > 'Super Admin' label should be changed to 'Make this user a superadmin'
|
Bug P2 Participant manager Process: Dev Process: Fixed Process: Tested QA Process: Tested dev
|
AR : Role is displaying as 'Super Admin'
ER : Add new user page > Role > 'Super Admin' label should be changed to 'Make this user a superadmin'

|
4.0
|
Add new user page > Role > 'Super Admin' label should be changed to 'Make this user a superadmin' - AR : Role is displaying as 'Super Admin'
ER : Add new user page > Role > 'Super Admin' label should be changed to 'Make this user a superadmin'

|
process
|
add new user page role super admin label should be changed to make this user a superadmin ar role is displaying as super admin er add new user page role super admin label should be changed to make this user a superadmin
| 1
|
2,283
| 5,108,317,835
|
IssuesEvent
|
2017-01-05 17:19:55
|
DynareTeam/dynare
|
https://api.github.com/repos/DynareTeam/dynare
|
closed
|
Depth issue
|
bug preprocessor
|
Looking into #1175 , by reverting commit 3c7e60b744567f6f39a9c611bce6dcaadcd52bc6, I obtained the following error from matlab when trying to run Christiano-Motto-Rostagno model (the one in subfolder `figure4` of the archive available [here](http://faculty.wcas.northwestern.edu/~lchrist/research/ECB/risk_shocks/20100922_data.zip)
```
Error: File: cmr_static.m Line: 1292 Column: 16331
Nesting of {, [, and ( cannot exceed a depth of 32.
```
@MichelJuillard May this be a consequence of your patch about on auxiliary variables in steady state and static files (see #1133)?
|
1.0
|
Depth issue - Looking into #1175 , by reverting commit 3c7e60b744567f6f39a9c611bce6dcaadcd52bc6, I obtained the following error from matlab when trying to run Christiano-Motto-Rostagno model (the one in subfolder `figure4` of the archive available [here](http://faculty.wcas.northwestern.edu/~lchrist/research/ECB/risk_shocks/20100922_data.zip)
```
Error: File: cmr_static.m Line: 1292 Column: 16331
Nesting of {, [, and ( cannot exceed a depth of 32.
```
@MichelJuillard May this be a consequence of your patch about on auxiliary variables in steady state and static files (see #1133)?
|
process
|
depth issue looking into by reverting commit i obtained the following error from matlab when trying to run christiano motto rostagno model the one in subfolder of the archive available error file cmr static m line column nesting of and cannot exceed a depth of micheljuillard may this be a consequence of your patch about on auxiliary variables in steady state and static files see
| 1
|
276,043
| 23,961,956,210
|
IssuesEvent
|
2022-09-12 19:59:18
|
Sheldonfrith/sst-scheduled-api-test-testee-repo
|
https://api.github.com/repos/Sheldonfrith/sst-scheduled-api-test-testee-repo
|
closed
|
Auto-Generated: g/json - get - 200 - Test: deliberateBug - Response body is not an empty object empty
|
sst-scheduled-api-testing
|
{"userDefinedErrorMessage":"Test: deliberateBug - Response body is not an empty object empty","endpointUrl":"https://api.sunrise-sunset.org/json","response":{"status":200,"headers":{"server":"nginx","date":"Tue, 30 Aug 2022 00:54:02 GMT","content-type":"application/json","transfer-encoding":"chunked","connection":"close","vary":"Accept-Encoding","access-control-allow-origin":"*"},"body":"{\"results\":{\"sunrise\":\"10:16:46 AM\",\"sunset\":\"11:28:22 PM\",\"solar_noon\":\"4:52:34 PM\",\"day_length\":\"13:11:36\",\"civil_twilight_begin\":\"9:50:18 AM\",\"civil_twilight_end\":\"11:54:50 PM\",\"nautical_twilight_begin\":\"9:17:01 AM\",\"nautical_twilight_end\":\"12:28:07 AM\",\"astronomical_twilight_begin\":\"8:42:14 AM\",\"astronomical_twilight_end\":\"1:02:54 AM\"},\"status\":\"OK\"}"},"request":{"headers":{"Accept":"application/json, text/plain, */*","User-Agent":"axios/0.27.2"}}}
|
1.0
|
Auto-Generated: g/json - get - 200 - Test: deliberateBug - Response body is not an empty object empty - {"userDefinedErrorMessage":"Test: deliberateBug - Response body is not an empty object empty","endpointUrl":"https://api.sunrise-sunset.org/json","response":{"status":200,"headers":{"server":"nginx","date":"Tue, 30 Aug 2022 00:54:02 GMT","content-type":"application/json","transfer-encoding":"chunked","connection":"close","vary":"Accept-Encoding","access-control-allow-origin":"*"},"body":"{\"results\":{\"sunrise\":\"10:16:46 AM\",\"sunset\":\"11:28:22 PM\",\"solar_noon\":\"4:52:34 PM\",\"day_length\":\"13:11:36\",\"civil_twilight_begin\":\"9:50:18 AM\",\"civil_twilight_end\":\"11:54:50 PM\",\"nautical_twilight_begin\":\"9:17:01 AM\",\"nautical_twilight_end\":\"12:28:07 AM\",\"astronomical_twilight_begin\":\"8:42:14 AM\",\"astronomical_twilight_end\":\"1:02:54 AM\"},\"status\":\"OK\"}"},"request":{"headers":{"Accept":"application/json, text/plain, */*","User-Agent":"axios/0.27.2"}}}
|
non_process
|
auto generated g json get test deliberatebug response body is not an empty object empty userdefinederrormessage test deliberatebug response body is not an empty object empty endpointurl aug gmt content type application json transfer encoding chunked connection close vary accept encoding access control allow origin body results sunrise am sunset pm solar noon pm day length civil twilight begin am civil twilight end pm nautical twilight begin am nautical twilight end am astronomical twilight begin am astronomical twilight end am status ok request headers accept application json text plain user agent axios
| 0
|
15,727
| 19,899,566,280
|
IssuesEvent
|
2022-01-25 05:46:29
|
SigNoz/signoz
|
https://api.github.com/repos/SigNoz/signoz
|
closed
|
Add Events data from OpenTelemetry spans
|
processors query-service
|
While flattening the data, we should also store `Events` data from spans to Apache Druid.
This contains `exception` data apart from reasons for erroneous spans like `Redis timeout`
|
1.0
|
Add Events data from OpenTelemetry spans - While flattening the data, we should also store `Events` data from spans to Apache Druid.
This contains `exception` data apart from reasons for erroneous spans like `Redis timeout`
|
process
|
add events data from opentelemetry spans while flattening the data we should also store events data from spans to apache druid this contains exception data apart from reasons for erroneous spans like redis timeout
| 1
|
368,348
| 25,789,442,856
|
IssuesEvent
|
2022-12-10 01:10:24
|
Navnedia/Link-Shortener
|
https://api.github.com/repos/Navnedia/Link-Shortener
|
opened
|
Add documentation for the project
|
documentation
|
Write an organized README and potentially other documents detailing information about the project:
- [ ] What is the project?
- [ ] What's the purpose of the project? Why was this project developed?
- [ ] What tools and concepts were used to develop this project?
- [ ] Skills and tools learned and developed along the way
- [ ] Document the components of the project
- [ ] The API endpoints
- [ ] Describe how to install and use the application
|
1.0
|
Add documentation for the project - Write an organized README and potentially other documents detailing information about the project:
- [ ] What is the project?
- [ ] What's the purpose of the project? Why was this project developed?
- [ ] What tools and concepts were used to develop this project?
- [ ] Skills and tools learned and developed along the way
- [ ] Document the components of the project
- [ ] The API endpoints
- [ ] Describe how to install and use the application
|
non_process
|
add documentation for the project write an organized readme and potentially other documents detailing information about the project what is the project what s the purpose of the project why was this project developed what tools and concepts were used to develop this project skills and tools learned and developed along the way document the components of the project the api endpoints describe how to install and use the application
| 0
|
11,420
| 14,246,229,201
|
IssuesEvent
|
2020-11-19 09:47:20
|
pystatgen/sgkit
|
https://api.github.com/repos/pystatgen/sgkit
|
closed
|
Release 0.1.0a1
|
process + tools
|
This is an umbrella issue for our first release.
This would be an alpha [pre-release](https://www.python.org/dev/peps/pep-0440/#pre-releases) (suggested version number: 0.1.0a1), which means that users would have to opt in to use it by specifying the `--pre` flag: `pip install --pre sgkit`.
Here are some of the tasks we need to do - please suggest more (or say if they are not needed).
- [ ] cyvcf2 wheel (should we wait for this?)
- [ ] Build release artifacts locally
- [ ] Add release number to documentation URL
- [ ] Use TestPyPI to test release process
- [ ] Update installation instructions for users
- [ ] Document release process
Note that these are all about what we need to do to make a release; since it's a pre-release, the exact set of features doesn't matter so much.
|
1.0
|
Release 0.1.0a1 - This is an umbrella issue for our first release.
This would be an alpha [pre-release](https://www.python.org/dev/peps/pep-0440/#pre-releases) (suggested version number: 0.1.0a1), which means that users would have to opt in to use it by specifying the `--pre` flag: `pip install --pre sgkit`.
Here are some of the tasks we need to do - please suggest more (or say if they are not needed).
- [ ] cyvcf2 wheel (should we wait for this?)
- [ ] Build release artifacts locally
- [ ] Add release number to documentation URL
- [ ] Use TestPyPI to test release process
- [ ] Update installation instructions for users
- [ ] Document release process
Note that these are all about what we need to do to make a release; since it's a pre-release, the exact set of features doesn't matter so much.
|
process
|
release this is an umbrella issue for our first release this would be an alpha suggested version number which means that users would have to opt in to use it by specifying the pre flag pip install pre sgkit here are some of the tasks we need to do please suggest more or say if they are not needed wheel should we wait for this build release artifacts locally add release number to documentation url use testpypi to test release process update installation instructions for users document release process note that these are all about what we need to do to make a release since it s a pre release the exact set of features doesn t matter so much
| 1
|
6,450
| 9,546,286,441
|
IssuesEvent
|
2019-05-01 19:29:41
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Bug: Lose education data when education added
|
Apply Process Bug State Dept.
|
Environment: Test
Browser: Chrome
Steps to reproduce:
1) Answer the 3 education questions and GPA
2) Select Add Education (when you don't already have education)
3) Enter education and save
4) Questions and GPA are lost
- Once saved, I can edit the questions and follow the same path and the edits will save. Appears to be an issue when blank and the page hasn't been saved.
|
1.0
|
Bug: Lose education data when education added - Environment: Test
Browser: Chrome
Steps to reproduce:
1) Answer the 3 education questions and GPA
2) Select Add Education (when you don't already have education)
3) Enter education and save
4) Questions and GPA are lost
- Once saved, I can edit the questions and follow the same path and the edits will save. Appears to be an issue when blank and the page hasn't been saved.
|
process
|
bug lose education data when education added environment test browser chrome steps to reproduce answer the education questions and gpa select add education when you don t already have education enter education and save questions and gpa are lost once saved i can edit the questions and follow the same path and the edits will save appears to be an issue when blank and the page hasn t been saved
| 1
|
11,786
| 14,617,358,501
|
IssuesEvent
|
2020-12-22 14:40:17
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
opened
|
[PM] Participant details > Consent document is not shown for closed study having Eligibility test
|
Bug P0 Participant manager Process: Dev
|
**Steps:**
1. Publish a study having token validation and Eligibility test
2. PM admin invites a user
3. Mobile participant enrolls into the study using valid token and passes the eligibility test
4. User enrolls successfully
5. Observe the participant details --> Consent History
A/R: Consent document is not shown for closed study having Eligibility test
E/R: Consent document should be shown for all type of studies

|
1.0
|
[PM] Participant details > Consent document is not shown for closed study having Eligibility test - **Steps:**
1. Publish a study having token validation and Eligibility test
2. PM admin invites a user
3. Mobile participant enrolls into the study using valid token and passes the eligibility test
4. User enrolls successfully
5. Observe the participant details --> Consent History
A/R: Consent document is not shown for closed study having Eligibility test
E/R: Consent document should be shown for all type of studies

|
process
|
participant details consent document is not shown for closed study having eligibility test steps publish a study having token validation and eligibility test pm admin invites a user mobile participant enrolls into the study using valid token and passes the eligibility test user enrolls successfully observe the participant details consent history a r consent document is not shown for closed study having eligibility test e r consent document should be shown for all type of studies
| 1
|
46,942
| 7,296,112,410
|
IssuesEvent
|
2018-02-26 09:41:22
|
alphagov/govuk-frontend
|
https://api.github.com/repos/alphagov/govuk-frontend
|
opened
|
Explain why we use the current CSS methodologies
|
documentation
|
We have detail on what we're doing, but not why here https://github.com/alphagov/govuk-frontend/blob/master/docs/coding-standards/css.md
This has come up already and it'd be good have something comprehensive to link people to.
Might be good as a blog post.
|
1.0
|
Explain why we use the current CSS methodologies - We have detail on what we're doing, but not why here https://github.com/alphagov/govuk-frontend/blob/master/docs/coding-standards/css.md
This has come up already and it'd be good have something comprehensive to link people to.
Might be good as a blog post.
|
non_process
|
explain why we use the current css methodologies we have detail on what we re doing but not why here this has come up already and it d be good have something comprehensive to link people to might be good as a blog post
| 0
|
186,046
| 14,394,554,906
|
IssuesEvent
|
2020-12-03 01:34:08
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
chengfangang/fabric-ca-gm-test: cmd/fabric-ca-client/main_test.go; 3 LoC
|
fresh test tiny
|
Found a possible issue in [chengfangang/fabric-ca-gm-test](https://www.github.com/chengfangang/fabric-ca-gm-test) at [cmd/fabric-ca-client/main_test.go](https://github.com/chengfangang/fabric-ca-gm-test/blob/6aca151d16f6978ce651134c28cb34f1a6c29c85/cmd/fabric-ca-client/main_test.go#L132-L134)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to e at line 133 may start a goroutine
[Click here to see the code in its original context.](https://github.com/chengfangang/fabric-ca-gm-test/blob/6aca151d16f6978ce651134c28cb34f1a6c29c85/cmd/fabric-ca-client/main_test.go#L132-L134)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, e := range errCases {
extraArgErrorTest(&e, t)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 6aca151d16f6978ce651134c28cb34f1a6c29c85
|
1.0
|
chengfangang/fabric-ca-gm-test: cmd/fabric-ca-client/main_test.go; 3 LoC -
Found a possible issue in [chengfangang/fabric-ca-gm-test](https://www.github.com/chengfangang/fabric-ca-gm-test) at [cmd/fabric-ca-client/main_test.go](https://github.com/chengfangang/fabric-ca-gm-test/blob/6aca151d16f6978ce651134c28cb34f1a6c29c85/cmd/fabric-ca-client/main_test.go#L132-L134)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> function call which takes a reference to e at line 133 may start a goroutine
[Click here to see the code in its original context.](https://github.com/chengfangang/fabric-ca-gm-test/blob/6aca151d16f6978ce651134c28cb34f1a6c29c85/cmd/fabric-ca-client/main_test.go#L132-L134)
<details>
<summary>Click here to show the 3 line(s) of Go which triggered the analyzer.</summary>
```go
for _, e := range errCases {
extraArgErrorTest(&e, t)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 6aca151d16f6978ce651134c28cb34f1a6c29c85
|
non_process
|
chengfangang fabric ca gm test cmd fabric ca client main test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to e at line may start a goroutine click here to show the line s of go which triggered the analyzer go for e range errcases extraargerrortest e t leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
14,230
| 4,856,810,590
|
IssuesEvent
|
2016-11-12 08:08:56
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
new menu item check
|
No Code Attached Yet
|
### Steps to reproduce the issue
It is not allowed that a menu item alias is the same as a sub-folder in your Joomla installation.
The message will be: "Save failed with the following error: A first level menu item alias can't be 'tmp' because 'tmp' is a sub-folder of your joomla installation folder."
My problem is that when you enter a new menu item this check is not done, before creating the menu item.
So you can create a menu item with the name of a sub-folder. The moment the item is creacted you can't save it again.
### Expected result
Check before the item is created.
### Actual result
No check and an error on the frontend of the site because your not allowed to browse a sub-folder.
### System information (as much as possible)
### Additional comments
|
1.0
|
new menu item check - ### Steps to reproduce the issue
It is not allowed that a menu item alias is the same as a sub-folder in your Joomla installation.
The message will be: "Save failed with the following error: A first level menu item alias can't be 'tmp' because 'tmp' is a sub-folder of your joomla installation folder."
My problem is that when you enter a new menu item this check is not done, before creating the menu item.
So you can create a menu item with the name of a sub-folder. The moment the item is creacted you can't save it again.
### Expected result
Check before the item is created.
### Actual result
No check and an error on the frontend of the site because your not allowed to browse a sub-folder.
### System information (as much as possible)
### Additional comments
|
non_process
|
new menu item check steps to reproduce the issue it is not allowed that a menu item alias is the same as a sub folder in your joomla installation the message will be save failed with the following error a first level menu item alias can t be tmp because tmp is a sub folder of your joomla installation folder my problem is that when you enter a new menu item this check is not done before creating the menu item so you can create a menu item with the name of a sub folder the moment the item is creacted you can t save it again expected result check before the item is created actual result no check and an error on the frontend of the site because your not allowed to browse a sub folder system information as much as possible additional comments
| 0
|
2,646
| 5,425,882,156
|
IssuesEvent
|
2017-03-03 08:13:46
|
rubberduck-vba/Rubberduck
|
https://api.github.com/repos/rubberduck-vba/Rubberduck
|
closed
|
Can't initiate Parse in Office XP
|
bug critical parse-tree-processing
|
The Parse button is enabled and clickable in Excel, Word and Access, but the Parse doesn't appear to commence, and so never completes.
In PowerPoint, the button is clickable, and the spinning circles appear in the Code Explorer and Code Inspections windows, but there is no other visible sign of progress, and parsing never completes.
|
1.0
|
Can't initiate Parse in Office XP - The Parse button is enabled and clickable in Excel, Word and Access, but the Parse doesn't appear to commence, and so never completes.
In PowerPoint, the button is clickable, and the spinning circles appear in the Code Explorer and Code Inspections windows, but there is no other visible sign of progress, and parsing never completes.
|
process
|
can t initiate parse in office xp the parse button is enabled and clickable in excel word and access but the parse doesn t appear to commence and so never completes in powerpoint the button is clickable and the spinning circles appear in the code explorer and code inspections windows but there is no other visible sign of progress and parsing never completes
| 1
|
21,944
| 30,446,799,875
|
IssuesEvent
|
2023-07-15 19:28:47
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
pyutils 0.0.1b7 has 2 GuardDog issues
|
guarddog typosquatting silent-process-execution
|
https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1b7",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils",
"silent-process-execution": [
{
"location": "pyutils/exec_utils.py/pyutils/exec_utils.py:204",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpylvasmk3/pyutils"
}
}```
|
1.0
|
pyutils 0.0.1b7 has 2 GuardDog issues - https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1b7",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils",
"silent-process-execution": [
{
"location": "pyutils/exec_utils.py/pyutils/exec_utils.py:204",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpylvasmk3/pyutils"
}
}```
|
process
|
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt python utils pytils silent process execution location pyutils exec utils py pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pyutils
| 1
|
16,350
| 21,009,456,998
|
IssuesEvent
|
2022-03-30 04:08:16
|
bitPogo/kmock
|
https://api.github.com/repos/bitPogo/kmock
|
opened
|
Add Argument types to invoke
|
enhancement kmock-processor
|
## Description
<!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug -->
Currently invoke accepts any value, since there is no generic value given.
|
1.0
|
Add Argument types to invoke - ## Description
<!--- Provide a detailed introduction to the issue itself, and why you consider it to be a bug -->
Currently invoke accepts any value, since there is no generic value given.
|
process
|
add argument types to invoke description currently invoke accepts any value since there is no generic value given
| 1
|
15,262
| 19,211,018,701
|
IssuesEvent
|
2021-12-07 01:56:31
|
googleapis/python-spanner
|
https://api.github.com/repos/googleapis/python-spanner
|
closed
|
tests.system.test_system_dbapi.TestTransactionsManagement: test_commit failed
|
api: spanner type: process flakybot: issue flakybot: flaky
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 2487800e31842a44dcc37937c325e130c8c926b0
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/306a2e02-87cb-4be9-be31-37456ec7a8a2), [Sponge](http://sponge2/306a2e02-87cb-4be9-be31-37456ec7a8a2)
status: failed
<details><summary>Test output</summary><br><pre>args = (parent: "projects/precise-truck-742"
instance_id: "google-cloud-1627550679627"
instance {
name: "projects/precise-t...1627550946"
}
labels {
key: "python-spanner-dbapi-systests"
value: "true"
}
processing_units: 1000
}
,)
kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/precise-truck-742/instances/google-cloud-1627550679627'), ('x...ms', 'parent=projects/precise-truck-742'), ('x-goog-api-client', 'gl-python/3.8.6 grpc/1.39.0 gax/1.31.1 gccl/3.6.0')]}
@six.wraps(callable_)
def error_remapped_callable(*args, **kwargs):
try:
> return callable_(*args, **kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:67:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7fe7942014f0>
request = parent: "projects/precise-truck-742"
instance_id: "google-cloud-1627550679627"
instance {
name: "projects/precise-tr... "1627550946"
}
labels {
key: "python-spanner-dbapi-systests"
value: "true"
}
processing_units: 1000
}
timeout = None
metadata = [('google-cloud-resource-prefix', 'projects/precise-truck-742/instances/google-cloud-1627550679627'), ('x-goog-request...ams', 'parent=projects/precise-truck-742'), ('x-goog-api-client', 'gl-python/3.8.6 grpc/1.39.0 gax/1.31.1 gccl/3.6.0')]
credentials = None, wait_for_ready = None, compression = None
def __call__(self,
request,
timeout=None,
metadata=None,
credentials=None,
wait_for_ready=None,
compression=None):
state, call, = self._blocking(request, timeout, metadata, credentials,
wait_for_ready, compression)
> return _end_unary_response_blocking(state, call, False, None)
.nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:946:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
state = <grpc._channel._RPCState object at 0x7fe78e5b66d0>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7fe79445bc00>
with_call = False, deadline = None
def _end_unary_response_blocking(state, call, with_call, deadline):
if state.code is grpc.StatusCode.OK:
if with_call:
rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
return state.response, rendezvous
else:
return state.response
else:
> raise _InactiveRpcError(state)
E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E status = StatusCode.ALREADY_EXISTS
E details = "Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627"
E debug_error_string = "{"created":"@1627550946.587652143","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1069,"grpc_message":"Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627","grpc_status":6}"
E >
.nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:849: _InactiveRpcError
The above exception was the direct cause of the following exception:
def setUpModule():
if USE_EMULATOR:
from google.auth.credentials import AnonymousCredentials
emulator_project = os.getenv("GCLOUD_PROJECT", "emulator-test-project")
Config.CLIENT = Client(
project=emulator_project, credentials=AnonymousCredentials()
)
else:
Config.CLIENT = Client()
retry = RetryErrors(exceptions.ServiceUnavailable)
configs = list(retry(Config.CLIENT.list_instance_configs)())
instances = retry(_list_instances)()
EXISTING_INSTANCES[:] = instances
# Delete test instances that are older than an hour.
cutoff = int(time.time()) - 1 * 60 * 60
for instance_pb in Config.CLIENT.list_instances(
"labels.python-spanner-dbapi-systests:true"
):
instance = Instance.from_pb(instance_pb, Config.CLIENT)
if "created" not in instance.labels:
continue
create_time = int(instance.labels["created"])
if create_time > cutoff:
continue
# Instance cannot be deleted while backups exist.
for backup_pb in instance.list_backups():
backup = Backup.from_pb(backup_pb, instance)
backup.delete()
instance.delete()
if CREATE_INSTANCE:
if not USE_EMULATOR:
# Defend against back-end returning configs for regions we aren't
# actually allowed to use.
configs = [config for config in configs if "-us-" in config.name]
if not configs:
raise ValueError("List instance configs failed in module set up.")
Config.INSTANCE_CONFIG = configs[0]
config_name = configs[0].name
create_time = str(int(time.time()))
labels = {"python-spanner-dbapi-systests": "true", "created": create_time}
Config.INSTANCE = Config.CLIENT.instance(
INSTANCE_ID, config_name, labels=labels
)
> created_op = Config.INSTANCE.create()
tests/system/test_system_dbapi.py:98:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/spanner_v1/instance.py:318: in create
future = api.create_instance(
google/cloud/spanner_admin_instance_v1/services/instance_admin/client.py:829: in create_instance
response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/gapic_v1/method.py:145: in __call__
return wrapped_func(*args, **kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = None
from_value = <_InactiveRpcError of RPC that terminated with:
status = StatusCode.ALREADY_EXISTS
details = "Instance already exist...message":"Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627","grpc_status":6}"
>
> ???
E google.api_core.exceptions.AlreadyExists: 409 Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627
<string>:3: AlreadyExists</pre></details>
|
1.0
|
tests.system.test_system_dbapi.TestTransactionsManagement: test_commit failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 2487800e31842a44dcc37937c325e130c8c926b0
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/306a2e02-87cb-4be9-be31-37456ec7a8a2), [Sponge](http://sponge2/306a2e02-87cb-4be9-be31-37456ec7a8a2)
status: failed
<details><summary>Test output</summary><br><pre>args = (parent: "projects/precise-truck-742"
instance_id: "google-cloud-1627550679627"
instance {
name: "projects/precise-t...1627550946"
}
labels {
key: "python-spanner-dbapi-systests"
value: "true"
}
processing_units: 1000
}
,)
kwargs = {'metadata': [('google-cloud-resource-prefix', 'projects/precise-truck-742/instances/google-cloud-1627550679627'), ('x...ms', 'parent=projects/precise-truck-742'), ('x-goog-api-client', 'gl-python/3.8.6 grpc/1.39.0 gax/1.31.1 gccl/3.6.0')]}
@six.wraps(callable_)
def error_remapped_callable(*args, **kwargs):
try:
> return callable_(*args, **kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:67:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7fe7942014f0>
request = parent: "projects/precise-truck-742"
instance_id: "google-cloud-1627550679627"
instance {
name: "projects/precise-tr... "1627550946"
}
labels {
key: "python-spanner-dbapi-systests"
value: "true"
}
processing_units: 1000
}
timeout = None
metadata = [('google-cloud-resource-prefix', 'projects/precise-truck-742/instances/google-cloud-1627550679627'), ('x-goog-request...ams', 'parent=projects/precise-truck-742'), ('x-goog-api-client', 'gl-python/3.8.6 grpc/1.39.0 gax/1.31.1 gccl/3.6.0')]
credentials = None, wait_for_ready = None, compression = None
def __call__(self,
request,
timeout=None,
metadata=None,
credentials=None,
wait_for_ready=None,
compression=None):
state, call, = self._blocking(request, timeout, metadata, credentials,
wait_for_ready, compression)
> return _end_unary_response_blocking(state, call, False, None)
.nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:946:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
state = <grpc._channel._RPCState object at 0x7fe78e5b66d0>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7fe79445bc00>
with_call = False, deadline = None
def _end_unary_response_blocking(state, call, with_call, deadline):
if state.code is grpc.StatusCode.OK:
if with_call:
rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
return state.response, rendezvous
else:
return state.response
else:
> raise _InactiveRpcError(state)
E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E status = StatusCode.ALREADY_EXISTS
E details = "Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627"
E debug_error_string = "{"created":"@1627550946.587652143","description":"Error received from peer ipv4:74.125.195.95:443","file":"src/core/lib/surface/call.cc","file_line":1069,"grpc_message":"Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627","grpc_status":6}"
E >
.nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:849: _InactiveRpcError
The above exception was the direct cause of the following exception:
def setUpModule():
if USE_EMULATOR:
from google.auth.credentials import AnonymousCredentials
emulator_project = os.getenv("GCLOUD_PROJECT", "emulator-test-project")
Config.CLIENT = Client(
project=emulator_project, credentials=AnonymousCredentials()
)
else:
Config.CLIENT = Client()
retry = RetryErrors(exceptions.ServiceUnavailable)
configs = list(retry(Config.CLIENT.list_instance_configs)())
instances = retry(_list_instances)()
EXISTING_INSTANCES[:] = instances
# Delete test instances that are older than an hour.
cutoff = int(time.time()) - 1 * 60 * 60
for instance_pb in Config.CLIENT.list_instances(
"labels.python-spanner-dbapi-systests:true"
):
instance = Instance.from_pb(instance_pb, Config.CLIENT)
if "created" not in instance.labels:
continue
create_time = int(instance.labels["created"])
if create_time > cutoff:
continue
# Instance cannot be deleted while backups exist.
for backup_pb in instance.list_backups():
backup = Backup.from_pb(backup_pb, instance)
backup.delete()
instance.delete()
if CREATE_INSTANCE:
if not USE_EMULATOR:
# Defend against back-end returning configs for regions we aren't
# actually allowed to use.
configs = [config for config in configs if "-us-" in config.name]
if not configs:
raise ValueError("List instance configs failed in module set up.")
Config.INSTANCE_CONFIG = configs[0]
config_name = configs[0].name
create_time = str(int(time.time()))
labels = {"python-spanner-dbapi-systests": "true", "created": create_time}
Config.INSTANCE = Config.CLIENT.instance(
INSTANCE_ID, config_name, labels=labels
)
> created_op = Config.INSTANCE.create()
tests/system/test_system_dbapi.py:98:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/spanner_v1/instance.py:318: in create
future = api.create_instance(
google/cloud/spanner_admin_instance_v1/services/instance_admin/client.py:829: in create_instance
response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/gapic_v1/method.py:145: in __call__
return wrapped_func(*args, **kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = None
from_value = <_InactiveRpcError of RPC that terminated with:
status = StatusCode.ALREADY_EXISTS
details = "Instance already exist...message":"Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627","grpc_status":6}"
>
> ???
E google.api_core.exceptions.AlreadyExists: 409 Instance already exists: projects/precise-truck-742/instances/google-cloud-1627550679627
<string>:3: AlreadyExists</pre></details>
|
process
|
tests system test system dbapi testtransactionsmanagement test commit failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output args parent projects precise truck instance id google cloud instance name projects precise t labels key python spanner dbapi systests value true processing units kwargs metadata six wraps callable def error remapped callable args kwargs try return callable args kwargs nox system lib site packages google api core grpc helpers py self request parent projects precise truck instance id google cloud instance name projects precise tr labels key python spanner dbapi systests value true processing units timeout none metadata credentials none wait for ready none compression none def call self request timeout none metadata none credentials none wait for ready none compression none state call self blocking request timeout metadata credentials wait for ready compression return end unary response blocking state call false none nox system lib site packages grpc channel py state call with call false deadline none def end unary response blocking state call with call deadline if state code is grpc statuscode ok if with call rendezvous multithreadedrendezvous state call none deadline return state response rendezvous else return state response else raise inactiverpcerror state e grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with e status statuscode already exists e details instance already exists projects precise truck instances google cloud e debug error string created description error received from peer file src core lib surface call cc file line grpc message instance already exists projects precise truck instances google cloud grpc status e nox system lib site packages grpc channel py inactiverpcerror the above exception was the direct cause of the following exception def setupmodule if use emulator from google auth credentials import anonymouscredentials emulator project os getenv gcloud project emulator test project config client client project emulator project credentials anonymouscredentials else config client client retry retryerrors exceptions serviceunavailable configs list retry config client list instance configs instances retry list instances existing instances instances delete test instances that are older than an hour cutoff int time time for instance pb in config client list instances labels python spanner dbapi systests true instance instance from pb instance pb config client if created not in instance labels continue create time int instance labels if create time cutoff continue instance cannot be deleted while backups exist for backup pb in instance list backups backup backup from pb backup pb instance backup delete instance delete if create instance if not use emulator defend against back end returning configs for regions we aren t actually allowed to use configs if not configs raise valueerror list instance configs failed in module set up config instance config configs config name configs name create time str int time time labels python spanner dbapi systests true created create time config instance config client instance instance id config name labels labels created op config instance create tests system test system dbapi py google cloud spanner instance py in create future api create instance google cloud spanner admin instance services instance admin client py in create instance response rpc request retry retry timeout timeout metadata metadata nox system lib site packages google api core gapic method py in call return wrapped func args kwargs nox system lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value inactiverpcerror of rpc that terminated with status statuscode already exists details instance already exist message instance already exists projects precise truck instances google cloud grpc status e google api core exceptions alreadyexists instance already exists projects precise truck instances google cloud alreadyexists
| 1
|
13,049
| 15,387,938,537
|
IssuesEvent
|
2021-03-03 10:08:57
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Introspection: Secret flag to disable guardrails
|
kind/feature process/candidate team/migrations tech/engines tech/typescript topic: introspection
|
Right now we have a bunch of guardrails for Introspection in place that make sure, we do not introspect things into our schema that do not pass validation or would not work in Prisma Client or Query Engine.
We should have a "secret" flag on `introspect` that let's you disable those guardrails, so we can observe and test the behavior if those guardrails are disabled. That way we could from time to time run Introspection CI with this flag enabled, and see what still breaks and what the effect of removing one of those guardrails would be.
|
1.0
|
Introspection: Secret flag to disable guardrails - Right now we have a bunch of guardrails for Introspection in place that make sure, we do not introspect things into our schema that do not pass validation or would not work in Prisma Client or Query Engine.
We should have a "secret" flag on `introspect` that let's you disable those guardrails, so we can observe and test the behavior if those guardrails are disabled. That way we could from time to time run Introspection CI with this flag enabled, and see what still breaks and what the effect of removing one of those guardrails would be.
|
process
|
introspection secret flag to disable guardrails right now we have a bunch of guardrails for introspection in place that make sure we do not introspect things into our schema that do not pass validation or would not work in prisma client or query engine we should have a secret flag on introspect that let s you disable those guardrails so we can observe and test the behavior if those guardrails are disabled that way we could from time to time run introspection ci with this flag enabled and see what still breaks and what the effect of removing one of those guardrails would be
| 1
|
11,611
| 14,478,980,146
|
IssuesEvent
|
2020-12-10 09:11:47
|
decidim/decidim
|
https://api.github.com/repos/decidim/decidim
|
closed
|
Statistics content block for Process Groups
|
contract: process-groups
|
Ref. PG02-4
**Is your feature request related to a problem? Please describe.**
As a visitor I want a quick way to see the activity of all the processes inside the PG.
**Describe the solution you'd like**
To have implemented Statistics for all the PG, with the sum of every one of the Resources/Components in every Participatory Process.
**Describe alternatives you've considered**
To use Metrics as they give more data and are more visual (with graphics), but the Metrics features not work well (they have issues with showing the last dates instead of what makes sense on a Participatory Process, such as taking in account the beginning and finishing of the PP).
To have the unique sum in the case of Participants, but that would be difficult to understand.
**Additional context**
Statistics on a Participatory Process:

**Does this issue could impact on users private data?**
No
**Acceptance criteria**
- [x] As a visitor I can see the Statistics for a PG in its homepage
- [x] As an administrator I want to be able to decide whether or not to display the statistics on this page.
|
1.0
|
Statistics content block for Process Groups - Ref. PG02-4
**Is your feature request related to a problem? Please describe.**
As a visitor I want a quick way to see the activity of all the processes inside the PG.
**Describe the solution you'd like**
To have implemented Statistics for all the PG, with the sum of every one of the Resources/Components in every Participatory Process.
**Describe alternatives you've considered**
To use Metrics as they give more data and are more visual (with graphics), but the Metrics features not work well (they have issues with showing the last dates instead of what makes sense on a Participatory Process, such as taking in account the beginning and finishing of the PP).
To have the unique sum in the case of Participants, but that would be difficult to understand.
**Additional context**
Statistics on a Participatory Process:

**Does this issue could impact on users private data?**
No
**Acceptance criteria**
- [x] As a visitor I can see the Statistics for a PG in its homepage
- [x] As an administrator I want to be able to decide whether or not to display the statistics on this page.
|
process
|
statistics content block for process groups ref is your feature request related to a problem please describe as a visitor i want a quick way to see the activity of all the processes inside the pg describe the solution you d like to have implemented statistics for all the pg with the sum of every one of the resources components in every participatory process describe alternatives you ve considered to use metrics as they give more data and are more visual with graphics but the metrics features not work well they have issues with showing the last dates instead of what makes sense on a participatory process such as taking in account the beginning and finishing of the pp to have the unique sum in the case of participants but that would be difficult to understand additional context statistics on a participatory process does this issue could impact on users private data no acceptance criteria as a visitor i can see the statistics for a pg in its homepage as an administrator i want to be able to decide whether or not to display the statistics on this page
| 1
|
6,834
| 9,977,591,930
|
IssuesEvent
|
2019-07-09 17:40:41
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
MP modulation by symbiont of host defense-related PCD
|
multi-species process
|
GO:0034053
modulation by symbiont of host defense-related programmed cell death
should be
is_a
regulation of
GO:0097300 programmed necrotic cell death
@CuzickA
|
1.0
|
MP modulation by symbiont of host defense-related PCD - GO:0034053
modulation by symbiont of host defense-related programmed cell death
should be
is_a
regulation of
GO:0097300 programmed necrotic cell death
@CuzickA
|
process
|
mp modulation by symbiont of host defense related pcd go modulation by symbiont of host defense related programmed cell death should be is a regulation of go programmed necrotic cell death cuzicka
| 1
|
8,544
| 11,717,252,876
|
IssuesEvent
|
2020-03-09 16:56:57
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
opened
|
AUTO BATCH - Support XML Requests
|
EPIC - Auto Batch Process :oncoming_automobile: HIGH PRIORITY :arrow_double_up:
|
Accenture require an XML interface to the `doc-index-updater` API.
**Acceptance Criteria**
- The API should check the `Content-Type` or `Accept` headers. If the header is `application/xml`, then we should deserialize body & serialize response assuming XML rather than JSON.
**Stages**
- [ ] Backlog
- [ ] Discovery
- [ ] Dev
- [ ] Review
- [ ] QA
- [ ] UAT
|
1.0
|
AUTO BATCH - Support XML Requests - Accenture require an XML interface to the `doc-index-updater` API.
**Acceptance Criteria**
- The API should check the `Content-Type` or `Accept` headers. If the header is `application/xml`, then we should deserialize body & serialize response assuming XML rather than JSON.
**Stages**
- [ ] Backlog
- [ ] Discovery
- [ ] Dev
- [ ] Review
- [ ] QA
- [ ] UAT
|
process
|
auto batch support xml requests accenture require an xml interface to the doc index updater api acceptance criteria the api should check the content type or accept headers if the header is application xml then we should deserialize body serialize response assuming xml rather than json stages backlog discovery dev review qa uat
| 1
|
10,020
| 3,082,097,318
|
IssuesEvent
|
2015-08-23 12:07:06
|
tntim96/JSCover
|
https://api.github.com/repos/tntim96/JSCover
|
closed
|
Syntax error bug in generated JavaScript
|
bug Fix applied - please re-test
|
When trying to run JSCover on a large JavaScript file I get an error in the browser developer console:
```
Uncaught SyntaxError: Unexpected token -
```
Here is the relevant block of code that JSCover generated, note the offending line that tries to create a JavaScript function with the (invalid) name `visit8876_-1_1`
``` javascript
function visit8878_59_15(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][15].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][14].init(25, 21, '0 < (f = this.data[n] >> u)');
function visit8877_59_14(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][14].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['-1'][1].init(23, 36, 'visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f))');
function visit8876_-1_1(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['-1'][1].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][13].init(14, 8, 'u < this.A');
function visit8875_59_13(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][13].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][12].init(218, 46, 'visit8870_59_8(u < this.A) && visit8871_59_9(visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f)))');
function visit8874_59_12(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][12].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][11].init(204, 1, 'visit8869_59_7(visit8870_59_8(u < this.A) && visit8871_59_9(visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f))))');
function visit8873_59_11(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][11].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][10].init(228, 21, '0 < (f = this.data[n] >> u)');
```
If you would like to see this error for yourself, here is a link to the HTML page containing the JavaScript that triggers this bug: https://bitc.github.io/tojs-benchmark/particles-ghcjs/index.min.html?barriers=l5&particles=500
|
1.0
|
Syntax error bug in generated JavaScript - When trying to run JSCover on a large JavaScript file I get an error in the browser developer console:
```
Uncaught SyntaxError: Unexpected token -
```
Here is the relevant block of code that JSCover generated, note the offending line that tries to create a JavaScript function with the (invalid) name `visit8876_-1_1`
``` javascript
function visit8878_59_15(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][15].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][14].init(25, 21, '0 < (f = this.data[n] >> u)');
function visit8877_59_14(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][14].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['-1'][1].init(23, 36, 'visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f))');
function visit8876_-1_1(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['-1'][1].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][13].init(14, 8, 'u < this.A');
function visit8875_59_13(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][13].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][12].init(218, 46, 'visit8870_59_8(u < this.A) && visit8871_59_9(visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f)))');
function visit8874_59_12(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][12].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][11].init(204, 1, 'visit8869_59_7(visit8870_59_8(u < this.A) && visit8871_59_9(visit8872_59_10(0 < (f = this.data[n] >> u)) && (k = !0 , l = m(f))))');
function visit8873_59_11(result) {
_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][11].ranCondition(result);
return result;
}_$jscoverage['/particles-ghcjs/_build/all.min.js'].branchData['59'][10].init(228, 21, '0 < (f = this.data[n] >> u)');
```
If you would like to see this error for yourself, here is a link to the HTML page containing the JavaScript that triggers this bug: https://bitc.github.io/tojs-benchmark/particles-ghcjs/index.min.html?barriers=l5&particles=500
|
non_process
|
syntax error bug in generated javascript when trying to run jscover on a large javascript file i get an error in the browser developer console uncaught syntaxerror unexpected token here is the relevant block of code that jscover generated note the offending line that tries to create a javascript function with the invalid name javascript function result jscoverage branchdata rancondition result return result jscoverage branchdata init u function result jscoverage branchdata rancondition result return result jscoverage branchdata init u k l m f function result jscoverage branchdata rancondition result return result jscoverage branchdata init u this a function result jscoverage branchdata rancondition result return result jscoverage branchdata init u u k l m f function result jscoverage branchdata rancondition result return result jscoverage branchdata init u u k l m f function result jscoverage branchdata rancondition result return result jscoverage branchdata init u if you would like to see this error for yourself here is a link to the html page containing the javascript that triggers this bug
| 0
|
4,317
| 2,849,431,231
|
IssuesEvent
|
2015-05-30 17:46:13
|
SoftverInzenjeringETFSA/SI2014Tim7
|
https://api.github.com/repos/SoftverInzenjeringETFSA/SI2014Tim7
|
closed
|
Nije prikazana klasa Ocitanja u Class diagramu
|
documentation wontfix
|
Komponenta: JKP Vodovod i Kanalizacija desktop
Commit ID: bb0e5cac45f01fadbfae2afe4829398aa4b231ae
Opis problema:
Class diagram iz dokumenta: https://github.com/SoftverInzenjeringETFSA/SI2014Tim7/blob/master/Dizajn%20Sistema.pdf
ne sadrzi klasu Ocitanja kao sto je to uradjeno u sistemu:
http://imgur.com/J02RFWu
|
1.0
|
Nije prikazana klasa Ocitanja u Class diagramu - Komponenta: JKP Vodovod i Kanalizacija desktop
Commit ID: bb0e5cac45f01fadbfae2afe4829398aa4b231ae
Opis problema:
Class diagram iz dokumenta: https://github.com/SoftverInzenjeringETFSA/SI2014Tim7/blob/master/Dizajn%20Sistema.pdf
ne sadrzi klasu Ocitanja kao sto je to uradjeno u sistemu:
http://imgur.com/J02RFWu
|
non_process
|
nije prikazana klasa ocitanja u class diagramu komponenta jkp vodovod i kanalizacija desktop commit id opis problema class diagram iz dokumenta ne sadrzi klasu ocitanja kao sto je to uradjeno u sistemu
| 0
|
237,253
| 18,154,763,019
|
IssuesEvent
|
2021-09-26 21:52:00
|
secheaper/slash
|
https://api.github.com/repos/secheaper/slash
|
opened
|
Add section links to README
|
documentation enhancement good first issue
|
Add links that scrolls you to particular sections in the README files. Something like a legend
|
1.0
|
Add section links to README - Add links that scrolls you to particular sections in the README files. Something like a legend
|
non_process
|
add section links to readme add links that scrolls you to particular sections in the readme files something like a legend
| 0
|
254,988
| 27,484,688,374
|
IssuesEvent
|
2023-03-04 01:08:20
|
panasalap/linux-4.1.15
|
https://api.github.com/repos/panasalap/linux-4.1.15
|
opened
|
CVE-2018-1120 (Medium) detected in linux-yocto-devv4.2.8
|
security vulnerability
|
## CVE-2018-1120 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary>
<p>
<p>Linux Embedded Kernel - tracks the next mainline release</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/gup.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/gup.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found affecting the Linux kernel before version 4.17. By mmap()ing a FUSE-backed file onto a process's memory containing command line arguments (or environment strings), an attacker can cause utilities from psutils or procps (such as ps, w) or any other program which makes a read() call to the /proc/<pid>/cmdline (or /proc/<pid>/environ) files to block indefinitely (denial of service) or for some controlled time (as a synchronization primitive for other attacks).
<p>Publish Date: 2018-06-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-1120>CVE-2018-1120</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-1120">https://nvd.nist.gov/vuln/detail/CVE-2018-1120</a></p>
<p>Release Date: 2018-06-20</p>
<p>Fix Resolution: 4.17</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-1120 (Medium) detected in linux-yocto-devv4.2.8 - ## CVE-2018-1120 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv4.2.8</b></p></summary>
<p>
<p>Linux Embedded Kernel - tracks the next mainline release</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/gup.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/mm/gup.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found affecting the Linux kernel before version 4.17. By mmap()ing a FUSE-backed file onto a process's memory containing command line arguments (or environment strings), an attacker can cause utilities from psutils or procps (such as ps, w) or any other program which makes a read() call to the /proc/<pid>/cmdline (or /proc/<pid>/environ) files to block indefinitely (denial of service) or for some controlled time (as a synchronization primitive for other attacks).
<p>Publish Date: 2018-06-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-1120>CVE-2018-1120</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-1120">https://nvd.nist.gov/vuln/detail/CVE-2018-1120</a></p>
<p>Release Date: 2018-06-20</p>
<p>Fix Resolution: 4.17</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux yocto cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel tracks the next mainline release library home page a href found in base branch master vulnerable source files mm gup c mm gup c vulnerability details a flaw was found affecting the linux kernel before version by mmap ing a fuse backed file onto a process s memory containing command line arguments or environment strings an attacker can cause utilities from psutils or procps such as ps w or any other program which makes a read call to the proc cmdline or proc environ files to block indefinitely denial of service or for some controlled time as a synchronization primitive for other attacks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
231,591
| 17,694,531,343
|
IssuesEvent
|
2021-08-24 13:59:06
|
MeisterOfPizza/file-variants
|
https://api.github.com/repos/MeisterOfPizza/file-variants
|
opened
|
Multiple file-variants of the same variant inside input
|
documentation enhancement
|
Output will be however many file-variants of variant found.
file-variant naming: <variant>(|.<part>)*.<extension>
outputName: "{name}.{part0}.{part1}" (extension is appended automatically)
|
1.0
|
Multiple file-variants of the same variant inside input - Output will be however many file-variants of variant found.
file-variant naming: <variant>(|.<part>)*.<extension>
outputName: "{name}.{part0}.{part1}" (extension is appended automatically)
|
non_process
|
multiple file variants of the same variant inside input output will be however many file variants of variant found file variant naming outputname name extension is appended automatically
| 0
|
227,599
| 18,073,015,186
|
IssuesEvent
|
2021-09-21 06:29:02
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
roachtest: sqlsmith/setup=empty/setting=no-mutations failed
|
C-test-failure O-robot O-roachtest branch-master release-blocker
|
roachtest.sqlsmith/setup=empty/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3473201&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3473201&tab=artifacts#/sqlsmith/setup=empty/setting=no-mutations) on master @ [676dc076dc6a137d88eac9217764bbb63beae86f](https://github.com/cockroachdb/cockroach/commits/676dc076dc6a137d88eac9217764bbb63beae86f):
```
('17:28:14.764522+14:43:00':::TIMETZ),
('07:23:03.812281+14:50:00':::TIMETZ),
(
parse_timetz(COALESCE(st_astext('01030000C0010000000600000070EBF2E3C6B0FCC12418761831E2FEC1BEF7A174620F014290AE7E7800F10042A0E5BFE035CDF8C1FBF9AEBA467F00C2955B7F1FADF900C22E37DC6732D7F9C1F8291D6E8EC7F941342DD7B74BECF6C12CDCD92FA44FFE415C8813B08A060142CA9A4CD716F2FAC160827E0A3AD0E441C2A16B9537A6F441E8E10023E840FF41BC4F7B62083AFBC1009D25D3324EECC1647CD166902AFB418C99946456D0014270EBF2E3C6B0FCC12418761831E2FEC1BEF7A174620F014290AE7E7800F10042':::GEOMETRY::GEOMETRY)::STRING, e'\x1c':::STRING)::STRING)::TIMETZ
),
('04:50:55.334281-00:35:00':::TIMETZ)
)
AS tab_97115 (col_167897)
LEFT JOIN (
VALUES
('ffc2e94c-474a-45b0-933c-cf76cc75bce3':::UUID),
('2c6a5785-6910-4180-a14e-71319f6862a5':::UUID),
(uuid_generate_v4()::UUID),
('203ac89f-0554-4aed-8f55-e9259898b6a9':::UUID),
(NULL)
)
AS tab_97116 (col_167898) ON false,
(
VALUES
('ffff:ffff:ffff:ffff:ffff:ffff:ffff:ffff':::INET),
('7a5f:be04:ed4a:c9ca:1e60:169f:8a5f:2ab/108':::INET)
)
AS tab_97117 (col_167899)
ORDER BY
tab_97117.col_167899, tab_97117.col_167899, tab_97117.col_167899
LIMIT
1:::INT8
)
),
('0104000060E610000000000000':::GEOGRAPHY, NULL)
)
AS tab_97118 (col_167901, col_167902)
)
SELECT
tab_97119.col_167904 AS col_167905
FROM
(
VALUES
(NULL),
('20 years 181 days 06:29:21.419479':::INTERVAL),
('59 years 524 days 02:49:03.527908':::INTERVAL),
('-51 years -5 mons -748 days -06:39:54.748217':::INTERVAL),
(parse_interval(e',1-\x04Bh``':::STRING::STRING, e'\x05\x04%R!)EbC':::STRING::STRING)::INTERVAL),
('-46 years -2 mons -829 days -13:15:27.684704':::INTERVAL)
)
AS tab_97119 (col_167904)
WHERE
true
LIMIT
87:::INT8;
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=empty/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: sqlsmith/setup=empty/setting=no-mutations failed - roachtest.sqlsmith/setup=empty/setting=no-mutations [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3473201&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3473201&tab=artifacts#/sqlsmith/setup=empty/setting=no-mutations) on master @ [676dc076dc6a137d88eac9217764bbb63beae86f](https://github.com/cockroachdb/cockroach/commits/676dc076dc6a137d88eac9217764bbb63beae86f):
```
('17:28:14.764522+14:43:00':::TIMETZ),
('07:23:03.812281+14:50:00':::TIMETZ),
(
parse_timetz(COALESCE(st_astext('01030000C0010000000600000070EBF2E3C6B0FCC12418761831E2FEC1BEF7A174620F014290AE7E7800F10042A0E5BFE035CDF8C1FBF9AEBA467F00C2955B7F1FADF900C22E37DC6732D7F9C1F8291D6E8EC7F941342DD7B74BECF6C12CDCD92FA44FFE415C8813B08A060142CA9A4CD716F2FAC160827E0A3AD0E441C2A16B9537A6F441E8E10023E840FF41BC4F7B62083AFBC1009D25D3324EECC1647CD166902AFB418C99946456D0014270EBF2E3C6B0FCC12418761831E2FEC1BEF7A174620F014290AE7E7800F10042':::GEOMETRY::GEOMETRY)::STRING, e'\x1c':::STRING)::STRING)::TIMETZ
),
('04:50:55.334281-00:35:00':::TIMETZ)
)
AS tab_97115 (col_167897)
LEFT JOIN (
VALUES
('ffc2e94c-474a-45b0-933c-cf76cc75bce3':::UUID),
('2c6a5785-6910-4180-a14e-71319f6862a5':::UUID),
(uuid_generate_v4()::UUID),
('203ac89f-0554-4aed-8f55-e9259898b6a9':::UUID),
(NULL)
)
AS tab_97116 (col_167898) ON false,
(
VALUES
('ffff:ffff:ffff:ffff:ffff:ffff:ffff:ffff':::INET),
('7a5f:be04:ed4a:c9ca:1e60:169f:8a5f:2ab/108':::INET)
)
AS tab_97117 (col_167899)
ORDER BY
tab_97117.col_167899, tab_97117.col_167899, tab_97117.col_167899
LIMIT
1:::INT8
)
),
('0104000060E610000000000000':::GEOGRAPHY, NULL)
)
AS tab_97118 (col_167901, col_167902)
)
SELECT
tab_97119.col_167904 AS col_167905
FROM
(
VALUES
(NULL),
('20 years 181 days 06:29:21.419479':::INTERVAL),
('59 years 524 days 02:49:03.527908':::INTERVAL),
('-51 years -5 mons -748 days -06:39:54.748217':::INTERVAL),
(parse_interval(e',1-\x04Bh``':::STRING::STRING, e'\x05\x04%R!)EbC':::STRING::STRING)::INTERVAL),
('-46 years -2 mons -829 days -13:15:27.684704':::INTERVAL)
)
AS tab_97119 (col_167904)
WHERE
true
LIMIT
87:::INT8;
```
<details><summary>Reproduce</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=empty/setting=no-mutations.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
non_process
|
roachtest sqlsmith setup empty setting no mutations failed roachtest sqlsmith setup empty setting no mutations with on master timetz timetz parse timetz coalesce st astext geometry geometry string e string string timetz timetz as tab col left join values uuid uuid uuid generate uuid uuid null as tab col on false values ffff ffff ffff ffff ffff ffff ffff ffff inet inet as tab col order by tab col tab col tab col limit geography null as tab col col select tab col as col from values null years days interval years days interval years mons days interval parse interval e string string e r ebc string string interval years mons days interval as tab col where true limit reproduce see cc cockroachdb sql queries
| 0
|
157,455
| 12,374,004,927
|
IssuesEvent
|
2020-05-19 00:12:02
|
rancher/dashboard
|
https://api.github.com/repos/rancher/dashboard
|
closed
|
Create buttons are missing for workload
|
[zube]: To Test kind/bug
|
Version: master-head(1f4a2c388)
Steps:
1. Navigate to the workload section from the dashboard UI
"Create from Yaml" and "Create" buttons are missing

|
1.0
|
Create buttons are missing for workload - Version: master-head(1f4a2c388)
Steps:
1. Navigate to the workload section from the dashboard UI
"Create from Yaml" and "Create" buttons are missing

|
non_process
|
create buttons are missing for workload version master head steps navigate to the workload section from the dashboard ui create from yaml and create buttons are missing
| 0
|
4,596
| 7,434,937,657
|
IssuesEvent
|
2018-03-26 12:48:31
|
decidim/decidim
|
https://api.github.com/repos/decidim/decidim
|
closed
|
Title and subtitle of a participatory process are required at model level but not at form level
|
space: processes type: bug
|
<!--
1. Please check if an issue already exists so there are no duplicates
2. Fill out the whole template so we have a good overview on the issue
3. Do not remove any section of the template. If something is not applicable leave it empty but leave it in the Issue
4. Please follow the template, otherwise we'll have to ask you to update it
-->
# This is a Bug Report
#### :tophat: Description
For bug reports:
* What went wrong?
Just a confusing behaviour when you create a participatory process (and it's maybe the case for other elements), the form doesn't have a 'required' asterisk for the title and subtitle fields. But when you try to save the element without the subtitle, you get an error message.
* What did you expect should have happened?
Add an asterisk on all the required fields
* What was the config you used?
Standard
* What stacktrace or error message from your provider did you see?
can't be blank
For feature proposals:
* What is the use case that should be solved. The more detail you describe this in the easier it is to understand for us.
* If there is additional config how would it look
#### :pushpin: Related issues
#3054 #3056
#### :clipboard: Additional Data
* ***Decidim deployment where you found the issue***:
* ***Browser & version***:
Chrome latest
* ***Screenshot***:

* ***Error messages***:
can't be blank
* ***URL to reproduce the error***:
localhost:3000/admin/participatory_processes/new
|
1.0
|
Title and subtitle of a participatory process are required at model level but not at form level - <!--
1. Please check if an issue already exists so there are no duplicates
2. Fill out the whole template so we have a good overview on the issue
3. Do not remove any section of the template. If something is not applicable leave it empty but leave it in the Issue
4. Please follow the template, otherwise we'll have to ask you to update it
-->
# This is a Bug Report
#### :tophat: Description
For bug reports:
* What went wrong?
Just a confusing behaviour when you create a participatory process (and it's maybe the case for other elements), the form doesn't have a 'required' asterisk for the title and subtitle fields. But when you try to save the element without the subtitle, you get an error message.
* What did you expect should have happened?
Add an asterisk on all the required fields
* What was the config you used?
Standard
* What stacktrace or error message from your provider did you see?
can't be blank
For feature proposals:
* What is the use case that should be solved. The more detail you describe this in the easier it is to understand for us.
* If there is additional config how would it look
#### :pushpin: Related issues
#3054 #3056
#### :clipboard: Additional Data
* ***Decidim deployment where you found the issue***:
* ***Browser & version***:
Chrome latest
* ***Screenshot***:

* ***Error messages***:
can't be blank
* ***URL to reproduce the error***:
localhost:3000/admin/participatory_processes/new
|
process
|
title and subtitle of a participatory process are required at model level but not at form level please check if an issue already exists so there are no duplicates fill out the whole template so we have a good overview on the issue do not remove any section of the template if something is not applicable leave it empty but leave it in the issue please follow the template otherwise we ll have to ask you to update it this is a bug report tophat description for bug reports what went wrong just a confusing behaviour when you create a participatory process and it s maybe the case for other elements the form doesn t have a required asterisk for the title and subtitle fields but when you try to save the element without the subtitle you get an error message what did you expect should have happened add an asterisk on all the required fields what was the config you used standard what stacktrace or error message from your provider did you see can t be blank for feature proposals what is the use case that should be solved the more detail you describe this in the easier it is to understand for us if there is additional config how would it look pushpin related issues clipboard additional data decidim deployment where you found the issue browser version chrome latest screenshot error messages can t be blank url to reproduce the error localhost admin participatory processes new
| 1
|
255,981
| 19,349,336,417
|
IssuesEvent
|
2021-12-15 14:12:48
|
NOAA-EMC/NCEPLIBS-grib_util
|
https://api.github.com/repos/NOAA-EMC/NCEPLIBS-grib_util
|
closed
|
update README to include authors of each utility
|
documentation
|
update README to include authors of each utility
|
1.0
|
update README to include authors of each utility - update README to include authors of each utility
|
non_process
|
update readme to include authors of each utility update readme to include authors of each utility
| 0
|
163,172
| 6,192,404,266
|
IssuesEvent
|
2017-07-05 01:33:44
|
ykoblick/Shabbat-Issue-Tracker
|
https://api.github.com/repos/ykoblick/Shabbat-Issue-Tracker
|
closed
|
About me being erased
|
app High Priority website
|
The "about me" message in my profile was deleted. I asked R. Klatzko about it and he said he doesn't know how it happened
|
1.0
|
About me being erased - The "about me" message in my profile was deleted. I asked R. Klatzko about it and he said he doesn't know how it happened
|
non_process
|
about me being erased the about me message in my profile was deleted i asked r klatzko about it and he said he doesn t know how it happened
| 0
|
38,867
| 10,257,960,274
|
IssuesEvent
|
2019-08-21 21:26:04
|
tmrowco/electricitymap-contrib
|
https://api.github.com/repos/tmrowco/electricitymap-contrib
|
opened
|
USA hourly generation by fuel type is now available
|
parser parser buildable!
|
The EIA have just released a new visualization for the USA Electric Operating System. One of the graphs is hourly generation by fuel type for the lower 48 states! From a quick glance it appears to lag behind by about a day but it's still hugely useful! :tada:
https://www.eia.gov/beta/electricity/gridmonitor/dashboard/electric_overview/US48/US48
It also seems like you can drill down into individual balancing authorities and get the exact same data (hourly by type D-1), for instance checkout [Tennessee Valley Authority](https://www.eia.gov/beta/electricity/gridmonitor/dashboard/electric_overview/balancing_authority/TVA)
|
1.0
|
USA hourly generation by fuel type is now available - The EIA have just released a new visualization for the USA Electric Operating System. One of the graphs is hourly generation by fuel type for the lower 48 states! From a quick glance it appears to lag behind by about a day but it's still hugely useful! :tada:
https://www.eia.gov/beta/electricity/gridmonitor/dashboard/electric_overview/US48/US48
It also seems like you can drill down into individual balancing authorities and get the exact same data (hourly by type D-1), for instance checkout [Tennessee Valley Authority](https://www.eia.gov/beta/electricity/gridmonitor/dashboard/electric_overview/balancing_authority/TVA)
|
non_process
|
usa hourly generation by fuel type is now available the eia have just released a new visualization for the usa electric operating system one of the graphs is hourly generation by fuel type for the lower states from a quick glance it appears to lag behind by about a day but it s still hugely useful tada it also seems like you can drill down into individual balancing authorities and get the exact same data hourly by type d for instance checkout
| 0
|
4,157
| 7,103,958,236
|
IssuesEvent
|
2018-01-16 08:09:19
|
Incubaid/crm
|
https://api.github.com/repos/Incubaid/crm
|
closed
|
FR: show node status of member
|
process_wontfix
|
At the moment there's no way to see the node status of a members node. Need to have this before PTO.
|
1.0
|
FR: show node status of member - At the moment there's no way to see the node status of a members node. Need to have this before PTO.
|
process
|
fr show node status of member at the moment there s no way to see the node status of a members node need to have this before pto
| 1
|
323,046
| 9,842,352,132
|
IssuesEvent
|
2019-06-18 09:05:57
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.twitch.tv - see bug description
|
browser-firefox engine-gecko priority-critical type-webrender-enabled
|
<!-- @browser: Firefox 69.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @extra_labels: type-webrender-enabled -->
**URL**: https://www.twitch.tv/riotgames
**Browser / Version**: Firefox 69.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: stream is stuttery even in safe mode
**Steps to Reproduce**:
1. click on a stream to watch it
[](https://webcompat.com/uploads/2019/6/8b3acc0c-5043-4c46-a0e1-674d45f58917.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190614100022</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: true</li><li>gfx.webrender.all: true</li><li>channel: nightly</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "Loading failed for the <script> with source https://c.amazon-adsystem.com/aax2/apstag.js." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://c.amazon-adsystem.com/aax2/apstag.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://www.gstatic.com/cv/js/sender/v1/cast_sender.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://cdn-gl.imrworldwide.com/conf/P34D56F9D-5684-4C83-8EE1-5EA7DE9CF45D.js#name=1a0UrMElc2nTFhA32wPD2WhXYc6RFiFi&ns=NOLBUNDLE." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://client-event-reporter.twitch.tv/v1/stats. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Request to access cookie or storage on https://cdn-gl.imrworldwide.com/conf/P34D56F9D-5684-4C83-8EE1-5EA7DE9CF45D.js#name=1a0UrMElc2nTFhA32wPD2WhXYc6RFiFi&ns=NOLBUNDLE was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://imasdk.googleapis.com/js/sdkloader/ima3.js." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted." {file: "https://static.twitchcdn.net/assets/player-core-base-e939ccadccca15e51973.js" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://s.amazon-adsystem.com/iui3?d=3p-hbg&ex-src=twitch.tv&ex-hargs=v%3D1.0%3Bc%3D8858214122683%3Bp%3De75425fb-5407-7bd5-fd20-f462e98a8777 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://cdn.krxd.net/controltag?confid=JYk8LCsC was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://ssl.google-analytics.com/ga.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/beacon.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted." {file: "https://static.twitchcdn.net/assets/player-core-base-e939ccadccca15e51973.js" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/p?c1=2&c2=6745306&ns_type=hidden&ns_st_sv=5.1.3.160420&ns_st_smv=5.1&ns_st_it=r&ns_st_id=1560539414288&ns_st_ec=1&ns_st_sp=1&ns_st_sc=1&ns_st_sq=1&ns_st_ppc=1&ns_st_apc=1&ns_st_spc=1&ns_st_cn=1&ns_st_ev=play&ns_st_po=0&ns_st_cl=0&ns_st_pb=1&ns_st_mp=js_api&ns_st_mv=5.1.3.160420&ns_st_pn=1&ns_st_tp=0&ns_st_ci=34526695408&ns_st_pt=0&ns_st_dpt=0&ns_st_ipt=0&ns_st_et=0&ns_st_det=0&ns_st_upc=0&ns_st_dupc=0&ns_st_iupc=0&ns_st_upa=0&ns_st_dupa=0&ns_st_iupa=0&ns_st_lpc=0&ns_st_dlpc=0&ns_st_lpa=0&ns_st_dlpa=0&ns_st_pa=0&ns_ts=1560539416234&ns_st_bc=0&ns_st_dbc=0&ns_st_bt=0&ns_st_dbt=0&ns_st_bp=0&ns_st_lt=1946&ns_st_skc=0&ns_st_dskc=0&ns_st_ska=0&ns_st_dska=0&ns_st_skd=0&ns_st_skt=0&ns_st_dskt=0&ns_st_pc=0&ns_st_dpc=0&ns_st_pp=0&ns_st_br=0&ns_st_ub=0&ns_st_ki=1200000&ns_st_pr=*null&ns_st_sn=*null&ns_st_en=*null&ns_st_ep=Week%202%20Day%201%20%7C%20LEC%20Summer%20Split%20(2019)%20-%20Splyce%20vs.%20Schalke%2004&ns_st_ct=vc13&ns_st_ge=Gaming&ns_st_st=Riot%20Games&ns_st_ce=1&ns_st_ia=0&ns_st_ddt=2019-06-14&ns_st_tdt=*null&ns_st_pu=*null&ns_st_ti=*null&c3=TWITCH&c4=twitch.tv&c6=*null&c7=https%3A%2F%2Fwww.twitch.tv%2Friotgames&c8=Riot%20Games%20-%20Twitch&c9=&cs_ucfr=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/p?c1=2&c2=6745306&ns_type=hidden&ns_st_sv=5.1.3.160420&ns_st_smv=5.1&ns_st_it=r&ns_st_id=1560539414288&ns_st_ec=2&ns_st_sp=1&ns_st_sc=1&ns_st_sq=1&ns_st_ppc=1&ns_st_apc=1&ns_st_spc=1&ns_st_cn=1&ns_st_ev=hb&ns_st_po=10003&ns_st_cl=0&ns_st_hc=1&ns_st_mp=js_api&ns_st_mv=5.1.3.160420&ns_st_pn=1&ns_st_tp=0&ns_st_ci=34526695408&ns_st_pt=10003&ns_st_dpt=10003&ns_st_ipt=10003&ns_st_et=10003&ns_st_det=10003&ns_st_upc=10003&ns_st_dupc=10003&ns_st_iupc=10003&ns_st_upa=10003&ns_st_dupa=10003&ns_st_iupa=10003&ns_st_lpc=10003&ns_st_dlpc=10003&ns_st_lpa=10003&ns_st_dlpa=10003&ns_st_pa=10003&ns_ts=1560539426237&ns_st_bc=0&ns_st_dbc=0&ns_st_bt=0&ns_st_dbt=0&ns_st_bp=0&ns_st_skc=0&ns_st_dskc=0&ns_st_ska=0&ns_st_dska=0&ns_st_skd=0&ns_st_skt=0&ns_st_dskt=0&ns_st_pc=0&ns_st_dpc=0&ns_st_pp=0&ns_st_br=0&ns_st_ub=0&ns_st_ki=1200000&ns_st_pr=*null&ns_st_sn=*null&ns_st_en=*null&ns_st_ep=Week%202%20Day%201%20%7C%20LEC%20Summer%20Split%20(2019)%20-%20Splyce%20vs.%20Schalke%2004&ns_st_ct=vc13&ns_st_ge=Gaming&ns_st_st=Riot%20Games&ns_st_ce=1&ns_st_ia=0&ns_st_ddt=2019-06-14&ns_st_tdt=*null&ns_st_pu=*null&ns_st_ti=*null&c3=TWITCH&c4=twitch.tv&c6=*null&c7=https%3A%2F%2Fwww.twitch.tv%2Friotgames&c8=Riot%20Games%20-%20Twitch&c9=&cs_ucfr=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.twitch.tv - see bug description - <!-- @browser: Firefox 69.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:69.0) Gecko/20100101 Firefox/69.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @extra_labels: type-webrender-enabled -->
**URL**: https://www.twitch.tv/riotgames
**Browser / Version**: Firefox 69.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: stream is stuttery even in safe mode
**Steps to Reproduce**:
1. click on a stream to watch it
[](https://webcompat.com/uploads/2019/6/8b3acc0c-5043-4c46-a0e1-674d45f58917.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190614100022</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: true</li><li>gfx.webrender.all: true</li><li>channel: nightly</li>
</ul>
<p>Console Messages:</p>
<pre>
[u'[JavaScript Warning: "Loading failed for the <script> with source https://c.amazon-adsystem.com/aax2/apstag.js." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://c.amazon-adsystem.com/aax2/apstag.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://www.gstatic.com/cv/js/sender/v1/cast_sender.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://cdn-gl.imrworldwide.com/conf/P34D56F9D-5684-4C83-8EE1-5EA7DE9CF45D.js#name=1a0UrMElc2nTFhA32wPD2WhXYc6RFiFi&ns=NOLBUNDLE." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://client-event-reporter.twitch.tv/v1/stats. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Request to access cookie or storage on https://cdn-gl.imrworldwide.com/conf/P34D56F9D-5684-4C83-8EE1-5EA7DE9CF45D.js#name=1a0UrMElc2nTFhA32wPD2WhXYc6RFiFi&ns=NOLBUNDLE was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Loading failed for the <script> with source https://imasdk.googleapis.com/js/sdkloader/ima3.js." {file: "https://www.twitch.tv/riotgames" line: 1}]', u'[JavaScript Warning: "Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted." {file: "https://static.twitchcdn.net/assets/player-core-base-e939ccadccca15e51973.js" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://s.amazon-adsystem.com/iui3?d=3p-hbg&ex-src=twitch.tv&ex-hargs=v%3D1.0%3Bc%3D8858214122683%3Bp%3De75425fb-5407-7bd5-fd20-f462e98a8777 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://cdn.krxd.net/controltag?confid=JYk8LCsC was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://ssl.google-analytics.com/ga.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/beacon.js was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Autoplay is only allowed when approved by the user, the site is activated by the user, or media is muted." {file: "https://static.twitchcdn.net/assets/player-core-base-e939ccadccca15e51973.js" line: 1}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/p?c1=2&c2=6745306&ns_type=hidden&ns_st_sv=5.1.3.160420&ns_st_smv=5.1&ns_st_it=r&ns_st_id=1560539414288&ns_st_ec=1&ns_st_sp=1&ns_st_sc=1&ns_st_sq=1&ns_st_ppc=1&ns_st_apc=1&ns_st_spc=1&ns_st_cn=1&ns_st_ev=play&ns_st_po=0&ns_st_cl=0&ns_st_pb=1&ns_st_mp=js_api&ns_st_mv=5.1.3.160420&ns_st_pn=1&ns_st_tp=0&ns_st_ci=34526695408&ns_st_pt=0&ns_st_dpt=0&ns_st_ipt=0&ns_st_et=0&ns_st_det=0&ns_st_upc=0&ns_st_dupc=0&ns_st_iupc=0&ns_st_upa=0&ns_st_dupa=0&ns_st_iupa=0&ns_st_lpc=0&ns_st_dlpc=0&ns_st_lpa=0&ns_st_dlpa=0&ns_st_pa=0&ns_ts=1560539416234&ns_st_bc=0&ns_st_dbc=0&ns_st_bt=0&ns_st_dbt=0&ns_st_bp=0&ns_st_lt=1946&ns_st_skc=0&ns_st_dskc=0&ns_st_ska=0&ns_st_dska=0&ns_st_skd=0&ns_st_skt=0&ns_st_dskt=0&ns_st_pc=0&ns_st_dpc=0&ns_st_pp=0&ns_st_br=0&ns_st_ub=0&ns_st_ki=1200000&ns_st_pr=*null&ns_st_sn=*null&ns_st_en=*null&ns_st_ep=Week%202%20Day%201%20%7C%20LEC%20Summer%20Split%20(2019)%20-%20Splyce%20vs.%20Schalke%2004&ns_st_ct=vc13&ns_st_ge=Gaming&ns_st_st=Riot%20Games&ns_st_ce=1&ns_st_ia=0&ns_st_ddt=2019-06-14&ns_st_tdt=*null&ns_st_pu=*null&ns_st_ti=*null&c3=TWITCH&c4=twitch.tv&c6=*null&c7=https%3A%2F%2Fwww.twitch.tv%2Friotgames&c8=Riot%20Games%20-%20Twitch&c9=&cs_ucfr=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]', u'[JavaScript Warning: "Request to access cookie or storage on https://sb.scorecardresearch.com/p?c1=2&c2=6745306&ns_type=hidden&ns_st_sv=5.1.3.160420&ns_st_smv=5.1&ns_st_it=r&ns_st_id=1560539414288&ns_st_ec=2&ns_st_sp=1&ns_st_sc=1&ns_st_sq=1&ns_st_ppc=1&ns_st_apc=1&ns_st_spc=1&ns_st_cn=1&ns_st_ev=hb&ns_st_po=10003&ns_st_cl=0&ns_st_hc=1&ns_st_mp=js_api&ns_st_mv=5.1.3.160420&ns_st_pn=1&ns_st_tp=0&ns_st_ci=34526695408&ns_st_pt=10003&ns_st_dpt=10003&ns_st_ipt=10003&ns_st_et=10003&ns_st_det=10003&ns_st_upc=10003&ns_st_dupc=10003&ns_st_iupc=10003&ns_st_upa=10003&ns_st_dupa=10003&ns_st_iupa=10003&ns_st_lpc=10003&ns_st_dlpc=10003&ns_st_lpa=10003&ns_st_dlpa=10003&ns_st_pa=10003&ns_ts=1560539426237&ns_st_bc=0&ns_st_dbc=0&ns_st_bt=0&ns_st_dbt=0&ns_st_bp=0&ns_st_skc=0&ns_st_dskc=0&ns_st_ska=0&ns_st_dska=0&ns_st_skd=0&ns_st_skt=0&ns_st_dskt=0&ns_st_pc=0&ns_st_dpc=0&ns_st_pp=0&ns_st_br=0&ns_st_ub=0&ns_st_ki=1200000&ns_st_pr=*null&ns_st_sn=*null&ns_st_en=*null&ns_st_ep=Week%202%20Day%201%20%7C%20LEC%20Summer%20Split%20(2019)%20-%20Splyce%20vs.%20Schalke%2004&ns_st_ct=vc13&ns_st_ge=Gaming&ns_st_st=Riot%20Games&ns_st_ce=1&ns_st_ia=0&ns_st_ddt=2019-06-14&ns_st_tdt=*null&ns_st_pu=*null&ns_st_ti=*null&c3=TWITCH&c4=twitch.tv&c6=*null&c7=https%3A%2F%2Fwww.twitch.tv%2Friotgames&c8=Riot%20Games%20-%20Twitch&c9=&cs_ucfr=1 was blocked because it came from a tracker and content blocking is enabled." {file: "https://www.twitch.tv/riotgames" line: 0}]']
</pre>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
see bug description url browser version firefox operating system windows tested another browser yes problem type something else description stream is stuttery even in safe mode steps to reproduce click on a stream to watch it browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled true gfx webrender all true channel nightly console messages u u u u u u u u u u u u u u from with ❤️
| 0
|
6,794
| 9,923,696,062
|
IssuesEvent
|
2019-07-01 07:55:26
|
pmem/pmemkv
|
https://api.github.com/repos/pmem/pmemkv
|
closed
|
Continuous Integration for bindings
|
Type: Process
|
We need to include checks in CI for our bindings, so we know if we didn't break anything by changes in the main pmemkv repo
|
1.0
|
Continuous Integration for bindings - We need to include checks in CI for our bindings, so we know if we didn't break anything by changes in the main pmemkv repo
|
process
|
continuous integration for bindings we need to include checks in ci for our bindings so we know if we didn t break anything by changes in the main pmemkv repo
| 1
|
55,052
| 13,507,515,196
|
IssuesEvent
|
2020-09-14 06:07:50
|
rsx-labs/aide-frontend
|
https://api.github.com/repos/rsx-labs/aide-frontend
|
closed
|
[Daily Workplace Audit] Application error on first access
|
Bug Fixed - ready for build For Next Build High Priority
|
**Describe the bug**
Application error on first access
[4:54 PM] Trilles, Marvin
2020-09-03 16:46:04.7696::ERROR::UI_AIDE_CommCellServices.DailyAuditPage::System.NullReferenceException: Object reference not set to an instance of an object. at UI_AIDE_CommCellServices.DailyAuditPage.GenerateQuestions()
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Version (please complete the following information):**
- Version 3.40
**Additional context**
Add any other context about the problem here.
|
2.0
|
[Daily Workplace Audit] Application error on first access - **Describe the bug**
Application error on first access
[4:54 PM] Trilles, Marvin
2020-09-03 16:46:04.7696::ERROR::UI_AIDE_CommCellServices.DailyAuditPage::System.NullReferenceException: Object reference not set to an instance of an object. at UI_AIDE_CommCellServices.DailyAuditPage.GenerateQuestions()
**To Reproduce**
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Version (please complete the following information):**
- Version 3.40
**Additional context**
Add any other context about the problem here.
|
non_process
|
application error on first access describe the bug application error on first access trilles marvin error ui aide commcellservices dailyauditpage system nullreferenceexception object reference not set to an instance of an object at ui aide commcellservices dailyauditpage generatequestions to reproduce steps to reproduce the behavior go to click on scroll down to see error expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem version please complete the following information version additional context add any other context about the problem here
| 0
|
177,776
| 13,748,232,269
|
IssuesEvent
|
2020-10-06 08:47:10
|
tgstation/tgstation
|
https://api.github.com/repos/tgstation/tgstation
|
closed
|
Metal false walls instantly disappear when clicked, and cannot be closed unless you alt click
|
Bug Tested/Reproduced
|
## Reproduction
1. Build a false wall.
2. Open it.

This only seems to happen with metal.

|
1.0
|
Metal false walls instantly disappear when clicked, and cannot be closed unless you alt click - ## Reproduction
1. Build a false wall.
2. Open it.

This only seems to happen with metal.

|
non_process
|
metal false walls instantly disappear when clicked and cannot be closed unless you alt click reproduction build a false wall open it this only seems to happen with metal
| 0
|
5,149
| 7,689,143,767
|
IssuesEvent
|
2018-05-17 11:47:52
|
unb-cic-esw/youtube-data-monitor
|
https://api.github.com/repos/unb-cic-esw/youtube-data-monitor
|
opened
|
Informar inexistência de certos dados
|
functional requirements
|
EU : desenvolvedor(a)
QUERO : Informar a inexistência de determinados dados:
* Criar um arquivo .txt informando os canais que não possuem vídeos criados a partir de 01.01.2018.
* Nos arquivos youtube.csv informar a inexistência dos dados.
PARA : Melhorar a visibilidade das informações obtidas da API.
Testes de aceitação:
* Criar um arquivo .txt informando os canais que não possuem vídeos criados a partir de 01.01.2018, bem como a criação de testes para validar essa informação;
* Nos arquivos youtube.csv informar a inexistência dos dados, bem como a criação de testes para validar essa informação.
|
1.0
|
Informar inexistência de certos dados - EU : desenvolvedor(a)
QUERO : Informar a inexistência de determinados dados:
* Criar um arquivo .txt informando os canais que não possuem vídeos criados a partir de 01.01.2018.
* Nos arquivos youtube.csv informar a inexistência dos dados.
PARA : Melhorar a visibilidade das informações obtidas da API.
Testes de aceitação:
* Criar um arquivo .txt informando os canais que não possuem vídeos criados a partir de 01.01.2018, bem como a criação de testes para validar essa informação;
* Nos arquivos youtube.csv informar a inexistência dos dados, bem como a criação de testes para validar essa informação.
|
non_process
|
informar inexistência de certos dados eu desenvolvedor a quero informar a inexistência de determinados dados criar um arquivo txt informando os canais que não possuem vídeos criados a partir de nos arquivos youtube csv informar a inexistência dos dados para melhorar a visibilidade das informações obtidas da api testes de aceitação criar um arquivo txt informando os canais que não possuem vídeos criados a partir de bem como a criação de testes para validar essa informação nos arquivos youtube csv informar a inexistência dos dados bem como a criação de testes para validar essa informação
| 0
|
15,297
| 26,890,456,833
|
IssuesEvent
|
2023-02-06 08:29:56
|
bytedance/bitsail
|
https://api.github.com/repos/bytedance/bitsail
|
closed
|
[BitSail][Connector] Support local file source connector
|
help wanted connector requirement difficulty-easy
|
## Is your feature request related to a problem? Please describe
We need to read local file as source. CSV file format could be a good start.
## Describe the solution you'd like
Use V1 source connector interface, read CSV file as source and able to write to `PrintSink`
## Additional context
|
1.0
|
[BitSail][Connector] Support local file source connector - ## Is your feature request related to a problem? Please describe
We need to read local file as source. CSV file format could be a good start.
## Describe the solution you'd like
Use V1 source connector interface, read CSV file as source and able to write to `PrintSink`
## Additional context
|
non_process
|
support local file source connector is your feature request related to a problem please describe we need to read local file as source csv file format could be a good start describe the solution you d like use source connector interface read csv file as source and able to write to printsink additional context
| 0
|
524,067
| 15,195,377,986
|
IssuesEvent
|
2021-02-16 06:12:37
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
[0.9.1.8] No fertilisers appear in the EcoPedia
|
Category: Usability Priority: Low Status: Fixed
|
Seems the various fertilisers are missing entirely from EcoPedia:

|
1.0
|
[0.9.1.8] No fertilisers appear in the EcoPedia - Seems the various fertilisers are missing entirely from EcoPedia:

|
non_process
|
no fertilisers appear in the ecopedia seems the various fertilisers are missing entirely from ecopedia
| 0
|
8,407
| 11,572,751,108
|
IssuesEvent
|
2020-02-21 01:06:03
|
kubeflow/testing
|
https://api.github.com/repos/kubeflow/testing
|
closed
|
Auto-deploy V2
|
area/engprod kind/feature kind/process priority/p1
|
This is the uber bug for updating the auto-deployment infrastructure.
The key feature that we'd like to achieve with V2 is to auto-deploy as soon as a PR is merged.
The current implementation doesn't facilitate that because we have a simple cron job which launches a new deployment every X hours.
If we decrease X in order to deploy sooner we will end up accumulating multiple instances of the auto-deployed infra and using up all our quota.
We want more of a rate limiting queue type pattern similar to what K8s controllers and reconcilers do.
This is critical to being able to rapidly verify changes to the deployments.
|
1.0
|
Auto-deploy V2 - This is the uber bug for updating the auto-deployment infrastructure.
The key feature that we'd like to achieve with V2 is to auto-deploy as soon as a PR is merged.
The current implementation doesn't facilitate that because we have a simple cron job which launches a new deployment every X hours.
If we decrease X in order to deploy sooner we will end up accumulating multiple instances of the auto-deployed infra and using up all our quota.
We want more of a rate limiting queue type pattern similar to what K8s controllers and reconcilers do.
This is critical to being able to rapidly verify changes to the deployments.
|
process
|
auto deploy this is the uber bug for updating the auto deployment infrastructure the key feature that we d like to achieve with is to auto deploy as soon as a pr is merged the current implementation doesn t facilitate that because we have a simple cron job which launches a new deployment every x hours if we decrease x in order to deploy sooner we will end up accumulating multiple instances of the auto deployed infra and using up all our quota we want more of a rate limiting queue type pattern similar to what controllers and reconcilers do this is critical to being able to rapidly verify changes to the deployments
| 1
|
389,045
| 11,496,609,682
|
IssuesEvent
|
2020-02-12 08:22:51
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
nrf52 telnet_shell panic. Mutex using in ISR.
|
bug has-pr priority: medium
|
Stack is
```
#0 assert_post_action (file=file@entry=0x4ab31 "ZEPHYR_BASE/kernel/include/ksched.h", line=line@entry=269) at ZEPHYR_BASE/lib/os/assert.c:45
#1 0x0003086a in z_sched_lock () at ZEPHYR_BASE/kernel/include/ksched.h:269
#2 z_impl_k_mutex_unlock (mutex=mutex@entry=0x200023d8 <contexts+248>) at ZEPHYR_BASE/kernel/mutex.c:232
#3 0x0003c556 in k_mutex_unlock (mutex=mutex@entry=0x200023d8 <contexts+248>) at zephyr/include/generated/syscalls/kernel.h:644
#4 0x0003c692 in net_context_send (context=0x200023d0 <contexts+240>, buf=0x20000938 <shell_transport_telnet_shell_telnet+8>, len=13, cb=cb@entry=0x5399 <telnet_sent_cb>, timeout=timeout@entry=-1, user_data=user_data@entry=0x0 <sys_pm_notify_power_state_entry>) at ZEPHYR_BASE/subsys/net/ip/net_context.c:1668
#5 0x0000525e in telnet_send () at ZEPHYR_BASE/subsys/shell/shell_telnet.c:138
#6 0x00032eea in z_timer_expiration_handler (t=0x200009a4 <shell_transport_telnet_shell_telnet+116>) at ZEPHYR_BASE/kernel/timer.c:65
#7 0x00032dc4 in z_clock_announce (ticks=<optimized out>) at ZEPHYR_BASE/kernel/timeout.c:204
#8 0x00007dee in _isr_wrapper () at ZEPHYR_BASE/arch/arm/core/aarch32/isr_wrapper.S:163
#9 <signal handler called>
```
Since timer handlers in nrf52 is in ISR and network uses mutexes we got panic. Maybe moving `telnet_send` in separate thread will resolve this issue.
|
1.0
|
nrf52 telnet_shell panic. Mutex using in ISR. - Stack is
```
#0 assert_post_action (file=file@entry=0x4ab31 "ZEPHYR_BASE/kernel/include/ksched.h", line=line@entry=269) at ZEPHYR_BASE/lib/os/assert.c:45
#1 0x0003086a in z_sched_lock () at ZEPHYR_BASE/kernel/include/ksched.h:269
#2 z_impl_k_mutex_unlock (mutex=mutex@entry=0x200023d8 <contexts+248>) at ZEPHYR_BASE/kernel/mutex.c:232
#3 0x0003c556 in k_mutex_unlock (mutex=mutex@entry=0x200023d8 <contexts+248>) at zephyr/include/generated/syscalls/kernel.h:644
#4 0x0003c692 in net_context_send (context=0x200023d0 <contexts+240>, buf=0x20000938 <shell_transport_telnet_shell_telnet+8>, len=13, cb=cb@entry=0x5399 <telnet_sent_cb>, timeout=timeout@entry=-1, user_data=user_data@entry=0x0 <sys_pm_notify_power_state_entry>) at ZEPHYR_BASE/subsys/net/ip/net_context.c:1668
#5 0x0000525e in telnet_send () at ZEPHYR_BASE/subsys/shell/shell_telnet.c:138
#6 0x00032eea in z_timer_expiration_handler (t=0x200009a4 <shell_transport_telnet_shell_telnet+116>) at ZEPHYR_BASE/kernel/timer.c:65
#7 0x00032dc4 in z_clock_announce (ticks=<optimized out>) at ZEPHYR_BASE/kernel/timeout.c:204
#8 0x00007dee in _isr_wrapper () at ZEPHYR_BASE/arch/arm/core/aarch32/isr_wrapper.S:163
#9 <signal handler called>
```
Since timer handlers in nrf52 is in ISR and network uses mutexes we got panic. Maybe moving `telnet_send` in separate thread will resolve this issue.
|
non_process
|
telnet shell panic mutex using in isr stack is assert post action file file entry zephyr base kernel include ksched h line line entry at zephyr base lib os assert c in z sched lock at zephyr base kernel include ksched h z impl k mutex unlock mutex mutex entry at zephyr base kernel mutex c in k mutex unlock mutex mutex entry at zephyr include generated syscalls kernel h in net context send context buf len cb cb entry timeout timeout entry user data user data entry at zephyr base subsys net ip net context c in telnet send at zephyr base subsys shell shell telnet c in z timer expiration handler t at zephyr base kernel timer c in z clock announce ticks at zephyr base kernel timeout c in isr wrapper at zephyr base arch arm core isr wrapper s since timer handlers in is in isr and network uses mutexes we got panic maybe moving telnet send in separate thread will resolve this issue
| 0
|
280
| 2,718,446,390
|
IssuesEvent
|
2015-04-12 09:45:57
|
tomatocart/TomatoCart-v1
|
https://api.github.com/repos/tomatocart/TomatoCart-v1
|
opened
|
Replace mootools popup cart with (jquery, zepto) popup cart in bootstrap template
|
Priority: High Status: In process
|
In order to support mobile device, we have to remove mootools framework from bootstrap template. Only one framework should be loaded. Jquery for pc and Zepto for mobile. In this way, the bootstrap store front will work smoothly in mobile devices.
This is the first step - Replace mootools popup cart with (jquery, zepto) popup cart in bootstrap template
|
1.0
|
Replace mootools popup cart with (jquery, zepto) popup cart in bootstrap template - In order to support mobile device, we have to remove mootools framework from bootstrap template. Only one framework should be loaded. Jquery for pc and Zepto for mobile. In this way, the bootstrap store front will work smoothly in mobile devices.
This is the first step - Replace mootools popup cart with (jquery, zepto) popup cart in bootstrap template
|
process
|
replace mootools popup cart with jquery zepto popup cart in bootstrap template in order to support mobile device we have to remove mootools framework from bootstrap template only one framework should be loaded jquery for pc and zepto for mobile in this way the bootstrap store front will work smoothly in mobile devices this is the first step replace mootools popup cart with jquery zepto popup cart in bootstrap template
| 1
|
514,020
| 14,931,620,596
|
IssuesEvent
|
2021-01-25 06:11:09
|
googleapis/google-cloud-go
|
https://api.github.com/repos/googleapis/google-cloud-go
|
closed
|
logging: missing assignment in function "toHTTPRequest"
|
api: logging priority: p2 type: question
|
In this [file](https://github.com/googleapis/google-cloud-go/blob/master/logging/logadmin/logadmin.go#L164) of line 164, we can see the assignment of logging.HTTPRequest's field "RemoteIP", and maybe the field "LocalIP" should be assigned also?
here shows the fixed:

|
1.0
|
logging: missing assignment in function "toHTTPRequest" - In this [file](https://github.com/googleapis/google-cloud-go/blob/master/logging/logadmin/logadmin.go#L164) of line 164, we can see the assignment of logging.HTTPRequest's field "RemoteIP", and maybe the field "LocalIP" should be assigned also?
here shows the fixed:

|
non_process
|
logging missing assignment in function tohttprequest in this of line we can see the assignment of logging httprequest s field remoteip and maybe the field localip should be assigned also here shows the fixed
| 0
|
111,549
| 24,146,142,141
|
IssuesEvent
|
2022-09-21 18:57:22
|
rkd103/Group-11
|
https://api.github.com/repos/rkd103/Group-11
|
closed
|
Implement User Story B
|
enhancement code
|
Taking the User Stories assigned in previous issues, begin brainstorming how to implement the requirements in code.
Namely, perform independent research regarding what packages, techniques, etc. are needed to satisfy the stakeholder's image.
User Story B: "compose, edit, and post statuses on personal timeline"
Original Assignee: HeFeng (ho170)
|
1.0
|
Implement User Story B - Taking the User Stories assigned in previous issues, begin brainstorming how to implement the requirements in code.
Namely, perform independent research regarding what packages, techniques, etc. are needed to satisfy the stakeholder's image.
User Story B: "compose, edit, and post statuses on personal timeline"
Original Assignee: HeFeng (ho170)
|
non_process
|
implement user story b taking the user stories assigned in previous issues begin brainstorming how to implement the requirements in code namely perform independent research regarding what packages techniques etc are needed to satisfy the stakeholder s image user story b compose edit and post statuses on personal timeline original assignee hefeng
| 0
|
17,442
| 23,265,968,540
|
IssuesEvent
|
2022-08-04 17:24:44
|
MPMG-DCC-UFMG/C01
|
https://api.github.com/repos/MPMG-DCC-UFMG/C01
|
opened
|
Transparência - Detalhes do coletor/Opções avançadas (extração de links e download de arquivos)
|
[1] Requisito [0] Desenvolvimento [2] Média Prioridade [3] Processamento Dinâmico
|
## Comportamento Esperado
Espera-se que as configurações de extração e download das Opções avançadas se apliquem também às coletas que usam processamento dinâmico.
## Comportamento Atual
Ao configurar um coletor dinâmico com essa ferramenta, as opções avançadas disponíveis na tela de detalhes do coletor não se aplicam da mesma maneira que funcionam na extração de links "estáticos" pelo Scrapy. Para usar essas ferramentas dinamicamente, apesar de haver opções extras para filtrar xpaths em alguns passos do mecanismo de processamento dinâmico, não há uma opção nos detalhes do coletor para aplicar as opções avançadas desejadas. Isso pode ser pouco intuitivo para o usuário.
## Passos para reproduzir o erro
Não se aplica.
## Sistema
- MP ou local: ambos
- Branch específica: master
- Sistema diferente: não
## Screenshots
Não se aplica.
|
1.0
|
Transparência - Detalhes do coletor/Opções avançadas (extração de links e download de arquivos) - ## Comportamento Esperado
Espera-se que as configurações de extração e download das Opções avançadas se apliquem também às coletas que usam processamento dinâmico.
## Comportamento Atual
Ao configurar um coletor dinâmico com essa ferramenta, as opções avançadas disponíveis na tela de detalhes do coletor não se aplicam da mesma maneira que funcionam na extração de links "estáticos" pelo Scrapy. Para usar essas ferramentas dinamicamente, apesar de haver opções extras para filtrar xpaths em alguns passos do mecanismo de processamento dinâmico, não há uma opção nos detalhes do coletor para aplicar as opções avançadas desejadas. Isso pode ser pouco intuitivo para o usuário.
## Passos para reproduzir o erro
Não se aplica.
## Sistema
- MP ou local: ambos
- Branch específica: master
- Sistema diferente: não
## Screenshots
Não se aplica.
|
process
|
transparência detalhes do coletor opções avançadas extração de links e download de arquivos comportamento esperado espera se que as configurações de extração e download das opções avançadas se apliquem também às coletas que usam processamento dinâmico comportamento atual ao configurar um coletor dinâmico com essa ferramenta as opções avançadas disponíveis na tela de detalhes do coletor não se aplicam da mesma maneira que funcionam na extração de links estáticos pelo scrapy para usar essas ferramentas dinamicamente apesar de haver opções extras para filtrar xpaths em alguns passos do mecanismo de processamento dinâmico não há uma opção nos detalhes do coletor para aplicar as opções avançadas desejadas isso pode ser pouco intuitivo para o usuário passos para reproduzir o erro não se aplica sistema mp ou local ambos branch específica master sistema diferente não screenshots não se aplica
| 1
|
37,563
| 10,029,748,102
|
IssuesEvent
|
2019-07-17 14:30:57
|
akka/akka-http
|
https://api.github.com/repos/akka/akka-http
|
closed
|
Release Akka HTTP 10.1.9
|
3 - in progress t:build
|
<!--
(Liberally copied and adopted from Scala itself https://github.com/scala/scala-dev/blob/b11cd2e4a4431de7867db6b39362bea8fa6650e7/notes/releases/template.md)
For every Akka HTTP release, make a copy of this file named after the release, and expand the variables.
Ideally replacing variables could become a script you can run on your local machine.
Variables to be expanded in this template:
- VERSION=???
Key links:
- akka/akka-http milestone: https://github.com/akka/akka-http/milestone/?
-->
### ~ 1 week before the release
- [x] Check that open PRs and issues assigned to the milestone are reasonable
- [x] Triage tickets that should be ready for this release, add "pick next" label and release milestone
- [x] Triage open PRs and apply "pick next" label and maybe add to release milestone. Some PRs might be explicitly scheduled for this release, others might be ready enough to bring them over the finish line
Wind down PR queue. There has to be enough time after the last (non-trivial) PR is merged and the next phase. The core of the eco-system needs time to prepare for the final!
- [x] Decide on planned release date
- [x] Notify depending projects (notably Play + cinnamon) about upcoming release
### 1 day before the release
- [x] Make sure all important / big PRs have been merged by now
- [x] Check that latest nightly (once we have that) / master still works with depending projects (notably Play + cinnamon)
- [x] Communicate that a release is about to be released in [Gitter Akka Dev Channel](https://gitter.im/akka/dev), so that no new Pull Requests are merged
### Preparing release notes in the documentation / announcement
- [x] If this is a new minor (not patch) release, rename the 'akka-http-x.x-stable' reporting project in [WhiteSource](https://saas.whitesourcesoftware.com/) accordingly
- [x] Add a release notes entry in docs/src/main/paradox/release-notes/. As a helper run `scripts/commits-for-release-notes.sh <last-version-tag>` which will output a list of commits grouped by submodule, and the closed issues for this milestone.
- [x] Create a PR on https://github.com/akka/akka.github.com with a news item draft, using the milestones and `scripts/authors.scala previousVersion thisVersion`
- [x] Create a PR on https://github.com/akka/akka.github.com updating Akka HTTP dependency for akka.io website: https://github.com/akka/akka.github.com/blob/master/_config.yml
- [x] Release notes PR has been merged
- [x] Create a new milestone for the next version at https://github.com/akka/akka-http/milestones
- [x] Move all unclosed issues to the newly created milestone (or remove milestone) and close the version you're releasing
### Cutting the release
- [x] Make sure there are no stray staging repos on sonatype
- [x] Wait until [master build finished](https://travis-ci.org/akka/akka-http/builds/) after merging the release notes
- [x] Create a [new release](https://github.com/akka/akka-http/releases/new) with the next tag version (e.g. `v13.3.7`), title and release description linking to announcement, release notes and milestone.
- [x] Check that the Travis CI release build has executed successfully
- [x] Go to https://bintray.com/akka/maven/com.typesafe.akka:akka-http_2.11 and select the just released version
- [x] Go to the Maven Central tab and sync with Sonatype
- [x] Log in to Sonatype to Close the staging repository (optional, should happen automatically if selected in Bintray)
- [x] Notify Telemetry / Play team to check against staged artifacts
- [x] Run a test against the staging repository to make sure the release went well, for example by using https://github.com/akka/akka-http-scala-seed.g8 and adding the sonatype staging repo with `resolvers += "Staging Repo" at "https://oss.sonatype.org/content/repositories/comtypesafe-xxx"`
- [x] Release the staging repository to Maven Central.
- [x] Checkout the newly created tag and run `sbt -Dakka.genjavadoc.enabled=true ++2.12.8 "deployRsync akkarepo@gustav.akka.io"` to deploy API and reference documentation.
### Check availability
- [x] Check release on sonatype: https://oss.sonatype.org/content/repositories/releases/com/typesafe/akka/akka-http-core_2.11/
- [x] Check the release on maven central: http://central.maven.org/maven2/com/typesafe/akka/akka-http-core_2.11/
### When everything is on maven central
- [x] Log into `gustav.akka.io` as `akkarepo`
- [ ] update the `10.1` and `current` links on `repo.akka.io` to point to the latest version with (**replace the minor appropriately**)
```
ln -nsf 10.1.9 www/docs/akka-http/10.1
ln -nsf 10.1.9 www/api/akka-http/10.1
ln -nsf 10.1.9 www/japi/akka-http/10.1
ln -nsf 10.1.9 www/docs/akka-http/current
ln -nsf 10.1.9 www/api/akka-http/current
ln -nsf 10.1.9 www/jap/akka-http/current
```
- [x] check changes and commit the new version to the local git repository
```
cd ~/www
git add docs/akka-http/ api/akka-http/ japi/akka-http/
git commit -m "Akka HTTP 10.1.9"
```
### Announcements
- [x] Merge draft news item at https://github.com/akka/akka.github.com
- [x] Send a release notification to https://discuss.akka.io
- [x] Tweet using the akka account (or ask someone to) about the new release
- [x] Announce on Gitter at https://gitter.im/akka/akka
### Afterwards
- [x] Add the released version to `project/MiMa.scala` to the `mimaPreviousArtifacts` key *of all current compatible branches*.
- [ ] Update Akka HTTP dependency in [akka-management](https://github.com/akka/akka-management/edit/master/project/Dependencies.scala)
- [x] Update Akka HTTP reference in [reactive-platform-docs](https://github.com/typesafehub/reactive-platform-docs/edit/master/build.sbt#L29)
- Close this issue
|
1.0
|
Release Akka HTTP 10.1.9 - <!--
(Liberally copied and adopted from Scala itself https://github.com/scala/scala-dev/blob/b11cd2e4a4431de7867db6b39362bea8fa6650e7/notes/releases/template.md)
For every Akka HTTP release, make a copy of this file named after the release, and expand the variables.
Ideally replacing variables could become a script you can run on your local machine.
Variables to be expanded in this template:
- VERSION=???
Key links:
- akka/akka-http milestone: https://github.com/akka/akka-http/milestone/?
-->
### ~ 1 week before the release
- [x] Check that open PRs and issues assigned to the milestone are reasonable
- [x] Triage tickets that should be ready for this release, add "pick next" label and release milestone
- [x] Triage open PRs and apply "pick next" label and maybe add to release milestone. Some PRs might be explicitly scheduled for this release, others might be ready enough to bring them over the finish line
Wind down PR queue. There has to be enough time after the last (non-trivial) PR is merged and the next phase. The core of the eco-system needs time to prepare for the final!
- [x] Decide on planned release date
- [x] Notify depending projects (notably Play + cinnamon) about upcoming release
### 1 day before the release
- [x] Make sure all important / big PRs have been merged by now
- [x] Check that latest nightly (once we have that) / master still works with depending projects (notably Play + cinnamon)
- [x] Communicate that a release is about to be released in [Gitter Akka Dev Channel](https://gitter.im/akka/dev), so that no new Pull Requests are merged
### Preparing release notes in the documentation / announcement
- [x] If this is a new minor (not patch) release, rename the 'akka-http-x.x-stable' reporting project in [WhiteSource](https://saas.whitesourcesoftware.com/) accordingly
- [x] Add a release notes entry in docs/src/main/paradox/release-notes/. As a helper run `scripts/commits-for-release-notes.sh <last-version-tag>` which will output a list of commits grouped by submodule, and the closed issues for this milestone.
- [x] Create a PR on https://github.com/akka/akka.github.com with a news item draft, using the milestones and `scripts/authors.scala previousVersion thisVersion`
- [x] Create a PR on https://github.com/akka/akka.github.com updating Akka HTTP dependency for akka.io website: https://github.com/akka/akka.github.com/blob/master/_config.yml
- [x] Release notes PR has been merged
- [x] Create a new milestone for the next version at https://github.com/akka/akka-http/milestones
- [x] Move all unclosed issues to the newly created milestone (or remove milestone) and close the version you're releasing
### Cutting the release
- [x] Make sure there are no stray staging repos on sonatype
- [x] Wait until [master build finished](https://travis-ci.org/akka/akka-http/builds/) after merging the release notes
- [x] Create a [new release](https://github.com/akka/akka-http/releases/new) with the next tag version (e.g. `v13.3.7`), title and release description linking to announcement, release notes and milestone.
- [x] Check that the Travis CI release build has executed successfully
- [x] Go to https://bintray.com/akka/maven/com.typesafe.akka:akka-http_2.11 and select the just released version
- [x] Go to the Maven Central tab and sync with Sonatype
- [x] Log in to Sonatype to Close the staging repository (optional, should happen automatically if selected in Bintray)
- [x] Notify Telemetry / Play team to check against staged artifacts
- [x] Run a test against the staging repository to make sure the release went well, for example by using https://github.com/akka/akka-http-scala-seed.g8 and adding the sonatype staging repo with `resolvers += "Staging Repo" at "https://oss.sonatype.org/content/repositories/comtypesafe-xxx"`
- [x] Release the staging repository to Maven Central.
- [x] Checkout the newly created tag and run `sbt -Dakka.genjavadoc.enabled=true ++2.12.8 "deployRsync akkarepo@gustav.akka.io"` to deploy API and reference documentation.
### Check availability
- [x] Check release on sonatype: https://oss.sonatype.org/content/repositories/releases/com/typesafe/akka/akka-http-core_2.11/
- [x] Check the release on maven central: http://central.maven.org/maven2/com/typesafe/akka/akka-http-core_2.11/
### When everything is on maven central
- [x] Log into `gustav.akka.io` as `akkarepo`
- [ ] update the `10.1` and `current` links on `repo.akka.io` to point to the latest version with (**replace the minor appropriately**)
```
ln -nsf 10.1.9 www/docs/akka-http/10.1
ln -nsf 10.1.9 www/api/akka-http/10.1
ln -nsf 10.1.9 www/japi/akka-http/10.1
ln -nsf 10.1.9 www/docs/akka-http/current
ln -nsf 10.1.9 www/api/akka-http/current
ln -nsf 10.1.9 www/jap/akka-http/current
```
- [x] check changes and commit the new version to the local git repository
```
cd ~/www
git add docs/akka-http/ api/akka-http/ japi/akka-http/
git commit -m "Akka HTTP 10.1.9"
```
### Announcements
- [x] Merge draft news item at https://github.com/akka/akka.github.com
- [x] Send a release notification to https://discuss.akka.io
- [x] Tweet using the akka account (or ask someone to) about the new release
- [x] Announce on Gitter at https://gitter.im/akka/akka
### Afterwards
- [x] Add the released version to `project/MiMa.scala` to the `mimaPreviousArtifacts` key *of all current compatible branches*.
- [ ] Update Akka HTTP dependency in [akka-management](https://github.com/akka/akka-management/edit/master/project/Dependencies.scala)
- [x] Update Akka HTTP reference in [reactive-platform-docs](https://github.com/typesafehub/reactive-platform-docs/edit/master/build.sbt#L29)
- Close this issue
|
non_process
|
release akka http liberally copied and adopted from scala itself for every akka http release make a copy of this file named after the release and expand the variables ideally replacing variables could become a script you can run on your local machine variables to be expanded in this template version key links akka akka http milestone week before the release check that open prs and issues assigned to the milestone are reasonable triage tickets that should be ready for this release add pick next label and release milestone triage open prs and apply pick next label and maybe add to release milestone some prs might be explicitly scheduled for this release others might be ready enough to bring them over the finish line wind down pr queue there has to be enough time after the last non trivial pr is merged and the next phase the core of the eco system needs time to prepare for the final decide on planned release date notify depending projects notably play cinnamon about upcoming release day before the release make sure all important big prs have been merged by now check that latest nightly once we have that master still works with depending projects notably play cinnamon communicate that a release is about to be released in so that no new pull requests are merged preparing release notes in the documentation announcement if this is a new minor not patch release rename the akka http x x stable reporting project in accordingly add a release notes entry in docs src main paradox release notes as a helper run scripts commits for release notes sh which will output a list of commits grouped by submodule and the closed issues for this milestone create a pr on with a news item draft using the milestones and scripts authors scala previousversion thisversion create a pr on updating akka http dependency for akka io website release notes pr has been merged create a new milestone for the next version at move all unclosed issues to the newly created milestone or remove milestone and close the version you re releasing cutting the release make sure there are no stray staging repos on sonatype wait until after merging the release notes create a with the next tag version e g title and release description linking to announcement release notes and milestone check that the travis ci release build has executed successfully go to and select the just released version go to the maven central tab and sync with sonatype log in to sonatype to close the staging repository optional should happen automatically if selected in bintray notify telemetry play team to check against staged artifacts run a test against the staging repository to make sure the release went well for example by using and adding the sonatype staging repo with resolvers staging repo at release the staging repository to maven central checkout the newly created tag and run sbt dakka genjavadoc enabled true deployrsync akkarepo gustav akka io to deploy api and reference documentation check availability check release on sonatype check the release on maven central when everything is on maven central log into gustav akka io as akkarepo update the and current links on repo akka io to point to the latest version with replace the minor appropriately ln nsf www docs akka http ln nsf www api akka http ln nsf www japi akka http ln nsf www docs akka http current ln nsf www api akka http current ln nsf www jap akka http current check changes and commit the new version to the local git repository cd www git add docs akka http api akka http japi akka http git commit m akka http announcements merge draft news item at send a release notification to tweet using the akka account or ask someone to about the new release announce on gitter at afterwards add the released version to project mima scala to the mimapreviousartifacts key of all current compatible branches update akka http dependency in update akka http reference in close this issue
| 0
|
11,247
| 14,015,567,798
|
IssuesEvent
|
2020-10-29 13:29:43
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - dcterms:modified
|
Class - Record-level Process - implement Term - change
|
## Change term
* Submitter: John Wieczorek
* Justification (why is this change necessary?): The recommended specification for dates has been deprecated.
* Proponents (who needs this change): Everyone
Proposed new attributes of the term:
* Usage comments (recommendations regarding content, etc.):
"Recommended best practice is to use a date that conforms to ISO 8601-1:2019."
|
1.0
|
Change term - dcterms:modified - ## Change term
* Submitter: John Wieczorek
* Justification (why is this change necessary?): The recommended specification for dates has been deprecated.
* Proponents (who needs this change): Everyone
Proposed new attributes of the term:
* Usage comments (recommendations regarding content, etc.):
"Recommended best practice is to use a date that conforms to ISO 8601-1:2019."
|
process
|
change term dcterms modified change term submitter john wieczorek justification why is this change necessary the recommended specification for dates has been deprecated proponents who needs this change everyone proposed new attributes of the term usage comments recommendations regarding content etc recommended best practice is to use a date that conforms to iso
| 1
|
22,234
| 3,777,304,118
|
IssuesEvent
|
2016-03-17 19:34:15
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
Elimination of the use of semicolon at the end of every statement in C# 7
|
Area-Language Design
|
**Version Used**: C# 6
**Steps to Reproduce**:
We all have to put the ; at the end of every statement. The compiler complain if we don't. Could it be possible to eliminate the need of adding the semicolon?
It is a dramatic request, but I think at this point we can get rid of it. The compiler knows when we missed it, it give us an error. Since it knows, could the IDE analyzer add it automatically or better just not expect us to type it anymore. F# do not have it, Swift do not have it. C# can get rid of it :)
1.
2.
3.
**Expected Behavior**:
Love to see this: Console.Write("No more semicolons, yeah!!")
**Actual Behavior**:
Console.Write("Hating the semicolon at the end, yeah!!");
|
1.0
|
Elimination of the use of semicolon at the end of every statement in C# 7 - **Version Used**: C# 6
**Steps to Reproduce**:
We all have to put the ; at the end of every statement. The compiler complain if we don't. Could it be possible to eliminate the need of adding the semicolon?
It is a dramatic request, but I think at this point we can get rid of it. The compiler knows when we missed it, it give us an error. Since it knows, could the IDE analyzer add it automatically or better just not expect us to type it anymore. F# do not have it, Swift do not have it. C# can get rid of it :)
1.
2.
3.
**Expected Behavior**:
Love to see this: Console.Write("No more semicolons, yeah!!")
**Actual Behavior**:
Console.Write("Hating the semicolon at the end, yeah!!");
|
non_process
|
elimination of the use of semicolon at the end of every statement in c version used c steps to reproduce we all have to put the at the end of every statement the compiler complain if we don t could it be possible to eliminate the need of adding the semicolon it is a dramatic request but i think at this point we can get rid of it the compiler knows when we missed it it give us an error since it knows could the ide analyzer add it automatically or better just not expect us to type it anymore f do not have it swift do not have it c can get rid of it expected behavior love to see this console write no more semicolons yeah actual behavior console write hating the semicolon at the end yeah
| 0
|
13,190
| 15,613,698,160
|
IssuesEvent
|
2021-03-19 16:47:31
|
Kreusada/Kreusada-Cogs
|
https://api.github.com/repos/Kreusada/Kreusada-Cogs
|
opened
|
[Dehoister] Some builtins do not have a `__file__` attr
|
Cog: Dehoister Status: Processing Type: Bug
|
**What cog are you reporting for?**
Dehoister
**What command(s) inside the cog are you reporting for?**
`[p]vinfo mod`
**Description of the issue**
Some builtin (or external) modules and/or dependencies do not have `__file__` attributes. This means that the module cannot be checked against `pypath` (`str(distutils.sysconfig.get_python_lib(standard_lib=True))`) because they do not have those attributes. Suggested fixes would probably be to check via `dep.__spec__.origin`. This seems like a more reliable way, although I noticed that one module `sys` returns None (perhaps because it is quite literally the python core, idk).
Examples of modules that threw this attribute error were math, and array. I'm sure there will probably be more.

I am making this issue so that I can get round to resolving it as soon as possible. Please do not open any subsequent issues for this bug.
|
1.0
|
[Dehoister] Some builtins do not have a `__file__` attr - **What cog are you reporting for?**
Dehoister
**What command(s) inside the cog are you reporting for?**
`[p]vinfo mod`
**Description of the issue**
Some builtin (or external) modules and/or dependencies do not have `__file__` attributes. This means that the module cannot be checked against `pypath` (`str(distutils.sysconfig.get_python_lib(standard_lib=True))`) because they do not have those attributes. Suggested fixes would probably be to check via `dep.__spec__.origin`. This seems like a more reliable way, although I noticed that one module `sys` returns None (perhaps because it is quite literally the python core, idk).
Examples of modules that threw this attribute error were math, and array. I'm sure there will probably be more.

I am making this issue so that I can get round to resolving it as soon as possible. Please do not open any subsequent issues for this bug.
|
process
|
some builtins do not have a file attr what cog are you reporting for dehoister what command s inside the cog are you reporting for vinfo mod description of the issue some builtin or external modules and or dependencies do not have file attributes this means that the module cannot be checked against pypath str distutils sysconfig get python lib standard lib true because they do not have those attributes suggested fixes would probably be to check via dep spec origin this seems like a more reliable way although i noticed that one module sys returns none perhaps because it is quite literally the python core idk examples of modules that threw this attribute error were math and array i m sure there will probably be more i am making this issue so that i can get round to resolving it as soon as possible please do not open any subsequent issues for this bug
| 1
|
290,397
| 32,076,510,509
|
IssuesEvent
|
2023-09-25 11:22:49
|
exasol/exasol-testcontainers
|
https://api.github.com/repos/exasol/exasol-testcontainers
|
closed
|
Fix vulnerability in dependency commons-compress
|
security
|
```
Error: Failed to execute goal org.sonatype.ossindex.maven:ossindex-maven-plugin:3.2.0:audit (default-cli) on project exasol-testcontainers: Detected 1 vulnerable components:
Error: org.apache.commons:commons-compress:jar:1.23.0:compile; https://ossindex.sonatype.org/component/pkg:maven/org.apache.commons/commons-compress@1.23.0?utm_source=ossindex-client&utm_medium=integration&utm_content=1.8.1
Error: * [CVE-2023-42503] CWE-20: Improper Input Validation (5.5); https://ossindex.sonatype.org/vulnerability/CVE-2023-42503?component-type=maven&component-name=org.apache.commons%2Fcommons-compress&utm_source=ossindex-client&utm_medium=integration&utm_content=1.8.1
```
|
True
|
Fix vulnerability in dependency commons-compress - ```
Error: Failed to execute goal org.sonatype.ossindex.maven:ossindex-maven-plugin:3.2.0:audit (default-cli) on project exasol-testcontainers: Detected 1 vulnerable components:
Error: org.apache.commons:commons-compress:jar:1.23.0:compile; https://ossindex.sonatype.org/component/pkg:maven/org.apache.commons/commons-compress@1.23.0?utm_source=ossindex-client&utm_medium=integration&utm_content=1.8.1
Error: * [CVE-2023-42503] CWE-20: Improper Input Validation (5.5); https://ossindex.sonatype.org/vulnerability/CVE-2023-42503?component-type=maven&component-name=org.apache.commons%2Fcommons-compress&utm_source=ossindex-client&utm_medium=integration&utm_content=1.8.1
```
|
non_process
|
fix vulnerability in dependency commons compress error failed to execute goal org sonatype ossindex maven ossindex maven plugin audit default cli on project exasol testcontainers detected vulnerable components error org apache commons commons compress jar compile error cwe improper input validation
| 0
|
95,486
| 10,881,096,189
|
IssuesEvent
|
2019-11-17 15:35:15
|
hsloot/rmo
|
https://api.github.com/repos/hsloot/rmo
|
closed
|
[DOCUMENTATION]: Package documentation in manual
|
documentation enhancement good first issue help wanted
|
## Summary
The documentation of the package-manual is missing.
## Proposal
Create a file`R/rmo.R` With a short description of the package and its main applications.
## Additional context
See [Object documentation - R packages](http://r-pkgs.had.co.nz/man.html#man-packages).
|
1.0
|
[DOCUMENTATION]: Package documentation in manual - ## Summary
The documentation of the package-manual is missing.
## Proposal
Create a file`R/rmo.R` With a short description of the package and its main applications.
## Additional context
See [Object documentation - R packages](http://r-pkgs.had.co.nz/man.html#man-packages).
|
non_process
|
package documentation in manual summary the documentation of the package manual is missing proposal create a file r rmo r with a short description of the package and its main applications additional context see
| 0
|
184,721
| 6,715,683,664
|
IssuesEvent
|
2017-10-13 22:29:15
|
FStarLang/FStar
|
https://api.github.com/repos/FStarLang/FStar
|
closed
|
Type annotations in match (and destructuring let) get completely ignored
|
area/syntax area/usability component/typechecker kind/bug priority/high
|
The following should in my eyes be refused (or at least trigger a warning).
```
module Test
type test : Type = | T : x:bool -> test
val f : test -> bool
let f h = match h with T (t : int) -> t
```
|
1.0
|
Type annotations in match (and destructuring let) get completely ignored - The following should in my eyes be refused (or at least trigger a warning).
```
module Test
type test : Type = | T : x:bool -> test
val f : test -> bool
let f h = match h with T (t : int) -> t
```
|
non_process
|
type annotations in match and destructuring let get completely ignored the following should in my eyes be refused or at least trigger a warning module test type test type t x bool test val f test bool let f h match h with t t int t
| 0
|
217,957
| 24,351,699,753
|
IssuesEvent
|
2022-10-03 01:11:03
|
jyothsna/votingapplication
|
https://api.github.com/repos/jyothsna/votingapplication
|
opened
|
CVE-2022-42004 (Medium) detected in jackson-databind-2.9.6.jar, jackson-databind-2.9.0.jar
|
security vulnerability
|
## CVE-2022-42004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.0.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.3.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.0.3.RELEASE.jar
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /epository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization.
<p>Publish Date: 2022-10-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-42004 (Medium) detected in jackson-databind-2.9.6.jar, jackson-databind-2.9.0.jar - ## CVE-2022-42004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.0.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.3.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.0.3.RELEASE.jar
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /epository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization.
<p>Publish Date: 2022-10-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jackson databind jar jackson databind jar cve medium severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library epository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details in fasterxml jackson databind before resource exhaustion can occur because of a lack of a check in beandeserializer deserializefromarray to prevent use of deeply nested arrays an application is vulnerable only with certain customized choices for deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with mend
| 0
|
31,906
| 6,017,146,993
|
IssuesEvent
|
2017-06-07 08:57:57
|
mysociety/alaveteli
|
https://api.github.com/repos/mysociety/alaveteli
|
closed
|
Create a diff of the English alaveteli documentation since we deployed the Spanish version
|
documentation re-user experience
|
Give it to @JenMysoc to create/update the translations.
|
1.0
|
Create a diff of the English alaveteli documentation since we deployed the Spanish version - Give it to @JenMysoc to create/update the translations.
|
non_process
|
create a diff of the english alaveteli documentation since we deployed the spanish version give it to jenmysoc to create update the translations
| 0
|
68,887
| 8,358,420,429
|
IssuesEvent
|
2018-10-03 02:44:18
|
ByteClubGames/YumiAndTheYokai
|
https://api.github.com/repos/ByteClubGames/YumiAndTheYokai
|
closed
|
Wall Clinging
|
Bug HIGH PRIORITY Level Design
|
They player can currently stick to walls by holding the direction of the wall.
-This could be an issue with either the tiles themselves
-The could also have to do with the shape of the player model's collider
-If the other approaches don't work, it could probably be fixed with a small script.
|
1.0
|
Wall Clinging - They player can currently stick to walls by holding the direction of the wall.
-This could be an issue with either the tiles themselves
-The could also have to do with the shape of the player model's collider
-If the other approaches don't work, it could probably be fixed with a small script.
|
non_process
|
wall clinging they player can currently stick to walls by holding the direction of the wall this could be an issue with either the tiles themselves the could also have to do with the shape of the player model s collider if the other approaches don t work it could probably be fixed with a small script
| 0
|
250,396
| 18,888,278,219
|
IssuesEvent
|
2021-11-15 10:20:47
|
do-mpc/do-mpc
|
https://api.github.com/repos/do-mpc/do-mpc
|
opened
|
Minor error in the MHE Formulation
|
Documentation
|
Regarding the documentation on our [do-mpc](do-mpc.com) homepage.
**Link to page / section of the issue**
https://www.do-mpc.com/en/latest/mhe_example.html
**What is wrong / missing?**
Some minor errors in the MHE formuation.
**How could it be improved / what should be added?**
1- The sum of the measurement noise term v_k should be from 0 to N
2- The model ode constraint should be from k = 0 to k = N-1
3- I am a bit skiptical about using the control input as a decision variable.
Thanks for offering such a useful toolbox!
|
1.0
|
Minor error in the MHE Formulation - Regarding the documentation on our [do-mpc](do-mpc.com) homepage.
**Link to page / section of the issue**
https://www.do-mpc.com/en/latest/mhe_example.html
**What is wrong / missing?**
Some minor errors in the MHE formuation.
**How could it be improved / what should be added?**
1- The sum of the measurement noise term v_k should be from 0 to N
2- The model ode constraint should be from k = 0 to k = N-1
3- I am a bit skiptical about using the control input as a decision variable.
Thanks for offering such a useful toolbox!
|
non_process
|
minor error in the mhe formulation regarding the documentation on our do mpc com homepage link to page section of the issue what is wrong missing some minor errors in the mhe formuation how could it be improved what should be added the sum of the measurement noise term v k should be from to n the model ode constraint should be from k to k n i am a bit skiptical about using the control input as a decision variable thanks for offering such a useful toolbox
| 0
|
137,724
| 20,203,861,273
|
IssuesEvent
|
2022-02-11 17:57:13
|
CDCgov/prime-reportstream
|
https://api.github.com/repos/CDCgov/prime-reportstream
|
closed
|
Submission History Design
|
design Epic experience
|
As the Experience Team, we need some design for the Submission History & Lineage feature as outlined in this Epic [https://app.zenhub.com/workspaces/experience-607d9d5e68b95200150fec37/issues/cdcgov/prime-reportstream/1652](url).
|
1.0
|
Submission History Design - As the Experience Team, we need some design for the Submission History & Lineage feature as outlined in this Epic [https://app.zenhub.com/workspaces/experience-607d9d5e68b95200150fec37/issues/cdcgov/prime-reportstream/1652](url).
|
non_process
|
submission history design as the experience team we need some design for the submission history lineage feature as outlined in this epic url
| 0
|
201
| 2,610,024,982
|
IssuesEvent
|
2015-02-26 18:05:00
|
numenta/htm.java
|
https://api.github.com/repos/numenta/htm.java
|
closed
|
Sync HTM.java Releases w/NuPIC
|
administrative help wanted P3 process wiki
|
The release numbering (in the grade.build and pom.xml files) and the Roadmap need to be updated with compatible numbering and sync'd up with the new release and Roadmap present in NuPIC.
@rhyolight
|
1.0
|
Sync HTM.java Releases w/NuPIC - The release numbering (in the grade.build and pom.xml files) and the Roadmap need to be updated with compatible numbering and sync'd up with the new release and Roadmap present in NuPIC.
@rhyolight
|
process
|
sync htm java releases w nupic the release numbering in the grade build and pom xml files and the roadmap need to be updated with compatible numbering and sync d up with the new release and roadmap present in nupic rhyolight
| 1
|
158,306
| 12,412,255,011
|
IssuesEvent
|
2020-05-22 10:11:10
|
aliasrobotics/RVD
|
https://api.github.com/repos/aliasrobotics/RVD
|
opened
|
716]
|
bug cppcheck static analysis testing triage
|
```yaml
{
"id": 1,
"title": "716]",
"type": "bug",
"description": "[src/opencv3/3rdparty/libjpeg/jdarith.c:718] -> [src/opencv3/3rdparty/libjpeg/jdarith.c:716]: (warning) Either the condition 'tbl>=16' is redundant or the array 'entropy->dc_stats[16]' is accessed at index 16, which is out of bounds.",
"cwe": "None",
"cve": "None",
"keywords": [
"cppcheck",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "src/opencv3/3rdparty/libjpeg/jdarith.c",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "N/A",
"architectural-location": "N/A",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-22 (10:10)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-22 (10:10)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "",
"reproduction": "See artifacts below (if available)",
"reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download"
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
```
|
1.0
|
716] - ```yaml
{
"id": 1,
"title": "716]",
"type": "bug",
"description": "[src/opencv3/3rdparty/libjpeg/jdarith.c:718] -> [src/opencv3/3rdparty/libjpeg/jdarith.c:716]: (warning) Either the condition 'tbl>=16' is redundant or the array 'entropy->dc_stats[16]' is accessed at index 16, which is out of bounds.",
"cwe": "None",
"cve": "None",
"keywords": [
"cppcheck",
"static analysis",
"testing",
"triage",
"bug"
],
"system": "src/opencv3/3rdparty/libjpeg/jdarith.c",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "N/A",
"architectural-location": "N/A",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-05-22 (10:10)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-05-22 (10:10)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "",
"reproduction": "See artifacts below (if available)",
"reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download"
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
```
|
non_process
|
yaml id title type bug description warning either the condition tbl is redundant or the array entropy dc stats is accessed at index which is out of bounds cwe none cve none keywords cppcheck static analysis testing triage bug system src libjpeg jdarith c vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity n a architectural location n a application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros kinetic jobs artifacts download exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
| 0
|
173,146
| 21,103,687,134
|
IssuesEvent
|
2022-04-04 16:36:07
|
opensearch-project/data-prepper
|
https://api.github.com/repos/opensearch-project/data-prepper
|
closed
|
CVE-2019-10782 (Medium) detected in checkstyle-8.27.jar - autoclosed
|
security vulnerability
|
## CVE-2019-10782 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>checkstyle-8.27.jar</b></p></summary>
<p>Checkstyle is a development tool to help programmers write Java code
that adheres to a coding standard</p>
<p>Library home page: <a href="https://checkstyle.org/">https://checkstyle.org/</a></p>
<p>
Dependency Hierarchy:
<p>Found in HEAD commit: <a href="https://github.com/opensearch-project/data-prepper/commit/022b333dc9be3548b8eb8bb73d0337fd26425056">022b333dc9be3548b8eb8bb73d0337fd26425056</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of com.puppycrawl.tools:checkstyle before 8.29 are vulnerable to XML External Entity (XXE) Injection due to an incomplete fix for CVE-2019-9658.
<p>Publish Date: 2020-01-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10782>CVE-2019-10782</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10782">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10782</a></p>
<p>Release Date: 2020-02-10</p>
<p>Fix Resolution: com.puppycrawl.tools:checkstyle:8.29</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.puppycrawl.tools","packageName":"checkstyle","packageVersion":"8.27","packageFilePaths":[],"isTransitiveDependency":false,"isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.puppycrawl.tools:checkstyle:8.29","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-10782","vulnerabilityDetails":"All versions of com.puppycrawl.tools:checkstyle before 8.29 are vulnerable to XML External Entity (XXE) Injection due to an incomplete fix for CVE-2019-9658.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10782","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-10782 (Medium) detected in checkstyle-8.27.jar - autoclosed - ## CVE-2019-10782 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>checkstyle-8.27.jar</b></p></summary>
<p>Checkstyle is a development tool to help programmers write Java code
that adheres to a coding standard</p>
<p>Library home page: <a href="https://checkstyle.org/">https://checkstyle.org/</a></p>
<p>
Dependency Hierarchy:
<p>Found in HEAD commit: <a href="https://github.com/opensearch-project/data-prepper/commit/022b333dc9be3548b8eb8bb73d0337fd26425056">022b333dc9be3548b8eb8bb73d0337fd26425056</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of com.puppycrawl.tools:checkstyle before 8.29 are vulnerable to XML External Entity (XXE) Injection due to an incomplete fix for CVE-2019-9658.
<p>Publish Date: 2020-01-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10782>CVE-2019-10782</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10782">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10782</a></p>
<p>Release Date: 2020-02-10</p>
<p>Fix Resolution: com.puppycrawl.tools:checkstyle:8.29</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.puppycrawl.tools","packageName":"checkstyle","packageVersion":"8.27","packageFilePaths":[],"isTransitiveDependency":false,"isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.puppycrawl.tools:checkstyle:8.29","isBinary":true}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-10782","vulnerabilityDetails":"All versions of com.puppycrawl.tools:checkstyle before 8.29 are vulnerable to XML External Entity (XXE) Injection due to an incomplete fix for CVE-2019-9658.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10782","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in checkstyle jar autoclosed cve medium severity vulnerability vulnerable library checkstyle jar checkstyle is a development tool to help programmers write java code that adheres to a coding standard library home page a href dependency hierarchy found in head commit a href found in base branch main vulnerability details all versions of com puppycrawl tools checkstyle before are vulnerable to xml external entity xxe injection due to an incomplete fix for cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com puppycrawl tools checkstyle isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false isminimumfixversionavailable true minimumfixversion com puppycrawl tools checkstyle isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails all versions of com puppycrawl tools checkstyle before are vulnerable to xml external entity xxe injection due to an incomplete fix for cve vulnerabilityurl
| 0
|
18,909
| 3,098,087,211
|
IssuesEvent
|
2015-08-28 08:40:22
|
mbj4668/pyang
|
https://api.github.com/repos/mbj4668/pyang
|
closed
|
Tree plugin: missing i_config attribute
|
bug Priority-High Type-Defect
|
Tree plugin gives the following stack trace for ietf-routing. The statement on which it fails is `net-hop-address` leaf.
```
$ git clone git@github.com:netmod-wg/routing-cfg.git
...
$ cd routing-cfg/
$ make yang
$ make hello.xml
$ pyang -f tree -o model.tree -L hello.xml
Traceback (most recent call last):
File "/Users/lhotka/bin/pyang", line 403, in <module>
run()
File "/Users/lhotka/bin/pyang", line 377, in run
emit_obj.emit(ctx, modules, fd)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 54, in emit
emit_tree(ctx, modules, fd, ctx.opts.tree_depth, path)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 117, in emit_tree
print_children(chs, module, fd, ' ', path, 'data', depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 259, in print_node
print_children(chs, module, fd, prefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 212, in print_node
flags = get_flags_str(s, mode)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 279, in get_flags_str
elif s.i_config == True:
AttributeError: 'Statement' object has no attribute 'i_config'
```
|
1.0
|
Tree plugin: missing i_config attribute - Tree plugin gives the following stack trace for ietf-routing. The statement on which it fails is `net-hop-address` leaf.
```
$ git clone git@github.com:netmod-wg/routing-cfg.git
...
$ cd routing-cfg/
$ make yang
$ make hello.xml
$ pyang -f tree -o model.tree -L hello.xml
Traceback (most recent call last):
File "/Users/lhotka/bin/pyang", line 403, in <module>
run()
File "/Users/lhotka/bin/pyang", line 377, in run
emit_obj.emit(ctx, modules, fd)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 54, in emit
emit_tree(ctx, modules, fd, ctx.opts.tree_depth, path)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 117, in emit_tree
print_children(chs, module, fd, ' ', path, 'data', depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 261, in print_node
print_children(chs, module, fd, prefix, path, mode, depth)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 259, in print_node
print_children(chs, module, fd, prefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 203, in print_children
print_node(ch, module, fd, newprefix, path, mode, depth, width)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 212, in print_node
flags = get_flags_str(s, mode)
File "/Users/lhotka/lib/python/pyang/plugins/tree.py", line 279, in get_flags_str
elif s.i_config == True:
AttributeError: 'Statement' object has no attribute 'i_config'
```
|
non_process
|
tree plugin missing i config attribute tree plugin gives the following stack trace for ietf routing the statement on which it fails is net hop address leaf git clone git github com netmod wg routing cfg git cd routing cfg make yang make hello xml pyang f tree o model tree l hello xml traceback most recent call last file users lhotka bin pyang line in run file users lhotka bin pyang line in run emit obj emit ctx modules fd file users lhotka lib python pyang plugins tree py line in emit emit tree ctx modules fd ctx opts tree depth path file users lhotka lib python pyang plugins tree py line in emit tree print children chs module fd path data depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node print children chs module fd prefix path mode depth width file users lhotka lib python pyang plugins tree py line in print children print node ch module fd newprefix path mode depth width file users lhotka lib python pyang plugins tree py line in print node flags get flags str s mode file users lhotka lib python pyang plugins tree py line in get flags str elif s i config true attributeerror statement object has no attribute i config
| 0
|
759,543
| 26,600,514,829
|
IssuesEvent
|
2023-01-23 15:27:34
|
Assignaly/assignaly-api
|
https://api.github.com/repos/Assignaly/assignaly-api
|
closed
|
Create router middleware
|
api feature update core mvp priority:high
|
As a developer I want to be able to execute code before reaching the callback method given to a route.
This could be something like user authentication or turning empty interfaces into strings.
- [ ] Create middleware interface and necessary methods and variables
|
1.0
|
Create router middleware - As a developer I want to be able to execute code before reaching the callback method given to a route.
This could be something like user authentication or turning empty interfaces into strings.
- [ ] Create middleware interface and necessary methods and variables
|
non_process
|
create router middleware as a developer i want to be able to execute code before reaching the callback method given to a route this could be something like user authentication or turning empty interfaces into strings create middleware interface and necessary methods and variables
| 0
|
98,934
| 12,379,149,314
|
IssuesEvent
|
2020-05-19 11:59:49
|
raiden-network/webui
|
https://api.github.com/repos/raiden-network/webui
|
closed
|
Add complete transfer history in subpage
|
Alderaan Redesign
|
### Description
Add a functionality / screen to see the complete transfer history. In #349 we decided to have more pages. This will be in a separate page.
### Tasks
- [x] Add complete transfer history
- [x] Add pagination for transfer history
|
1.0
|
Add complete transfer history in subpage - ### Description
Add a functionality / screen to see the complete transfer history. In #349 we decided to have more pages. This will be in a separate page.
### Tasks
- [x] Add complete transfer history
- [x] Add pagination for transfer history
|
non_process
|
add complete transfer history in subpage description add a functionality screen to see the complete transfer history in we decided to have more pages this will be in a separate page tasks add complete transfer history add pagination for transfer history
| 0
|
150,686
| 19,610,137,307
|
IssuesEvent
|
2022-01-06 14:30:23
|
ghc-dev/Michelle-Jones
|
https://api.github.com/repos/ghc-dev/Michelle-Jones
|
opened
|
CVE-2020-9488 (Low) detected in log4j-core-2.8.2.jar
|
security vulnerability
|
## CVE-2020-9488 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-core-2.8.2.jar</b></p></summary>
<p>The Apache Log4j Implementation</p>
<p>Library home page: <a href="https://logging.apache.org/log4j/2.x/log4j-core/">https://logging.apache.org/log4j/2.x/log4j-core/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /sitory/org/apache/logging/log4j/log4j-core/2.8.2/log4j-core-2.8.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-core-2.8.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Michelle-Jones/commit/48bc7ec513d624a25e5109795f26c020f7b0241a">48bc7ec513d624a25e5109795f26c020f7b0241a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.8.2","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.13.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-9488","vulnerabilityDetails":"Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-9488 (Low) detected in log4j-core-2.8.2.jar - ## CVE-2020-9488 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-core-2.8.2.jar</b></p></summary>
<p>The Apache Log4j Implementation</p>
<p>Library home page: <a href="https://logging.apache.org/log4j/2.x/log4j-core/">https://logging.apache.org/log4j/2.x/log4j-core/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /sitory/org/apache/logging/log4j/log4j-core/2.8.2/log4j-core-2.8.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-core-2.8.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Michelle-Jones/commit/48bc7ec513d624a25e5109795f26c020f7b0241a">48bc7ec513d624a25e5109795f26c020f7b0241a</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.8.2","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.8.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.13.2","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-9488","vulnerabilityDetails":"Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve low detected in core jar cve low severity vulnerability vulnerable library core jar the apache implementation library home page a href path to dependency file pom xml path to vulnerable library sitory org apache logging core core jar dependency hierarchy x core jar vulnerable library found in head commit a href found in base branch master vulnerability details improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache logging core isminimumfixversionavailable true minimumfixversion org apache logging core isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender vulnerabilityurl
| 0
|
11,649
| 14,503,020,924
|
IssuesEvent
|
2020-12-11 21:58:07
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
[automl] no information about `AUTOML_PROJECT_ID` and `OBJECT_DETECTION_MODEL_ID`
|
api: automl priority: p2 type: process
|
I tried to run `automl/cloud-client/vision_object_detection_predict_test.py` against my project then it errored out because I don't set `AUTOML_PROJECT_ID` and `OBJECT_DETECTION_MODEL_ID`.
It's great if we have those environment variables in [`testing/test-env.tmpl.sh`](https://github.com/GoogleCloudPlatform/python-docs-samples/blob/master/testing/test-env.tmpl.sh) with some explanations for why you need a different cloud project and how to prepare the model.
|
1.0
|
[automl] no information about `AUTOML_PROJECT_ID` and `OBJECT_DETECTION_MODEL_ID` - I tried to run `automl/cloud-client/vision_object_detection_predict_test.py` against my project then it errored out because I don't set `AUTOML_PROJECT_ID` and `OBJECT_DETECTION_MODEL_ID`.
It's great if we have those environment variables in [`testing/test-env.tmpl.sh`](https://github.com/GoogleCloudPlatform/python-docs-samples/blob/master/testing/test-env.tmpl.sh) with some explanations for why you need a different cloud project and how to prepare the model.
|
process
|
no information about automl project id and object detection model id i tried to run automl cloud client vision object detection predict test py against my project then it errored out because i don t set automl project id and object detection model id it s great if we have those environment variables in with some explanations for why you need a different cloud project and how to prepare the model
| 1
|
289,034
| 8,854,304,687
|
IssuesEvent
|
2019-01-09 00:43:47
|
jstanden/cerb
|
https://api.github.com/repos/jstanden/cerb
|
closed
|
Bulk delete
|
bug priority-support
|
In the latest 9.1 release, bulk delete _tickets_ or attachments does not appear to work.
It lets you submit the task but the worklist just refreshes without any action taken.
If you delete the _tickets_ individually on the worklist, this works.
|
1.0
|
Bulk delete - In the latest 9.1 release, bulk delete _tickets_ or attachments does not appear to work.
It lets you submit the task but the worklist just refreshes without any action taken.
If you delete the _tickets_ individually on the worklist, this works.
|
non_process
|
bulk delete in the latest release bulk delete tickets or attachments does not appear to work it lets you submit the task but the worklist just refreshes without any action taken if you delete the tickets individually on the worklist this works
| 0
|
453,645
| 13,086,232,602
|
IssuesEvent
|
2020-08-02 05:14:13
|
Journaly/journaly
|
https://api.github.com/repos/Journaly/journaly
|
closed
|
Visual Bugs & Improvements
|
bug high priority visual
|
#### Visual Bugs To Fix
- [x] My Feed search box isn't the right width [#181]
- [x] Chevron/down arrow is in wrong position on My Feed filters [#181]
- [x] Chevron/down arrow is in wrong position on Settings Add Languages
- [x] Chevron/down arrow is in wrong position on New Post drop downs
- [x] Settings > languages drop downs are off on mobile view
|
1.0
|
Visual Bugs & Improvements - #### Visual Bugs To Fix
- [x] My Feed search box isn't the right width [#181]
- [x] Chevron/down arrow is in wrong position on My Feed filters [#181]
- [x] Chevron/down arrow is in wrong position on Settings Add Languages
- [x] Chevron/down arrow is in wrong position on New Post drop downs
- [x] Settings > languages drop downs are off on mobile view
|
non_process
|
visual bugs improvements visual bugs to fix my feed search box isn t the right width chevron down arrow is in wrong position on my feed filters chevron down arrow is in wrong position on settings add languages chevron down arrow is in wrong position on new post drop downs settings languages drop downs are off on mobile view
| 0
|
17,972
| 23,984,015,533
|
IssuesEvent
|
2022-09-13 17:22:19
|
Open-Data-Product-Initiative/open-data-product-spec
|
https://api.github.com/repos/Open-Data-Product-Initiative/open-data-product-spec
|
closed
|
New elements to dataPipeline: platform, technology, storageType
|
enhancement processed
|
`"platform": "GCP"`
Options: Azure, AWS, GCP, or probably just Terraform.
`"technology": "Amazon S3" `
Describes the internal storage area technology, such as: Amazon S3, Google Cloud Storage, Azure Blob Storage.
`"storageType": "files"`
Describes the internal storage type: files, sql, events, MQTT, etc.
All types are string.
The purpose is to model the principles of data mesh
|
1.0
|
New elements to dataPipeline: platform, technology, storageType - `"platform": "GCP"`
Options: Azure, AWS, GCP, or probably just Terraform.
`"technology": "Amazon S3" `
Describes the internal storage area technology, such as: Amazon S3, Google Cloud Storage, Azure Blob Storage.
`"storageType": "files"`
Describes the internal storage type: files, sql, events, MQTT, etc.
All types are string.
The purpose is to model the principles of data mesh
|
process
|
new elements to datapipeline platform technology storagetype platform gcp options azure aws gcp or probably just terraform technology amazon describes the internal storage area technology such as amazon google cloud storage azure blob storage storagetype files describes the internal storage type files sql events mqtt etc all types are string the purpose is to model the principles of data mesh
| 1
|
6,127
| 8,996,635,507
|
IssuesEvent
|
2019-02-02 03:04:04
|
bow-simulation/virtualbow
|
https://api.github.com/repos/bow-simulation/virtualbow
|
closed
|
Use Gitlab-CI to build and run tests automatically
|
area: software process prio: normal type: improvement
|
In GitLab by **spfeifer** on Dec 7, 2018, 11:52
This is a first step towards #33
Optional Bonus: Provide development build artifacts (AppImage, deb, rpm)
|
1.0
|
Use Gitlab-CI to build and run tests automatically - In GitLab by **spfeifer** on Dec 7, 2018, 11:52
This is a first step towards #33
Optional Bonus: Provide development build artifacts (AppImage, deb, rpm)
|
process
|
use gitlab ci to build and run tests automatically in gitlab by spfeifer on dec this is a first step towards optional bonus provide development build artifacts appimage deb rpm
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.