Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
520,233 | 15,082,093,977 | IssuesEvent | 2021-02-05 14:07:39 | OpenSRP/opensrp-client-reveal | https://api.github.com/repos/OpenSRP/opensrp-client-reveal | closed | Mark residential structures as ineligible | Android Client Pilot Critical Priority: High enhancement | The Thailand team would like the ability to mark an unenumerated residential structure as ineligible so that they can state the location was visited and is not eligible for the intervention. An ineligible structure is a valid structure that is ineligible to participate in an intervention for a valid reason (e.g., a residential structure where the inhabitants are not currently residing but may return in the future).
### Changes
Update Register Family Form (both English and Thai)
- [ ] Add a question to the register family form that asks if this structure is eligible with a default value = Eligible (https://docs.google.com/spreadsheets/d/1kGPgyF0eJdag19FFL1Oog2gdphdPtKobY4EQ0gMgtgw/edit#gid=951684107&range=14:14)
- [ ] Hide all other fields in the form if the value of this is changed to Ineligible.
- [ ] If a user is editing the Register Family form, the eligible field is not editable, do not ask any other questions and allow them to save the form immediately (do not go to any other steps of the form)
Update Register Family Form Post Save Action:
If the field eligible = ineligible
1) We do not create the family in the family module.
2) We mark the RACD Register Family task with businessStatus = Ineligible
3) If any tasks are assigned to that location, we mark those tasks as Ineligible
Update colour in map view and change on tap behaviour
- [ ] We need to display a black icon in the map view when a structure has been marked with businessStatus = ineligible
- [ ] When a user taps the icon, they need to be presented with a card that displays the date that the location was marked as ineligible and the user who completed the form. (See Add Card View Below)
Add Card View
- [ ] We need to add a card view to a structure that has the same fields as the non-residential structures (mosquito collection point and larval dipping sites)
- [ ] The card does not have any buttons
- [ ] The user has the ability to hide the card | 1.0 | Mark residential structures as ineligible - The Thailand team would like the ability to mark an unenumerated residential structure as ineligible so that they can state the location was visited and is not eligible for the intervention. An ineligible structure is a valid structure that is ineligible to participate in an intervention for a valid reason (e.g., a residential structure where the inhabitants are not currently residing but may return in the future).
### Changes
Update Register Family Form (both English and Thai)
- [ ] Add a question to the register family form that asks if this structure is eligible with a default value = Eligible (https://docs.google.com/spreadsheets/d/1kGPgyF0eJdag19FFL1Oog2gdphdPtKobY4EQ0gMgtgw/edit#gid=951684107&range=14:14)
- [ ] Hide all other fields in the form if the value of this is changed to Ineligible.
- [ ] If a user is editing the Register Family form, the eligible field is not editable, do not ask any other questions and allow them to save the form immediately (do not go to any other steps of the form)
Update Register Family Form Post Save Action:
If the field eligible = ineligible
1) We do not create the family in the family module.
2) We mark the RACD Register Family task with businessStatus = Ineligible
3) If any tasks are assigned to that location, we mark those tasks as Ineligible
Update colour in map view and change on tap behaviour
- [ ] We need to display a black icon in the map view when a structure has been marked with businessStatus = ineligible
- [ ] When a user taps the icon, they need to be presented with a card that displays the date that the location was marked as ineligible and the user who completed the form. (See Add Card View Below)
Add Card View
- [ ] We need to add a card view to a structure that has the same fields as the non-residential structures (mosquito collection point and larval dipping sites)
- [ ] The card does not have any buttons
- [ ] The user has the ability to hide the card | priority | mark residential structures as ineligible the thailand team would like the ability to mark an unenumerated residential structure as ineligible so that they can state the location was visited and is not eligible for the intervention an ineligible structure is a valid structure that is ineligible to participate in an intervention for a valid reason e g a residential structure where the inhabitants are not currently residing but may return in the future changes update register family form both english and thai add a question to the register family form that asks if this structure is eligible with a default value eligible hide all other fields in the form if the value of this is changed to ineligible if a user is editing the register family form the eligible field is not editable do not ask any other questions and allow them to save the form immediately do not go to any other steps of the form update register family form post save action if the field eligible ineligible we do not create the family in the family module we mark the racd register family task with businessstatus ineligible if any tasks are assigned to that location we mark those tasks as ineligible update colour in map view and change on tap behaviour we need to display a black icon in the map view when a structure has been marked with businessstatus ineligible when a user taps the icon they need to be presented with a card that displays the date that the location was marked as ineligible and the user who completed the form see add card view below add card view we need to add a card view to a structure that has the same fields as the non residential structures mosquito collection point and larval dipping sites the card does not have any buttons the user has the ability to hide the card | 1 |
300,499 | 9,211,291,451 | IssuesEvent | 2019-03-09 14:08:34 | qgisissuebot/QGIS | https://api.github.com/repos/qgisissuebot/QGIS | closed | 'processing' extension cannot be loaded. | Bug Priority: normal Regression | ---
Author Name: **Robin Schwemmle** (Robin Schwemmle)
Original Redmine Issue: 21411, https://issues.qgis.org/issues/21411
Original Assignee: Denis Rouzaud
---
The processing toolbox is not available. I followed the instructions in the README and added the path of the GDAL framework, but this doesn't help. It seems like, that the osgeo package for python is missing. Using pip3 to install gdal failed with error code 1.
This is the error message I get when launching QGIS 3.6:
Could not load 'processing' extension.
ModuleNotFoundError: No module named 'osgeo'
Traceback (most recent call last):
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 309, in loadPlugin
__import__(packageName)
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 672, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/plugins/processing/__init__.py", line 28, in
from processing.tools.dataobjects import * # NOQA
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 672, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/plugins/processing/tools/dataobjects.py", line 51, in
from processing.algs.gdal.GdalUtils import GdalUtils
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 672, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/plugins/processing/algs/gdal/GdalUtils.py", line 38, in
from osgeo import ogr
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 672, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
ModuleNotFoundError: No module named 'osgeo'
Python-Version: 3.6.8 (v3.6.8:3c6b436a57, Dec 24 2018, 02:04:31) [GCC 4.2.1 Compatible Apple LLVM 6.0 (clang-600.0.57)]
QGIS-Version: 3.6.0-Noosa Noosa, exported
| 1.0 | 'processing' extension cannot be loaded. - ---
Author Name: **Robin Schwemmle** (Robin Schwemmle)
Original Redmine Issue: 21411, https://issues.qgis.org/issues/21411
Original Assignee: Denis Rouzaud
---
The processing toolbox is not available. I followed the instructions in the README and added the path of the GDAL framework, but this doesn't help. It seems like, that the osgeo package for python is missing. Using pip3 to install gdal failed with error code 1.
This is the error message I get when launching QGIS 3.6:
Could not load 'processing' extension.
ModuleNotFoundError: No module named 'osgeo'
Traceback (most recent call last):
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 309, in loadPlugin
__import__(packageName)
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 672, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/plugins/processing/__init__.py", line 28, in
from processing.tools.dataobjects import * # NOQA
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 672, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/plugins/processing/tools/dataobjects.py", line 51, in
from processing.algs.gdal.GdalUtils import GdalUtils
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 672, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/plugins/processing/algs/gdal/GdalUtils.py", line 38, in
from osgeo import ogr
File "/Applications/QGIS3.6.app/Contents/MacOS/../Resources/python/qgis/utils.py", line 672, in _import
mod = _builtin_import(name, globals, locals, fromlist, level)
ModuleNotFoundError: No module named 'osgeo'
Python-Version: 3.6.8 (v3.6.8:3c6b436a57, Dec 24 2018, 02:04:31) [GCC 4.2.1 Compatible Apple LLVM 6.0 (clang-600.0.57)]
QGIS-Version: 3.6.0-Noosa Noosa, exported
| priority | processing extension cannot be loaded author name robin schwemmle robin schwemmle original redmine issue original assignee denis rouzaud the processing toolbox is not available i followed the instructions in the readme and added the path of the gdal framework but this doesn t help it seems like that the osgeo package for python is missing using to install gdal failed with error code this is the error message i get when launching qgis could not load processing extension modulenotfounderror no module named osgeo traceback most recent call last file applications app contents macos resources python qgis utils py line in loadplugin import packagename file applications app contents macos resources python qgis utils py line in import mod builtin import name globals locals fromlist level file applications app contents macos resources python plugins processing init py line in from processing tools dataobjects import noqa file applications app contents macos resources python qgis utils py line in import mod builtin import name globals locals fromlist level file applications app contents macos resources python plugins processing tools dataobjects py line in from processing algs gdal gdalutils import gdalutils file applications app contents macos resources python qgis utils py line in import mod builtin import name globals locals fromlist level file applications app contents macos resources python plugins processing algs gdal gdalutils py line in from osgeo import ogr file applications app contents macos resources python qgis utils py line in import mod builtin import name globals locals fromlist level modulenotfounderror no module named osgeo python version dec qgis version noosa noosa exported | 1 |
26,464 | 2,684,555,930 | IssuesEvent | 2015-03-29 03:30:59 | gtcasl/gpuocelot | https://api.github.com/repos/gtcasl/gpuocelot | closed | ret instruction for global entry kernel not supported in llvm bacend (cuda 4.1) | bug imported Priority-High | _From [wangjin....@gmail.com](https://code.google.com/u/115604721152422776435/) on January 21, 2012 19:42:26_
What steps will reproduce the problem? 1.Use cuda toolkit 4.1
2.Compile helloWorld.cu attached with nvcc, then link against libocelot
3.Set llvm as backend in configure.ocelot
4.Run helloWolrd What is the expected output? What do you see instead? Error infomration:
helloWorld: ocelot/executive/implementation/LLVMCooperativeThreadArray.cpp:495: bool executive::LLVMCooperativeThreadArray::_finishContext(unsigned int): Assertion `nextFunction \< _queuedThreads.size()' failed. Please use labels and text to provide additional information. This problem exists for all apps compiled with cuda toolkit 4.1. I traced the problem a little and found that cuda 4.1 use PTX instruction "ret" instead of "exit" (as cuda 4.0 or before does) in the exit point of the kernel with __global__ entry. However, ocelot (llvm backend) does not detect this situation and simply assumes that "ret" means the subkernel returns to its caller. It fails when it tries to pop the stack to get the caller function which actually does not exist in this case.
**Attachment:** [helloWorld.cu](http://code.google.com/p/gpuocelot/issues/detail?id=62)
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=62_ | 1.0 | ret instruction for global entry kernel not supported in llvm bacend (cuda 4.1) - _From [wangjin....@gmail.com](https://code.google.com/u/115604721152422776435/) on January 21, 2012 19:42:26_
What steps will reproduce the problem? 1.Use cuda toolkit 4.1
2.Compile helloWorld.cu attached with nvcc, then link against libocelot
3.Set llvm as backend in configure.ocelot
4.Run helloWolrd What is the expected output? What do you see instead? Error infomration:
helloWorld: ocelot/executive/implementation/LLVMCooperativeThreadArray.cpp:495: bool executive::LLVMCooperativeThreadArray::_finishContext(unsigned int): Assertion `nextFunction \< _queuedThreads.size()' failed. Please use labels and text to provide additional information. This problem exists for all apps compiled with cuda toolkit 4.1. I traced the problem a little and found that cuda 4.1 use PTX instruction "ret" instead of "exit" (as cuda 4.0 or before does) in the exit point of the kernel with __global__ entry. However, ocelot (llvm backend) does not detect this situation and simply assumes that "ret" means the subkernel returns to its caller. It fails when it tries to pop the stack to get the caller function which actually does not exist in this case.
**Attachment:** [helloWorld.cu](http://code.google.com/p/gpuocelot/issues/detail?id=62)
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=62_ | priority | ret instruction for global entry kernel not supported in llvm bacend cuda from on january what steps will reproduce the problem use cuda toolkit compile helloworld cu attached with nvcc then link against libocelot set llvm as backend in configure ocelot run hellowolrd what is the expected output what do you see instead error infomration helloworld ocelot executive implementation llvmcooperativethreadarray cpp bool executive llvmcooperativethreadarray finishcontext unsigned int assertion nextfunction queuedthreads size failed please use labels and text to provide additional information this problem exists for all apps compiled with cuda toolkit i traced the problem a little and found that cuda use ptx instruction ret instead of exit as cuda or before does in the exit point of the kernel with global entry however ocelot llvm backend does not detect this situation and simply assumes that ret means the subkernel returns to its caller it fails when it tries to pop the stack to get the caller function which actually does not exist in this case attachment original issue | 1 |
425,608 | 12,342,984,881 | IssuesEvent | 2020-05-15 02:32:33 | Matteas-Eden/roll-for-reaction | https://api.github.com/repos/Matteas-Eden/roll-for-reaction | closed | Add more spells to the spellbook | Medium Priority enhancement | **User Story**
As a game designer, I'd like to offer spellcasting players a wide variety of spells and provide an element of progression with those spells.
**Acceptance Criteria**
- More spells are added to the spellbook
- Spells are unlocked based on level
**Notes**
---
**Why is this feature needed? Please describe the problem your requested feature wants to solve**
There are only 2 spells available, *Fireball* and *Mend*.
**Describe the solution you'd like**
Additional spells should be added, and these spells should be unlocked as the player reaches certain levels.
| 1.0 | Add more spells to the spellbook - **User Story**
As a game designer, I'd like to offer spellcasting players a wide variety of spells and provide an element of progression with those spells.
**Acceptance Criteria**
- More spells are added to the spellbook
- Spells are unlocked based on level
**Notes**
---
**Why is this feature needed? Please describe the problem your requested feature wants to solve**
There are only 2 spells available, *Fireball* and *Mend*.
**Describe the solution you'd like**
Additional spells should be added, and these spells should be unlocked as the player reaches certain levels.
| priority | add more spells to the spellbook user story as a game designer i d like to offer spellcasting players a wide variety of spells and provide an element of progression with those spells acceptance criteria more spells are added to the spellbook spells are unlocked based on level notes why is this feature needed please describe the problem your requested feature wants to solve there are only spells available fireball and mend describe the solution you d like additional spells should be added and these spells should be unlocked as the player reaches certain levels | 1 |
465,182 | 13,358,227,770 | IssuesEvent | 2020-08-31 11:18:19 | webpack-contrib/file-loader | https://api.github.com/repos/webpack-contrib/file-loader | closed | Pass additional assetInfo object when calling emitFile | flag: Community help wanted priority: 5 (nice to have) semver: Minor type: Feature | <!--
Issues are so 🔥
If you remove or skip this template, you'll make the 🐼 sad and the mighty god
of Github will appear and pile-drive the close button from a great height
while making animal noises.
👉🏽 Need support, advice, or help? Don't open an issue!
Head to StackOverflow or https://gitter.im/webpack/webpack.
-->
- Operating System: MacOS
- Node Version: 12.6.1
- NPM Version: 6.14.7
- webpack Version: 4.29.6
- file-loader Version: 3.0.1
### Feature Proposal
Webpack loader's [emitFile](https://github.com/webpack-contrib/file-loader/blob/master/src/index.js#L59) call takes a 4th param that can provide additional info about an asset, like whether it is immutable or not. `file-loader` should pass in `{ immutable: true }` when the file name includes `[contenthash]` so that other plugins can benefit from this info.
The new `assetInfo` param was added in https://github.com/webpack/webpack/issues/9038.
### Feature Use Case
I'm writing a plugin that uses this info to determine what cache-control headers a file can be given. | 1.0 | Pass additional assetInfo object when calling emitFile - <!--
Issues are so 🔥
If you remove or skip this template, you'll make the 🐼 sad and the mighty god
of Github will appear and pile-drive the close button from a great height
while making animal noises.
👉🏽 Need support, advice, or help? Don't open an issue!
Head to StackOverflow or https://gitter.im/webpack/webpack.
-->
- Operating System: MacOS
- Node Version: 12.6.1
- NPM Version: 6.14.7
- webpack Version: 4.29.6
- file-loader Version: 3.0.1
### Feature Proposal
Webpack loader's [emitFile](https://github.com/webpack-contrib/file-loader/blob/master/src/index.js#L59) call takes a 4th param that can provide additional info about an asset, like whether it is immutable or not. `file-loader` should pass in `{ immutable: true }` when the file name includes `[contenthash]` so that other plugins can benefit from this info.
The new `assetInfo` param was added in https://github.com/webpack/webpack/issues/9038.
### Feature Use Case
I'm writing a plugin that uses this info to determine what cache-control headers a file can be given. | priority | pass additional assetinfo object when calling emitfile issues are so 🔥 if you remove or skip this template you ll make the 🐼 sad and the mighty god of github will appear and pile drive the close button from a great height while making animal noises 👉🏽 need support advice or help don t open an issue head to stackoverflow or operating system macos node version npm version webpack version file loader version feature proposal webpack loader s call takes a param that can provide additional info about an asset like whether it is immutable or not file loader should pass in immutable true when the file name includes so that other plugins can benefit from this info the new assetinfo param was added in feature use case i m writing a plugin that uses this info to determine what cache control headers a file can be given | 1 |
92,995 | 11,730,391,184 | IssuesEvent | 2020-03-10 21:16:39 | ibm-openbmc/dev | https://api.github.com/repos/ibm-openbmc/dev | closed | GUI : Design : Revisit Header | GUI UI Design | ## Stakeholders
**SME**: n/a
**Design Researcher**: @nicoleconser
**UX Designer**: TBD
**FED**: TBD
## Use Case
Through talking with end users, several questions involving the header have surfaced, including:
• **Firmware Level** — "It's very helpful that the ASMI interface displays the firmware level of the server in the upper right corner of the GUI." (feedback from 6 Sales Reps during Masters 2019)
• **Process pending / in progress** — Would including a visual indication of a process that is *in progress* or *pending* improve the user experience? How? Indication might be a simple notification OR may be a warning that the user should not perform certain operations during this time.
• **Server Power On / Off / Standby** — For consistency, should we update the Server Power On / Off / Standby / Error labels in the header, to match the verbiage on the Firmware page?
• **Refresh button** — Do users' understanding of what this button does match its intended use?
## InVision Prototype
n/a
## Design Issue (phosphor-webui)
n/a
## Development Issue
n/a
## References/Resources
- feature discovery: https://ibm.box.com/s/xmq7u0dvqqvl72xng3dvop10s0ki72yd
- user research*: TBD
*Restricted access in accordance with GDPR guidelines.
| 1.0 | GUI : Design : Revisit Header - ## Stakeholders
**SME**: n/a
**Design Researcher**: @nicoleconser
**UX Designer**: TBD
**FED**: TBD
## Use Case
Through talking with end users, several questions involving the header have surfaced, including:
• **Firmware Level** — "It's very helpful that the ASMI interface displays the firmware level of the server in the upper right corner of the GUI." (feedback from 6 Sales Reps during Masters 2019)
• **Process pending / in progress** — Would including a visual indication of a process that is *in progress* or *pending* improve the user experience? How? Indication might be a simple notification OR may be a warning that the user should not perform certain operations during this time.
• **Server Power On / Off / Standby** — For consistency, should we update the Server Power On / Off / Standby / Error labels in the header, to match the verbiage on the Firmware page?
• **Refresh button** — Do users' understanding of what this button does match its intended use?
## InVision Prototype
n/a
## Design Issue (phosphor-webui)
n/a
## Development Issue
n/a
## References/Resources
- feature discovery: https://ibm.box.com/s/xmq7u0dvqqvl72xng3dvop10s0ki72yd
- user research*: TBD
*Restricted access in accordance with GDPR guidelines.
| non_priority | gui design revisit header stakeholders sme n a design researcher nicoleconser ux designer tbd fed tbd use case through talking with end users several questions involving the header have surfaced including • firmware level — it s very helpful that the asmi interface displays the firmware level of the server in the upper right corner of the gui feedback from sales reps during masters • process pending in progress — would including a visual indication of a process that is in progress or pending improve the user experience how indication might be a simple notification or may be a warning that the user should not perform certain operations during this time • server power on off standby — for consistency should we update the server power on off standby error labels in the header to match the verbiage on the firmware page • refresh button — do users understanding of what this button does match its intended use invision prototype n a design issue phosphor webui n a development issue n a references resources feature discovery user research tbd restricted access in accordance with gdpr guidelines | 0 |
420,838 | 12,244,511,322 | IssuesEvent | 2020-05-05 11:16:44 | minio/mc | https://api.github.com/repos/minio/mc | closed | [Proposal] Support STS based access to MinIO | priority: medium stale | To access MinIO via the STS API, we can add a command to mc such as:
`mc sts ENDPOINT <sts-type> <sts-specific-args>`
to print out access key, secret key and session token.
For example for LDAP, this could be:
`mc sts ENDPOINT ldap ldapuser=myusername ldappasswd=yyy`
In addition to this, we can add a mode to mc to fetch and configure MinIO/S3 aliases with STS APIs.
For example, with `mc config sts ldap add ALIAS URL ldapuser=xx ldappasswd=yy`,
fetches temporary credentials via STS from the MinIO server and then configures the alias to use the returned access credentials. Now the user can make use of other mc commands to access the MinIO server. | 1.0 | [Proposal] Support STS based access to MinIO - To access MinIO via the STS API, we can add a command to mc such as:
`mc sts ENDPOINT <sts-type> <sts-specific-args>`
to print out access key, secret key and session token.
For example for LDAP, this could be:
`mc sts ENDPOINT ldap ldapuser=myusername ldappasswd=yyy`
In addition to this, we can add a mode to mc to fetch and configure MinIO/S3 aliases with STS APIs.
For example, with `mc config sts ldap add ALIAS URL ldapuser=xx ldappasswd=yy`,
fetches temporary credentials via STS from the MinIO server and then configures the alias to use the returned access credentials. Now the user can make use of other mc commands to access the MinIO server. | priority | support sts based access to minio to access minio via the sts api we can add a command to mc such as mc sts endpoint to print out access key secret key and session token for example for ldap this could be mc sts endpoint ldap ldapuser myusername ldappasswd yyy in addition to this we can add a mode to mc to fetch and configure minio aliases with sts apis for example with mc config sts ldap add alias url ldapuser xx ldappasswd yy fetches temporary credentials via sts from the minio server and then configures the alias to use the returned access credentials now the user can make use of other mc commands to access the minio server | 1 |
28,514 | 23,307,862,799 | IssuesEvent | 2022-08-08 04:25:55 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Create a system for deriving a pmf model from a generic base model. | newfeature interface/infrastructure | There are a lot of plant models in APSIM that are (or should be) very similar to other models. e.g. barley and oats should be similar to wheat. At the moment we don't know what differences there are between wheat/barley/oats or between the different legume models. Some consistency is needed.
We would like to create the concept of a base (generic) model that other models can derive from. e.g. we could create a winter cereal model that wheat, barley and oats derive from.
One way to implement this is to use the property overriding system that cultivars and factorials use. A plant model could then be the name of the base plant model and 1 or more property overrides. This would restrict derived models to simple re-parameterisation and not allow different model structures, which is one of the objectives for doing this. We already have code in APSIM to do parameter setting so the task should be straight-forward, although some refactoring will be needed first. | 1.0 | Create a system for deriving a pmf model from a generic base model. - There are a lot of plant models in APSIM that are (or should be) very similar to other models. e.g. barley and oats should be similar to wheat. At the moment we don't know what differences there are between wheat/barley/oats or between the different legume models. Some consistency is needed.
We would like to create the concept of a base (generic) model that other models can derive from. e.g. we could create a winter cereal model that wheat, barley and oats derive from.
One way to implement this is to use the property overriding system that cultivars and factorials use. A plant model could then be the name of the base plant model and 1 or more property overrides. This would restrict derived models to simple re-parameterisation and not allow different model structures, which is one of the objectives for doing this. We already have code in APSIM to do parameter setting so the task should be straight-forward, although some refactoring will be needed first. | non_priority | create a system for deriving a pmf model from a generic base model there are a lot of plant models in apsim that are or should be very similar to other models e g barley and oats should be similar to wheat at the moment we don t know what differences there are between wheat barley oats or between the different legume models some consistency is needed we would like to create the concept of a base generic model that other models can derive from e g we could create a winter cereal model that wheat barley and oats derive from one way to implement this is to use the property overriding system that cultivars and factorials use a plant model could then be the name of the base plant model and or more property overrides this would restrict derived models to simple re parameterisation and not allow different model structures which is one of the objectives for doing this we already have code in apsim to do parameter setting so the task should be straight forward although some refactoring will be needed first | 0 |
246,298 | 7,894,388,593 | IssuesEvent | 2018-06-28 21:18:22 | minetest/minetest | https://api.github.com/repos/minetest/minetest | closed | Android 0.4.17.1: Google Play Store build crash while loading singleplayer world | Android Bug High priority | ##### Issue type
<!-- Pick one below and delete others -->
- Bug report
##### Minetest version
<!--
Paste Minetest version between quotes below
If you are on a devel version, please add git commit hash
You can use `minetest --version` to find it.
-->
Google Play Store store build of `0.4.17.1` (4a48cd57, Android), apk version `0.4.17.19`
##### OS / Hardware
<!-- General information about your hardware and operating system -->
Operating system: Android 7.0
CPU: HiSilicon Kirin 650 2.11 Ghz (aarch64), 8 cores
##### Summary
<!-- Describe your problem here -->
When loading singleplayer on Android, Minetest Play Store build crash to the phone interface while no crash log is written to `debug.txt` file, happens a bit later after creating `worlds/singleplayerworld/ipban.txt` file.
Adding `debug_log_level = verbose` to `minetest.conf` produces more information but still no crash log, showing that the last operation is only to hash some files, after that game exits to the phone interface w/o any meaningful message in the log file (`debug.txt`).
from debug.txt (w/o verbose flag):
```
2018-06-12 07:19:46: ERROR[Main]: Subgame specified in default_game [minetest] is invalid.
2018-06-12 07:19:51: WARNING[Main]: Irrlicht: Could not open file of texture: /storage/emulated/0/Minetest/textures/base/pack/server_flags_.png
2018-06-12 07:19:51: WARNING[Main]: BanManager: creating /storage/emulated/0/Minetest/worlds/singleplayerworld/ipban.txt
```
verbose debug.txt: https://pastebin.com/S8YUEq16
**P.s.** There's only `map_meta.txt` and `world.mt` in `worlds/singleplayerworld` folder, there're no subfolder `minimal` in directory `games`.
**Update:** There's a working beta version on Play Store. After confirming that the future non-beta release would be working this issue should be closed.
##### Steps to reproduce
<!-- For bug reports or build issues, explain how the problem happened -->
1. Clean install Minetest from Google Play Store or delete data for Minetest app if it's already installed, update it if app version is older than `0.4.17.19`.
2. Remove or rename/move (preferred) `Minetest` folder in your storage directory if it exists.
3. Start Minetest app, click on Start Singleplayer button.
4. Wait a moment and watch how Minetest crashes to phone interface.
| 1.0 | Android 0.4.17.1: Google Play Store build crash while loading singleplayer world - ##### Issue type
<!-- Pick one below and delete others -->
- Bug report
##### Minetest version
<!--
Paste Minetest version between quotes below
If you are on a devel version, please add git commit hash
You can use `minetest --version` to find it.
-->
Google Play Store store build of `0.4.17.1` (4a48cd57, Android), apk version `0.4.17.19`
##### OS / Hardware
<!-- General information about your hardware and operating system -->
Operating system: Android 7.0
CPU: HiSilicon Kirin 650 2.11 Ghz (aarch64), 8 cores
##### Summary
<!-- Describe your problem here -->
When loading singleplayer on Android, Minetest Play Store build crash to the phone interface while no crash log is written to `debug.txt` file, happens a bit later after creating `worlds/singleplayerworld/ipban.txt` file.
Adding `debug_log_level = verbose` to `minetest.conf` produces more information but still no crash log, showing that the last operation is only to hash some files, after that game exits to the phone interface w/o any meaningful message in the log file (`debug.txt`).
from debug.txt (w/o verbose flag):
```
2018-06-12 07:19:46: ERROR[Main]: Subgame specified in default_game [minetest] is invalid.
2018-06-12 07:19:51: WARNING[Main]: Irrlicht: Could not open file of texture: /storage/emulated/0/Minetest/textures/base/pack/server_flags_.png
2018-06-12 07:19:51: WARNING[Main]: BanManager: creating /storage/emulated/0/Minetest/worlds/singleplayerworld/ipban.txt
```
verbose debug.txt: https://pastebin.com/S8YUEq16
**P.s.** There's only `map_meta.txt` and `world.mt` in `worlds/singleplayerworld` folder, there're no subfolder `minimal` in directory `games`.
**Update:** There's a working beta version on Play Store. After confirming that the future non-beta release would be working this issue should be closed.
##### Steps to reproduce
<!-- For bug reports or build issues, explain how the problem happened -->
1. Clean install Minetest from Google Play Store or delete data for Minetest app if it's already installed, update it if app version is older than `0.4.17.19`.
2. Remove or rename/move (preferred) `Minetest` folder in your storage directory if it exists.
3. Start Minetest app, click on Start Singleplayer button.
4. Wait a moment and watch how Minetest crashes to phone interface.
| priority | android google play store build crash while loading singleplayer world issue type bug report minetest version paste minetest version between quotes below if you are on a devel version please add git commit hash you can use minetest version to find it google play store store build of android apk version os hardware operating system android cpu hisilicon kirin ghz cores summary when loading singleplayer on android minetest play store build crash to the phone interface while no crash log is written to debug txt file happens a bit later after creating worlds singleplayerworld ipban txt file adding debug log level verbose to minetest conf produces more information but still no crash log showing that the last operation is only to hash some files after that game exits to the phone interface w o any meaningful message in the log file debug txt from debug txt w o verbose flag error subgame specified in default game is invalid warning irrlicht could not open file of texture storage emulated minetest textures base pack server flags png warning banmanager creating storage emulated minetest worlds singleplayerworld ipban txt verbose debug txt p s there s only map meta txt and world mt in worlds singleplayerworld folder there re no subfolder minimal in directory games update there s a working beta version on play store after confirming that the future non beta release would be working this issue should be closed steps to reproduce clean install minetest from google play store or delete data for minetest app if it s already installed update it if app version is older than remove or rename move preferred minetest folder in your storage directory if it exists start minetest app click on start singleplayer button wait a moment and watch how minetest crashes to phone interface | 1 |
14,022 | 5,536,632,959 | IssuesEvent | 2017-03-21 20:07:38 | SFTtech/openage | https://api.github.com/repos/SFTtech/openage | closed | Normalize install prefixes in cmake | buildsystem easy improvement just do it | Please normalize the paths that are shown after the buildsystem configuration step:
observed:
```
install prefix | /usr/local/../asdf/../test
py install prefix | /usr/lib/python-exec/python3.5/../../../lib64/python3.5/site-packages
```
expected:
```
install prefix | /usr/test
py install prefix | /usr/lib64/python3.5/site-packages
```
| 1.0 | Normalize install prefixes in cmake - Please normalize the paths that are shown after the buildsystem configuration step:
observed:
```
install prefix | /usr/local/../asdf/../test
py install prefix | /usr/lib/python-exec/python3.5/../../../lib64/python3.5/site-packages
```
expected:
```
install prefix | /usr/test
py install prefix | /usr/lib64/python3.5/site-packages
```
| non_priority | normalize install prefixes in cmake please normalize the paths that are shown after the buildsystem configuration step observed install prefix usr local asdf test py install prefix usr lib python exec site packages expected install prefix usr test py install prefix usr site packages | 0 |
609,532 | 18,876,064,299 | IssuesEvent | 2021-11-14 02:16:26 | massenergize/api | https://api.github.com/repos/massenergize/api | closed | Sandbox for all communities page | bug priority 1 | Make ?sandbox=true show non-published communities on all communities page | 1.0 | Sandbox for all communities page - Make ?sandbox=true show non-published communities on all communities page | priority | sandbox for all communities page make sandbox true show non published communities on all communities page | 1 |
72,438 | 9,592,942,391 | IssuesEvent | 2019-05-09 10:09:38 | gama-platform/gama | https://api.github.com/repos/gama-platform/gama | opened | introduce ?help in interactive console to bring help into console | > Enhancement Affects Usability Concerns Documentation Concerns GAML Concerns Interface Concerns Modeling | **Is your request related to a problem? Please describe.**
Currently, the search box in GAMA can be used to lookup for operators and hovering over the results would show @doc text in a small purple window. This window can be missed by the user because it does not appear close to the search result and neither does it stay visible for long. Its visibility is subject to mouse hovering on the correct search result. This is inconvenient as is normal tendency to move your mouse towards this purple window or the mouse sometimes moves unexpectedly and we lose this purple window.
**Describe the solution you'd like**
A more common query system such as it exists in matlab, R, python where user can use queries starting with `?` and the query string (usually the method or function required).
So you can write e.g., `?intersects` , `covers` and press enter, this is parsed as a query similar to the search box and the results of the **exact** match are shown in the interactive console. I would prefer that along with @doc value, @example is also shown.
I do not request all kinds of wild card queries. The idea is largely to help users who know the name of the operator, but do not fully remember the number and type of arguments or the return type.
thank you
| 1.0 | introduce ?help in interactive console to bring help into console - **Is your request related to a problem? Please describe.**
Currently, the search box in GAMA can be used to lookup for operators and hovering over the results would show @doc text in a small purple window. This window can be missed by the user because it does not appear close to the search result and neither does it stay visible for long. Its visibility is subject to mouse hovering on the correct search result. This is inconvenient as is normal tendency to move your mouse towards this purple window or the mouse sometimes moves unexpectedly and we lose this purple window.
**Describe the solution you'd like**
A more common query system such as it exists in matlab, R, python where user can use queries starting with `?` and the query string (usually the method or function required).
So you can write e.g., `?intersects` , `covers` and press enter, this is parsed as a query similar to the search box and the results of the **exact** match are shown in the interactive console. I would prefer that along with @doc value, @example is also shown.
I do not request all kinds of wild card queries. The idea is largely to help users who know the name of the operator, but do not fully remember the number and type of arguments or the return type.
thank you
| non_priority | introduce help in interactive console to bring help into console is your request related to a problem please describe currently the search box in gama can be used to lookup for operators and hovering over the results would show doc text in a small purple window this window can be missed by the user because it does not appear close to the search result and neither does it stay visible for long its visibility is subject to mouse hovering on the correct search result this is inconvenient as is normal tendency to move your mouse towards this purple window or the mouse sometimes moves unexpectedly and we lose this purple window describe the solution you d like a more common query system such as it exists in matlab r python where user can use queries starting with and the query string usually the method or function required so you can write e g intersects covers and press enter this is parsed as a query similar to the search box and the results of the exact match are shown in the interactive console i would prefer that along with doc value example is also shown i do not request all kinds of wild card queries the idea is largely to help users who know the name of the operator but do not fully remember the number and type of arguments or the return type thank you | 0 |
238,914 | 7,784,454,938 | IssuesEvent | 2018-06-06 13:20:24 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | The notifications API endpoint returns the notification named "announcements" twice | component: api priority: p2 | ### Describe the problem and steps to reproduce it:
From: https://github.com/mozilla/addons-server/issues/8286#issuecomment-392537562
Retrieve the user notifications for any user.
### What happened?
The notifications API endpoint returns the notification named "announcements" twice.
### What did you expect to happen?
The notifications API endpoint returns the notification named "announcements" once, NOT twice or more.
### Anything else we should know?
This prevents the new frontend to work correctly because notification names should be unique (at least that's my assumption).
| 1.0 | The notifications API endpoint returns the notification named "announcements" twice - ### Describe the problem and steps to reproduce it:
From: https://github.com/mozilla/addons-server/issues/8286#issuecomment-392537562
Retrieve the user notifications for any user.
### What happened?
The notifications API endpoint returns the notification named "announcements" twice.
### What did you expect to happen?
The notifications API endpoint returns the notification named "announcements" once, NOT twice or more.
### Anything else we should know?
This prevents the new frontend to work correctly because notification names should be unique (at least that's my assumption).
| priority | the notifications api endpoint returns the notification named announcements twice describe the problem and steps to reproduce it from retrieve the user notifications for any user what happened the notifications api endpoint returns the notification named announcements twice what did you expect to happen the notifications api endpoint returns the notification named announcements once not twice or more anything else we should know this prevents the new frontend to work correctly because notification names should be unique at least that s my assumption | 1 |
440,052 | 30,727,279,021 | IssuesEvent | 2023-07-27 20:49:11 | Nalini1998/Database | https://api.github.com/repos/Nalini1998/Database | closed | What is a NULL value? | documentation enhancement help wanted good first issue question | ** A. A value that represents missing or unknown data.
B. A known value.
C. An outdated value.
D. A numerical value. | 1.0 | What is a NULL value? - ** A. A value that represents missing or unknown data.
B. A known value.
C. An outdated value.
D. A numerical value. | non_priority | what is a null value a a value that represents missing or unknown data b a known value c an outdated value d a numerical value | 0 |
16,168 | 20,605,072,275 | IssuesEvent | 2022-03-06 21:13:27 | hoprnet/hoprnet | https://api.github.com/repos/hoprnet/hoprnet | opened | Update development processes | new issue processes | <!--- Please DO NOT remove the automatically added 'new issue' label -->
<!--- Provide a general summary of the issue in the Title above -->
After our call on 07/03/22 the new development processes were announced, update the development processes.
Updates: TBA | 1.0 | Update development processes - <!--- Please DO NOT remove the automatically added 'new issue' label -->
<!--- Provide a general summary of the issue in the Title above -->
After our call on 07/03/22 the new development processes were announced, update the development processes.
Updates: TBA | non_priority | update development processes after our call on the new development processes were announced update the development processes updates tba | 0 |
573,805 | 17,023,711,206 | IssuesEvent | 2021-07-03 03:26:04 | tomhughes/trac-tickets | https://api.github.com/repos/tomhughes/trac-tickets | closed | [roads] Render passage ways | Component: mapnik Priority: major Resolution: fixed Type: enhancement | **[Submitted to the original trac issue database at 7.57am, Wednesday, 11th May 2011]**
passage ways are ways that go through buildings on the same level as the ground and are therefore not tunnels. example: http://wiki.openstreetmap.org/wiki/File:Durchgang.jpg
according to the (german) wiki page for "how to map". those passages ("Hausdurchfahrten") should be tagged with covered=yes for the part that goes through the building.
currently mapnik renders the highway on top of the building and ignores the covered=yes part completely. it would be better to render these parts of the highway differently. maybe a 50% transparency should do trick but without the dashed lines, otherwise people could mistake it for tunnels. | 1.0 | [roads] Render passage ways - **[Submitted to the original trac issue database at 7.57am, Wednesday, 11th May 2011]**
passage ways are ways that go through buildings on the same level as the ground and are therefore not tunnels. example: http://wiki.openstreetmap.org/wiki/File:Durchgang.jpg
according to the (german) wiki page for "how to map". those passages ("Hausdurchfahrten") should be tagged with covered=yes for the part that goes through the building.
currently mapnik renders the highway on top of the building and ignores the covered=yes part completely. it would be better to render these parts of the highway differently. maybe a 50% transparency should do trick but without the dashed lines, otherwise people could mistake it for tunnels. | priority | render passage ways passage ways are ways that go through buildings on the same level as the ground and are therefore not tunnels example according to the german wiki page for how to map those passages hausdurchfahrten should be tagged with covered yes for the part that goes through the building currently mapnik renders the highway on top of the building and ignores the covered yes part completely it would be better to render these parts of the highway differently maybe a transparency should do trick but without the dashed lines otherwise people could mistake it for tunnels | 1 |
653,611 | 21,608,508,394 | IssuesEvent | 2022-05-04 07:35:21 | bounswe/bounswe2022group2 | https://api.github.com/repos/bounswe/bounswe2022group2 | opened | Practice App: Set up the base structure of the back-end | priority-medium status-new practice-app practice-app:back-end | ### Issue Description
I will set up the base structure of the back-end of the practice app project. We decided to use Node.js and Express.js in our back-end. So, I will integrate the required packages and set up the base project. Afterward, we will develop the endpoints as the whole team one by one.
### Step Details
Steps that will be performed:
- [ ] Create a folder for back-end
- [ ] Create the readme file of the back-end part
- [ ] Create a new .gitignore file to ignore node.js related files under the back-end folder
- [ ] Crete the "package.json" file
- [ ] Integrate "express" framework
- [ ] Integrate "joi" package for model creation
- [ ] Integrate other small functional packages (if there is any)
- [ ] Create the "app.js" file
- [ ] Implement express loaders inside "express.js" file
- [ ] Implement the "index.js" file to combine loaders
- [ ] Create the "config.js" file
- [ ] Create a base folder structure as following:
* api
* controllers
* middlewares
* routes
* validators
* config
* loaders
* models
* utils
* app.js
### Final Actions
After I created the base structure, files, folders, and package integrations, we will start to develop the endpoints one by one.
### Deadline of the Issue
06.05.2022 - Friday - 23:59
### Reviewer
Mehmet Batuhan Çelik
### Deadline for the Review
07.05.2022 - Saturday - 23:59 | 1.0 | Practice App: Set up the base structure of the back-end - ### Issue Description
I will set up the base structure of the back-end of the practice app project. We decided to use Node.js and Express.js in our back-end. So, I will integrate the required packages and set up the base project. Afterward, we will develop the endpoints as the whole team one by one.
### Step Details
Steps that will be performed:
- [ ] Create a folder for back-end
- [ ] Create the readme file of the back-end part
- [ ] Create a new .gitignore file to ignore node.js related files under the back-end folder
- [ ] Crete the "package.json" file
- [ ] Integrate "express" framework
- [ ] Integrate "joi" package for model creation
- [ ] Integrate other small functional packages (if there is any)
- [ ] Create the "app.js" file
- [ ] Implement express loaders inside "express.js" file
- [ ] Implement the "index.js" file to combine loaders
- [ ] Create the "config.js" file
- [ ] Create a base folder structure as following:
* api
* controllers
* middlewares
* routes
* validators
* config
* loaders
* models
* utils
* app.js
### Final Actions
After I created the base structure, files, folders, and package integrations, we will start to develop the endpoints one by one.
### Deadline of the Issue
06.05.2022 - Friday - 23:59
### Reviewer
Mehmet Batuhan Çelik
### Deadline for the Review
07.05.2022 - Saturday - 23:59 | priority | practice app set up the base structure of the back end issue description i will set up the base structure of the back end of the practice app project we decided to use node js and express js in our back end so i will integrate the required packages and set up the base project afterward we will develop the endpoints as the whole team one by one step details steps that will be performed create a folder for back end create the readme file of the back end part create a new gitignore file to ignore node js related files under the back end folder crete the package json file integrate express framework integrate joi package for model creation integrate other small functional packages if there is any create the app js file implement express loaders inside express js file implement the index js file to combine loaders create the config js file create a base folder structure as following api controllers middlewares routes validators config loaders models utils app js final actions after i created the base structure files folders and package integrations we will start to develop the endpoints one by one deadline of the issue friday reviewer mehmet batuhan çelik deadline for the review saturday | 1 |
9,853 | 3,973,458,515 | IssuesEvent | 2016-05-04 18:39:51 | teotidev/guide | https://api.github.com/repos/teotidev/guide | closed | MetaData - use sub scroll pane and do not scroll the path label | code work | - make the label one size larger | 1.0 | MetaData - use sub scroll pane and do not scroll the path label - - make the label one size larger | non_priority | metadata use sub scroll pane and do not scroll the path label make the label one size larger | 0 |
238,739 | 7,782,751,724 | IssuesEvent | 2018-06-06 07:44:42 | javaee/servlet-spec | https://api.github.com/repos/javaee/servlet-spec | closed | Async Request parameters | Component: I/O Priority: Major Type: Improvement | Section 9.7.2 describes a set of request attributes that contain the path values of the original request, so that they may be accessed by a servlet called as a result of a AsyncContext.dispatch(...)
However, this implies that these attributes are only set after a AsyncContext.dispatch(...), which means that they are not available to a thread that might be acting as part of a startAsync().... AsyncContext.complete() pattern.
Note that a thread cannot access the original request paths via AsyncContext.getRequest().getServletPath() because the value returned from that can be affected by forwards that happen before and/or after the startAsync call, or even a forward after an async dispatch. The path methods are inherently volatile.
I think that the ASYNC request parameters should be set when startAsync is called, so that those values are available for the entire async life cycle and not only during async dispatch. | 1.0 | Async Request parameters - Section 9.7.2 describes a set of request attributes that contain the path values of the original request, so that they may be accessed by a servlet called as a result of a AsyncContext.dispatch(...)
However, this implies that these attributes are only set after a AsyncContext.dispatch(...), which means that they are not available to a thread that might be acting as part of a startAsync().... AsyncContext.complete() pattern.
Note that a thread cannot access the original request paths via AsyncContext.getRequest().getServletPath() because the value returned from that can be affected by forwards that happen before and/or after the startAsync call, or even a forward after an async dispatch. The path methods are inherently volatile.
I think that the ASYNC request parameters should be set when startAsync is called, so that those values are available for the entire async life cycle and not only during async dispatch. | priority | async request parameters section describes a set of request attributes that contain the path values of the original request so that they may be accessed by a servlet called as a result of a asynccontext dispatch however this implies that these attributes are only set after a asynccontext dispatch which means that they are not available to a thread that might be acting as part of a startasync asynccontext complete pattern note that a thread cannot access the original request paths via asynccontext getrequest getservletpath because the value returned from that can be affected by forwards that happen before and or after the startasync call or even a forward after an async dispatch the path methods are inherently volatile i think that the async request parameters should be set when startasync is called so that those values are available for the entire async life cycle and not only during async dispatch | 1 |
641,712 | 20,832,780,130 | IssuesEvent | 2022-03-19 18:18:48 | Frontesque/VueTube | https://api.github.com/repos/Frontesque/VueTube | closed | Keyboard Enter to search not working | bug Priority: Low | **Describe the bug**
The enter key on the keyboard should allow for the same function as the search button
**To Reproduce**
1. Go to search
2. Type in search
3. Press enter and nothing will happen
**Expected behavior**
1. Go to search
2. Type in search
3. Press enter and search result should be retuned
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Device Information**
- OS: Android 11
- App Version: Latest github action build
**Additional context**
Add any other context about the problem here.
| 1.0 | Keyboard Enter to search not working - **Describe the bug**
The enter key on the keyboard should allow for the same function as the search button
**To Reproduce**
1. Go to search
2. Type in search
3. Press enter and nothing will happen
**Expected behavior**
1. Go to search
2. Type in search
3. Press enter and search result should be retuned
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Device Information**
- OS: Android 11
- App Version: Latest github action build
**Additional context**
Add any other context about the problem here.
| priority | keyboard enter to search not working describe the bug the enter key on the keyboard should allow for the same function as the search button to reproduce go to search type in search press enter and nothing will happen expected behavior go to search type in search press enter and search result should be retuned screenshots if applicable add screenshots to help explain your problem device information os android app version latest github action build additional context add any other context about the problem here | 1 |
121,576 | 25,994,735,429 | IssuesEvent | 2022-12-20 10:36:08 | Open-Telecoms-Data/open-fibre-data-standard | https://api.github.com/repos/Open-Telecoms-Data/open-fibre-data-standard | closed | Multiple network providers per span and node | Schema Codelist | Currently, `Span.networkProvider` and `Node.networkProvider` are objects, i.e. there is a 1:1 relationship between `Span`/`Node` and `networkProvider`.
Therefore, to represent a span with many network providers (a common situation), implementers would need to publish multiple spans - one for each network provider.
This is a problem because properties of `Span` and `Node` that relate to the physical infrastructure would need to be repeated for each span/node, which could lead to double-counting, e.g. `fibreCount`.
## Proposal
* Replace `Span.networkProvider` (an object) with `Span.networkProviders` (an array).
* Replace `Node.networkProvider` (an object) with `Node.networkProviders` (an array)
## Related issues
The following issues are related, but not addressed by the proposal:
* https://github.com/Open-Telecoms-Data/open-fibre-data-standard/issues/199
* https://github.com/Open-Telecoms-Data/open-fibre-data-standard/issues/192
* https://github.com/Open-Telecoms-Data/open-fibre-data-standard/issues/193
## Analysis
### Schema and codelist changes
This section provides an analysis of the changes to schema and codelists resulting from the proposed change.
#### Span
The following fields relate to the active layer so descriptions may need to be updated. Proposed changes are shown in *italics*:
* `status`('operational' code):
> 'The active network infrastructure for *at least one network provider on* the span is live and carrying traffic.'
* `readyForServiceDate`:
> 'The date this span was ready to carry traffic. *For spans with multiple network providers, the earliest of the dates at which each provider's network was ready to carry traffic on this span*'.
* `technologies`:
> `The active technologies used on this span,from the open [spanTechnologies codelist](https://open-fibre-data-standard.readthedocs.io/en/latest/reference/codelists.html#spanTechnologies). *For spans with multiple network providers, the union of the active technologies used by each provider.*
* `capacity`:
> 'The transmission rate, or throughput, of this span, expressed in Gbit/sec (Gbps). The equipped capacity is the total capacity of the circuits (e.g. E1, DS3, STM-1 etc.) which have been activated in the network transmission equipment of the span. *For spans with multiple network providers, the sum of the capacity of each provider's network on this span.* Further details of this span’s capacity can be provided in `.capacityDetails`. '
* `capacityDetails`: No change required.
The following fields relate to the passive (physical) layer so they are not affected by the cardinality of `.networkProvider`:
* `id`
* `name`
* `phase`
* `start`
* `end`
* `directed`
* `route`
* `physicalInfrastructureProvider`
* `supplier`
* `transmissionMedium`
* `deployment`
* `deploymentDetails`
* `darkFibre`
* `fibreType`
* `fibreTypeDetails`
* `fibreCount`
* `fibreLength`
* `countries`
#### Node
The following fields relate to the active layer so descriptions may need to be updated. Proposed changes are shown in *italics*:
* `status` ('operational' code):
> 'The active network infrastructure for *at least one network provider at* the node is live and carrying traffic.'
* `type`:
> 'The type of this node, from the open [nodeType codelist](https://open-fibre-data-standard.readthedocs.io/en/latest/reference/codelists.html#nodeType). *For nodes with multiple network providers, the union of the node types for each provider.*'
* `accessPoint`: No change required.
* `technologies`:
> 'The active technologies used at this node, from the open [nodeTechnologies codelist](https://open-fibre-data-standard.readthedocs.io/en/latest/reference/codelists.html#nodeTechnologies). *For nodes with multiple network providers, the union of the active technologies used by each provider.*'
The following fields relate to the passive (physical) layer so they are not affected by the cardinality of `.networkProvider`:
* `id`
* `name`
* `phase`
* `location`
* `address`
* `internationalConnections`
* `power`
* `physicalInfrastructureProvider`
### Publication models
This section provides an analysis of how the proposed changes would apply to aggregated and distributed publication models in the following scenario:
> FibreCo provides the physical infrastructure, operates their own network and leases dark fibre to ProviderCo.
#### Aggregated publication
FibreCo publishes details of their own network and names ProviderCo as a network provider:
```json
{
"networks": [
{
"publisher": {
"name": "FibreCo"
},
"spans": [
{
"physicalInfrastructureProvider": {
"name": "FibreCo"
},
"networkProviders": [
{
"name": "FibreCo"
},
{
"name": "FastWeb"
}
],
"darkFibre": true
}
]
}
]
}
```
#### Distributed publication
FibreCo publishes details of their own network without naming ProviderCo as a network provider. ProviderCo publishes details of their own network and names FibreCo as the physical infrastructure provider:
```json
{
"networks": [
{
"publisher": {
"name": "FibreCo"
},
"id": "a096d627-72e1-4f9b-b129-951b1737bff4",
"spans": [
{
"id": "1",
"physicalInfrastructureProvider": {
"name": "FibreCo"
},
"networkProviders": [
{
"name": "FibreCo"
}
],
"darkFibre": true
}
]
},
{
"publisher": {
"name": "FastWeb"
},
"id": "6187511d-2ec4-4d24-bc97-6289fa6fd4a8",
"spans": [
{
"id": "A",
"physicalInfrastructureProvider": {
"name": "FibreCo"
},
"networkProviders": [
{
"name": "FastWeb"
}
]
}
]
}
]
}
```
### Publication formats
This section provides an analysis of how the proposed changes would affect the GeoJSON and CSV publication formats.
#### GeoJSON
The following fields will change:
* `nodes.geojson`: `features/properties/networkProvider` (object) -> `features/properties/networkProviders` (array)
* `spans.geojson`: `features/properties/networkProvider` (object) -> `features/properties/networkProviders` (array)
#### CSV
The CSV publication format will include two additional tables:
* `node_networkProviders` (`id`, `nodes/0/id`, `nodes/0/networkProviders/0/id`, `nodes/0/networkProviders/0/name`)
* `span_networkProviders` (`id`, `spans/0/id`, `spans/0/networkProviders/0/id`, `spans/0/networkProviders/0/name`)
### Data use
This section provides an analysis of the impact on data use.
#### Mapping
If a user wants to colour the spans and nodes in a network by the network provider, there are several options:
1. Categorise according to the set of network providers, e.g.
* `[{"name": "Org A"}]`: Red
* `[{"name": "Org B"}]`: Blue
* `[{"name": "Org A"}, {"name": "Org B"}]`: Green
* `[{"name": "Org B"}, {"name": "Org A"}]`: Green
2. Categorise according to the presence of a particular network provider, e.g. for 'Org A':
* `[{"name": "Org A"}]`: Red
* `[{"name": "Org B"}]`: Blue
* `[{"name": "Org A"}, {"name": "Org B"}]`: Red
* `[{"name": "Org B"}, {"name": "Org A"}]`: Red
3. Create a separate layer for each network provider, e.g.
* 'Org A':
* `[{"name": "Org A"}]`: Red
* `[{"name": "Org B"}]`: Discard
* `[{"name": "Org A"}, {"name": "Org B"}]`: Red
* `[{"name": "Org B"}, {"name": "Org A"}]`: Red
* 'Org B':
* `[{"name": "Org A"}]`: Discard
* `[{"name": "Org B"}]`: Blue
* `[{"name": "Org A"}, {"name": "Org B"}]`: Blue
* `[{"name": "Org B"}, {"name": "Org A"}]`: Blue
Option 3. is essentially equivalent to the current model of a 1:1 relationship between `Span`/`Node` and `networkProvider`.
#### Reconciliation
In the distributed publication model discussed above, if the spans declared by each publisher share similar start and end locations and similar routes, a user could infer that they are the same span. In the aggregated distribution model, that fact could be declared explicitly. | 1.0 | Multiple network providers per span and node - Currently, `Span.networkProvider` and `Node.networkProvider` are objects, i.e. there is a 1:1 relationship between `Span`/`Node` and `networkProvider`.
Therefore, to represent a span with many network providers (a common situation), implementers would need to publish multiple spans - one for each network provider.
This is a problem because properties of `Span` and `Node` that relate to the physical infrastructure would need to be repeated for each span/node, which could lead to double-counting, e.g. `fibreCount`.
## Proposal
* Replace `Span.networkProvider` (an object) with `Span.networkProviders` (an array).
* Replace `Node.networkProvider` (an object) with `Node.networkProviders` (an array)
## Related issues
The following issues are related, but not addressed by the proposal:
* https://github.com/Open-Telecoms-Data/open-fibre-data-standard/issues/199
* https://github.com/Open-Telecoms-Data/open-fibre-data-standard/issues/192
* https://github.com/Open-Telecoms-Data/open-fibre-data-standard/issues/193
## Analysis
### Schema and codelist changes
This section provides an analysis of the changes to schema and codelists resulting from the proposed change.
#### Span
The following fields relate to the active layer so descriptions may need to be updated. Proposed changes are shown in *italics*:
* `status`('operational' code):
> 'The active network infrastructure for *at least one network provider on* the span is live and carrying traffic.'
* `readyForServiceDate`:
> 'The date this span was ready to carry traffic. *For spans with multiple network providers, the earliest of the dates at which each provider's network was ready to carry traffic on this span*'.
* `technologies`:
> `The active technologies used on this span,from the open [spanTechnologies codelist](https://open-fibre-data-standard.readthedocs.io/en/latest/reference/codelists.html#spanTechnologies). *For spans with multiple network providers, the union of the active technologies used by each provider.*
* `capacity`:
> 'The transmission rate, or throughput, of this span, expressed in Gbit/sec (Gbps). The equipped capacity is the total capacity of the circuits (e.g. E1, DS3, STM-1 etc.) which have been activated in the network transmission equipment of the span. *For spans with multiple network providers, the sum of the capacity of each provider's network on this span.* Further details of this span’s capacity can be provided in `.capacityDetails`. '
* `capacityDetails`: No change required.
The following fields relate to the passive (physical) layer so they are not affected by the cardinality of `.networkProvider`:
* `id`
* `name`
* `phase`
* `start`
* `end`
* `directed`
* `route`
* `physicalInfrastructureProvider`
* `supplier`
* `transmissionMedium`
* `deployment`
* `deploymentDetails`
* `darkFibre`
* `fibreType`
* `fibreTypeDetails`
* `fibreCount`
* `fibreLength`
* `countries`
#### Node
The following fields relate to the active layer so descriptions may need to be updated. Proposed changes are shown in *italics*:
* `status` ('operational' code):
> 'The active network infrastructure for *at least one network provider at* the node is live and carrying traffic.'
* `type`:
> 'The type of this node, from the open [nodeType codelist](https://open-fibre-data-standard.readthedocs.io/en/latest/reference/codelists.html#nodeType). *For nodes with multiple network providers, the union of the node types for each provider.*'
* `accessPoint`: No change required.
* `technologies`:
> 'The active technologies used at this node, from the open [nodeTechnologies codelist](https://open-fibre-data-standard.readthedocs.io/en/latest/reference/codelists.html#nodeTechnologies). *For nodes with multiple network providers, the union of the active technologies used by each provider.*'
The following fields relate to the passive (physical) layer so they are not affected by the cardinality of `.networkProvider`:
* `id`
* `name`
* `phase`
* `location`
* `address`
* `internationalConnections`
* `power`
* `physicalInfrastructureProvider`
### Publication models
This section provides an analysis of how the proposed changes would apply to aggregated and distributed publication models in the following scenario:
> FibreCo provides the physical infrastructure, operates their own network and leases dark fibre to ProviderCo.
#### Aggregated publication
FibreCo publishes details of their own network and names ProviderCo as a network provider:
```json
{
"networks": [
{
"publisher": {
"name": "FibreCo"
},
"spans": [
{
"physicalInfrastructureProvider": {
"name": "FibreCo"
},
"networkProviders": [
{
"name": "FibreCo"
},
{
"name": "FastWeb"
}
],
"darkFibre": true
}
]
}
]
}
```
#### Distributed publication
FibreCo publishes details of their own network without naming ProviderCo as a network provider. ProviderCo publishes details of their own network and names FibreCo as the physical infrastructure provider:
```json
{
"networks": [
{
"publisher": {
"name": "FibreCo"
},
"id": "a096d627-72e1-4f9b-b129-951b1737bff4",
"spans": [
{
"id": "1",
"physicalInfrastructureProvider": {
"name": "FibreCo"
},
"networkProviders": [
{
"name": "FibreCo"
}
],
"darkFibre": true
}
]
},
{
"publisher": {
"name": "FastWeb"
},
"id": "6187511d-2ec4-4d24-bc97-6289fa6fd4a8",
"spans": [
{
"id": "A",
"physicalInfrastructureProvider": {
"name": "FibreCo"
},
"networkProviders": [
{
"name": "FastWeb"
}
]
}
]
}
]
}
```
### Publication formats
This section provides an analysis of how the proposed changes would affect the GeoJSON and CSV publication formats.
#### GeoJSON
The following fields will change:
* `nodes.geojson`: `features/properties/networkProvider` (object) -> `features/properties/networkProviders` (array)
* `spans.geojson`: `features/properties/networkProvider` (object) -> `features/properties/networkProviders` (array)
#### CSV
The CSV publication format will include two additional tables:
* `node_networkProviders` (`id`, `nodes/0/id`, `nodes/0/networkProviders/0/id`, `nodes/0/networkProviders/0/name`)
* `span_networkProviders` (`id`, `spans/0/id`, `spans/0/networkProviders/0/id`, `spans/0/networkProviders/0/name`)
### Data use
This section provides an analysis of the impact on data use.
#### Mapping
If a user wants to colour the spans and nodes in a network by the network provider, there are several options:
1. Categorise according to the set of network providers, e.g.
* `[{"name": "Org A"}]`: Red
* `[{"name": "Org B"}]`: Blue
* `[{"name": "Org A"}, {"name": "Org B"}]`: Green
* `[{"name": "Org B"}, {"name": "Org A"}]`: Green
2. Categorise according to the presence of a particular network provider, e.g. for 'Org A':
* `[{"name": "Org A"}]`: Red
* `[{"name": "Org B"}]`: Blue
* `[{"name": "Org A"}, {"name": "Org B"}]`: Red
* `[{"name": "Org B"}, {"name": "Org A"}]`: Red
3. Create a separate layer for each network provider, e.g.
* 'Org A':
* `[{"name": "Org A"}]`: Red
* `[{"name": "Org B"}]`: Discard
* `[{"name": "Org A"}, {"name": "Org B"}]`: Red
* `[{"name": "Org B"}, {"name": "Org A"}]`: Red
* 'Org B':
* `[{"name": "Org A"}]`: Discard
* `[{"name": "Org B"}]`: Blue
* `[{"name": "Org A"}, {"name": "Org B"}]`: Blue
* `[{"name": "Org B"}, {"name": "Org A"}]`: Blue
Option 3. is essentially equivalent to the current model of a 1:1 relationship between `Span`/`Node` and `networkProvider`.
#### Reconciliation
In the distributed publication model discussed above, if the spans declared by each publisher share similar start and end locations and similar routes, a user could infer that they are the same span. In the aggregated distribution model, that fact could be declared explicitly. | non_priority | multiple network providers per span and node currently span networkprovider and node networkprovider are objects i e there is a relationship between span node and networkprovider therefore to represent a span with many network providers a common situation implementers would need to publish multiple spans one for each network provider this is a problem because properties of span and node that relate to the physical infrastructure would need to be repeated for each span node which could lead to double counting e g fibrecount proposal replace span networkprovider an object with span networkproviders an array replace node networkprovider an object with node networkproviders an array related issues the following issues are related but not addressed by the proposal analysis schema and codelist changes this section provides an analysis of the changes to schema and codelists resulting from the proposed change span the following fields relate to the active layer so descriptions may need to be updated proposed changes are shown in italics status operational code the active network infrastructure for at least one network provider on the span is live and carrying traffic readyforservicedate the date this span was ready to carry traffic for spans with multiple network providers the earliest of the dates at which each provider s network was ready to carry traffic on this span technologies the active technologies used on this span from the open for spans with multiple network providers the union of the active technologies used by each provider capacity the transmission rate or throughput of this span expressed in gbit sec gbps the equipped capacity is the total capacity of the circuits e g stm etc which have been activated in the network transmission equipment of the span for spans with multiple network providers the sum of the capacity of each provider s network on this span further details of this span’s capacity can be provided in capacitydetails capacitydetails no change required the following fields relate to the passive physical layer so they are not affected by the cardinality of networkprovider id name phase start end directed route physicalinfrastructureprovider supplier transmissionmedium deployment deploymentdetails darkfibre fibretype fibretypedetails fibrecount fibrelength countries node the following fields relate to the active layer so descriptions may need to be updated proposed changes are shown in italics status operational code the active network infrastructure for at least one network provider at the node is live and carrying traffic type the type of this node from the open for nodes with multiple network providers the union of the node types for each provider accesspoint no change required technologies the active technologies used at this node from the open for nodes with multiple network providers the union of the active technologies used by each provider the following fields relate to the passive physical layer so they are not affected by the cardinality of networkprovider id name phase location address internationalconnections power physicalinfrastructureprovider publication models this section provides an analysis of how the proposed changes would apply to aggregated and distributed publication models in the following scenario fibreco provides the physical infrastructure operates their own network and leases dark fibre to providerco aggregated publication fibreco publishes details of their own network and names providerco as a network provider json networks publisher name fibreco spans physicalinfrastructureprovider name fibreco networkproviders name fibreco name fastweb darkfibre true distributed publication fibreco publishes details of their own network without naming providerco as a network provider providerco publishes details of their own network and names fibreco as the physical infrastructure provider json networks publisher name fibreco id spans id physicalinfrastructureprovider name fibreco networkproviders name fibreco darkfibre true publisher name fastweb id spans id a physicalinfrastructureprovider name fibreco networkproviders name fastweb publication formats this section provides an analysis of how the proposed changes would affect the geojson and csv publication formats geojson the following fields will change nodes geojson features properties networkprovider object features properties networkproviders array spans geojson features properties networkprovider object features properties networkproviders array csv the csv publication format will include two additional tables node networkproviders id nodes id nodes networkproviders id nodes networkproviders name span networkproviders id spans id spans networkproviders id spans networkproviders name data use this section provides an analysis of the impact on data use mapping if a user wants to colour the spans and nodes in a network by the network provider there are several options categorise according to the set of network providers e g red blue green green categorise according to the presence of a particular network provider e g for org a red blue red red create a separate layer for each network provider e g org a red discard red red org b discard blue blue blue option is essentially equivalent to the current model of a relationship between span node and networkprovider reconciliation in the distributed publication model discussed above if the spans declared by each publisher share similar start and end locations and similar routes a user could infer that they are the same span in the aggregated distribution model that fact could be declared explicitly | 0 |
710,237 | 24,411,738,197 | IssuesEvent | 2022-10-05 12:57:20 | OptimistikSAS/OIBus | https://api.github.com/repos/OptimistikSAS/OIBus | closed | [UI - Home page] - Remove connection to engine when a connector is disabled | enhancement good first issue priority:high | When a connector is disabled, remove its connection to the engine
 | 1.0 | [UI - Home page] - Remove connection to engine when a connector is disabled - When a connector is disabled, remove its connection to the engine
 | priority | remove connection to engine when a connector is disabled when a connector is disabled remove its connection to the engine | 1 |
77,090 | 21,665,229,133 | IssuesEvent | 2022-05-07 04:07:14 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | Document packaging for Windows builds | tool platform-windows a: desktop documentation a: build P4 | See https://github.com/google/flutter-desktop-embedding/issues/587 for earlier discussion.
We should very likely either:
- switch the templates for apps and plugins to use the static runtime, or
- update the app bundling script to copy in the runtime libraries into the output directory for ease of packaging.
From my initial investigation in the original bug:
> I did a quick experiment, and using the static runtime increases the size of each of the current FDE plugins from ~50-70K to ~300-500K, and Testbed.exe from ~100K to ~500K.
>
> While that's not all that much in current desktop machine size terms, it's not clear to me that the benefit of doing that vs. having a simple path to bundling the necessary runtime libraries (e.g., adding to bundle_assets_and_deps) is the right approach. We should look more into the tradeoffs in terms of potential interactions between plugins, which would be more isolated with a static runtime. | 1.0 | Document packaging for Windows builds - See https://github.com/google/flutter-desktop-embedding/issues/587 for earlier discussion.
We should very likely either:
- switch the templates for apps and plugins to use the static runtime, or
- update the app bundling script to copy in the runtime libraries into the output directory for ease of packaging.
From my initial investigation in the original bug:
> I did a quick experiment, and using the static runtime increases the size of each of the current FDE plugins from ~50-70K to ~300-500K, and Testbed.exe from ~100K to ~500K.
>
> While that's not all that much in current desktop machine size terms, it's not clear to me that the benefit of doing that vs. having a simple path to bundling the necessary runtime libraries (e.g., adding to bundle_assets_and_deps) is the right approach. We should look more into the tradeoffs in terms of potential interactions between plugins, which would be more isolated with a static runtime. | non_priority | document packaging for windows builds see for earlier discussion we should very likely either switch the templates for apps and plugins to use the static runtime or update the app bundling script to copy in the runtime libraries into the output directory for ease of packaging from my initial investigation in the original bug i did a quick experiment and using the static runtime increases the size of each of the current fde plugins from to and testbed exe from to while that s not all that much in current desktop machine size terms it s not clear to me that the benefit of doing that vs having a simple path to bundling the necessary runtime libraries e g adding to bundle assets and deps is the right approach we should look more into the tradeoffs in terms of potential interactions between plugins which would be more isolated with a static runtime | 0 |
151,033 | 19,648,308,952 | IssuesEvent | 2022-01-10 01:24:28 | ThomasRuemmler/openspaceplanner | https://api.github.com/repos/ThomasRuemmler/openspaceplanner | closed | WS-2022-0007 (Medium) detected in node-forge-0.9.0.tgz - autoclosed | security vulnerability | ## WS-2022-0007 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /Web/package.json</p>
<p>Path to vulnerable library: /Web/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.803.20.tgz (Root Library)
- webpack-dev-server-3.9.0.tgz
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In node-forge before 1.0.0 he regex used for the forge.util.parseUrl API would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/db8016c805371e72b06d8e2edfe0ace0df934a5e>WS-2022-0007</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-gf8q-jrpm-jvxq">https://github.com/advisories/GHSA-gf8q-jrpm-jvxq</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution: node-forge - 1.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2022-0007 (Medium) detected in node-forge-0.9.0.tgz - autoclosed - ## WS-2022-0007 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /Web/package.json</p>
<p>Path to vulnerable library: /Web/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.803.20.tgz (Root Library)
- webpack-dev-server-3.9.0.tgz
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In node-forge before 1.0.0 he regex used for the forge.util.parseUrl API would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior.
<p>Publish Date: 2022-01-08
<p>URL: <a href=https://github.com/digitalbazaar/forge/commit/db8016c805371e72b06d8e2edfe0ace0df934a5e>WS-2022-0007</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-gf8q-jrpm-jvxq">https://github.com/advisories/GHSA-gf8q-jrpm-jvxq</a></p>
<p>Release Date: 2022-01-08</p>
<p>Fix Resolution: node-forge - 1.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in node forge tgz autoclosed ws medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file web package json path to vulnerable library web node modules node forge package json dependency hierarchy build angular tgz root library webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library vulnerability details in node forge before he regex used for the forge util parseurl api would not properly parse certain inputs resulting in a parsed data structure that could lead to undesired behavior publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge step up your open source security game with whitesource | 0 |
777,914 | 27,297,519,485 | IssuesEvent | 2023-02-23 21:47:06 | bcgov/entity | https://api.github.com/repos/bcgov/entity | closed | NameX - Restoration/Reinstatement NRs should be 1 year 56 days for Societies | bug Priority2 ENTITY SRE | We probably missed this last time when we fixed the expiry dates for other entity types. | 1.0 | NameX - Restoration/Reinstatement NRs should be 1 year 56 days for Societies - We probably missed this last time when we fixed the expiry dates for other entity types. | priority | namex restoration reinstatement nrs should be year days for societies we probably missed this last time when we fixed the expiry dates for other entity types | 1 |
74,906 | 3,452,191,069 | IssuesEvent | 2015-12-17 02:03:37 | YetiForceCompany/YetiForceCRM | https://api.github.com/repos/YetiForceCompany/YetiForceCRM | closed | [PL][BUG] Kalendarz - wyświetlanie zakończonych elementów jako zaplanowane | Label::Logic Priority::#2 Normal Type::Bug | Przy dodawaniu rekordu do kalendarza pojawiają się nam zadania dodane wcześniej, które są zaplanowane na ten sam dzień, ale wyświetlają się tam również el. wykonane.
Np:

| 1.0 | [PL][BUG] Kalendarz - wyświetlanie zakończonych elementów jako zaplanowane - Przy dodawaniu rekordu do kalendarza pojawiają się nam zadania dodane wcześniej, które są zaplanowane na ten sam dzień, ale wyświetlają się tam również el. wykonane.
Np:

| priority | kalendarz wyświetlanie zakończonych elementów jako zaplanowane przy dodawaniu rekordu do kalendarza pojawiają się nam zadania dodane wcześniej które są zaplanowane na ten sam dzień ale wyświetlają się tam również el wykonane np | 1 |
49,559 | 7,522,073,924 | IssuesEvent | 2018-04-12 19:13:00 | acquia/blt | https://api.github.com/repos/acquia/blt | closed | Update WSL/Ubuntu Bash-based documentation to account for Creators Update | documentation | My system information:
* Operating system type: Windows 10
* Operating system version: Creators Update (April 11, 2017)
* BLT version: 8.7.0
For reference:
- [Using Ubuntu Bash in Windows Creators' Update with Vagrant](https://www.jeffgeerling.com/blog/2017/using-ubuntu-bash-windows-creators-update-vagrant)
- [Update docs for Windows with WSL Windows Interoperability improvements](https://github.com/geerlingguy/drupal-vm/issues/1012)
We may be able to tighten up our docs for Windows users a bit, now that WSL can call out to `vagrant.exe` from the Ubuntu bash environment.
So far I've just focused on raw vagrant usage... but it might be possible to run all the `blt` commands within WSL (assuming you install composer and PHP) against a VM that's also managed from within the WSL. | 1.0 | Update WSL/Ubuntu Bash-based documentation to account for Creators Update - My system information:
* Operating system type: Windows 10
* Operating system version: Creators Update (April 11, 2017)
* BLT version: 8.7.0
For reference:
- [Using Ubuntu Bash in Windows Creators' Update with Vagrant](https://www.jeffgeerling.com/blog/2017/using-ubuntu-bash-windows-creators-update-vagrant)
- [Update docs for Windows with WSL Windows Interoperability improvements](https://github.com/geerlingguy/drupal-vm/issues/1012)
We may be able to tighten up our docs for Windows users a bit, now that WSL can call out to `vagrant.exe` from the Ubuntu bash environment.
So far I've just focused on raw vagrant usage... but it might be possible to run all the `blt` commands within WSL (assuming you install composer and PHP) against a VM that's also managed from within the WSL. | non_priority | update wsl ubuntu bash based documentation to account for creators update my system information operating system type windows operating system version creators update april blt version for reference we may be able to tighten up our docs for windows users a bit now that wsl can call out to vagrant exe from the ubuntu bash environment so far i ve just focused on raw vagrant usage but it might be possible to run all the blt commands within wsl assuming you install composer and php against a vm that s also managed from within the wsl | 0 |
620,339 | 19,559,559,865 | IssuesEvent | 2022-01-03 14:28:56 | bounswe/2021SpringGroup12 | https://api.github.com/repos/bounswe/2021SpringGroup12 | opened | Implementation of privacy policy and terms and conditions | priority: high frontend | **Description**
- Add privacy policy and terms and conditions forms to the register page
**Deadline**
- 04.01.2022
| 1.0 | Implementation of privacy policy and terms and conditions - **Description**
- Add privacy policy and terms and conditions forms to the register page
**Deadline**
- 04.01.2022
| priority | implementation of privacy policy and terms and conditions description add privacy policy and terms and conditions forms to the register page deadline | 1 |
303,615 | 26,218,325,637 | IssuesEvent | 2023-01-04 12:55:18 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: follower-reads/mixed-version/single-region failed | C-test-failure O-robot O-roachtest branch-master release-blocker T-kv | roachtest.follower-reads/mixed-version/single-region [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8153407?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8153407?buildTab=artifacts#/follower-reads/mixed-version/single-region) on master @ [0725273ac7f789ba8ed78aacaf73cc953ca47fe8](https://github.com/cockroachdb/cockroach/commits/0725273ac7f789ba8ed78aacaf73cc953ca47fe8):
```
test artifacts and logs in: /artifacts/follower-reads/mixed-version/single-region/run_1
(follower_reads.go:765).verifyHighFollowerReadRatios: too many intervals with more than 1 nodes with low follower read ratios: 23 intervals > 4 threshold. Bad intervals:
interval 10:41:30-10:41:40: n1 ratio: 0.004 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:41:40-10:41:50: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:41:50-10:42:00: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:00-10:42:10: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:10-10:42:20: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:20-10:42:30: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:30-10:42:40: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:40-10:42:50: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:50-10:43:00: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:00-10:43:10: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:10-10:43:20: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:20-10:43:30: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:30-10:43:40: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:40-10:43:50: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:50-10:44:00: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:00-10:44:10: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:10-10:44:20: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:20-10:44:30: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:30-10:44:40: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:40-10:44:50: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:50-10:45:00: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:45:00-10:45:10: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:45:10-10:45:20: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=2</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #94577 roachtest: follower-reads/mixed-version/single-region failed [C-test-failure O-roachtest O-robot T-kv branch-release-22.2 release-blocker]
- #94506 roachtest: follower-reads/mixed-version/single-region failed [C-test-failure O-roachtest O-robot T-kv branch-release-22.1 release-blocker]
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*follower-reads/mixed-version/single-region.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-23014 | 2.0 | roachtest: follower-reads/mixed-version/single-region failed - roachtest.follower-reads/mixed-version/single-region [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8153407?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8153407?buildTab=artifacts#/follower-reads/mixed-version/single-region) on master @ [0725273ac7f789ba8ed78aacaf73cc953ca47fe8](https://github.com/cockroachdb/cockroach/commits/0725273ac7f789ba8ed78aacaf73cc953ca47fe8):
```
test artifacts and logs in: /artifacts/follower-reads/mixed-version/single-region/run_1
(follower_reads.go:765).verifyHighFollowerReadRatios: too many intervals with more than 1 nodes with low follower read ratios: 23 intervals > 4 threshold. Bad intervals:
interval 10:41:30-10:41:40: n1 ratio: 0.004 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:41:40-10:41:50: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:41:50-10:42:00: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:00-10:42:10: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:10-10:42:20: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:20-10:42:30: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:30-10:42:40: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:40-10:42:50: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:42:50-10:43:00: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:00-10:43:10: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:10-10:43:20: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:20-10:43:30: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:30-10:43:40: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:40-10:43:50: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:43:50-10:44:00: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:00-10:44:10: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:10-10:44:20: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:20-10:44:30: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:30-10:44:40: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:40-10:44:50: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:44:50-10:45:00: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:45:00-10:45:10: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
interval 10:45:10-10:45:20: n1 ratio: 0.000 n2 ratio: 0.000 n3 ratio: 0.000
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=2</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #94577 roachtest: follower-reads/mixed-version/single-region failed [C-test-failure O-roachtest O-robot T-kv branch-release-22.2 release-blocker]
- #94506 roachtest: follower-reads/mixed-version/single-region failed [C-test-failure O-roachtest O-robot T-kv branch-release-22.1 release-blocker]
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*follower-reads/mixed-version/single-region.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-23014 | non_priority | roachtest follower reads mixed version single region failed roachtest follower reads mixed version single region with on master test artifacts and logs in artifacts follower reads mixed version single region run follower reads go verifyhighfollowerreadratios too many intervals with more than nodes with low follower read ratios intervals threshold bad intervals interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio interval ratio ratio ratio parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see same failure on other branches roachtest follower reads mixed version single region failed roachtest follower reads mixed version single region failed cc cockroachdb kv triage jira issue crdb | 0 |
105,119 | 16,624,126,435 | IssuesEvent | 2021-06-03 07:25:48 | Thanraj/OpenSSL_1.0.1b | https://api.github.com/repos/Thanraj/OpenSSL_1.0.1b | opened | CVE-2014-3566 (Low) detected in multiple libraries | security vulnerability | ## CVE-2014-3566 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The SSL protocol 3.0, as used in OpenSSL through 1.0.1i and other products, uses nondeterministic CBC padding, which makes it easier for man-in-the-middle attackers to obtain cleartext data via a padding-oracle attack, aka the "POODLE" issue.
<p>Publish Date: 2014-10-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3566>CVE-2014-3566</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security.gentoo.org/glsa/201507-14">https://security.gentoo.org/glsa/201507-14</a></p>
<p>Release Date: 2015-07-10</p>
<p>Fix Resolution: All Oracle JRE 8 users should upgrade to the latest stable version >= oracle-jre-bin-1.8.0.31
All Oracle JDK 8 users should upgrade to the latest stable version >= oracle-jdk-bin-1.8.0.31
All Oracle JRE 7 users should upgrade to the latest version >= oracle-jre-bin-1.7.0.76
All Oracle JDK 7 users should upgrade to the latest stable version >= oracle-jdk-bin-1.7.0.76
</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2014-3566 (Low) detected in multiple libraries - ## CVE-2014-3566 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b>, <b>opensslOpenSSL_1_0_1b</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The SSL protocol 3.0, as used in OpenSSL through 1.0.1i and other products, uses nondeterministic CBC padding, which makes it easier for man-in-the-middle attackers to obtain cleartext data via a padding-oracle attack, aka the "POODLE" issue.
<p>Publish Date: 2014-10-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2014-3566>CVE-2014-3566</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security.gentoo.org/glsa/201507-14">https://security.gentoo.org/glsa/201507-14</a></p>
<p>Release Date: 2015-07-10</p>
<p>Fix Resolution: All Oracle JRE 8 users should upgrade to the latest stable version >= oracle-jre-bin-1.8.0.31
All Oracle JDK 8 users should upgrade to the latest stable version >= oracle-jdk-bin-1.8.0.31
All Oracle JRE 7 users should upgrade to the latest version >= oracle-jre-bin-1.7.0.76
All Oracle JDK 7 users should upgrade to the latest stable version >= oracle-jdk-bin-1.7.0.76
</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve low detected in multiple libraries cve low severity vulnerability vulnerable libraries opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl opensslopenssl vulnerability details the ssl protocol as used in openssl through and other products uses nondeterministic cbc padding which makes it easier for man in the middle attackers to obtain cleartext data via a padding oracle attack aka the poodle issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution all oracle jre users should upgrade to the latest stable version oracle jre bin all oracle jdk users should upgrade to the latest stable version oracle jdk bin all oracle jre users should upgrade to the latest version oracle jre bin all oracle jdk users should upgrade to the latest stable version oracle jdk bin step up your open source security game with whitesource | 0 |
367,368 | 25,734,822,690 | IssuesEvent | 2022-12-07 23:34:32 | hashicorp/terraform | https://api.github.com/repos/hashicorp/terraform | closed | Mistake/ Manage Resources in Terraform State Doc | documentation new | ### Terraform Version
```shell
terraform -v
Terraform v1.3.6
on linux_amd64
```
### Affected Pages
https://developer.hashicorp.com/terraform/tutorials/certification-associate-tutorials-003/state-cli
### What is the docs issue?


A simple mistake to be corrected :)
### Proposal
Replace one of the output "public_ip" with output "instance_id"
### References
_No response_ | 1.0 | Mistake/ Manage Resources in Terraform State Doc - ### Terraform Version
```shell
terraform -v
Terraform v1.3.6
on linux_amd64
```
### Affected Pages
https://developer.hashicorp.com/terraform/tutorials/certification-associate-tutorials-003/state-cli
### What is the docs issue?


A simple mistake to be corrected :)
### Proposal
Replace one of the output "public_ip" with output "instance_id"
### References
_No response_ | non_priority | mistake manage resources in terraform state doc terraform version shell terraform v terraform on linux affected pages what is the docs issue a simple mistake to be corrected proposal replace one of the output public ip with output instance id references no response | 0 |
30,124 | 8,482,756,797 | IssuesEvent | 2018-10-25 19:28:41 | cdnjs/cdnjs | https://api.github.com/repos/cdnjs/cdnjs | closed | Add latest semantic-ui v2.4.0 | :building_construction: Needs-review | [Semantic UI v2.4.0](https://github.com/Semantic-Org/Semantic-UI/releases/tag/2.4.0) was released a few days ago but still doesn't appear on cdnjs. Is this pending review or is there some other reason it doesn't appear on the site?
/cc @jlukic | 1.0 | Add latest semantic-ui v2.4.0 - [Semantic UI v2.4.0](https://github.com/Semantic-Org/Semantic-UI/releases/tag/2.4.0) was released a few days ago but still doesn't appear on cdnjs. Is this pending review or is there some other reason it doesn't appear on the site?
/cc @jlukic | non_priority | add latest semantic ui was released a few days ago but still doesn t appear on cdnjs is this pending review or is there some other reason it doesn t appear on the site cc jlukic | 0 |
16,648 | 9,479,363,683 | IssuesEvent | 2019-04-20 07:41:47 | apache/incubator-mxnet | https://api.github.com/repos/apache/incubator-mxnet | closed | `mx.gluon.data.DataLoader` is slow when the size of batch is large | Gluon Performance | Hi, there.
I found `mx.gluon.data.DataLoader` is slow when the size of batch is large, since it calls `pickle.load` in `__next__` method of `_MultiWorkerIter` class. [[Code]](https://github.com/apache/incubator-mxnet/blob/master/python/mxnet/gluon/data/dataloader.py#L450)
I will upload the re-procedure code. | True | `mx.gluon.data.DataLoader` is slow when the size of batch is large - Hi, there.
I found `mx.gluon.data.DataLoader` is slow when the size of batch is large, since it calls `pickle.load` in `__next__` method of `_MultiWorkerIter` class. [[Code]](https://github.com/apache/incubator-mxnet/blob/master/python/mxnet/gluon/data/dataloader.py#L450)
I will upload the re-procedure code. | non_priority | mx gluon data dataloader is slow when the size of batch is large hi there i found mx gluon data dataloader is slow when the size of batch is large since it calls pickle load in next method of multiworkeriter class i will upload the re procedure code | 0 |
532,615 | 15,560,086,220 | IssuesEvent | 2021-03-16 12:19:56 | AY2021S2-CS2103-T14-3/tp | https://api.github.com/repos/AY2021S2-CS2103-T14-3/tp | closed | Update User Guide | priority.Medium | Update user guide to change all appointment-related command.
1. remove `to/TIME_TO`
2. make `l/LOCATION compulsory` | 1.0 | Update User Guide - Update user guide to change all appointment-related command.
1. remove `to/TIME_TO`
2. make `l/LOCATION compulsory` | priority | update user guide update user guide to change all appointment related command remove to time to make l location compulsory | 1 |
326,042 | 9,942,063,804 | IssuesEvent | 2019-07-03 13:07:17 | cypress-io/cypress | https://api.github.com/repos/cypress-io/cypress | closed | Remove snap CI provider, add GoCD provider | difficulty: 1️⃣ first-timers-only pkg/server priority: low 🎗 stage: work in progress type: bug | Snap CI is no longer available: https://snap-ci.com/
They recommend looking at GoCD: https://www.gocd.org
Should be updated in our CI Provider file here: https://github.com/cypress-io/cypress/blob/develop/packages/server/lib/util/ci_provider.coffee#L31
Standard GoCD environment variables listed here: https://docs.gocd.org/current/faq/environment_variables.html#standard-gocd-environment-variables | 1.0 | Remove snap CI provider, add GoCD provider - Snap CI is no longer available: https://snap-ci.com/
They recommend looking at GoCD: https://www.gocd.org
Should be updated in our CI Provider file here: https://github.com/cypress-io/cypress/blob/develop/packages/server/lib/util/ci_provider.coffee#L31
Standard GoCD environment variables listed here: https://docs.gocd.org/current/faq/environment_variables.html#standard-gocd-environment-variables | priority | remove snap ci provider add gocd provider snap ci is no longer available they recommend looking at gocd should be updated in our ci provider file here standard gocd environment variables listed here | 1 |
302,157 | 22,792,060,866 | IssuesEvent | 2022-07-10 06:25:37 | ohmtech-rdi/eurorack-blocks | https://api.github.com/repos/ohmtech-rdi/eurorack-blocks | closed | Make an overview video and "Getting Started" guide | documentation | The current state of the project doesn't include a call-to-action that would lead to a successful feeling in the 5-minutes and first-hour of evaluation milestones.
- Direct the user to first watch an overview video and then read the "Getting Started" guide,
- Add a less than 5 minutes overview video to show the end-to-end process of making a Eurorack module from conception to use in a Eurorack system,
- Add a less than 1 hour complete guide, to show in details the end-to-end process of making a Eurorack module from conception to use in a Eurorack system, with at least text and pictures, video at most.
As for strategy, first start by writing the guide and getting feedback about it, *before* we get into shooting the video, as it will be easier to integrate feedback for some text and pictures rather than a video edit.
## Chapters
- Overview
- Setting up dev environment
- Developing the software and testing in the simulator
- Ordering PCB/BOM and assembling `kivu12`
- Ordering PCB/BOM and assembling the module
## Tasks
- [x] Write "Getting Started" guide
- [x] Get feedback for "Getting Started" guide
- [x] Iterate "Getting Started" guide
- [ ] Write overview video script
- [ ] Get feedback for overview video script
- [ ] Shoot overview video draft
- [ ] Review draft & script
- [ ] Shoot overview video
- [ ] Integrate to project | 1.0 | Make an overview video and "Getting Started" guide - The current state of the project doesn't include a call-to-action that would lead to a successful feeling in the 5-minutes and first-hour of evaluation milestones.
- Direct the user to first watch an overview video and then read the "Getting Started" guide,
- Add a less than 5 minutes overview video to show the end-to-end process of making a Eurorack module from conception to use in a Eurorack system,
- Add a less than 1 hour complete guide, to show in details the end-to-end process of making a Eurorack module from conception to use in a Eurorack system, with at least text and pictures, video at most.
As for strategy, first start by writing the guide and getting feedback about it, *before* we get into shooting the video, as it will be easier to integrate feedback for some text and pictures rather than a video edit.
## Chapters
- Overview
- Setting up dev environment
- Developing the software and testing in the simulator
- Ordering PCB/BOM and assembling `kivu12`
- Ordering PCB/BOM and assembling the module
## Tasks
- [x] Write "Getting Started" guide
- [x] Get feedback for "Getting Started" guide
- [x] Iterate "Getting Started" guide
- [ ] Write overview video script
- [ ] Get feedback for overview video script
- [ ] Shoot overview video draft
- [ ] Review draft & script
- [ ] Shoot overview video
- [ ] Integrate to project | non_priority | make an overview video and getting started guide the current state of the project doesn t include a call to action that would lead to a successful feeling in the minutes and first hour of evaluation milestones direct the user to first watch an overview video and then read the getting started guide add a less than minutes overview video to show the end to end process of making a eurorack module from conception to use in a eurorack system add a less than hour complete guide to show in details the end to end process of making a eurorack module from conception to use in a eurorack system with at least text and pictures video at most as for strategy first start by writing the guide and getting feedback about it before we get into shooting the video as it will be easier to integrate feedback for some text and pictures rather than a video edit chapters overview setting up dev environment developing the software and testing in the simulator ordering pcb bom and assembling ordering pcb bom and assembling the module tasks write getting started guide get feedback for getting started guide iterate getting started guide write overview video script get feedback for overview video script shoot overview video draft review draft script shoot overview video integrate to project | 0 |
766,771 | 26,898,084,224 | IssuesEvent | 2023-02-06 13:53:39 | anoma/juvix | https://api.github.com/repos/anoma/juvix | closed | Special syntax for `case` | enhancement syntax priority:medium | As we discussed, we want to add the syntax:
```haskell
case v of
| p1 := b1
| p2 := b2
```
or
```haskell
case v of {
p1 := b1;
p2 := b2;
}
```
where one doesn't need parentheses around complex patterns on the left-hand side.
- The first version works better with #1639, I think.
| 1.0 | Special syntax for `case` - As we discussed, we want to add the syntax:
```haskell
case v of
| p1 := b1
| p2 := b2
```
or
```haskell
case v of {
p1 := b1;
p2 := b2;
}
```
where one doesn't need parentheses around complex patterns on the left-hand side.
- The first version works better with #1639, I think.
| priority | special syntax for case as we discussed we want to add the syntax haskell case v of or haskell case v of where one doesn t need parentheses around complex patterns on the left hand side the first version works better with i think | 1 |
68,444 | 21,664,559,307 | IssuesEvent | 2022-05-07 01:46:00 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Impossible to backfill over federation on joining a room on a new server | T-Defect P1 S-Major A-Timeline A-Federation | Repro steps (I hope):
* A room exists like #seaglass:matrix.org
* A user joins it from a virginal HS (e.g. `@neil:sandbox.modular.im`)
* They do not yet backpaginate any history for that room over federation for whatever reason
* Another user joins the room from the same HS (e.g. `@matthew:sandbox.modular.im`)
* The /sync as the new user joins it does not include a backwards pagination token at all (we think, assuming this is not a js-sdk bug).
* Therefore the new user can never paginate backwards in the room to pull in old history.
| 1.0 | Impossible to backfill over federation on joining a room on a new server - Repro steps (I hope):
* A room exists like #seaglass:matrix.org
* A user joins it from a virginal HS (e.g. `@neil:sandbox.modular.im`)
* They do not yet backpaginate any history for that room over federation for whatever reason
* Another user joins the room from the same HS (e.g. `@matthew:sandbox.modular.im`)
* The /sync as the new user joins it does not include a backwards pagination token at all (we think, assuming this is not a js-sdk bug).
* Therefore the new user can never paginate backwards in the room to pull in old history.
| non_priority | impossible to backfill over federation on joining a room on a new server repro steps i hope a room exists like seaglass matrix org a user joins it from a virginal hs e g neil sandbox modular im they do not yet backpaginate any history for that room over federation for whatever reason another user joins the room from the same hs e g matthew sandbox modular im the sync as the new user joins it does not include a backwards pagination token at all we think assuming this is not a js sdk bug therefore the new user can never paginate backwards in the room to pull in old history | 0 |
68,211 | 3,284,911,315 | IssuesEvent | 2015-10-28 18:20:36 | cs2103aug2015-t10-3j/main | https://api.github.com/repos/cs2103aug2015-t10-3j/main | closed | A user can view the history of a past task (completed or not) | priority.low type.story | ...so that the user can look at it for auditing and tracking purposes.
| 1.0 | A user can view the history of a past task (completed or not) - ...so that the user can look at it for auditing and tracking purposes.
| priority | a user can view the history of a past task completed or not so that the user can look at it for auditing and tracking purposes | 1 |
170,722 | 14,268,494,082 | IssuesEvent | 2020-11-20 22:35:21 | Titane73/it115-a5-group2-stackproject | https://api.github.com/repos/Titane73/it115-a5-group2-stackproject | opened | Install NPM | NPM documentation | Lets install NPM which will help us install the rest
https://linuxconfig.org/install-the-mean-stack-on-ubuntu-18-04-bionic-beaver-linux
if it works lets document it!
Thanks | 1.0 | Install NPM - Lets install NPM which will help us install the rest
https://linuxconfig.org/install-the-mean-stack-on-ubuntu-18-04-bionic-beaver-linux
if it works lets document it!
Thanks | non_priority | install npm lets install npm which will help us install the rest if it works lets document it thanks | 0 |
605,121 | 18,725,213,584 | IssuesEvent | 2021-11-03 15:38:57 | kubernetes/release | https://api.github.com/repos/kubernetes/release | closed | Collaborative Release Notes review | kind/feature sig/release lifecycle/rotten area/release-eng needs-priority | <!-- Please only use this template for submitting feature requests -->
### Intro
[Initial design document](https://docs.google.com/document/d/13SCA5HfjdJy5ZiEMPfhhZ2w25sqMS_QOcsxrJEwSLn4/edit?usp=sharing) was shared to `#release-notes`.
#### What would you like to be added:
Ability for release notes review and workflow to be more collaborative, instead of individuals running the tool and reviewing by themselves.
#### Why is this needed:
Increasing collaboration in team so that work does not happen in isolation
#### Proof of concept (ish?)
Ignoring the code, I have created [a sample of what this can look like](https://github.com/wilsonehusin/relnotesd-test/pull/1). However, instead of having the preview rendered on a comment (which makes the noise to PRs being linked), we can create a dummy repository or deploy to some page over Netlify / GitHub pages.
This means our workflow of reviewing release notes can be clearly seen through PR and commits, linked to each and every step.
### Roadmap
#### Prefactor
- [x] reduce API calls to GitHub #1888
- [ ] allow release notes process to pause / continue #2258
- [ ] increase test coverage #2272
... to be determined / continued / discussed ... | 1.0 | Collaborative Release Notes review - <!-- Please only use this template for submitting feature requests -->
### Intro
[Initial design document](https://docs.google.com/document/d/13SCA5HfjdJy5ZiEMPfhhZ2w25sqMS_QOcsxrJEwSLn4/edit?usp=sharing) was shared to `#release-notes`.
#### What would you like to be added:
Ability for release notes review and workflow to be more collaborative, instead of individuals running the tool and reviewing by themselves.
#### Why is this needed:
Increasing collaboration in team so that work does not happen in isolation
#### Proof of concept (ish?)
Ignoring the code, I have created [a sample of what this can look like](https://github.com/wilsonehusin/relnotesd-test/pull/1). However, instead of having the preview rendered on a comment (which makes the noise to PRs being linked), we can create a dummy repository or deploy to some page over Netlify / GitHub pages.
This means our workflow of reviewing release notes can be clearly seen through PR and commits, linked to each and every step.
### Roadmap
#### Prefactor
- [x] reduce API calls to GitHub #1888
- [ ] allow release notes process to pause / continue #2258
- [ ] increase test coverage #2272
... to be determined / continued / discussed ... | priority | collaborative release notes review intro was shared to release notes what would you like to be added ability for release notes review and workflow to be more collaborative instead of individuals running the tool and reviewing by themselves why is this needed increasing collaboration in team so that work does not happen in isolation proof of concept ish ignoring the code i have created however instead of having the preview rendered on a comment which makes the noise to prs being linked we can create a dummy repository or deploy to some page over netlify github pages this means our workflow of reviewing release notes can be clearly seen through pr and commits linked to each and every step roadmap prefactor reduce api calls to github allow release notes process to pause continue increase test coverage to be determined continued discussed | 1 |
245,770 | 7,890,693,403 | IssuesEvent | 2018-06-28 09:35:28 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | opened | viewer process still running after exit | Bug Likelihood: 3 - Occasional OS: All Priority: High Severity: 5 - Very Serious Support Group: Any version: 2.8.0 | I notice this most when using the cli. I've seen it on trunk and 2.8RC. Still working on a reliable reproducer and will post a script as soon as I can.
In the meantime, developers may want to check for runaway processes when finished with a VisIt session.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Kathleen Biagas
Original creation: 09/10/2014 05:33 pm
Original update: 09/24/2014 03:50 pm
Ticket number: 1978 | 1.0 | viewer process still running after exit - I notice this most when using the cli. I've seen it on trunk and 2.8RC. Still working on a reliable reproducer and will post a script as soon as I can.
In the meantime, developers may want to check for runaway processes when finished with a VisIt session.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Kathleen Biagas
Original creation: 09/10/2014 05:33 pm
Original update: 09/24/2014 03:50 pm
Ticket number: 1978 | priority | viewer process still running after exit i notice this most when using the cli i ve seen it on trunk and still working on a reliable reproducer and will post a script as soon as i can in the meantime developers may want to check for runaway processes when finished with a visit session redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author kathleen biagas original creation pm original update pm ticket number | 1 |
778,446 | 27,316,866,506 | IssuesEvent | 2023-02-24 16:21:42 | grpc/grpc | https://api.github.com/repos/grpc/grpc | closed | in c++ async stream mode, can write and read run simultaneously | kind/question priority/P3 untriaged | PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
When my program calls ::grpc::ClientAsyncReaderWriter<R, W>::write more than once and didn't wait the first write complete. The program assert false and quit. I can see the reason is that I am doing write while another write is already pending on the same RPC.
I am wondering, if I can call ::grpc::ClientAsyncReaderWriter<R, W>::write when ::grpc::ClientAsyncReaderWriter<R, W>::read is pending at the same RPC ?
This issue will be closed down once seen by the repo managers.
| 1.0 | in c++ async stream mode, can write and read run simultaneously - PLEASE DO NOT POST A QUESTION HERE.
This form is for bug reports and feature requests ONLY!
For general questions and troubleshooting, please ask/look for answers at StackOverflow, with "grpc" tag: https://stackoverflow.com/questions/tagged/grpc
For questions that specifically need to be answered by gRPC team members, please ask/look for answers at grpc.io mailing list: https://groups.google.com/forum/#!forum/grpc-io
When my program calls ::grpc::ClientAsyncReaderWriter<R, W>::write more than once and didn't wait the first write complete. The program assert false and quit. I can see the reason is that I am doing write while another write is already pending on the same RPC.
I am wondering, if I can call ::grpc::ClientAsyncReaderWriter<R, W>::write when ::grpc::ClientAsyncReaderWriter<R, W>::read is pending at the same RPC ?
This issue will be closed down once seen by the repo managers.
| priority | in c async stream mode can write and read run simultaneously please do not post a question here this form is for bug reports and feature requests only for general questions and troubleshooting please ask look for answers at stackoverflow with grpc tag for questions that specifically need to be answered by grpc team members please ask look for answers at grpc io mailing list when my program calls grpc clientasyncreaderwriter write more than once and didn t wait the first write complete the program assert false and quit i can see the reason is that i am doing write while another write is already pending on the same rpc i am wondering if i can call grpc clientasyncreaderwriter write when grpc clientasyncreaderwriter read is pending at the same rpc this issue will be closed down once seen by the repo managers | 1 |
329,080 | 10,012,181,912 | IssuesEvent | 2019-07-15 12:37:12 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.reddit.com - video or audio doesn't play | browser-fenix engine-gecko priority-critical | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.reddit.com/r/interestingasfuck/comments/cd70m2/red_bull_just_made_the_fastest_pit_stop_in_f1/
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes
**Problem type**: Video or audio doesn't play
**Description**: "No video with supported format and MIME type found"
**Steps to Reproduce**:
The video will not play when viewed on Reddit. It will play if viewed on the hosting page (https://gfycat.com/genuineignorantiaerismetalmark). The same issue occurs in regular Firefox for Android. I see this often on Reddit. Not sure if it corresponds to specific content hosts.
[](https://webcompat.com/uploads/2019/7/c67245bb-5257-4885-bfb3-fe7a6b940dd6.jpg)
[](https://webcompat.com/uploads/2019/7/4f4ddf84-64b4-488f-b62e-28da77caa34a.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.reddit.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://www.reddit.com/r/interestingasfuck/comments/cd70m2/red_bull_just_made_the_fastest_pit_stop_in_f1/
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes
**Problem type**: Video or audio doesn't play
**Description**: "No video with supported format and MIME type found"
**Steps to Reproduce**:
The video will not play when viewed on Reddit. It will play if viewed on the hosting page (https://gfycat.com/genuineignorantiaerismetalmark). The same issue occurs in regular Firefox for Android. I see this often on Reddit. Not sure if it corresponds to specific content hosts.
[](https://webcompat.com/uploads/2019/7/c67245bb-5257-4885-bfb3-fe7a6b940dd6.jpg)
[](https://webcompat.com/uploads/2019/7/4f4ddf84-64b4-488f-b62e-28da77caa34a.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | video or audio doesn t play url browser version firefox mobile operating system android tested another browser yes problem type video or audio doesn t play description no video with supported format and mime type found steps to reproduce the video will not play when viewed on reddit it will play if viewed on the hosting page the same issue occurs in regular firefox for android i see this often on reddit not sure if it corresponds to specific content hosts browser configuration none from with ❤️ | 1 |
342,208 | 10,313,377,077 | IssuesEvent | 2019-08-29 22:34:23 | metabase/metabase | https://api.github.com/repos/metabase/metabase | opened | X-axis on saved charts is obscured by the footer until it's resized | Priority:P2 Type:Bug Visualization/ | Whenever I open a saved question on `release-0.33.x` if it's an x/y chart of any kind, the x-axis labels are covered up by the footer UI elements until I either resize the browser window or cause the chart to rerender by opening up a sidebar.

| 1.0 | X-axis on saved charts is obscured by the footer until it's resized - Whenever I open a saved question on `release-0.33.x` if it's an x/y chart of any kind, the x-axis labels are covered up by the footer UI elements until I either resize the browser window or cause the chart to rerender by opening up a sidebar.

| priority | x axis on saved charts is obscured by the footer until it s resized whenever i open a saved question on release x if it s an x y chart of any kind the x axis labels are covered up by the footer ui elements until i either resize the browser window or cause the chart to rerender by opening up a sidebar | 1 |
414,502 | 12,103,957,487 | IssuesEvent | 2020-04-20 19:20:27 | Noah2610/LD46 | https://api.github.com/repos/Noah2610/LD46 | closed | State between MainMenu and Ingame - "Play with Headphones" | enhancement priority:high | Also loads level.
`LoadIngame` state. | 1.0 | State between MainMenu and Ingame - "Play with Headphones" - Also loads level.
`LoadIngame` state. | priority | state between mainmenu and ingame play with headphones also loads level loadingame state | 1 |
375,272 | 11,102,235,719 | IssuesEvent | 2019-12-16 23:22:28 | kubernetes/minikube | https://api.github.com/repos/kubernetes/minikube | closed | minikube status should show status for all profiles | good first issue help wanted kind/feature priority/important-longterm | minikube status shows "Stopped" when starting a minikube in a different profile.
```
$ minikube start -p p1
$ minikube status
host: Stopped
kubelet:
apiserver:
kubectl:
$ minikube status -p p1
host: Running
kubelet: Running
apiserver: Running
kubectl: Correctly Configured: pointing to minikube-vm at 192.168.99.132
```
To recreate :
- Start minikube default profile `minikube start`
- Stop minikube `minikube stop`
- Start minikube in a different profile `minikube start -p p1`
- Get miniube status `minikube status`
Expected :
I would like minikube status, show status of all profiles, if I don't provide a profile name. | 1.0 | minikube status should show status for all profiles - minikube status shows "Stopped" when starting a minikube in a different profile.
```
$ minikube start -p p1
$ minikube status
host: Stopped
kubelet:
apiserver:
kubectl:
$ minikube status -p p1
host: Running
kubelet: Running
apiserver: Running
kubectl: Correctly Configured: pointing to minikube-vm at 192.168.99.132
```
To recreate :
- Start minikube default profile `minikube start`
- Stop minikube `minikube stop`
- Start minikube in a different profile `minikube start -p p1`
- Get miniube status `minikube status`
Expected :
I would like minikube status, show status of all profiles, if I don't provide a profile name. | priority | minikube status should show status for all profiles minikube status shows stopped when starting a minikube in a different profile minikube start p minikube status host stopped kubelet apiserver kubectl minikube status p host running kubelet running apiserver running kubectl correctly configured pointing to minikube vm at to recreate start minikube default profile minikube start stop minikube minikube stop start minikube in a different profile minikube start p get miniube status minikube status expected i would like minikube status show status of all profiles if i don t provide a profile name | 1 |
154,668 | 19,751,425,662 | IssuesEvent | 2022-01-15 05:08:46 | b-tomi/100DaysOfCode | https://api.github.com/repos/b-tomi/100DaysOfCode | opened | CVE-2018-19827 (High) detected in node-sass-4.14.1.tgz | security vulnerability | ## CVE-2018-19827 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>
Dependency Hierarchy:
- gulp-sass-4.1.0.tgz (Root Library)
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/b-tomi/100DaysOfCode/commit/c88b9429eb68a85b22f0e39cac7bf20b89cb6709">c88b9429eb68a85b22f0e39cac7bf20b89cb6709</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a use-after-free vulnerability exists in the SharedPtr class in SharedPtr.cpp (or SharedPtr.hpp) that may cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19827>CVE-2018-19827</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.6.0">https://github.com/sass/libsass/releases/tag/3.6.0</a></p>
<p>Release Date: 2018-12-03</p>
<p>Fix Resolution: libsass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-19827 (High) detected in node-sass-4.14.1.tgz - ## CVE-2018-19827 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>
Dependency Hierarchy:
- gulp-sass-4.1.0.tgz (Root Library)
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/b-tomi/100DaysOfCode/commit/c88b9429eb68a85b22f0e39cac7bf20b89cb6709">c88b9429eb68a85b22f0e39cac7bf20b89cb6709</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a use-after-free vulnerability exists in the SharedPtr class in SharedPtr.cpp (or SharedPtr.hpp) that may cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19827>CVE-2018-19827</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/releases/tag/3.6.0">https://github.com/sass/libsass/releases/tag/3.6.0</a></p>
<p>Release Date: 2018-12-03</p>
<p>Fix Resolution: libsass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in node sass tgz cve high severity vulnerability vulnerable library node sass tgz wrapper around libsass library home page a href dependency hierarchy gulp sass tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details in libsass a use after free vulnerability exists in the sharedptr class in sharedptr cpp or sharedptr hpp that may cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
221,490 | 7,389,146,901 | IssuesEvent | 2018-03-16 07:19:36 | llmhyy/microbat | https://api.github.com/repos/llmhyy/microbat | opened | [Instrumentation] StackOverflowError | high priority | hi @lylytran
Please kindly check the following command line, I generate a lot of include classes, it wind up with an exception as `Exception: java.lang.StackOverflowError thrown from the UncaughtExceptionHandler in thread "Thread-0"`. The bug happens for Closure-14. Would you please kindly have a check? Thanks!
```
E:\linyun\bug_repo\jdk1.7.0_800\bin\java -Xmx30g -ea -noverify -javaagent:E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\instrumentator.jar=printProgress=true,precheck=false,excludes=java.util.Properties*,java_home=E:\linyun\bug_repo\jdk1.7.0_800,stepLimit=100000,working_dir=E:\linyun\bug_repo\Closure\14\fix,varLayer=1,includes=com.google.javascript.jscomp.ControlFlowAnalysisTest;java.util.List;java.util.LinkedList;java.util.AbstractSequentialList;java.util.Collections$SynchronizedRandomAccessList;java.util.Collections$SynchronizedList;java.util.concurrent.CopyOnWriteArrayList;com.google.common.collect.ImmutableList;com.google.common.collect.EmptyImmutableList;java.lang.Object;com.google.javascript.jscomp.graph.LinkedDirectedGraph;com.google.javascript.jscomp.graph.Graph;java.util.Map;java.util.Collections$SingletonMap;com.google.common.collect.ImmutableMap;com.google.common.collect.RegularImmutableMap;java.util.TreeMap;java.util.NavigableMap;java.util.SortedMap;java.util.TreeMap$DescendingSubMap;java.util.TreeMap$NavigableSubMap;com.google.javascript.jscomp.graph.DiGraph;com.google.common.base.Predicates;com.google.javascript.jscomp.graph.DiGraph$DiGraphNode;com.google.javascript.jscomp.graph.LinkedDirectedGraph$AnnotatedLinkedDirectedGraphNode;com.google.javascript.jscomp.graph.LinkedDirectedGraph$LinkedDirectedGraphNode;java.util.Iterator;java.util.ServiceLoader$LazyIterator;java.util.ServiceLoader$1;java.util.LinkedHashMap$EntryIterator;java.util.LinkedHashMap$LinkedHashIterator;java.util.HashMap$ValueIterator;java.util.Collections$EmptyIterator;java.util.Vector$Itr;com.google.common.collect.EmptyImmutableList$1;com.google.common.collect.UnmodifiableListIterator;com.google.common.collect.UnmodifiableIterator;com.google.common.collect.Iterators$1;com.google.common.collect.Iterators$2;com.google.common.collect.Iterators$11;com.google.common.collect.AbstractIndexedListIterator;java.util.TreeMap$KeyIterator;java.util.TreeMap$PrivateEntryIterator;java.util.TreeMap$NavigableSubMap$DescendingSubMapKeyIterator;java.util.TreeMap$NavigableSubMap$SubMapIterator;com.google.javascript.rhino.head.Node$NodeIterator;java.util.Collections$UnmodifiableCollection$1;java.util.ArrayDeque$DeqIterator;com.google.common.collect.AbstractMultimap$WrappedCollection$WrappedIterator;java.util.IdentityHashMap$KeyIterator;java.util.IdentityHashMap$IdentityHashMapIterator;com.google.common.collect.Lists;com.google.javascript.jscomp.ControlFlowAnalysis;com.google.javascript.rhino.Node;com.google.javascript.jscomp.NodeTraversal$Callback;com.google.javascript.jscomp.CheckGlobalThis;com.google.javascript.jscomp.CheckAccessControls;com.google.javascript.jscomp.NodeTraversal$ScopedCallback;com.google.javascript.jscomp.VarCheck;com.google.javascript.jscomp.NodeTraversal$AbstractPostOrderCallback;com.google.javascript.jscomp.TypeCheck;com.google.javascript.jscomp.CheckDebuggerStatement;com.google.javascript.jscomp.CheckRegExp;com.google.javascript.jscomp.CheckSideEffects;com.google.javascript.jscomp.CheckUnreachableCode;com.google.javascript.jscomp.ConstCheck;com.google.javascript.jscomp.StrictModeCheck;com.google.javascript.jscomp.JsMessageVisitor;com.google.javascript.jscomp.PrepareAst$PrepareAnnotations;java.lang.Integer;com.google.common.base.Preconditions;java.util.Deque;com.google.javascript.jscomp.graph.DiGraph$DiGraphEdge;com.google.javascript.jscomp.graph.LinkedDirectedGraph$AnnotatedLinkedDirectedGraphEdge;com.google.javascript.jscomp.graph.LinkedDirectedGraph$LinkedDirectedGraphEdge;com.google.common.base.Predicate;com.google.javascript.jscomp.SourceMap$DetailLevel;com.google.javascript.jscomp.SourceMap$DetailLevel$1;com.google.javascript.jscomp.SourceMap$DetailLevel$2;com.google.common.base.CharMatcher;com.google.common.base.CharMatcher$11;com.google.common.base.CharMatcher$12;com.google.common.base.CharMatcher$Or;com.google.common.base.CharMatcher$15;com.google.common.base.CharMatcher$8;com.google.common.base.CharMatcher$1;com.google.common.base.CharMatcher$2;com.google.common.base.CharMatcher$3;com.google.common.base.CharMatcher$4;com.google.common.base.CharMatcher$5;com.google.common.base.CharMatcher$6;com.google.common.base.CharMatcher$7;com.google.javascript.jscomp.NodeUtil$NumbericResultPredicate;com.google.javascript.jscomp.NodeUtil$BooleanResultPredicate;com.google.javascript.jscomp.NodeUtil$MayBeStringResultPredicate;com.google.javascript.jscomp.NodeUtil$MatchNotFunction;com.google.common.base.Predicates$IsEqualToPredicate;com.google.common.collect.Multimap;com.google.common.collect.HashMultimap;com.google.common.collect.AbstractSetMultimap;com.google.common.collect.SetMultimap;com.google.common.collect.AbstractMultimap;com.google.javascript.jscomp.NodeUtil;java.util.Collection;java.util.LinkedHashSet;java.util.HashMap$Values;com.google.common.collect.ImmutableSet;com.google.common.collect.ImmutableCollection;com.google.common.collect.ImmutableCollection$EmptyImmutableCollection;com.google.common.collect.RegularImmutableSet;com.google.common.collect.ImmutableSet$ArrayImmutableSet;java.util.TreeSet;java.util.NavigableSet;java.util.SortedSet;com.google.common.collect.SingletonImmutableSet;java.util.TreeMap$KeySet;com.google.common.collect.RegularImmutableMap$EntrySet;com.google.common.collect.AbstractMultimap$WrappedSet;com.google.common.collect.AbstractMultimap$WrappedCollection;java.util.PriorityQueue;java.util.AbstractQueue;java.util.IdentityHashMap$KeySet,launch_class=com.google.javascript.jscomp.ControlFlowAnalysisTest,class_path=E:\linyun\bug_repo\Closure\14\fix\lib\caja-r4314.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\junit.jar;E:\linyun\bug_repo\Closure\14\fix\lib\junit.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\testrunner.jar;E:\linyun\bug_repo\Closure\14\fix\lib\jarjar.jar;E:\linyun\bug_repo\Closure\14\fix\build\lib\rhino1_7R4pre\js.jar;E:\linyun\bug_repo\Closure\14\fix\build\test;E:\linyun\bug_repo\Closure\14\fix\lib\jsr305.jar;E:\linyun\bug_repo\Closure\14\fix\lib\json.jar;E:\linyun\bug_repo\Closure\14\fix\build\lib\rhino.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\org.hamcrest.core.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\javassist.jar;E:\linyun\bug_repo\Closure\14\fix\lib\args4j.jar;E:\linyun\bug_repo\Closure\14\fix\lib\protobuf-java.jar;E:\linyun\bug_repo\Closure\14\fix\lib\rhino\testsrc\org\mozilla\javascript\tests\commonjs\module\modules.jar;E:\linyun\bug_repo\Closure\14\fix\lib\ant-launcher.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\bcel-6.0.jar;E:\linyun\bug_repo\Closure\14\fix\lib\ant.jar;E:\linyun\bug_repo\Closure\14\fix\lib\guava.jar;E:\linyun\bug_repo\Closure\14\fix\build\classes,dump_file_path=E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\trace\Closure\14\fix.exec,expectedSteps=20924 -cp E:\linyun\bug_repo\Closure\14\fix\lib\caja-r4314.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\junit.jar;E:\linyun\bug_repo\Closure\14\fix\lib\junit.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\testrunner.jar;E:\linyun\bug_repo\Closure\14\fix\lib\jarjar.jar;E:\linyun\bug_repo\Closure\14\fix\build\lib\rhino1_7R4pre\js.jar;E:\linyun\bug_repo\Closure\14\fix\build\test;E:\linyun\bug_repo\Closure\14\fix\lib\jsr305.jar;E:\linyun\bug_repo\Closure\14\fix\lib\json.jar;E:\linyun\bug_repo\Closure\14\fix\build\lib\rhino.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\org.hamcrest.core.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\javassist.jar;E:\linyun\bug_repo\Closure\14\fix\lib\args4j.jar;E:\linyun\bug_repo\Closure\14\fix\lib\protobuf-java.jar;E:\linyun\bug_repo\Closure\14\fix\lib\rhino\testsrc\org\mozilla\javascript\tests\commonjs\module\modules.jar;E:\linyun\bug_repo\Closure\14\fix\lib\ant-launcher.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\bcel-6.0.jar;E:\linyun\bug_repo\Closure\14\fix\lib\ant.jar;E:\linyun\bug_repo\Closure\14\fix\lib\guava.jar;E:\linyun\bug_repo\Closure\14\fix\build\classes microbat.evaluation.junit.MicroBatTestRunner com.google.javascript.jscomp.ControlFlowAnalysisTest testDeepNestedFinally
``` | 1.0 | [Instrumentation] StackOverflowError - hi @lylytran
Please kindly check the following command line, I generate a lot of include classes, it wind up with an exception as `Exception: java.lang.StackOverflowError thrown from the UncaughtExceptionHandler in thread "Thread-0"`. The bug happens for Closure-14. Would you please kindly have a check? Thanks!
```
E:\linyun\bug_repo\jdk1.7.0_800\bin\java -Xmx30g -ea -noverify -javaagent:E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\instrumentator.jar=printProgress=true,precheck=false,excludes=java.util.Properties*,java_home=E:\linyun\bug_repo\jdk1.7.0_800,stepLimit=100000,working_dir=E:\linyun\bug_repo\Closure\14\fix,varLayer=1,includes=com.google.javascript.jscomp.ControlFlowAnalysisTest;java.util.List;java.util.LinkedList;java.util.AbstractSequentialList;java.util.Collections$SynchronizedRandomAccessList;java.util.Collections$SynchronizedList;java.util.concurrent.CopyOnWriteArrayList;com.google.common.collect.ImmutableList;com.google.common.collect.EmptyImmutableList;java.lang.Object;com.google.javascript.jscomp.graph.LinkedDirectedGraph;com.google.javascript.jscomp.graph.Graph;java.util.Map;java.util.Collections$SingletonMap;com.google.common.collect.ImmutableMap;com.google.common.collect.RegularImmutableMap;java.util.TreeMap;java.util.NavigableMap;java.util.SortedMap;java.util.TreeMap$DescendingSubMap;java.util.TreeMap$NavigableSubMap;com.google.javascript.jscomp.graph.DiGraph;com.google.common.base.Predicates;com.google.javascript.jscomp.graph.DiGraph$DiGraphNode;com.google.javascript.jscomp.graph.LinkedDirectedGraph$AnnotatedLinkedDirectedGraphNode;com.google.javascript.jscomp.graph.LinkedDirectedGraph$LinkedDirectedGraphNode;java.util.Iterator;java.util.ServiceLoader$LazyIterator;java.util.ServiceLoader$1;java.util.LinkedHashMap$EntryIterator;java.util.LinkedHashMap$LinkedHashIterator;java.util.HashMap$ValueIterator;java.util.Collections$EmptyIterator;java.util.Vector$Itr;com.google.common.collect.EmptyImmutableList$1;com.google.common.collect.UnmodifiableListIterator;com.google.common.collect.UnmodifiableIterator;com.google.common.collect.Iterators$1;com.google.common.collect.Iterators$2;com.google.common.collect.Iterators$11;com.google.common.collect.AbstractIndexedListIterator;java.util.TreeMap$KeyIterator;java.util.TreeMap$PrivateEntryIterator;java.util.TreeMap$NavigableSubMap$DescendingSubMapKeyIterator;java.util.TreeMap$NavigableSubMap$SubMapIterator;com.google.javascript.rhino.head.Node$NodeIterator;java.util.Collections$UnmodifiableCollection$1;java.util.ArrayDeque$DeqIterator;com.google.common.collect.AbstractMultimap$WrappedCollection$WrappedIterator;java.util.IdentityHashMap$KeyIterator;java.util.IdentityHashMap$IdentityHashMapIterator;com.google.common.collect.Lists;com.google.javascript.jscomp.ControlFlowAnalysis;com.google.javascript.rhino.Node;com.google.javascript.jscomp.NodeTraversal$Callback;com.google.javascript.jscomp.CheckGlobalThis;com.google.javascript.jscomp.CheckAccessControls;com.google.javascript.jscomp.NodeTraversal$ScopedCallback;com.google.javascript.jscomp.VarCheck;com.google.javascript.jscomp.NodeTraversal$AbstractPostOrderCallback;com.google.javascript.jscomp.TypeCheck;com.google.javascript.jscomp.CheckDebuggerStatement;com.google.javascript.jscomp.CheckRegExp;com.google.javascript.jscomp.CheckSideEffects;com.google.javascript.jscomp.CheckUnreachableCode;com.google.javascript.jscomp.ConstCheck;com.google.javascript.jscomp.StrictModeCheck;com.google.javascript.jscomp.JsMessageVisitor;com.google.javascript.jscomp.PrepareAst$PrepareAnnotations;java.lang.Integer;com.google.common.base.Preconditions;java.util.Deque;com.google.javascript.jscomp.graph.DiGraph$DiGraphEdge;com.google.javascript.jscomp.graph.LinkedDirectedGraph$AnnotatedLinkedDirectedGraphEdge;com.google.javascript.jscomp.graph.LinkedDirectedGraph$LinkedDirectedGraphEdge;com.google.common.base.Predicate;com.google.javascript.jscomp.SourceMap$DetailLevel;com.google.javascript.jscomp.SourceMap$DetailLevel$1;com.google.javascript.jscomp.SourceMap$DetailLevel$2;com.google.common.base.CharMatcher;com.google.common.base.CharMatcher$11;com.google.common.base.CharMatcher$12;com.google.common.base.CharMatcher$Or;com.google.common.base.CharMatcher$15;com.google.common.base.CharMatcher$8;com.google.common.base.CharMatcher$1;com.google.common.base.CharMatcher$2;com.google.common.base.CharMatcher$3;com.google.common.base.CharMatcher$4;com.google.common.base.CharMatcher$5;com.google.common.base.CharMatcher$6;com.google.common.base.CharMatcher$7;com.google.javascript.jscomp.NodeUtil$NumbericResultPredicate;com.google.javascript.jscomp.NodeUtil$BooleanResultPredicate;com.google.javascript.jscomp.NodeUtil$MayBeStringResultPredicate;com.google.javascript.jscomp.NodeUtil$MatchNotFunction;com.google.common.base.Predicates$IsEqualToPredicate;com.google.common.collect.Multimap;com.google.common.collect.HashMultimap;com.google.common.collect.AbstractSetMultimap;com.google.common.collect.SetMultimap;com.google.common.collect.AbstractMultimap;com.google.javascript.jscomp.NodeUtil;java.util.Collection;java.util.LinkedHashSet;java.util.HashMap$Values;com.google.common.collect.ImmutableSet;com.google.common.collect.ImmutableCollection;com.google.common.collect.ImmutableCollection$EmptyImmutableCollection;com.google.common.collect.RegularImmutableSet;com.google.common.collect.ImmutableSet$ArrayImmutableSet;java.util.TreeSet;java.util.NavigableSet;java.util.SortedSet;com.google.common.collect.SingletonImmutableSet;java.util.TreeMap$KeySet;com.google.common.collect.RegularImmutableMap$EntrySet;com.google.common.collect.AbstractMultimap$WrappedSet;com.google.common.collect.AbstractMultimap$WrappedCollection;java.util.PriorityQueue;java.util.AbstractQueue;java.util.IdentityHashMap$KeySet,launch_class=com.google.javascript.jscomp.ControlFlowAnalysisTest,class_path=E:\linyun\bug_repo\Closure\14\fix\lib\caja-r4314.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\junit.jar;E:\linyun\bug_repo\Closure\14\fix\lib\junit.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\testrunner.jar;E:\linyun\bug_repo\Closure\14\fix\lib\jarjar.jar;E:\linyun\bug_repo\Closure\14\fix\build\lib\rhino1_7R4pre\js.jar;E:\linyun\bug_repo\Closure\14\fix\build\test;E:\linyun\bug_repo\Closure\14\fix\lib\jsr305.jar;E:\linyun\bug_repo\Closure\14\fix\lib\json.jar;E:\linyun\bug_repo\Closure\14\fix\build\lib\rhino.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\org.hamcrest.core.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\javassist.jar;E:\linyun\bug_repo\Closure\14\fix\lib\args4j.jar;E:\linyun\bug_repo\Closure\14\fix\lib\protobuf-java.jar;E:\linyun\bug_repo\Closure\14\fix\lib\rhino\testsrc\org\mozilla\javascript\tests\commonjs\module\modules.jar;E:\linyun\bug_repo\Closure\14\fix\lib\ant-launcher.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\bcel-6.0.jar;E:\linyun\bug_repo\Closure\14\fix\lib\ant.jar;E:\linyun\bug_repo\Closure\14\fix\lib\guava.jar;E:\linyun\bug_repo\Closure\14\fix\build\classes,dump_file_path=E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\trace\Closure\14\fix.exec,expectedSteps=20924 -cp E:\linyun\bug_repo\Closure\14\fix\lib\caja-r4314.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\junit.jar;E:\linyun\bug_repo\Closure\14\fix\lib\junit.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\testrunner.jar;E:\linyun\bug_repo\Closure\14\fix\lib\jarjar.jar;E:\linyun\bug_repo\Closure\14\fix\build\lib\rhino1_7R4pre\js.jar;E:\linyun\bug_repo\Closure\14\fix\build\test;E:\linyun\bug_repo\Closure\14\fix\lib\jsr305.jar;E:\linyun\bug_repo\Closure\14\fix\lib\json.jar;E:\linyun\bug_repo\Closure\14\fix\build\lib\rhino.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\org.hamcrest.core.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\javassist.jar;E:\linyun\bug_repo\Closure\14\fix\lib\args4j.jar;E:\linyun\bug_repo\Closure\14\fix\lib\protobuf-java.jar;E:\linyun\bug_repo\Closure\14\fix\lib\rhino\testsrc\org\mozilla\javascript\tests\commonjs\module\modules.jar;E:\linyun\bug_repo\Closure\14\fix\lib\ant-launcher.jar;E:\linyun\software\eclipse-java-mars\eclipse-java-mars-clean\eclipse\dropins\junit_lib\bcel-6.0.jar;E:\linyun\bug_repo\Closure\14\fix\lib\ant.jar;E:\linyun\bug_repo\Closure\14\fix\lib\guava.jar;E:\linyun\bug_repo\Closure\14\fix\build\classes microbat.evaluation.junit.MicroBatTestRunner com.google.javascript.jscomp.ControlFlowAnalysisTest testDeepNestedFinally
``` | priority | stackoverflowerror hi lylytran please kindly check the following command line i generate a lot of include classes it wind up with an exception as exception java lang stackoverflowerror thrown from the uncaughtexceptionhandler in thread thread the bug happens for closure would you please kindly have a check thanks e linyun bug repo bin java ea noverify javaagent e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib instrumentator jar printprogress true precheck false excludes java util properties java home e linyun bug repo steplimit working dir e linyun bug repo closure fix varlayer includes com google javascript jscomp controlflowanalysistest java util list java util linkedlist java util abstractsequentiallist java util collections synchronizedrandomaccesslist java util collections synchronizedlist java util concurrent copyonwritearraylist com google common collect immutablelist com google common collect emptyimmutablelist java lang object com google javascript jscomp graph linkeddirectedgraph com google javascript jscomp graph graph java util map java util collections singletonmap com google common collect immutablemap com google common collect regularimmutablemap java util treemap java util navigablemap java util sortedmap java util treemap descendingsubmap java util treemap navigablesubmap com google javascript jscomp graph digraph com google common base predicates com google javascript jscomp graph digraph digraphnode com google javascript jscomp graph linkeddirectedgraph annotatedlinkeddirectedgraphnode com google javascript jscomp graph linkeddirectedgraph linkeddirectedgraphnode java util iterator java util serviceloader lazyiterator java util serviceloader java util linkedhashmap entryiterator java util linkedhashmap linkedhashiterator java util hashmap valueiterator java util collections emptyiterator java util vector itr com google common collect emptyimmutablelist com google common collect unmodifiablelistiterator com google common collect unmodifiableiterator com google common collect iterators com google common collect iterators com google common collect iterators com google common collect abstractindexedlistiterator java util treemap keyiterator java util treemap privateentryiterator java util treemap navigablesubmap descendingsubmapkeyiterator java util treemap navigablesubmap submapiterator com google javascript rhino head node nodeiterator java util collections unmodifiablecollection java util arraydeque deqiterator com google common collect abstractmultimap wrappedcollection wrappediterator java util identityhashmap keyiterator java util identityhashmap identityhashmapiterator com google common collect lists com google javascript jscomp controlflowanalysis com google javascript rhino node com google javascript jscomp nodetraversal callback com google javascript jscomp checkglobalthis com google javascript jscomp checkaccesscontrols com google javascript jscomp nodetraversal scopedcallback com google javascript jscomp varcheck com google javascript jscomp nodetraversal abstractpostordercallback com google javascript jscomp typecheck com google javascript jscomp checkdebuggerstatement com google javascript jscomp checkregexp com google javascript jscomp checksideeffects com google javascript jscomp checkunreachablecode com google javascript jscomp constcheck com google javascript jscomp strictmodecheck com google javascript jscomp jsmessagevisitor com google javascript jscomp prepareast prepareannotations java lang integer com google common base preconditions java util deque com google javascript jscomp graph digraph digraphedge com google javascript jscomp graph linkeddirectedgraph annotatedlinkeddirectedgraphedge com google javascript jscomp graph linkeddirectedgraph linkeddirectedgraphedge com google common base predicate com google javascript jscomp sourcemap detaillevel com google javascript jscomp sourcemap detaillevel com google javascript jscomp sourcemap detaillevel com google common base charmatcher com google common base charmatcher com google common base charmatcher com google common base charmatcher or com google common base charmatcher com google common base charmatcher com google common base charmatcher com google common base charmatcher com google common base charmatcher com google common base charmatcher com google common base charmatcher com google common base charmatcher com google common base charmatcher com google javascript jscomp nodeutil numbericresultpredicate com google javascript jscomp nodeutil booleanresultpredicate com google javascript jscomp nodeutil maybestringresultpredicate com google javascript jscomp nodeutil matchnotfunction com google common base predicates isequaltopredicate com google common collect multimap com google common collect hashmultimap com google common collect abstractsetmultimap com google common collect setmultimap com google common collect abstractmultimap com google javascript jscomp nodeutil java util collection java util linkedhashset java util hashmap values com google common collect immutableset com google common collect immutablecollection com google common collect immutablecollection emptyimmutablecollection com google common collect regularimmutableset com google common collect immutableset arrayimmutableset java util treeset java util navigableset java util sortedset com google common collect singletonimmutableset java util treemap keyset com google common collect regularimmutablemap entryset com google common collect abstractmultimap wrappedset com google common collect abstractmultimap wrappedcollection java util priorityqueue java util abstractqueue java util identityhashmap keyset launch class com google javascript jscomp controlflowanalysistest class path e linyun bug repo closure fix lib caja jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib junit jar e linyun bug repo closure fix lib junit jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib testrunner jar e linyun bug repo closure fix lib jarjar jar e linyun bug repo closure fix build lib js jar e linyun bug repo closure fix build test e linyun bug repo closure fix lib jar e linyun bug repo closure fix lib json jar e linyun bug repo closure fix build lib rhino jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib org hamcrest core jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib javassist jar e linyun bug repo closure fix lib jar e linyun bug repo closure fix lib protobuf java jar e linyun bug repo closure fix lib rhino testsrc org mozilla javascript tests commonjs module modules jar e linyun bug repo closure fix lib ant launcher jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib bcel jar e linyun bug repo closure fix lib ant jar e linyun bug repo closure fix lib guava jar e linyun bug repo closure fix build classes dump file path e linyun software eclipse java mars eclipse java mars clean eclipse trace closure fix exec expectedsteps cp e linyun bug repo closure fix lib caja jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib junit jar e linyun bug repo closure fix lib junit jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib testrunner jar e linyun bug repo closure fix lib jarjar jar e linyun bug repo closure fix build lib js jar e linyun bug repo closure fix build test e linyun bug repo closure fix lib jar e linyun bug repo closure fix lib json jar e linyun bug repo closure fix build lib rhino jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib org hamcrest core jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib javassist jar e linyun bug repo closure fix lib jar e linyun bug repo closure fix lib protobuf java jar e linyun bug repo closure fix lib rhino testsrc org mozilla javascript tests commonjs module modules jar e linyun bug repo closure fix lib ant launcher jar e linyun software eclipse java mars eclipse java mars clean eclipse dropins junit lib bcel jar e linyun bug repo closure fix lib ant jar e linyun bug repo closure fix lib guava jar e linyun bug repo closure fix build classes microbat evaluation junit microbattestrunner com google javascript jscomp controlflowanalysistest testdeepnestedfinally | 1 |
325,653 | 9,934,165,028 | IssuesEvent | 2019-07-02 13:55:53 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | auth.geeksforgeeks.org - site is not usable | browser-fenix engine-gecko priority-important | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://auth.geeksforgeeks.org/fb-login.php?to=https://auth.geeksforgeeks.org/?to=https://practice.geeksforgeeks.org/courses/
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 7.1.2
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: I cannot sign up.
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | auth.geeksforgeeks.org - site is not usable - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-fenix -->
**URL**: https://auth.geeksforgeeks.org/fb-login.php?to=https://auth.geeksforgeeks.org/?to=https://practice.geeksforgeeks.org/courses/
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android 7.1.2
**Tested Another Browser**: Yes
**Problem type**: Site is not usable
**Description**: I cannot sign up.
**Steps to Reproduce**:
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | auth geeksforgeeks org site is not usable url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description i cannot sign up steps to reproduce browser configuration none from with ❤️ | 1 |
43,927 | 5,577,655,990 | IssuesEvent | 2017-03-28 10:13:01 | khartec/waltz | https://api.github.com/repos/khartec/waltz | closed | Change initiative id selector factory should support identity selector | fixed (test & close) | To go from change initiative to change initiative
Useful for creating survey runs for a single change initiative | 1.0 | Change initiative id selector factory should support identity selector - To go from change initiative to change initiative
Useful for creating survey runs for a single change initiative | non_priority | change initiative id selector factory should support identity selector to go from change initiative to change initiative useful for creating survey runs for a single change initiative | 0 |
398,863 | 27,215,444,463 | IssuesEvent | 2023-02-20 21:09:49 | mawada-sweis/Clustering-Analysis | https://api.github.com/repos/mawada-sweis/Clustering-Analysis | closed | Readme file | documentation enhancement | # The Citywide Mobility Survey (CMS)
The Citywide Mobility Survey (CMS) is a survey conducted by the New York City Department of Transportation (DOT) to gather information about the travel behavior, preferences, and attitudes of New York City residents. The survey is conducted periodically, and the data collected is used to inform transportation planning and policy decisions.
## Table of Contents
* [Data Objectives](#data-objectives)
* [Project Objectives](#project-objective)
## Data Objectives
The primary objectives of the CMS data collection are:
* Identify the factors and experiences that drive transportation choices for New York City residents.
* Understand current views on the state of transportation within the City.
* Measure attitudes toward current transportation issues and topics in New York City.
## Project Objective
The objective of this project is to analyze the CMS dataset to gain insights into the travel behavior, preferences, and attitudes of New York City residents. Specifically, we aim to:
* Understand the relationship between residents' behaviors, preferences, attitudes, and traveling methods.
* Identify any trends or patterns in the data that may be relevant to transportation planning and policy decisions.
By achieving these objectives, we hope to contribute to the ongoing efforts to improve transportation in New York City and enhance the mobility of its residents. | 1.0 | Readme file - # The Citywide Mobility Survey (CMS)
The Citywide Mobility Survey (CMS) is a survey conducted by the New York City Department of Transportation (DOT) to gather information about the travel behavior, preferences, and attitudes of New York City residents. The survey is conducted periodically, and the data collected is used to inform transportation planning and policy decisions.
## Table of Contents
* [Data Objectives](#data-objectives)
* [Project Objectives](#project-objective)
## Data Objectives
The primary objectives of the CMS data collection are:
* Identify the factors and experiences that drive transportation choices for New York City residents.
* Understand current views on the state of transportation within the City.
* Measure attitudes toward current transportation issues and topics in New York City.
## Project Objective
The objective of this project is to analyze the CMS dataset to gain insights into the travel behavior, preferences, and attitudes of New York City residents. Specifically, we aim to:
* Understand the relationship between residents' behaviors, preferences, attitudes, and traveling methods.
* Identify any trends or patterns in the data that may be relevant to transportation planning and policy decisions.
By achieving these objectives, we hope to contribute to the ongoing efforts to improve transportation in New York City and enhance the mobility of its residents. | non_priority | readme file the citywide mobility survey cms the citywide mobility survey cms is a survey conducted by the new york city department of transportation dot to gather information about the travel behavior preferences and attitudes of new york city residents the survey is conducted periodically and the data collected is used to inform transportation planning and policy decisions table of contents data objectives project objective data objectives the primary objectives of the cms data collection are identify the factors and experiences that drive transportation choices for new york city residents understand current views on the state of transportation within the city measure attitudes toward current transportation issues and topics in new york city project objective the objective of this project is to analyze the cms dataset to gain insights into the travel behavior preferences and attitudes of new york city residents specifically we aim to understand the relationship between residents behaviors preferences attitudes and traveling methods identify any trends or patterns in the data that may be relevant to transportation planning and policy decisions by achieving these objectives we hope to contribute to the ongoing efforts to improve transportation in new york city and enhance the mobility of its residents | 0 |
281,762 | 24,416,981,670 | IssuesEvent | 2022-10-05 16:45:19 | dotnet/source-build | https://api.github.com/repos/dotnet/source-build | closed | Test template smoke test failures in CI | area-ci-testing untriaged | All test template smoke-test projects are failing in CI (https://dev.azure.com/dnceng/internal/_build/results?buildId=2008513&view=results). This is likely related to https://github.com/dotnet/installer/pull/14634.
```
System.InvalidOperationException : Failed to execute /tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/.dotnet/dotnet test /bl:/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/logs/BasicScenarioTests_NUnit_CSharp-test.binlog
Exit code: 1
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/.dotnet/sdk/7.0.100-rtm.22480.1/MSBuild.dll -nologo -maxcpucount -nodereuse:false -property:VSTestArtifactsProcessingMode=collect -property:VSTestSessionCorrelationId=2624_c11366ee-1afb-411f-8daa-e75f30eb0db9 -restore -target:VSTest -verbosity:m /bl:/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/logs/BasicScenarioTests_NUnit_CSharp-test.binlog ./BasicScenarioTests_NUnit_CSharp.csproj
Determining projects to restore...
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency coverlet.collector does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency Microsoft.NET.Test.Sdk does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit.Analyzers does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit3TestAdapter does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1701: Package 'NUnit 2.5.7.10213' was restored using '.NETFramework,Version=v4.6.1, .NETFramework,Version=v4.6.2, .NETFramework,Version=v4.7, .NETFramework,Version=v4.7.1, .NETFramework,Version=v4.7.2, .NETFramework,Version=v4.8, .NETFramework,Version=v4.8.1' instead of the project target framework 'net7.0'. This package may not be fully compatible with your project.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1701: Package 'NUnit3TestAdapter 3.0.10' was restored using '.NETFramework,Version=v4.6.1, .NETFramework,Version=v4.6.2, .NETFramework,Version=v4.7, .NETFramework,Version=v4.7.1, .NETFramework,Version=v4.7.2, .NETFramework,Version=v4.8, .NETFramework,Version=v4.8.1' instead of the project target framework 'net7.0'. This package may not be fully compatible with your project.
All projects are up-to-date for restore.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/.dotnet/sdk/7.0.100-rtm.22480.1/Sdks/Microsoft.NET.Sdk/targets/Microsoft.NET.RuntimeIdentifierInference.targets(257,5): message NETSDK1057: You are using a preview version of .NET. See: https://aka.ms/dotnet-support-policy [/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj]
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency coverlet.collector does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency Microsoft.NET.Test.Sdk does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit.Analyzers does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit3TestAdapter does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1701: Package 'NUnit 2.5.7.10213' was restored using '.NETFramework,Version=v4.6.1, .NETFramework,Version=v4.6.2, .NETFramework,Version=v4.7, .NETFramework,Version=v4.7.1, .NETFramework,Version=v4.7.2, .NETFramework,Version=v4.8, .NETFramework,Version=v4.8.1' instead of the project target framework 'net7.0'. This package may not be fully compatible with your project.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1701: Package 'NUnit3TestAdapter 3.0.10' was restored using '.NETFramework,Version=v4.6.1, .NETFramework,Version=v4.6.2, .NETFramework,Version=v4.7, .NETFramework,Version=v4.7.1, .NETFramework,Version=v4.7.2, .NETFramework,Version=v4.8, .NETFramework,Version=v4.8.1' instead of the project target framework 'net7.0'. This package may not be fully compatible with your project.
BasicScenarioTests_NUnit_CSharp -> /tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/bin/Debug/net7.0/BasicScenarioTests_NUnit_CSharp.dll
Test run for /tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/bin/Debug/net7.0/BasicScenarioTests_NUnit_CSharp.dll (.NETCoreApp,Version=v7.0)
Microsoft (R) Test Execution Command Line Tool Version 17.4.0+576a7cc1b259976393cf576a6a9c5df967bb2b7d (x64)
Copyright (c) Microsoft Corporation. All rights reserved.
Starting test execution, please wait...
A total of 1 test files matched the specified pattern.
Failed to negotiate protocol, waiting for response timed out after 90 seconds. This may occur due to machine slowness, please set environment variable VSTEST_CONNECTION_TIMEOUT to increase timeout.
Test Run Aborted.
```
This is blocking PR validation in dotnet/installer | 1.0 | Test template smoke test failures in CI - All test template smoke-test projects are failing in CI (https://dev.azure.com/dnceng/internal/_build/results?buildId=2008513&view=results). This is likely related to https://github.com/dotnet/installer/pull/14634.
```
System.InvalidOperationException : Failed to execute /tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/.dotnet/dotnet test /bl:/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/logs/BasicScenarioTests_NUnit_CSharp-test.binlog
Exit code: 1
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/.dotnet/sdk/7.0.100-rtm.22480.1/MSBuild.dll -nologo -maxcpucount -nodereuse:false -property:VSTestArtifactsProcessingMode=collect -property:VSTestSessionCorrelationId=2624_c11366ee-1afb-411f-8daa-e75f30eb0db9 -restore -target:VSTest -verbosity:m /bl:/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/logs/BasicScenarioTests_NUnit_CSharp-test.binlog ./BasicScenarioTests_NUnit_CSharp.csproj
Determining projects to restore...
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency coverlet.collector does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency Microsoft.NET.Test.Sdk does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit.Analyzers does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit3TestAdapter does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1701: Package 'NUnit 2.5.7.10213' was restored using '.NETFramework,Version=v4.6.1, .NETFramework,Version=v4.6.2, .NETFramework,Version=v4.7, .NETFramework,Version=v4.7.1, .NETFramework,Version=v4.7.2, .NETFramework,Version=v4.8, .NETFramework,Version=v4.8.1' instead of the project target framework 'net7.0'. This package may not be fully compatible with your project.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1701: Package 'NUnit3TestAdapter 3.0.10' was restored using '.NETFramework,Version=v4.6.1, .NETFramework,Version=v4.6.2, .NETFramework,Version=v4.7, .NETFramework,Version=v4.7.1, .NETFramework,Version=v4.7.2, .NETFramework,Version=v4.8, .NETFramework,Version=v4.8.1' instead of the project target framework 'net7.0'. This package may not be fully compatible with your project.
All projects are up-to-date for restore.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/.dotnet/sdk/7.0.100-rtm.22480.1/Sdks/Microsoft.NET.Sdk/targets/Microsoft.NET.RuntimeIdentifierInference.targets(257,5): message NETSDK1057: You are using a preview version of .NET. See: https://aka.ms/dotnet-support-policy [/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj]
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency coverlet.collector does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency Microsoft.NET.Test.Sdk does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit.Analyzers does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1604: Project dependency NUnit3TestAdapter does not contain an inclusive lower bound. Include a lower bound in the dependency version to ensure consistent restore results.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1701: Package 'NUnit 2.5.7.10213' was restored using '.NETFramework,Version=v4.6.1, .NETFramework,Version=v4.6.2, .NETFramework,Version=v4.7, .NETFramework,Version=v4.7.1, .NETFramework,Version=v4.7.2, .NETFramework,Version=v4.8, .NETFramework,Version=v4.8.1' instead of the project target framework 'net7.0'. This package may not be fully compatible with your project.
/tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/BasicScenarioTests_NUnit_CSharp.csproj : warning NU1701: Package 'NUnit3TestAdapter 3.0.10' was restored using '.NETFramework,Version=v4.6.1, .NETFramework,Version=v4.6.2, .NETFramework,Version=v4.7, .NETFramework,Version=v4.7.1, .NETFramework,Version=v4.7.2, .NETFramework,Version=v4.8, .NETFramework,Version=v4.8.1' instead of the project target framework 'net7.0'. This package may not be fully compatible with your project.
BasicScenarioTests_NUnit_CSharp -> /tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/bin/Debug/net7.0/BasicScenarioTests_NUnit_CSharp.dll
Test run for /tarball/test/Microsoft.DotNet.SourceBuild.SmokeTests/bin/Release/net7.0/projects-202209302226513478/BasicScenarioTests_NUnit_CSharp/bin/Debug/net7.0/BasicScenarioTests_NUnit_CSharp.dll (.NETCoreApp,Version=v7.0)
Microsoft (R) Test Execution Command Line Tool Version 17.4.0+576a7cc1b259976393cf576a6a9c5df967bb2b7d (x64)
Copyright (c) Microsoft Corporation. All rights reserved.
Starting test execution, please wait...
A total of 1 test files matched the specified pattern.
Failed to negotiate protocol, waiting for response timed out after 90 seconds. This may occur due to machine slowness, please set environment variable VSTEST_CONNECTION_TIMEOUT to increase timeout.
Test Run Aborted.
```
This is blocking PR validation in dotnet/installer | non_priority | test template smoke test failures in ci all test template smoke test projects are failing in ci this is likely related to system invalidoperationexception failed to execute tarball test microsoft dotnet sourcebuild smoketests bin release dotnet dotnet test bl tarball test microsoft dotnet sourcebuild smoketests bin release logs basicscenariotests nunit csharp test binlog exit code tarball test microsoft dotnet sourcebuild smoketests bin release dotnet sdk rtm msbuild dll nologo maxcpucount nodereuse false property vstestartifactsprocessingmode collect property vstestsessioncorrelationid restore target vstest verbosity m bl tarball test microsoft dotnet sourcebuild smoketests bin release logs basicscenariotests nunit csharp test binlog basicscenariotests nunit csharp csproj determining projects to restore tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency coverlet collector does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency microsoft net test sdk does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency nunit does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency nunit analyzers does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning package nunit was restored using netframework version netframework version netframework version netframework version netframework version netframework version netframework version instead of the project target framework this package may not be fully compatible with your project tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning package was restored using netframework version netframework version netframework version netframework version netframework version netframework version netframework version instead of the project target framework this package may not be fully compatible with your project all projects are up to date for restore tarball test microsoft dotnet sourcebuild smoketests bin release dotnet sdk rtm sdks microsoft net sdk targets microsoft net runtimeidentifierinference targets message you are using a preview version of net see tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency coverlet collector does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency microsoft net test sdk does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency nunit does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency nunit analyzers does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning project dependency does not contain an inclusive lower bound include a lower bound in the dependency version to ensure consistent restore results tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning package nunit was restored using netframework version netframework version netframework version netframework version netframework version netframework version netframework version instead of the project target framework this package may not be fully compatible with your project tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp basicscenariotests nunit csharp csproj warning package was restored using netframework version netframework version netframework version netframework version netframework version netframework version netframework version instead of the project target framework this package may not be fully compatible with your project basicscenariotests nunit csharp tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp bin debug basicscenariotests nunit csharp dll test run for tarball test microsoft dotnet sourcebuild smoketests bin release projects basicscenariotests nunit csharp bin debug basicscenariotests nunit csharp dll netcoreapp version microsoft r test execution command line tool version copyright c microsoft corporation all rights reserved starting test execution please wait a total of test files matched the specified pattern failed to negotiate protocol waiting for response timed out after seconds this may occur due to machine slowness please set environment variable vstest connection timeout to increase timeout test run aborted this is blocking pr validation in dotnet installer | 0 |
167,458 | 13,026,028,405 | IssuesEvent | 2020-07-27 14:24:51 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | Update rendering/localized_fonts golden images | a: internationalization a: tests a: text input framework | When #17700 is fixed, these golden files will need to be updated:
bin/cache/pkg/goldens/packages/flutter/test/rendering/localized_fonts.rich_text.styled_text_span.png
bin/cache/pkg/goldens/packages/flutter/test/rendering/localized_fonts.text_ambient_locale.chars.png
bin/cache/pkg/goldens/packages/flutter/test/rendering/localized_fonts.text_explicit_locale.chars.png
They were introduced in: https://github.com/flutter/flutter/pull/18222
Each of the golden images should display a pair of characters:

| 1.0 | Update rendering/localized_fonts golden images - When #17700 is fixed, these golden files will need to be updated:
bin/cache/pkg/goldens/packages/flutter/test/rendering/localized_fonts.rich_text.styled_text_span.png
bin/cache/pkg/goldens/packages/flutter/test/rendering/localized_fonts.text_ambient_locale.chars.png
bin/cache/pkg/goldens/packages/flutter/test/rendering/localized_fonts.text_explicit_locale.chars.png
They were introduced in: https://github.com/flutter/flutter/pull/18222
Each of the golden images should display a pair of characters:

| non_priority | update rendering localized fonts golden images when is fixed these golden files will need to be updated bin cache pkg goldens packages flutter test rendering localized fonts rich text styled text span png bin cache pkg goldens packages flutter test rendering localized fonts text ambient locale chars png bin cache pkg goldens packages flutter test rendering localized fonts text explicit locale chars png they were introduced in each of the golden images should display a pair of characters | 0 |
217,806 | 24,351,601,595 | IssuesEvent | 2022-10-03 01:00:39 | billmcchesney1/hadoop | https://api.github.com/repos/billmcchesney1/hadoop | opened | CVE-2022-42004 (Medium) detected in jackson-databind-2.9.10.1.jar | security vulnerability | ## CVE-2022-42004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib/jackson-databind-2.9.10.1.jar,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/target/lib/jackson-databind-2.9.10.1.jar,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client/target/lib/jackson-databind-2.9.10.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p>
<p>Found in base branch: <b>trunk</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization.
<p>Publish Date: 2022-10-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p>
</p>
</details>
<p></p>
| True | CVE-2022-42004 (Medium) detected in jackson-databind-2.9.10.1.jar - ## CVE-2022-42004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: /hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-documentstore/target/lib/jackson-databind-2.9.10.1.jar,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-common/target/lib/jackson-databind-2.9.10.1.jar,/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client/target/lib/jackson-databind-2.9.10.1.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p>
<p>Found in base branch: <b>trunk</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In FasterXML jackson-databind before 2.13.4, resource exhaustion can occur because of a lack of a check in BeanDeserializer._deserializeFromArray to prevent use of deeply nested arrays. An application is vulnerable only with certain customized choices for deserialization.
<p>Publish Date: 2022-10-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42004>CVE-2022-42004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.13.4</p>
</p>
</details>
<p></p>
| non_priority | cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice documentstore target lib jackson databind jar hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice hbase hadoop yarn server timelineservice hbase common target lib jackson databind jar hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server timelineservice hbase hadoop yarn server timelineservice hbase client target lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch trunk vulnerability details in fasterxml jackson databind before resource exhaustion can occur because of a lack of a check in beandeserializer deserializefromarray to prevent use of deeply nested arrays an application is vulnerable only with certain customized choices for deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind | 0 |
123,831 | 4,876,737,864 | IssuesEvent | 2016-11-16 13:49:10 | odalic/odalic-ui | https://api.github.com/repos/odalic/odalic-ui | opened | No processing results display for the first result paged until i hit Next and then Previous to get back | priority: High | This needs to be fixed. | 1.0 | No processing results display for the first result paged until i hit Next and then Previous to get back - This needs to be fixed. | priority | no processing results display for the first result paged until i hit next and then previous to get back this needs to be fixed | 1 |
28,607 | 2,708,056,175 | IssuesEvent | 2015-04-08 05:26:51 | leo-project/leofs | https://api.github.com/repos/leo-project/leofs | closed | leo_gateway can't stop (kill -0) properly on OSX. | Priority-LOW _leo_gateway | leo_gateway can't stop properly on OSX.
This is used ``kill -0`` command, but it does not work on OSX.
leo_gateway/bin/leo_gateway line 157.
```bash
134 stop)
135 # Wait for the node to completely stop...
136 case `uname -s` in
137 Linux|Darwin|FreeBSD|DragonFly|NetBSD|OpenBSD)
138 # PID COMMAND
139 PID=`ps ax -o pid= -o command=|\
140 grep "$RUNNER_BASE_DIR/.*/[b]eam"|awk '{print $1}'`
141 ;;
142 SunOS)
143 # PID COMMAND
144 PID=`ps -ef -o pid= -o args=|\
145 grep "$RUNNER_BASE_DIR/.*/[b]eam"|awk '{print $1}'`
146 ;;
147 CYGWIN*)
148 # UID PID PPID TTY STIME COMMAND
149 PID=`ps -efW|grep "$RUNNER_BASE_DIR/.*/[b]eam"|awk '{print $2}'`
150 ;;
151 esac
152 $NODETOOL stop
153 ES=$?
154 if [ "$ES" -ne 0 ]; then
155 exit $ES
156 fi
157 while `kill -0 $PID 2>/dev/null`;
158 do
159 sleep 1
160 done
161 ;;
```
execute result.
```
++ ps ax -o pid= -o command=
++ grep '/tmp/leofs-1.2.7/leo_gateway/.*/[b]eam'
++ awk '{print $1}'
+ PID=64355
+ /tmp/leofs-1.2.7/leo_gateway/erts-6.3/bin/escript /tmp/leofs-1.2.7/leo_gateway/erts-6.3/bin/nodetool -name gateway_0@127.0.0.1 -setcookie 401321b4 stop
ok
+ ES=0
+ '[' 0 -ne 0 ']'
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
```
test program.
```
cat /tmp/a
while [ 1 ] ; do
sleep 30
done
% ps auxww | grep '\/tmp\/a'
hiroysato 73941 0.0 0.0 2435416 992 s005 S+ 9:00AM 0:00.01 sh /tmp/a
% kill -0 73941
% ps auxww | grep '\/tmp\/a'
hiroysato 73941 0.0 0.0 2435416 992 s005 S+ 9:00AM 0:00.01 sh /tmp/a
```
| 1.0 | leo_gateway can't stop (kill -0) properly on OSX. - leo_gateway can't stop properly on OSX.
This is used ``kill -0`` command, but it does not work on OSX.
leo_gateway/bin/leo_gateway line 157.
```bash
134 stop)
135 # Wait for the node to completely stop...
136 case `uname -s` in
137 Linux|Darwin|FreeBSD|DragonFly|NetBSD|OpenBSD)
138 # PID COMMAND
139 PID=`ps ax -o pid= -o command=|\
140 grep "$RUNNER_BASE_DIR/.*/[b]eam"|awk '{print $1}'`
141 ;;
142 SunOS)
143 # PID COMMAND
144 PID=`ps -ef -o pid= -o args=|\
145 grep "$RUNNER_BASE_DIR/.*/[b]eam"|awk '{print $1}'`
146 ;;
147 CYGWIN*)
148 # UID PID PPID TTY STIME COMMAND
149 PID=`ps -efW|grep "$RUNNER_BASE_DIR/.*/[b]eam"|awk '{print $2}'`
150 ;;
151 esac
152 $NODETOOL stop
153 ES=$?
154 if [ "$ES" -ne 0 ]; then
155 exit $ES
156 fi
157 while `kill -0 $PID 2>/dev/null`;
158 do
159 sleep 1
160 done
161 ;;
```
execute result.
```
++ ps ax -o pid= -o command=
++ grep '/tmp/leofs-1.2.7/leo_gateway/.*/[b]eam'
++ awk '{print $1}'
+ PID=64355
+ /tmp/leofs-1.2.7/leo_gateway/erts-6.3/bin/escript /tmp/leofs-1.2.7/leo_gateway/erts-6.3/bin/nodetool -name gateway_0@127.0.0.1 -setcookie 401321b4 stop
ok
+ ES=0
+ '[' 0 -ne 0 ']'
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
++ kill -0 64355
+ sleep 1
```
test program.
```
cat /tmp/a
while [ 1 ] ; do
sleep 30
done
% ps auxww | grep '\/tmp\/a'
hiroysato 73941 0.0 0.0 2435416 992 s005 S+ 9:00AM 0:00.01 sh /tmp/a
% kill -0 73941
% ps auxww | grep '\/tmp\/a'
hiroysato 73941 0.0 0.0 2435416 992 s005 S+ 9:00AM 0:00.01 sh /tmp/a
```
| priority | leo gateway can t stop kill properly on osx leo gateway can t stop properly on osx this is used kill command but it does not work on osx leo gateway bin leo gateway line bash stop wait for the node to completely stop case uname s in linux darwin freebsd dragonfly netbsd openbsd pid command pid ps ax o pid o command grep runner base dir eam awk print sunos pid command pid ps ef o pid o args grep runner base dir eam awk print cygwin uid pid ppid tty stime command pid ps efw grep runner base dir eam awk print esac nodetool stop es if then exit es fi while kill pid dev null do sleep done execute result ps ax o pid o command grep tmp leofs leo gateway eam awk print pid tmp leofs leo gateway erts bin escript tmp leofs leo gateway erts bin nodetool name gateway setcookie stop ok es kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep kill sleep test program cat tmp a while do sleep done ps auxww grep tmp a hiroysato s sh tmp a kill ps auxww grep tmp a hiroysato s sh tmp a | 1 |
186,267 | 21,920,324,108 | IssuesEvent | 2022-05-22 13:21:54 | rengert/bolzplatzarena.blog | https://api.github.com/repos/rengert/bolzplatzarena.blog | reopened | CVE-2018-8292 (High) detected in system.net.http.4.3.0.nupkg | security vulnerability | ## CVE-2018-8292 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /bolzplatzarena.blog.csproj</p>
<p>Path to vulnerable library: /usr/share/dotnet/sdk/NuGetFallbackFolder/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- piranha.aspnetcore.9.2.0.nupkg (Root Library)
- xsitemap.2.0.7.nupkg
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rengert/bolzplatzarena.blog/commit/dfdf3c8036bf586f4825d703adbbb7ae84375391">dfdf3c8036bf586f4825d703adbbb7ae84375391</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/dotnet/announcements/issues/88">https://github.com/dotnet/announcements/issues/88</a></p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-8292 (High) detected in system.net.http.4.3.0.nupkg - ## CVE-2018-8292 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /bolzplatzarena.blog.csproj</p>
<p>Path to vulnerable library: /usr/share/dotnet/sdk/NuGetFallbackFolder/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- piranha.aspnetcore.9.2.0.nupkg (Root Library)
- xsitemap.2.0.7.nupkg
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rengert/bolzplatzarena.blog/commit/dfdf3c8036bf586f4825d703adbbb7ae84375391">dfdf3c8036bf586f4825d703adbbb7ae84375391</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An information disclosure vulnerability exists in .NET Core when authentication information is inadvertently exposed in a redirect, aka ".NET Core Information Disclosure Vulnerability." This affects .NET Core 2.1, .NET Core 1.0, .NET Core 1.1, PowerShell Core 6.0.
<p>Publish Date: 2018-10-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-8292>CVE-2018-8292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/dotnet/announcements/issues/88">https://github.com/dotnet/announcements/issues/88</a></p>
<p>Release Date: 2018-10-10</p>
<p>Fix Resolution: System.Net.Http - 4.3.4;Microsoft.PowerShell.Commands.Utility - 6.1.0-rc.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in system net http nupkg cve high severity vulnerability vulnerable library system net http nupkg provides a programming interface for modern http applications including http client components that library home page a href path to dependency file bolzplatzarena blog csproj path to vulnerable library usr share dotnet sdk nugetfallbackfolder system net http system net http nupkg dependency hierarchy piranha aspnetcore nupkg root library xsitemap nupkg netstandard library nupkg x system net http nupkg vulnerable library found in head commit a href found in base branch develop vulnerability details an information disclosure vulnerability exists in net core when authentication information is inadvertently exposed in a redirect aka net core information disclosure vulnerability this affects net core net core net core powershell core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system net http microsoft powershell commands utility rc step up your open source security game with whitesource | 0 |
8,150 | 2,869,058,542 | IssuesEvent | 2015-06-05 22:59:55 | dart-lang/mock | https://api.github.com/repos/dart-lang/mock | closed | Mocks Package is not friendly to use | AsDesigned bug | <a href="https://github.com/TedSander"><img src="https://avatars.githubusercontent.com/u/5643740?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [TedSander](https://github.com/TedSander)**
_Originally opened as dart-lang/sdk#16147_
----
The mocks package is not user friendly. It can't be used with code completion, and refactoring is difficult with it.
An interface much like mockito would be ideal. As it stands I tend to avoid mocking in dart because the mocking package is not easy to use. | 1.0 | Mocks Package is not friendly to use - <a href="https://github.com/TedSander"><img src="https://avatars.githubusercontent.com/u/5643740?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [TedSander](https://github.com/TedSander)**
_Originally opened as dart-lang/sdk#16147_
----
The mocks package is not user friendly. It can't be used with code completion, and refactoring is difficult with it.
An interface much like mockito would be ideal. As it stands I tend to avoid mocking in dart because the mocking package is not easy to use. | non_priority | mocks package is not friendly to use issue by originally opened as dart lang sdk the mocks package is not user friendly it can t be used with code completion and refactoring is difficult with it an interface much like mockito would be ideal as it stands i tend to avoid mocking in dart because the mocking package is not easy to use | 0 |
105,129 | 22,916,206,009 | IssuesEvent | 2022-07-17 01:56:52 | Arcanorum/rogueworld | https://api.github.com/repos/Arcanorum/rogueworld | closed | Damage particles triggered when a creature sprite goes off screen | bug good first issue code | When a creature goes outside of the player view bounds, it is destroyed, which is also causing the death effect of spraying a bunch of damage particles from the point the entity was destroyed.
Need some check in the Character destroy event for if the entity was actually killed or not.
Hint:
https://github.com/Arcanorum/dungeonz/blob/master/client/src/entities/characters/Character.js#L56 | 1.0 | Damage particles triggered when a creature sprite goes off screen - When a creature goes outside of the player view bounds, it is destroyed, which is also causing the death effect of spraying a bunch of damage particles from the point the entity was destroyed.
Need some check in the Character destroy event for if the entity was actually killed or not.
Hint:
https://github.com/Arcanorum/dungeonz/blob/master/client/src/entities/characters/Character.js#L56 | non_priority | damage particles triggered when a creature sprite goes off screen when a creature goes outside of the player view bounds it is destroyed which is also causing the death effect of spraying a bunch of damage particles from the point the entity was destroyed need some check in the character destroy event for if the entity was actually killed or not hint | 0 |
489,109 | 14,101,302,735 | IssuesEvent | 2020-11-06 06:32:24 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | open.spotify.com - design is broken | browser-fixme ml-needsdiagnosis-false ml-probability-high priority-critical | <!-- @browser: Dragon 65.0.2 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0 IceDragon/65.0.2 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/61193 -->
**URL**: https://open.spotify.com/
**Browser / Version**: Dragon 65.0.2
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/11/5d653da5-d47d-44a1-acee-637b9bf00f01.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190318120942</li><li>channel: default</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/11/07d2dfd6-96ae-470a-a0c4-dc28835cef90)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | open.spotify.com - design is broken - <!-- @browser: Dragon 65.0.2 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:65.0) Gecko/20100101 Firefox/65.0 IceDragon/65.0.2 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/61193 -->
**URL**: https://open.spotify.com/
**Browser / Version**: Dragon 65.0.2
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Design is broken
**Description**: Images not loaded
**Steps to Reproduce**:
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/11/5d653da5-d47d-44a1-acee-637b9bf00f01.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190318120942</li><li>channel: default</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/11/07d2dfd6-96ae-470a-a0c4-dc28835cef90)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | open spotify com design is broken url browser version dragon operating system windows tested another browser yes edge problem type design is broken description images not loaded steps to reproduce view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
372,192 | 11,010,776,542 | IssuesEvent | 2019-12-04 15:10:14 | geosolutions-it/geotools | https://api.github.com/repos/geosolutions-it/geotools | closed | Parse user provided mapping and store it in an appropriate structure | C009-2016-MONGODB Priority: Medium enhancement in progress | Using the user provided schema and provided mapping we should store that information in a way that can be easealy used to build a complex feature type, app-schema already existing support for this should be used (including their defined DSL).
| 1.0 | Parse user provided mapping and store it in an appropriate structure - Using the user provided schema and provided mapping we should store that information in a way that can be easealy used to build a complex feature type, app-schema already existing support for this should be used (including their defined DSL).
| priority | parse user provided mapping and store it in an appropriate structure using the user provided schema and provided mapping we should store that information in a way that can be easealy used to build a complex feature type app schema already existing support for this should be used including their defined dsl | 1 |
620,532 | 19,564,321,929 | IssuesEvent | 2022-01-03 21:06:42 | stake-house/eth-wizard | https://api.github.com/repos/stake-house/eth-wizard | closed | Uncaught exception with Teku on Windows | bug priority - high Windows 10 | Handle this better with eth-wizard 0.7.3.
```
2022-01-03 13:18:13,924 - ethwizard.platforms.windows10 - CRITICAL - Uncaught exception
Traceback (most recent call last):
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_transports\default.py", line 62, in map_httpcore_exceptions
yield
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_transports\default.py", line 177, in handle_request
status_code, headers, byte_stream, extensions = self._pool.handle_request(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_sync\connection_pool.py", line 234, in handle_request
response = connection.handle_request(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_sync\connection.py", line 136, in handle_request
self.socket = self._open_socket(timeout)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_sync\connection.py", line 163, in _open_socket
return self._backend.open_tcp_stream(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_backends\sync.py", line 144, in open_tcp_stream
return SyncSocketStream(sock=sock)
File "contextlib.py", line 137, in __exit__
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_exceptions.py", line 12, in map_exceptions
raise to_exc(exc) from None
httpcore.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\windows10.py", line 2364, in install_teku
response = httpx.get(teku_query_url, headers=headers)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_api.py", line 189, in get
return request(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_api.py", line 100, in request
return client.request(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 792, in request
return self.send(request, auth=auth, follow_redirects=follow_redirects)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 877, in send
response = self._send_handling_auth(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 905, in _send_handling_auth
response = self._send_handling_redirects(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 942, in _send_handling_redirects
response = self._send_single_request(request)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 978, in _send_single_request
response = transport.handle_request(request)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_transports\default.py", line 177, in handle_request
status_code, headers, byte_stream, extensions = self._pool.handle_request(
File "contextlib.py", line 137, in __exit__
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_transports\default.py", line 79, in map_httpcore_exceptions
raise mapped_exc(message) from exc
httpx.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "runpy.py", line 197, in _run_module_as_main
File "runpy.py", line 87, in _run_code
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\__main__.py", line 4, in <module>
wizard.run()
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\wizard.py", line 88, in run
sequence.run_from_start()
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\common.py", line 69, in run_from_start
return self._run_from_index(0, context)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\common.py", line 119, in _run_from_index
context = current_step.exc_function(current_step, context, self)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\windows10.py", line 263, in install_teku_function
if not install_teku(context[selected_directory], context[selected_network],
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\windows10.py", line 2371, in install_teku
log_file.seek(-1024, 2)
io.UnsupportedOperation: can't do nonzero end-relative seeks
``` | 1.0 | Uncaught exception with Teku on Windows - Handle this better with eth-wizard 0.7.3.
```
2022-01-03 13:18:13,924 - ethwizard.platforms.windows10 - CRITICAL - Uncaught exception
Traceback (most recent call last):
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_transports\default.py", line 62, in map_httpcore_exceptions
yield
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_transports\default.py", line 177, in handle_request
status_code, headers, byte_stream, extensions = self._pool.handle_request(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_sync\connection_pool.py", line 234, in handle_request
response = connection.handle_request(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_sync\connection.py", line 136, in handle_request
self.socket = self._open_socket(timeout)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_sync\connection.py", line 163, in _open_socket
return self._backend.open_tcp_stream(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_backends\sync.py", line 144, in open_tcp_stream
return SyncSocketStream(sock=sock)
File "contextlib.py", line 137, in __exit__
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpcore\_exceptions.py", line 12, in map_exceptions
raise to_exc(exc) from None
httpcore.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\windows10.py", line 2364, in install_teku
response = httpx.get(teku_query_url, headers=headers)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_api.py", line 189, in get
return request(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_api.py", line 100, in request
return client.request(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 792, in request
return self.send(request, auth=auth, follow_redirects=follow_redirects)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 877, in send
response = self._send_handling_auth(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 905, in _send_handling_auth
response = self._send_handling_redirects(
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 942, in _send_handling_redirects
response = self._send_single_request(request)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_client.py", line 978, in _send_single_request
response = transport.handle_request(request)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_transports\default.py", line 177, in handle_request
status_code, headers, byte_stream, extensions = self._pool.handle_request(
File "contextlib.py", line 137, in __exit__
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\httpx\_transports\default.py", line 79, in map_httpcore_exceptions
raise mapped_exc(message) from exc
httpx.ConnectError: [WinError 10061] No connection could be made because the target machine actively refused it
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "runpy.py", line 197, in _run_module_as_main
File "runpy.py", line 87, in _run_code
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\__main__.py", line 4, in <module>
wizard.run()
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\wizard.py", line 88, in run
sequence.run_from_start()
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\common.py", line 69, in run_from_start
return self._run_from_index(0, context)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\common.py", line 119, in _run_from_index
context = current_step.exc_function(current_step, context, self)
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\windows10.py", line 263, in install_teku_function
if not install_teku(context[selected_directory], context[selected_network],
File "C:\Users\Test\AppData\Local\Temp\7zS805612E6\ethwizard\platforms\windows10.py", line 2371, in install_teku
log_file.seek(-1024, 2)
io.UnsupportedOperation: can't do nonzero end-relative seeks
``` | priority | uncaught exception with teku on windows handle this better with eth wizard ethwizard platforms critical uncaught exception traceback most recent call last file c users test appdata local temp httpx transports default py line in map httpcore exceptions yield file c users test appdata local temp httpx transports default py line in handle request status code headers byte stream extensions self pool handle request file c users test appdata local temp httpcore sync connection pool py line in handle request response connection handle request file c users test appdata local temp httpcore sync connection py line in handle request self socket self open socket timeout file c users test appdata local temp httpcore sync connection py line in open socket return self backend open tcp stream file c users test appdata local temp httpcore backends sync py line in open tcp stream return syncsocketstream sock sock file contextlib py line in exit file c users test appdata local temp httpcore exceptions py line in map exceptions raise to exc exc from none httpcore connecterror no connection could be made because the target machine actively refused it the above exception was the direct cause of the following exception traceback most recent call last file c users test appdata local temp ethwizard platforms py line in install teku response httpx get teku query url headers headers file c users test appdata local temp httpx api py line in get return request file c users test appdata local temp httpx api py line in request return client request file c users test appdata local temp httpx client py line in request return self send request auth auth follow redirects follow redirects file c users test appdata local temp httpx client py line in send response self send handling auth file c users test appdata local temp httpx client py line in send handling auth response self send handling redirects file c users test appdata local temp httpx client py line in send handling redirects response self send single request request file c users test appdata local temp httpx client py line in send single request response transport handle request request file c users test appdata local temp httpx transports default py line in handle request status code headers byte stream extensions self pool handle request file contextlib py line in exit file c users test appdata local temp httpx transports default py line in map httpcore exceptions raise mapped exc message from exc httpx connecterror no connection could be made because the target machine actively refused it during handling of the above exception another exception occurred traceback most recent call last file runpy py line in run module as main file runpy py line in run code file c users test appdata local temp ethwizard main py line in wizard run file c users test appdata local temp ethwizard wizard py line in run sequence run from start file c users test appdata local temp ethwizard platforms common py line in run from start return self run from index context file c users test appdata local temp ethwizard platforms common py line in run from index context current step exc function current step context self file c users test appdata local temp ethwizard platforms py line in install teku function if not install teku context context file c users test appdata local temp ethwizard platforms py line in install teku log file seek io unsupportedoperation can t do nonzero end relative seeks | 1 |
188,703 | 15,168,151,276 | IssuesEvent | 2021-02-12 18:56:06 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | opened | [Github] Request for changes to /products/identity-personalization/direct-deposit folder | direct deposit documentation-support vsa-authenticated-exp | ## Background
We need some restructuring in the `/products/identity-personalization/direct-deposit` folder in order to better account for different versions of direct deposit and to improve for overall organization.
## Tasks
Please help us with the following:
- [ ] Under `/products/identity-personalization/direct-deposit/design` please move all files that are currently there under a folder called profile-1.0, so `/products/identity-personalization/direct-deposit/profile-1.0`.
- [ ] Please move `/products/identity-personalization/direct-deposit/error-handling` under `/products/identity-personalization/direct-deposit/design`. | 1.0 | [Github] Request for changes to /products/identity-personalization/direct-deposit folder - ## Background
We need some restructuring in the `/products/identity-personalization/direct-deposit` folder in order to better account for different versions of direct deposit and to improve for overall organization.
## Tasks
Please help us with the following:
- [ ] Under `/products/identity-personalization/direct-deposit/design` please move all files that are currently there under a folder called profile-1.0, so `/products/identity-personalization/direct-deposit/profile-1.0`.
- [ ] Please move `/products/identity-personalization/direct-deposit/error-handling` under `/products/identity-personalization/direct-deposit/design`. | non_priority | request for changes to products identity personalization direct deposit folder background we need some restructuring in the products identity personalization direct deposit folder in order to better account for different versions of direct deposit and to improve for overall organization tasks please help us with the following under products identity personalization direct deposit design please move all files that are currently there under a folder called profile so products identity personalization direct deposit profile please move products identity personalization direct deposit error handling under products identity personalization direct deposit design | 0 |
297,771 | 9,181,501,010 | IssuesEvent | 2019-03-05 10:22:41 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | CONFIG_BT_HCI_TX_STACK_SIZE is too small | area: Bluetooth bug priority: low | **Describe the bug**
I have pca10059 configured as central (running ncs nrf_desktop). The device crashes during connection attempt with a peripheral.
`CONFIG_BT_HCI_TX_STACK_SIZE` is set to 640B in Zephyr. If extend it to 1024B program works (I have not tried lower values yet).
We have not seen it in the past. The thing is that due to recent bugs we have MPU stack guard enabled by default now.
**To Reproduce**
Steps to reproduce the behavior:
Build nrf_desktop for pca10059 and try to connect to peripheral (DK working as mouse from the same project should work).
Try to connect and observe crash.
**Expected behavior**
Should not crash.
**Impact**
Showstopper.
**Screenshots or console output**
```
[00000204] <inf> ble_scan: Scan started
--
[00000213] <inf> event_manager: e: usb_state_event id:0x200000a9 state:POWERED
[00008655] <inf> ble_scan: Filters matched. fd:a0:61:de:7d:07 (random) connectable
[00008679] <inf> ble_scan: Connecting done
[00009066] <inf> usb_state: USB HID enabled
[00009069] <inf> event_manager: e: usb_state_event id:0x200000a9 state:ACTIVE
[00009072] <inf> event_manager: e: hid_report_subscription_event MOUSE report notification enabled by 0x200000a9
[00009944] <wrn> usb_hid: len 143 doesn't match Report Descriptor size
[00012038] <inf> ble_state: Connected to fd:a0:61:de:7d:07 (random)
[00012044] <wrn> bt_ctlr_hci: handle: 0x0000, CSA: 1.
[00039955] <inf> ble_state: Security with fd:a0:61:de:7d:07 (random) level 2
[00039959] <inf> event_manager: e: ble_peer_event id=0x2000227c SECURED
***** MPU FAULT *****
Stacking error
Data Access Violation
MMFAR Address: 0x20006e98
***** Hardware exception *****
Current thread ID = 0x200021f0
Faulting instruction address = 0x2aa7c
Fatal fault in thread 0x200021f0! Aborting.
[ASSERTION FAIL [out_ctx->control_block->offset <= out_ctx->size] @ /home/pdunaj/work/ncs/zephyr/subsys/logging/log_output.c:101
```
**Environment (please complete the following information):**
rev: a76f833d2475af488fd86583142a79580e986a4c (ncs) from ec424b7a455cc98c36e3cab80be7e152ef2c5217 (zephyr)
**Additional context**
N/A
| 1.0 | CONFIG_BT_HCI_TX_STACK_SIZE is too small - **Describe the bug**
I have pca10059 configured as central (running ncs nrf_desktop). The device crashes during connection attempt with a peripheral.
`CONFIG_BT_HCI_TX_STACK_SIZE` is set to 640B in Zephyr. If extend it to 1024B program works (I have not tried lower values yet).
We have not seen it in the past. The thing is that due to recent bugs we have MPU stack guard enabled by default now.
**To Reproduce**
Steps to reproduce the behavior:
Build nrf_desktop for pca10059 and try to connect to peripheral (DK working as mouse from the same project should work).
Try to connect and observe crash.
**Expected behavior**
Should not crash.
**Impact**
Showstopper.
**Screenshots or console output**
```
[00000204] <inf> ble_scan: Scan started
--
[00000213] <inf> event_manager: e: usb_state_event id:0x200000a9 state:POWERED
[00008655] <inf> ble_scan: Filters matched. fd:a0:61:de:7d:07 (random) connectable
[00008679] <inf> ble_scan: Connecting done
[00009066] <inf> usb_state: USB HID enabled
[00009069] <inf> event_manager: e: usb_state_event id:0x200000a9 state:ACTIVE
[00009072] <inf> event_manager: e: hid_report_subscription_event MOUSE report notification enabled by 0x200000a9
[00009944] <wrn> usb_hid: len 143 doesn't match Report Descriptor size
[00012038] <inf> ble_state: Connected to fd:a0:61:de:7d:07 (random)
[00012044] <wrn> bt_ctlr_hci: handle: 0x0000, CSA: 1.
[00039955] <inf> ble_state: Security with fd:a0:61:de:7d:07 (random) level 2
[00039959] <inf> event_manager: e: ble_peer_event id=0x2000227c SECURED
***** MPU FAULT *****
Stacking error
Data Access Violation
MMFAR Address: 0x20006e98
***** Hardware exception *****
Current thread ID = 0x200021f0
Faulting instruction address = 0x2aa7c
Fatal fault in thread 0x200021f0! Aborting.
[ASSERTION FAIL [out_ctx->control_block->offset <= out_ctx->size] @ /home/pdunaj/work/ncs/zephyr/subsys/logging/log_output.c:101
```
**Environment (please complete the following information):**
rev: a76f833d2475af488fd86583142a79580e986a4c (ncs) from ec424b7a455cc98c36e3cab80be7e152ef2c5217 (zephyr)
**Additional context**
N/A
| priority | config bt hci tx stack size is too small describe the bug i have configured as central running ncs nrf desktop the device crashes during connection attempt with a peripheral config bt hci tx stack size is set to in zephyr if extend it to program works i have not tried lower values yet we have not seen it in the past the thing is that due to recent bugs we have mpu stack guard enabled by default now to reproduce steps to reproduce the behavior build nrf desktop for and try to connect to peripheral dk working as mouse from the same project should work try to connect and observe crash expected behavior should not crash impact showstopper screenshots or console output ble scan scan started event manager e usb state event id state powered ble scan filters matched fd de random connectable ble scan connecting done usb state usb hid enabled event manager e usb state event id state active event manager e hid report subscription event mouse report notification enabled by usb hid len doesn t match report descriptor size ble state connected to fd de random bt ctlr hci handle csa ble state security with fd de random level event manager e ble peer event id secured mpu fault stacking error data access violation mmfar address hardware exception current thread id faulting instruction address fatal fault in thread aborting home pdunaj work ncs zephyr subsys logging log output c environment please complete the following information rev ncs from zephyr additional context n a | 1 |
392,274 | 11,589,334,328 | IssuesEvent | 2020-02-24 01:38:44 | omou-org/mainframe | https://api.github.com/repos/omou-org/mainframe | closed | POST Instructor Availability Not Working | 3 hours bug priority | When we post an instructor availability, the response we get is "null" for all the times
<img width="941" alt="Screen Shot 2020-01-16 at 10 55 39 PM" src="https://user-images.githubusercontent.com/12959959/72590756-5daad680-38b3-11ea-8fc7-cf960847d88e.png">
| 1.0 | POST Instructor Availability Not Working - When we post an instructor availability, the response we get is "null" for all the times
<img width="941" alt="Screen Shot 2020-01-16 at 10 55 39 PM" src="https://user-images.githubusercontent.com/12959959/72590756-5daad680-38b3-11ea-8fc7-cf960847d88e.png">
| priority | post instructor availability not working when we post an instructor availability the response we get is null for all the times img width alt screen shot at pm src | 1 |
500,667 | 14,503,816,856 | IssuesEvent | 2020-12-11 23:30:04 | googleapis/nodejs-recommender | https://api.github.com/repos/googleapis/nodejs-recommender | closed | Synthesis failed for nodejs-recommender | api: recommender autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate nodejs-recommender. :broken_heart:
Here's the output from running `synth.py`:
```
e7e855aed5a29e6e83c39beba2a
2020-12-11 21:37:59,042 autosynth [DEBUG] > Running: git log -1 --pretty=%at c2de32114ec484aa708d32012d1fa8d75232daf5
2020-12-11 21:37:59,045 autosynth [DEBUG] > Running: git log -1 --pretty=%at 32af6da519a6b042e3da62008e2a75e991efb6b4
2020-12-11 21:37:59,048 autosynth [DEBUG] > Running: git log -1 --pretty=%at 69629b64b83c6421d616be2b8e11795738ec8a6c
2020-12-11 21:37:59,051 autosynth [DEBUG] > Running: git log -1 --pretty=%at f94318521f63085b9ccb43d42af89f153fb39f15
2020-12-11 21:37:59,054 autosynth [DEBUG] > Running: git log -1 --pretty=%at 18c5dbdb4ac8cf75d4d8174e7b4558f48e76f8a1
2020-12-11 21:37:59,057 autosynth [DEBUG] > Running: git log -1 --pretty=%at 5d11bd2888c38ce1fb6fa6bf25494a4219a73928
2020-12-11 21:37:59,060 autosynth [DEBUG] > Running: git log -1 --pretty=%at 3559dd8cb634347737632776ef4089443eacdb4f
2020-12-11 21:37:59,062 autosynth [DEBUG] > Running: git log -1 --pretty=%at 7d1ee3207371ef72a1ca931a0964c4007e687280
2020-12-11 21:37:59,065 autosynth [DEBUG] > Running: git log -1 --pretty=%at 352f25621c28d2fa0784984ea0ed5ea697aa1525
2020-12-11 21:37:59,068 autosynth [DEBUG] > Running: git log -1 --pretty=%at 996775eca5fd934edac3c2ae34b80ff0395b1717
2020-12-11 21:37:59,071 autosynth [DEBUG] > Running: git checkout e06347ac86941705801aad1e3c3b6b87092aa535
Note: checking out 'e06347ac86941705801aad1e3c3b6b87092aa535'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at e06347a chore: generate GAPIC metadata JSON file (#136)
2020-12-11 21:37:59,086 autosynth [DEBUG] > Running: git checkout 996775eca5fd934edac3c2ae34b80ff0395b1717
Note: checking out '996775eca5fd934edac3c2ae34b80ff0395b1717'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 996775e fix(python): skip samples readmegen in synthtool (#875)
2020-12-11 21:37:59,092 autosynth [DEBUG] > Running: git checkout dd372aa22ded7a8ba6f0e03a80e06358a3fa0907
Note: checking out 'dd372aa22ded7a8ba6f0e03a80e06358a3fa0907'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at dd372aa2 fix: remove client recv msg limit fix: add enums to `types/__init__.py`
2020-12-11 21:37:59,108 autosynth [DEBUG] > Running: git branch -f autosynth-60
2020-12-11 21:37:59,112 autosynth [DEBUG] > Running: git checkout autosynth-60
Switched to branch 'autosynth-60'
2020-12-11 21:37:59,116 autosynth [INFO] > Running synthtool
2020-12-11 21:37:59,116 autosynth [INFO] > ['/usr/local/bin/python', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-12-11 21:37:59,116 autosynth [DEBUG] > log_file_path: /kokoro/artifacts/logs/nodejs-recommender/60/sponge_log.log
2020-12-11 21:37:59,118 autosynth [DEBUG] > Running: /usr/local/bin/python -m synthtool --metadata synth.metadata synth.py --
2020-12-11 21:37:59,325 synthtool [DEBUG] > Executing /root/.cache/synthtool/nodejs-recommender/synth.py.
On branch autosynth-60
nothing to commit, working tree clean
2020-12-11 21:37:59,455 synthtool [DEBUG] > Ensuring dependencies.
DEBUG:synthtool:Ensuring dependencies.
2020-12-11 21:37:59,464 synthtool [DEBUG] > Cloning googleapis.
DEBUG:synthtool:Cloning googleapis.
2020-12-11 21:37:59,465 synthtool [DEBUG] > Using precloned repo /root/.cache/synthtool/googleapis
DEBUG:synthtool:Using precloned repo /root/.cache/synthtool/googleapis
2020-12-11 21:37:59,470 synthtool [DEBUG] > Generating code for: //google/cloud/recommender/v1:recommender-v1-nodejs.
DEBUG:synthtool:Generating code for: //google/cloud/recommender/v1:recommender-v1-nodejs.
2020-12-11 21:38:05,546 synthtool [SUCCESS] > Generated code into /tmp/tmphsvxy_9h.
SUCCESS:synthtool:Generated code into /tmp/tmphsvxy_9h.
2020-12-11 21:38:05,576 synthtool [DEBUG] > Using precloned repo /root/.cache/synthtool/synthtool
DEBUG:synthtool:Using precloned repo /root/.cache/synthtool/synthtool
.eslintignore
.eslintrc.json
.gitattributes
.github/ISSUE_TEMPLATE/bug_report.md
.github/ISSUE_TEMPLATE/feature_request.md
.github/ISSUE_TEMPLATE/support_request.md
.github/PULL_REQUEST_TEMPLATE.md
.github/release-please.yml
.github/workflows/ci.yaml
.kokoro/.gitattributes
.kokoro/common.cfg
.kokoro/continuous/node10/common.cfg
.kokoro/continuous/node10/docs.cfg
.kokoro/continuous/node10/test.cfg
.kokoro/continuous/node12/common.cfg
.kokoro/continuous/node12/lint.cfg
.kokoro/continuous/node12/samples-test.cfg
.kokoro/continuous/node12/system-test.cfg
.kokoro/continuous/node12/test.cfg
.kokoro/docs.sh
.kokoro/lint.sh
.kokoro/populate-secrets.sh
.kokoro/presubmit/node10/common.cfg
.kokoro/presubmit/node12/common.cfg
.kokoro/presubmit/node12/samples-test.cfg
.kokoro/presubmit/node12/system-test.cfg
.kokoro/presubmit/node12/test.cfg
.kokoro/publish.sh
.kokoro/release/docs-devsite.cfg
.kokoro/release/docs-devsite.sh
.kokoro/release/docs.cfg
.kokoro/release/docs.sh
.kokoro/release/publish.cfg
.kokoro/samples-test.sh
.kokoro/system-test.sh
.kokoro/test.bat
.kokoro/test.sh
.kokoro/trampoline.sh
.kokoro/trampoline_v2.sh
.mocharc.js
.nycrc
.prettierignore
.prettierrc.js
.trampolinerc
CODE_OF_CONDUCT.md
CONTRIBUTING.md
LICENSE
README.md
api-extractor.json
renovate.json
samples/README.md
2020-12-11 21:38:05,736 synthtool [DEBUG] > Post-processing GAPIC library...
DEBUG:synthtool:Post-processing GAPIC library...
2020-12-11 21:38:05,736 synthtool [DEBUG] > Installing dependencies...
DEBUG:synthtool:Installing dependencies...
Traceback (most recent call last):
File "/usr/local/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/usr/local/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/root/.cache/synthtool/nodejs-recommender/synth.py", line 40, in <module>
node.postprocess_gapic_library()
File "/synthtool/synthtool/languages/node.py", line 193, in postprocess_gapic_library
install(hide_output=hide_output)
File "/synthtool/synthtool/languages/node.py", line 167, in install
shell.run(["npm", "install"], hide_output=hide_output)
File "/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/usr/local/lib/python3.6/subprocess.py", line 423, in run
with Popen(*popenargs, **kwargs) as process:
File "/usr/local/lib/python3.6/subprocess.py", line 729, in __init__
restore_signals, start_new_session)
File "/usr/local/lib/python3.6/subprocess.py", line 1364, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: 'npm': 'npm'
2020-12-11 21:38:05,802 autosynth [ERROR] > Synthesis failed
2020-12-11 21:38:05,802 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at e06347a chore: generate GAPIC metadata JSON file (#136)
2020-12-11 21:38:05,812 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2020-12-11 21:38:05,819 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/usr/local/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/usr/local/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/synthtool/autosynth/synth.py", line 354, in <module>
main()
File "/synthtool/autosynth/synth.py", line 189, in main
return _inner_main(temp_dir)
File "/synthtool/autosynth/synth.py", line 334, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/synthtool/autosynth/synth.py", line 65, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/usr/local/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/usr/local/bin/python', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/None/targets/github%2Fsynthtool;config=default/tests;query=nodejs-recommender;failed=false).
| 1.0 | Synthesis failed for nodejs-recommender - Hello! Autosynth couldn't regenerate nodejs-recommender. :broken_heart:
Here's the output from running `synth.py`:
```
e7e855aed5a29e6e83c39beba2a
2020-12-11 21:37:59,042 autosynth [DEBUG] > Running: git log -1 --pretty=%at c2de32114ec484aa708d32012d1fa8d75232daf5
2020-12-11 21:37:59,045 autosynth [DEBUG] > Running: git log -1 --pretty=%at 32af6da519a6b042e3da62008e2a75e991efb6b4
2020-12-11 21:37:59,048 autosynth [DEBUG] > Running: git log -1 --pretty=%at 69629b64b83c6421d616be2b8e11795738ec8a6c
2020-12-11 21:37:59,051 autosynth [DEBUG] > Running: git log -1 --pretty=%at f94318521f63085b9ccb43d42af89f153fb39f15
2020-12-11 21:37:59,054 autosynth [DEBUG] > Running: git log -1 --pretty=%at 18c5dbdb4ac8cf75d4d8174e7b4558f48e76f8a1
2020-12-11 21:37:59,057 autosynth [DEBUG] > Running: git log -1 --pretty=%at 5d11bd2888c38ce1fb6fa6bf25494a4219a73928
2020-12-11 21:37:59,060 autosynth [DEBUG] > Running: git log -1 --pretty=%at 3559dd8cb634347737632776ef4089443eacdb4f
2020-12-11 21:37:59,062 autosynth [DEBUG] > Running: git log -1 --pretty=%at 7d1ee3207371ef72a1ca931a0964c4007e687280
2020-12-11 21:37:59,065 autosynth [DEBUG] > Running: git log -1 --pretty=%at 352f25621c28d2fa0784984ea0ed5ea697aa1525
2020-12-11 21:37:59,068 autosynth [DEBUG] > Running: git log -1 --pretty=%at 996775eca5fd934edac3c2ae34b80ff0395b1717
2020-12-11 21:37:59,071 autosynth [DEBUG] > Running: git checkout e06347ac86941705801aad1e3c3b6b87092aa535
Note: checking out 'e06347ac86941705801aad1e3c3b6b87092aa535'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at e06347a chore: generate GAPIC metadata JSON file (#136)
2020-12-11 21:37:59,086 autosynth [DEBUG] > Running: git checkout 996775eca5fd934edac3c2ae34b80ff0395b1717
Note: checking out '996775eca5fd934edac3c2ae34b80ff0395b1717'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 996775e fix(python): skip samples readmegen in synthtool (#875)
2020-12-11 21:37:59,092 autosynth [DEBUG] > Running: git checkout dd372aa22ded7a8ba6f0e03a80e06358a3fa0907
Note: checking out 'dd372aa22ded7a8ba6f0e03a80e06358a3fa0907'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at dd372aa2 fix: remove client recv msg limit fix: add enums to `types/__init__.py`
2020-12-11 21:37:59,108 autosynth [DEBUG] > Running: git branch -f autosynth-60
2020-12-11 21:37:59,112 autosynth [DEBUG] > Running: git checkout autosynth-60
Switched to branch 'autosynth-60'
2020-12-11 21:37:59,116 autosynth [INFO] > Running synthtool
2020-12-11 21:37:59,116 autosynth [INFO] > ['/usr/local/bin/python', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']
2020-12-11 21:37:59,116 autosynth [DEBUG] > log_file_path: /kokoro/artifacts/logs/nodejs-recommender/60/sponge_log.log
2020-12-11 21:37:59,118 autosynth [DEBUG] > Running: /usr/local/bin/python -m synthtool --metadata synth.metadata synth.py --
2020-12-11 21:37:59,325 synthtool [DEBUG] > Executing /root/.cache/synthtool/nodejs-recommender/synth.py.
On branch autosynth-60
nothing to commit, working tree clean
2020-12-11 21:37:59,455 synthtool [DEBUG] > Ensuring dependencies.
DEBUG:synthtool:Ensuring dependencies.
2020-12-11 21:37:59,464 synthtool [DEBUG] > Cloning googleapis.
DEBUG:synthtool:Cloning googleapis.
2020-12-11 21:37:59,465 synthtool [DEBUG] > Using precloned repo /root/.cache/synthtool/googleapis
DEBUG:synthtool:Using precloned repo /root/.cache/synthtool/googleapis
2020-12-11 21:37:59,470 synthtool [DEBUG] > Generating code for: //google/cloud/recommender/v1:recommender-v1-nodejs.
DEBUG:synthtool:Generating code for: //google/cloud/recommender/v1:recommender-v1-nodejs.
2020-12-11 21:38:05,546 synthtool [SUCCESS] > Generated code into /tmp/tmphsvxy_9h.
SUCCESS:synthtool:Generated code into /tmp/tmphsvxy_9h.
2020-12-11 21:38:05,576 synthtool [DEBUG] > Using precloned repo /root/.cache/synthtool/synthtool
DEBUG:synthtool:Using precloned repo /root/.cache/synthtool/synthtool
.eslintignore
.eslintrc.json
.gitattributes
.github/ISSUE_TEMPLATE/bug_report.md
.github/ISSUE_TEMPLATE/feature_request.md
.github/ISSUE_TEMPLATE/support_request.md
.github/PULL_REQUEST_TEMPLATE.md
.github/release-please.yml
.github/workflows/ci.yaml
.kokoro/.gitattributes
.kokoro/common.cfg
.kokoro/continuous/node10/common.cfg
.kokoro/continuous/node10/docs.cfg
.kokoro/continuous/node10/test.cfg
.kokoro/continuous/node12/common.cfg
.kokoro/continuous/node12/lint.cfg
.kokoro/continuous/node12/samples-test.cfg
.kokoro/continuous/node12/system-test.cfg
.kokoro/continuous/node12/test.cfg
.kokoro/docs.sh
.kokoro/lint.sh
.kokoro/populate-secrets.sh
.kokoro/presubmit/node10/common.cfg
.kokoro/presubmit/node12/common.cfg
.kokoro/presubmit/node12/samples-test.cfg
.kokoro/presubmit/node12/system-test.cfg
.kokoro/presubmit/node12/test.cfg
.kokoro/publish.sh
.kokoro/release/docs-devsite.cfg
.kokoro/release/docs-devsite.sh
.kokoro/release/docs.cfg
.kokoro/release/docs.sh
.kokoro/release/publish.cfg
.kokoro/samples-test.sh
.kokoro/system-test.sh
.kokoro/test.bat
.kokoro/test.sh
.kokoro/trampoline.sh
.kokoro/trampoline_v2.sh
.mocharc.js
.nycrc
.prettierignore
.prettierrc.js
.trampolinerc
CODE_OF_CONDUCT.md
CONTRIBUTING.md
LICENSE
README.md
api-extractor.json
renovate.json
samples/README.md
2020-12-11 21:38:05,736 synthtool [DEBUG] > Post-processing GAPIC library...
DEBUG:synthtool:Post-processing GAPIC library...
2020-12-11 21:38:05,736 synthtool [DEBUG] > Installing dependencies...
DEBUG:synthtool:Installing dependencies...
Traceback (most recent call last):
File "/usr/local/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/usr/local/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "<frozen importlib._bootstrap_external>", line 678, in exec_module
File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed
File "/root/.cache/synthtool/nodejs-recommender/synth.py", line 40, in <module>
node.postprocess_gapic_library()
File "/synthtool/synthtool/languages/node.py", line 193, in postprocess_gapic_library
install(hide_output=hide_output)
File "/synthtool/synthtool/languages/node.py", line 167, in install
shell.run(["npm", "install"], hide_output=hide_output)
File "/synthtool/synthtool/shell.py", line 33, in run
encoding="utf-8",
File "/usr/local/lib/python3.6/subprocess.py", line 423, in run
with Popen(*popenargs, **kwargs) as process:
File "/usr/local/lib/python3.6/subprocess.py", line 729, in __init__
restore_signals, start_new_session)
File "/usr/local/lib/python3.6/subprocess.py", line 1364, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: 'npm': 'npm'
2020-12-11 21:38:05,802 autosynth [ERROR] > Synthesis failed
2020-12-11 21:38:05,802 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at e06347a chore: generate GAPIC metadata JSON file (#136)
2020-12-11 21:38:05,812 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2020-12-11 21:38:05,819 autosynth [DEBUG] > Running: git clean -fdx
Removing __pycache__/
Traceback (most recent call last):
File "/usr/local/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/usr/local/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/synthtool/autosynth/synth.py", line 354, in <module>
main()
File "/synthtool/autosynth/synth.py", line 189, in main
return _inner_main(temp_dir)
File "/synthtool/autosynth/synth.py", line 334, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/synthtool/autosynth/synth.py", line 65, in synthesize_loop
has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest)
File "/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch
synthesizer.synthesize(synth_log_path, self.environ)
File "/synthtool/autosynth/synthesizer.py", line 120, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/usr/local/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/usr/local/bin/python', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/None/targets/github%2Fsynthtool;config=default/tests;query=nodejs-recommender;failed=false).
| priority | synthesis failed for nodejs recommender hello autosynth couldn t regenerate nodejs recommender broken heart here s the output from running synth py autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git log pretty at autosynth running git checkout note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at chore generate gapic metadata json file autosynth running git checkout note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at fix python skip samples readmegen in synthtool autosynth running git checkout note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at fix remove client recv msg limit fix add enums to types init py autosynth running git branch f autosynth autosynth running git checkout autosynth switched to branch autosynth autosynth running synthtool autosynth autosynth log file path kokoro artifacts logs nodejs recommender sponge log log autosynth running usr local bin python m synthtool metadata synth metadata synth py synthtool executing root cache synthtool nodejs recommender synth py on branch autosynth nothing to commit working tree clean synthtool ensuring dependencies debug synthtool ensuring dependencies synthtool cloning googleapis debug synthtool cloning googleapis synthtool using precloned repo root cache synthtool googleapis debug synthtool using precloned repo root cache synthtool googleapis synthtool generating code for google cloud recommender recommender nodejs debug synthtool generating code for google cloud recommender recommender nodejs synthtool generated code into tmp tmphsvxy success synthtool generated code into tmp tmphsvxy synthtool using precloned repo root cache synthtool synthtool debug synthtool using precloned repo root cache synthtool synthtool eslintignore eslintrc json gitattributes github issue template bug report md github issue template feature request md github issue template support request md github pull request template md github release please yml github workflows ci yaml kokoro gitattributes kokoro common cfg kokoro continuous common cfg kokoro continuous docs cfg kokoro continuous test cfg kokoro continuous common cfg kokoro continuous lint cfg kokoro continuous samples test cfg kokoro continuous system test cfg kokoro continuous test cfg kokoro docs sh kokoro lint sh kokoro populate secrets sh kokoro presubmit common cfg kokoro presubmit common cfg kokoro presubmit samples test cfg kokoro presubmit system test cfg kokoro presubmit test cfg kokoro publish sh kokoro release docs devsite cfg kokoro release docs devsite sh kokoro release docs cfg kokoro release docs sh kokoro release publish cfg kokoro samples test sh kokoro system test sh kokoro test bat kokoro test sh kokoro trampoline sh kokoro trampoline sh mocharc js nycrc prettierignore prettierrc js trampolinerc code of conduct md contributing md license readme md api extractor json renovate json samples readme md synthtool post processing gapic library debug synthtool post processing gapic library synthtool installing dependencies debug synthtool installing dependencies traceback most recent call last file usr local lib runpy py line in run module as main main mod spec file usr local lib runpy py line in run code exec code run globals file synthtool synthtool main py line in main file usr local lib site packages click core py line in call return self main args kwargs file usr local lib site packages click core py line in main rv self invoke ctx file usr local lib site packages click core py line in invoke return ctx invoke self callback ctx params file usr local lib site packages click core py line in invoke return callback args kwargs file synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file root cache synthtool nodejs recommender synth py line in node postprocess gapic library file synthtool synthtool languages node py line in postprocess gapic library install hide output hide output file synthtool synthtool languages node py line in install shell run hide output hide output file synthtool synthtool shell py line in run encoding utf file usr local lib subprocess py line in run with popen popenargs kwargs as process file usr local lib subprocess py line in init restore signals start new session file usr local lib subprocess py line in execute child raise child exception type errno num err msg err filename filenotfounderror no such file or directory npm npm autosynth synthesis failed autosynth running git reset hard head head is now at chore generate gapic metadata json file autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache traceback most recent call last file usr local lib runpy py line in run module as main main mod spec file usr local lib runpy py line in run code exec code run globals file synthtool autosynth synth py line in main file synthtool autosynth synth py line in main return inner main temp dir file synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file synthtool autosynth synth toolbox py line in synthesize version in new branch synthesizer synthesize synth log path self environ file synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file usr local lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 1 |
312,122 | 9,543,589,544 | IssuesEvent | 2019-05-01 10:49:10 | aau-giraf/weekplanner | https://api.github.com/repos/aau-giraf/weekplanner | closed | Change buttons on ConfirmDialog | priority: high type: refactor | The buttons on GirafConfirmDialog should be a button with gradient colors.
Button can be refactored after issue #85 has been merged.
| 1.0 | Change buttons on ConfirmDialog - The buttons on GirafConfirmDialog should be a button with gradient colors.
Button can be refactored after issue #85 has been merged.
| priority | change buttons on confirmdialog the buttons on girafconfirmdialog should be a button with gradient colors button can be refactored after issue has been merged | 1 |
66,821 | 16,725,814,689 | IssuesEvent | 2021-06-10 12:53:28 | opengeospatial/ogcapi-features | https://api.github.com/repos/opengeospatial/ogcapi-features | closed | CQL: requirement /req/filter/filter-crs-param: HTTP status code | API building blocks Part 3: CQL | Requirement /req/filter/filter-crs-param, subrequirement C:
> The server SHALL return an error, if it does not support the CRS identified in `filter-crs` for the resource.
Should that be the following in order to be consistent with /req/crs/fc-bbox-crs-valid-value and /req/crs/fc-crs-valid-value?
The server SHALL return HTTP status code 400, if it does not support the CRS identified in `filter-crs` for the resource.
| 1.0 | CQL: requirement /req/filter/filter-crs-param: HTTP status code - Requirement /req/filter/filter-crs-param, subrequirement C:
> The server SHALL return an error, if it does not support the CRS identified in `filter-crs` for the resource.
Should that be the following in order to be consistent with /req/crs/fc-bbox-crs-valid-value and /req/crs/fc-crs-valid-value?
The server SHALL return HTTP status code 400, if it does not support the CRS identified in `filter-crs` for the resource.
| non_priority | cql requirement req filter filter crs param http status code requirement req filter filter crs param subrequirement c the server shall return an error if it does not support the crs identified in filter crs for the resource should that be the following in order to be consistent with req crs fc bbox crs valid value and req crs fc crs valid value the server shall return http status code if it does not support the crs identified in filter crs for the resource | 0 |
379,980 | 11,252,614,164 | IssuesEvent | 2020-01-11 10:04:34 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.facebook.com - video or audio doesn't play | browser-firefox engine-gecko form-v2-experiment priority-critical | <!-- @browser: Firefox 72.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:72.0) Gecko/20100101 Firefox/72.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: form-v2-experiment -->
**URL**: http://www.facebook.com
**Browser / Version**: Firefox 72.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Other
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Video not playing properly at any cost
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.facebook.com - video or audio doesn't play - <!-- @browser: Firefox 72.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:72.0) Gecko/20100101 Firefox/72.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: form-v2-experiment -->
**URL**: http://www.facebook.com
**Browser / Version**: Firefox 72.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Other
**Problem type**: Video or audio doesn't play
**Description**: The video or audio does not play
**Steps to Reproduce**:
Video not playing properly at any cost
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | video or audio doesn t play url browser version firefox operating system windows tested another browser yes other problem type video or audio doesn t play description the video or audio does not play steps to reproduce video not playing properly at any cost browser configuration none from with ❤️ | 1 |
52,865 | 3,030,585,955 | IssuesEvent | 2015-08-04 18:13:19 | Financial-Times/next-front-page | https://api.github.com/repos/Financial-Times/next-front-page | closed | Mobile News/fastFT toggle styling wrong | bug HIGH PRIORITY | - wrong font family
- missing the initial status underline | 1.0 | Mobile News/fastFT toggle styling wrong - - wrong font family
- missing the initial status underline | priority | mobile news fastft toggle styling wrong wrong font family missing the initial status underline | 1 |
805,762 | 29,667,387,041 | IssuesEvent | 2023-06-11 00:53:08 | davidfstr/Crystal-Web-Archiver | https://api.github.com/repos/davidfstr/Crystal-Web-Archiver | opened | Allow treating URLs found inside <script> tags with an image file extension as NON-embedded | priority-low type-feature topic-sitespecific | URLs found inside `<script>` tags with an image file extension like .jpg or .png are assumed Crystal to be **embedded** by default, meaning that they will be automatically downloaded as an embedded-subresource when its referencing HTML resource is downloaded.
Sometimes when downloading a particular Resource Group within a Project, there is a desire to NOT assume such URLs are embedded, because (1) they are in fact not embedded for the pages in the group and (2) the user doesn't want to spend time downloading these resources up front.
Suggested remediations:
* [ ] Make it possible, on a per group basis, to customize whether "URLs found inside `<script>` tags with an image file extension like .jpg or .png are assumed to be embedded" = YES or NO.
* [ ] For any "Script Reference" links in the Entity Tree corresponding to the above type of URL, provide some kind of indication that (1) the link type was assumed to be embedded, (2) that it's possible to change the link type to assume it's NOT embedded (for links in the same group), and (3) some way to quickly toggle this assumption.
Priority: Low
* This is (mostly) a site-specific feature request, for the YRE site
See related screenshots at:
* CATHODE: 2023-05-22 - Assume URLs inside <script> with image file extension are embedded | 1.0 | Allow treating URLs found inside <script> tags with an image file extension as NON-embedded - URLs found inside `<script>` tags with an image file extension like .jpg or .png are assumed Crystal to be **embedded** by default, meaning that they will be automatically downloaded as an embedded-subresource when its referencing HTML resource is downloaded.
Sometimes when downloading a particular Resource Group within a Project, there is a desire to NOT assume such URLs are embedded, because (1) they are in fact not embedded for the pages in the group and (2) the user doesn't want to spend time downloading these resources up front.
Suggested remediations:
* [ ] Make it possible, on a per group basis, to customize whether "URLs found inside `<script>` tags with an image file extension like .jpg or .png are assumed to be embedded" = YES or NO.
* [ ] For any "Script Reference" links in the Entity Tree corresponding to the above type of URL, provide some kind of indication that (1) the link type was assumed to be embedded, (2) that it's possible to change the link type to assume it's NOT embedded (for links in the same group), and (3) some way to quickly toggle this assumption.
Priority: Low
* This is (mostly) a site-specific feature request, for the YRE site
See related screenshots at:
* CATHODE: 2023-05-22 - Assume URLs inside <script> with image file extension are embedded | priority | allow treating urls found inside tags with an image file extension as non embedded urls found inside tags with an image file extension like jpg or png are assumed crystal to be embedded by default meaning that they will be automatically downloaded as an embedded subresource when its referencing html resource is downloaded sometimes when downloading a particular resource group within a project there is a desire to not assume such urls are embedded because they are in fact not embedded for the pages in the group and the user doesn t want to spend time downloading these resources up front suggested remediations make it possible on a per group basis to customize whether urls found inside tags with an image file extension like jpg or png are assumed to be embedded yes or no for any script reference links in the entity tree corresponding to the above type of url provide some kind of indication that the link type was assumed to be embedded that it s possible to change the link type to assume it s not embedded for links in the same group and some way to quickly toggle this assumption priority low this is mostly a site specific feature request for the yre site see related screenshots at cathode assume urls inside with image file extension are embedded | 1 |
276,665 | 24,009,581,629 | IssuesEvent | 2022-09-14 17:32:55 | onc-healthit/onc-certification-g10-test-kit | https://api.github.com/repos/onc-healthit/onc-certification-g10-test-kit | closed | Add practitioner-scoped EHR Launch tests | g10-test-kit add constraint v3.0.0 | The SMART App Launch 2.0.0 IG [contains the following language](http://hl7.org/fhir/smart-app-launch/STU2/scopes-and-launch-context.html#apps-that-launch-from-the-ehr) in relation to both EHR and standalone launches:
> If an application requests a clinical scope which is restricted to a single patient (e.g., patient/*.rs), and the authorization results in the EHR is granting that scope, the EHR SHALL establish a patient in context.
We understand this to be a clarification rather than a new requirement in SMART 2.0.0. As a result, tests should be added for an EHR launch with patient scopes for both SMART 1.0.0 and 2.0.0.
`permission-patient` and `context-ehr-patient` are both included in the Clinicial Access for EHR Launch Capability Set [STU1](http://hl7.org/fhir/smart-app-launch/1.0.0/conformance/index.html#clinician-access-for-ehr-launch)/[STU2](http://hl7.org/fhir/smart-app-launch/STU2/conformance.html#clinician-access-for-ehr-launch), and the [2015 Edition Cures Update CCG](https://www.healthit.gov/test-method/standardized-api-patient-and-population-services#ccg) states that this capability set is valid to test:
> Health IT Modules will only be tested for the "Patient Access for Standalone Apps" and "Clinician Access for EHR Launch" scenarios described in the standard adopted at § 170.215(a)(3).
| 1.0 | Add practitioner-scoped EHR Launch tests - The SMART App Launch 2.0.0 IG [contains the following language](http://hl7.org/fhir/smart-app-launch/STU2/scopes-and-launch-context.html#apps-that-launch-from-the-ehr) in relation to both EHR and standalone launches:
> If an application requests a clinical scope which is restricted to a single patient (e.g., patient/*.rs), and the authorization results in the EHR is granting that scope, the EHR SHALL establish a patient in context.
We understand this to be a clarification rather than a new requirement in SMART 2.0.0. As a result, tests should be added for an EHR launch with patient scopes for both SMART 1.0.0 and 2.0.0.
`permission-patient` and `context-ehr-patient` are both included in the Clinicial Access for EHR Launch Capability Set [STU1](http://hl7.org/fhir/smart-app-launch/1.0.0/conformance/index.html#clinician-access-for-ehr-launch)/[STU2](http://hl7.org/fhir/smart-app-launch/STU2/conformance.html#clinician-access-for-ehr-launch), and the [2015 Edition Cures Update CCG](https://www.healthit.gov/test-method/standardized-api-patient-and-population-services#ccg) states that this capability set is valid to test:
> Health IT Modules will only be tested for the "Patient Access for Standalone Apps" and "Clinician Access for EHR Launch" scenarios described in the standard adopted at § 170.215(a)(3).
| non_priority | add practitioner scoped ehr launch tests the smart app launch ig in relation to both ehr and standalone launches if an application requests a clinical scope which is restricted to a single patient e g patient rs and the authorization results in the ehr is granting that scope the ehr shall establish a patient in context we understand this to be a clarification rather than a new requirement in smart as a result tests should be added for an ehr launch with patient scopes for both smart and permission patient and context ehr patient are both included in the clinicial access for ehr launch capability set and the states that this capability set is valid to test health it modules will only be tested for the patient access for standalone apps and clinician access for ehr launch scenarios described in the standard adopted at § a | 0 |
469,810 | 13,526,390,103 | IssuesEvent | 2020-09-15 14:11:39 | mozilla/addons-server | https://api.github.com/repos/mozilla/addons-server | closed | Optimize the storage in mlbf directory | priority: p4 | ### Describe the problem and steps to reproduce it:
Currently dumped blocked/notblocked json files are saved in `<storage_root>/mlbf` directory. In chatting with @eviljeff [[1]](https://mozilla.slack.com/archives/C4MTGNZ7S/p1589560549015900), we think that there are two possible issues with the current design.
First is that there is no hierarchy under `mlbf/`, thus this directory can grow very big, and cause the same management headache as the `files/` directory does (e.g. nearly impossible to do `ls`; hard to find files, and etc.).
Second is that the directory can grow very big given each dump is over 150 MB, and at a maximum rate of 28 dumps per week. This puts pressure on costs and backup operations.
### What happened?
As above.
### What did you expect to happen?
As mentioned in the slack thread [[1]](https://mozilla.slack.com/archives/C4MTGNZ7S/p1589560549015900), there are two things we can do to improve this:
One is to delete old dumps periodically and keep only dumps that are X days/months old on the disk.
Two is to implement a hierarchical directory structure under `mlbf/` similar to what's in `<storage root>/git-storage/` directory. That will make management of this directory much easier.
Out of the two, if we don't keep too many days worth of dumps, then I think the second measure is probably optional. Because if we prune the `mlbf/` directory periodically, then the sub-directories won't grow to a unmanageable level anyways.
### Anything else we should know?
n/a
| 1.0 | Optimize the storage in mlbf directory - ### Describe the problem and steps to reproduce it:
Currently dumped blocked/notblocked json files are saved in `<storage_root>/mlbf` directory. In chatting with @eviljeff [[1]](https://mozilla.slack.com/archives/C4MTGNZ7S/p1589560549015900), we think that there are two possible issues with the current design.
First is that there is no hierarchy under `mlbf/`, thus this directory can grow very big, and cause the same management headache as the `files/` directory does (e.g. nearly impossible to do `ls`; hard to find files, and etc.).
Second is that the directory can grow very big given each dump is over 150 MB, and at a maximum rate of 28 dumps per week. This puts pressure on costs and backup operations.
### What happened?
As above.
### What did you expect to happen?
As mentioned in the slack thread [[1]](https://mozilla.slack.com/archives/C4MTGNZ7S/p1589560549015900), there are two things we can do to improve this:
One is to delete old dumps periodically and keep only dumps that are X days/months old on the disk.
Two is to implement a hierarchical directory structure under `mlbf/` similar to what's in `<storage root>/git-storage/` directory. That will make management of this directory much easier.
Out of the two, if we don't keep too many days worth of dumps, then I think the second measure is probably optional. Because if we prune the `mlbf/` directory periodically, then the sub-directories won't grow to a unmanageable level anyways.
### Anything else we should know?
n/a
| priority | optimize the storage in mlbf directory describe the problem and steps to reproduce it currently dumped blocked notblocked json files are saved in mlbf directory in chatting with eviljeff we think that there are two possible issues with the current design first is that there is no hierarchy under mlbf thus this directory can grow very big and cause the same management headache as the files directory does e g nearly impossible to do ls hard to find files and etc second is that the directory can grow very big given each dump is over mb and at a maximum rate of dumps per week this puts pressure on costs and backup operations what happened as above what did you expect to happen as mentioned in the slack thread there are two things we can do to improve this one is to delete old dumps periodically and keep only dumps that are x days months old on the disk two is to implement a hierarchical directory structure under mlbf similar to what s in git storage directory that will make management of this directory much easier out of the two if we don t keep too many days worth of dumps then i think the second measure is probably optional because if we prune the mlbf directory periodically then the sub directories won t grow to a unmanageable level anyways anything else we should know n a | 1 |
112,307 | 9,560,075,560 | IssuesEvent | 2019-05-03 18:31:47 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Monitoring - can't completely clean up after disabling cluster monitoring | area/monitoring kind/bug status/ready-for-review status/resolved status/to-test status/triaged team/cn | **What kind of request is this (question/bug/enhancement/feature request):**
Bug
**Steps to reproduce (least amount of steps as possible):**
1. Enable Monitoring with large size requirement CPU & RAM
1. Disable Monitoring without enabling Monitoring completely
**Result:**
Can't clean up the `cluster-monitoring` App of `System` Project
**Other details that may be helpful:**
- Enable - Disable (quickly) may output the same result.
- Enable Monitoring in a limited resource environment also output the same result.
- Workaround:
+ Re-enable the Monitoring, assign 1m CPU & 1m RAM to Prometheus
+ Disable Monitoring again after starting Prometheus
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): v2.2.2-rc11
- Installation option (single install/HA): single | 1.0 | Monitoring - can't completely clean up after disabling cluster monitoring - **What kind of request is this (question/bug/enhancement/feature request):**
Bug
**Steps to reproduce (least amount of steps as possible):**
1. Enable Monitoring with large size requirement CPU & RAM
1. Disable Monitoring without enabling Monitoring completely
**Result:**
Can't clean up the `cluster-monitoring` App of `System` Project
**Other details that may be helpful:**
- Enable - Disable (quickly) may output the same result.
- Enable Monitoring in a limited resource environment also output the same result.
- Workaround:
+ Re-enable the Monitoring, assign 1m CPU & 1m RAM to Prometheus
+ Disable Monitoring again after starting Prometheus
**Environment information**
- Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): v2.2.2-rc11
- Installation option (single install/HA): single | non_priority | monitoring can t completely clean up after disabling cluster monitoring what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible enable monitoring with large size requirement cpu ram disable monitoring without enabling monitoring completely result can t clean up the cluster monitoring app of system project other details that may be helpful enable disable quickly may output the same result enable monitoring in a limited resource environment also output the same result workaround re enable the monitoring assign cpu ram to prometheus disable monitoring again after starting prometheus environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui installation option single install ha single | 0 |
113,070 | 11,786,994,136 | IssuesEvent | 2020-03-17 13:18:59 | scholaempirica/reschola | https://api.github.com/repos/scholaempirica/reschola | closed | Update install instructions for drat repo | bug documentation | install.packages("reschola", repos = "https://scholaempirica.github.io/drat/", dependencies = T)
install.packages treats dependencies differently if repos is set | 1.0 | Update install instructions for drat repo - install.packages("reschola", repos = "https://scholaempirica.github.io/drat/", dependencies = T)
install.packages treats dependencies differently if repos is set | non_priority | update install instructions for drat repo install packages reschola repos dependencies t install packages treats dependencies differently if repos is set | 0 |
590,725 | 17,786,172,422 | IssuesEvent | 2021-08-31 11:20:23 | IgniteUI/ignite-ui | https://api.github.com/repos/IgniteUI/ignite-ui | closed | igCombo - Duplicate items are displayed if dataSource contains both half-width and full-width characters and virtualization is enabled. | bug combo status: resolved priority: high | ## Description
Duplicate items are displayed if dataSource contains both half-width and full-width characters and virtualization is enabled.
IE works as expected.
* ignite-ui version: 20.2.20.2.17
* browser: Chrome, Edge, FireFox
## Steps to reproduce
1. Run the attached sample in Chrome.
For both combos:
2. Open the drop down.
3. Scroll to the bottom of the list.
## Result
The last item is duplicated.
(You will see a different item gets duplicated if you scroll down quickly.)
## Expected result
No duplicate items.
## Attachments
[sample.zip](https://github.com/IgniteUI/ignite-ui/files/6456234/sample.zip)
| 1.0 | igCombo - Duplicate items are displayed if dataSource contains both half-width and full-width characters and virtualization is enabled. - ## Description
Duplicate items are displayed if dataSource contains both half-width and full-width characters and virtualization is enabled.
IE works as expected.
* ignite-ui version: 20.2.20.2.17
* browser: Chrome, Edge, FireFox
## Steps to reproduce
1. Run the attached sample in Chrome.
For both combos:
2. Open the drop down.
3. Scroll to the bottom of the list.
## Result
The last item is duplicated.
(You will see a different item gets duplicated if you scroll down quickly.)
## Expected result
No duplicate items.
## Attachments
[sample.zip](https://github.com/IgniteUI/ignite-ui/files/6456234/sample.zip)
| priority | igcombo duplicate items are displayed if datasource contains both half width and full width characters and virtualization is enabled description duplicate items are displayed if datasource contains both half width and full width characters and virtualization is enabled ie works as expected ignite ui version browser chrome edge firefox steps to reproduce run the attached sample in chrome for both combos open the drop down scroll to the bottom of the list result the last item is duplicated you will see a different item gets duplicated if you scroll down quickly expected result no duplicate items attachments | 1 |
381,671 | 11,278,283,091 | IssuesEvent | 2020-01-15 06:13:55 | qlcchain/go-qlc | https://api.github.com/repos/qlcchain/go-qlc | closed | minble-winble on QLCChain proposal | Priority: High Type: Enhancement | ### Description of the issue
### Issue-Type
- [ ] bug report
- [x] feature request
- [ ] Documentation improvement | 1.0 | minble-winble on QLCChain proposal - ### Description of the issue
### Issue-Type
- [ ] bug report
- [x] feature request
- [ ] Documentation improvement | priority | minble winble on qlcchain proposal description of the issue issue type bug report feature request documentation improvement | 1 |
738,146 | 25,547,093,333 | IssuesEvent | 2022-11-29 19:49:05 | idom-team/idom | https://api.github.com/repos/idom-team/idom | closed | add VDOM to HTML util function | type: feature flag: good first issue priority: 3 (low) | Sample implementation:
```python
from html import escape as html_escape
def vdom_to_html(vdom: Union[str, VdomDict]) -> str:
"""Convert a VDOM dictionary into an HTML string
Only the following keys are translated to HTML:
- ``tagName``
- ``attributes``
- ``children`` (must be strings or more VDOM dicts)
"""
if isinstance(vdom, str):
return vdom
try:
tag = vdom["tagName"]
except TypeError as error:
raise TypeError(f"Expected a VDOM dictionary or string, not {vdom}") from error
if "attributes" in vdom:
attributes = " " + " ".join(
f'{k}="{html_escape(v)}"' for k, v in vdom["attributes"].items()
)
else:
attributes = ""
if "children" in vdom:
children = "".join(map(vdom_to_html, vdom["children"]))
else:
children = ""
return f"<{tag}{attributes}>{children}</{tag}>"
``` | 1.0 | add VDOM to HTML util function - Sample implementation:
```python
from html import escape as html_escape
def vdom_to_html(vdom: Union[str, VdomDict]) -> str:
"""Convert a VDOM dictionary into an HTML string
Only the following keys are translated to HTML:
- ``tagName``
- ``attributes``
- ``children`` (must be strings or more VDOM dicts)
"""
if isinstance(vdom, str):
return vdom
try:
tag = vdom["tagName"]
except TypeError as error:
raise TypeError(f"Expected a VDOM dictionary or string, not {vdom}") from error
if "attributes" in vdom:
attributes = " " + " ".join(
f'{k}="{html_escape(v)}"' for k, v in vdom["attributes"].items()
)
else:
attributes = ""
if "children" in vdom:
children = "".join(map(vdom_to_html, vdom["children"]))
else:
children = ""
return f"<{tag}{attributes}>{children}</{tag}>"
``` | priority | add vdom to html util function sample implementation python from html import escape as html escape def vdom to html vdom union str convert a vdom dictionary into an html string only the following keys are translated to html tagname attributes children must be strings or more vdom dicts if isinstance vdom str return vdom try tag vdom except typeerror as error raise typeerror f expected a vdom dictionary or string not vdom from error if attributes in vdom attributes join f k html escape v for k v in vdom items else attributes if children in vdom children join map vdom to html vdom else children return f children | 1 |
319,585 | 27,386,872,107 | IssuesEvent | 2023-02-28 13:54:55 | DD2480G19/teammates2 | https://api.github.com/repos/DD2480G19/teammates2 | closed | `execute`: Logs exception while creating response (ex16) | invalid wontfix New test | At the for-each loop at line 211: None of the exceptions should be happening as the responses have been pre-validated (i.e., `log.severe("Encountered exception when creating response: " + e.getMessage(), e)` when encountering exception).
| 1.0 | `execute`: Logs exception while creating response (ex16) - At the for-each loop at line 211: None of the exceptions should be happening as the responses have been pre-validated (i.e., `log.severe("Encountered exception when creating response: " + e.getMessage(), e)` when encountering exception).
| non_priority | execute logs exception while creating response at the for each loop at line none of the exceptions should be happening as the responses have been pre validated i e log severe encountered exception when creating response e getmessage e when encountering exception | 0 |
432,141 | 12,489,646,133 | IssuesEvent | 2020-05-31 19:50:03 | OpenMined/PyDP | https://api.github.com/repos/OpenMined/PyDP | closed | Provide version output | Good first issue :mortar_board: Priority: 2 - High :cold_sweat: Severity: 3 - Medium :unamused: Status: Available :wave: Type: New Feature :heavy_plus_sign: | ## Feature Description
I would like to get the version of the package when I run `pydp.__version__`
## Is your feature request related to a problem?
Is common practice to provide a way to access the version of each package after importing it using .\_\_version\_\_ after the name of the imported pa
## What alternatives have you considered?
You can look into the version of the package from outside the python prompt | 1.0 | Provide version output - ## Feature Description
I would like to get the version of the package when I run `pydp.__version__`
## Is your feature request related to a problem?
Is common practice to provide a way to access the version of each package after importing it using .\_\_version\_\_ after the name of the imported pa
## What alternatives have you considered?
You can look into the version of the package from outside the python prompt | priority | provide version output feature description i would like to get the version of the package when i run pydp version is your feature request related to a problem is common practice to provide a way to access the version of each package after importing it using version after the name of the imported pa what alternatives have you considered you can look into the version of the package from outside the python prompt | 1 |
630,491 | 20,110,978,364 | IssuesEvent | 2022-02-07 15:04:00 | oslopride/skeivt-kulturaar | https://api.github.com/repos/oslopride/skeivt-kulturaar | opened | Flytt filter-tags til å være under tekst "om arrangement". | priority-medium | Behold type arrangement over CTA-knapp.
Men tags som går på filter (universell utforming, tegnespråktolk og wc) kan god legges under teksten når man ekspanderer kortet.
<img width="414" alt="Skjermbilde 2022-02-07 kl 16 02 41" src="https://user-images.githubusercontent.com/93185620/152813614-a2bc8142-77e1-4e83-b245-64dd434277bb.png">
| 1.0 | Flytt filter-tags til å være under tekst "om arrangement". - Behold type arrangement over CTA-knapp.
Men tags som går på filter (universell utforming, tegnespråktolk og wc) kan god legges under teksten når man ekspanderer kortet.
<img width="414" alt="Skjermbilde 2022-02-07 kl 16 02 41" src="https://user-images.githubusercontent.com/93185620/152813614-a2bc8142-77e1-4e83-b245-64dd434277bb.png">
| priority | flytt filter tags til å være under tekst om arrangement behold type arrangement over cta knapp men tags som går på filter universell utforming tegnespråktolk og wc kan god legges under teksten når man ekspanderer kortet img width alt skjermbilde kl src | 1 |
20,524 | 30,443,419,911 | IssuesEvent | 2023-07-15 11:03:50 | SuperPWA/Super-Progressive-Web-Apps | https://api.github.com/repos/SuperPWA/Super-Progressive-Web-Apps | closed | Conflcit with the bunny.net plugin | high Compatibility feature | While using the bunny.net plugin, the Add to Home Screen banner is not showing, and it shows the error and a warning in the manifest.
Ref screenshoot: https://prnt.sc/0a9pXCnQjmE3
Ref Ticket: https://magazine3.in/conversation/154649?folder_id=29 | True | Conflcit with the bunny.net plugin - While using the bunny.net plugin, the Add to Home Screen banner is not showing, and it shows the error and a warning in the manifest.
Ref screenshoot: https://prnt.sc/0a9pXCnQjmE3
Ref Ticket: https://magazine3.in/conversation/154649?folder_id=29 | non_priority | conflcit with the bunny net plugin while using the bunny net plugin the add to home screen banner is not showing and it shows the error and a warning in the manifest ref screenshoot ref ticket | 0 |
98,436 | 20,733,152,961 | IssuesEvent | 2022-03-14 11:18:13 | Regalis11/Barotrauma | https://api.github.com/repos/Regalis11/Barotrauma | closed | multiplayer campaign keeps crashing | Bug Need more info Code Networking | *Sent by WordPress (issue@barotraumagame.com). Created by [fire](https://fire.fundersclub.com/).*
---
From: plaque
**Description**
everytime i start the campaign in multiplayer it crashes and sends me to the server browsing idk why it just says ||| error while reading a message from server (object reference not set to an instance of an object. void updateloadmenu |||
**Steps To Reproduce**
just click campaign gamemode when hosting a server
**Version**
0.15.13.0
**Operating System**
Windows
--
This issue was sent from a contact form on https://barotraumagame.com/bug
| 1.0 | multiplayer campaign keeps crashing - *Sent by WordPress (issue@barotraumagame.com). Created by [fire](https://fire.fundersclub.com/).*
---
From: plaque
**Description**
everytime i start the campaign in multiplayer it crashes and sends me to the server browsing idk why it just says ||| error while reading a message from server (object reference not set to an instance of an object. void updateloadmenu |||
**Steps To Reproduce**
just click campaign gamemode when hosting a server
**Version**
0.15.13.0
**Operating System**
Windows
--
This issue was sent from a contact form on https://barotraumagame.com/bug
| non_priority | multiplayer campaign keeps crashing sent by wordpress issue barotraumagame com created by from plaque description everytime i start the campaign in multiplayer it crashes and sends me to the server browsing idk why it just says error while reading a message from server object reference not set to an instance of an object void updateloadmenu steps to reproduce just click campaign gamemode when hosting a server version operating system windows this issue was sent from a contact form on | 0 |
82,384 | 3,606,234,698 | IssuesEvent | 2016-02-04 10:20:15 | molgenis/molgenis | https://api.github.com/repos/molgenis/molgenis | closed | Table attribute expand not working if attribute in compound | 1.10.0 1.11.0 1.17.0-RC 1.9.0 bug molgenis-dataexplorer priority-first | - Load Chr6 questionnair:
https://drive.google.com/open?id=0B4_AOpOI8V7GZV8wSll6emxjYW8
- Goto the questionnair plugin and fill (part of it) in.
- Goto the dataexplorer
- Clcik on a 'expand column' button in a column header
Expected: Column is expanded.
Observed: nothing happens

| 1.0 | Table attribute expand not working if attribute in compound - - Load Chr6 questionnair:
https://drive.google.com/open?id=0B4_AOpOI8V7GZV8wSll6emxjYW8
- Goto the questionnair plugin and fill (part of it) in.
- Goto the dataexplorer
- Clcik on a 'expand column' button in a column header
Expected: Column is expanded.
Observed: nothing happens

| priority | table attribute expand not working if attribute in compound load questionnair goto the questionnair plugin and fill part of it in goto the dataexplorer clcik on a expand column button in a column header expected column is expanded observed nothing happens | 1 |
118,438 | 25,307,929,088 | IssuesEvent | 2022-11-17 15:24:29 | sourcegraph/sourcegraph | https://api.github.com/repos/sourcegraph/sourcegraph | closed | [Accessibility Audit] Code Intel: Inspecting LSIF uploads for a repository | team/code-intelligence accessibility wcag/2.1/auditing wcag/2.1 | ### Steps to replicate journey
Taken from the [user journey list document](https://docs.google.com/document/d/1kA6aVOAgID_uPm-d6uEC1DG6WHjJ8fPT0UAXQcML_KQ/edit?usp=sharing). Use this for further context.
- Navigate to a repository in Sourcegraph.
- Click on the 🧠 Code Intelligence button.
- Scroll through the list of uploads.
- Optional: Adjust Upload State filter.
- Optional: Search uploads for specific uploads.
- Click on > next to an upload to inspect it.
- Potential options:
- Click the 🗑 Delete upload button.
- Expand Dependencies section
- Optional: Show dependents.
- Optional: Search dependencies.
- Click on > next to an upload to inspect it.
- Expand Retention overview section.
- Optional: Click Show matching only.
- Optional: Type filter in “Search Matches…”
- Click on a retention policy.
### How to audit
Follow the instructions here: [Auditing a user journey](https://docs.sourcegraph.com/dev/background-information/web/accessibility/how-to-audit#auditing-a-user-journey)
**Note:** We are planning on arranging thorough third-party audit in the future, so our focus here is just to find the _majority_ of accessibility issues. Don't worry if you feel you haven't found 100% of potential issues, it is better to focus on the core essentials to complete the journey rather than spending lots of time going through every possible problem.
### How to raise discovered problems
Follow the instructions here: [Raising an accessibility bug](https://docs.sourcegraph.com/dev/background-information/web/accessibility/how-to-audit#raising-a-bug)
### Any other questions?
See the [audit FAQ](https://docs.google.com/document/d/1nCi7TvYRpj9eG0_QtgqK1c4L28EiJKVuxyuvg1lAGgo/edit#) - add your question if it isn't there already! | 1.0 | [Accessibility Audit] Code Intel: Inspecting LSIF uploads for a repository - ### Steps to replicate journey
Taken from the [user journey list document](https://docs.google.com/document/d/1kA6aVOAgID_uPm-d6uEC1DG6WHjJ8fPT0UAXQcML_KQ/edit?usp=sharing). Use this for further context.
- Navigate to a repository in Sourcegraph.
- Click on the 🧠 Code Intelligence button.
- Scroll through the list of uploads.
- Optional: Adjust Upload State filter.
- Optional: Search uploads for specific uploads.
- Click on > next to an upload to inspect it.
- Potential options:
- Click the 🗑 Delete upload button.
- Expand Dependencies section
- Optional: Show dependents.
- Optional: Search dependencies.
- Click on > next to an upload to inspect it.
- Expand Retention overview section.
- Optional: Click Show matching only.
- Optional: Type filter in “Search Matches…”
- Click on a retention policy.
### How to audit
Follow the instructions here: [Auditing a user journey](https://docs.sourcegraph.com/dev/background-information/web/accessibility/how-to-audit#auditing-a-user-journey)
**Note:** We are planning on arranging thorough third-party audit in the future, so our focus here is just to find the _majority_ of accessibility issues. Don't worry if you feel you haven't found 100% of potential issues, it is better to focus on the core essentials to complete the journey rather than spending lots of time going through every possible problem.
### How to raise discovered problems
Follow the instructions here: [Raising an accessibility bug](https://docs.sourcegraph.com/dev/background-information/web/accessibility/how-to-audit#raising-a-bug)
### Any other questions?
See the [audit FAQ](https://docs.google.com/document/d/1nCi7TvYRpj9eG0_QtgqK1c4L28EiJKVuxyuvg1lAGgo/edit#) - add your question if it isn't there already! | non_priority | code intel inspecting lsif uploads for a repository steps to replicate journey taken from the use this for further context navigate to a repository in sourcegraph click on the 🧠 code intelligence button scroll through the list of uploads optional adjust upload state filter optional search uploads for specific uploads click on next to an upload to inspect it potential options click the 🗑 delete upload button expand dependencies section optional show dependents optional search dependencies click on next to an upload to inspect it expand retention overview section optional click show matching only optional type filter in “search matches…” click on a retention policy how to audit follow the instructions here note we are planning on arranging thorough third party audit in the future so our focus here is just to find the majority of accessibility issues don t worry if you feel you haven t found of potential issues it is better to focus on the core essentials to complete the journey rather than spending lots of time going through every possible problem how to raise discovered problems follow the instructions here any other questions see the add your question if it isn t there already | 0 |
271,489 | 29,511,455,670 | IssuesEvent | 2023-06-04 01:01:35 | AttestationLegale/api-pdf-generator | https://api.github.com/repos/AttestationLegale/api-pdf-generator | closed | spring-cloud-starter-openfeign-2.0.0.RC1.jar: 11 vulnerabilities (highest severity is: 9.1) - autoclosed | Mend: dependency security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-cloud-starter-openfeign-2.0.0.RC1.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (spring-cloud-starter-openfeign version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2019-20445](https://www.mend.io/vulnerability-database/CVE-2019-20445) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.1 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2019-20444](https://www.mend.io/vulnerability-database/CVE-2019-20444) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.1 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2020-7238](https://www.mend.io/vulnerability-database/CVE-2020-7238) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2020-11612](https://www.mend.io/vulnerability-database/CVE-2020-11612) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2021-37136](https://www.mend.io/vulnerability-database/CVE-2021-37136) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2021-37137](https://www.mend.io/vulnerability-database/CVE-2021-37137) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2019-16869](https://www.mend.io/vulnerability-database/CVE-2019-16869) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2021-43797](https://www.mend.io/vulnerability-database/CVE-2021-43797) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2021-21295](https://www.mend.io/vulnerability-database/CVE-2021-21295) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.9 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2022-24823](https://www.mend.io/vulnerability-database/CVE-2022-24823) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.5 | netty-common-4.1.24.Final.jar | Transitive | N/A* | ❌ |
| [CVE-2021-21290](https://www.mend.io/vulnerability-database/CVE-2021-21290) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.5 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2019-20445</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
HttpObjectDecoder.java in Netty before 4.1.44 allows a Content-Length header to be accompanied by a second Content-Length header, or by a Transfer-Encoding header.
<p>Publish Date: 2020-01-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20445>CVE-2019-20445</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-20445">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-20445</a></p>
<p>Release Date: 2020-01-29</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.44.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2019-20444</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
HttpObjectDecoder.java in Netty before 4.1.44 allows an HTTP header that lacks a colon, which might be interpreted as a separate header with an incorrect syntax, or might be interpreted as an "invalid fold."
<p>Publish Date: 2020-01-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20444>CVE-2019-20444</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-20444">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-20444</a></p>
<p>Release Date: 2020-01-29</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.44.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2020-7238</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869.
<p>Publish Date: 2020-01-27
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7238>CVE-2020-7238</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-01-27</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.44.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2020-11612</summary>
### Vulnerable Library - <b>netty-codec-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec/4.1.24.Final/netty-codec-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- netty-codec-http-4.1.24.Final.jar
- :x: **netty-codec-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The ZlibDecoders in Netty 4.1.x before 4.1.46 allow for unbounded memory allocation while decoding a ZlibEncoded byte stream. An attacker could send a large ZlibEncoded byte stream to the Netty server, forcing the server to allocate all of its free memory to a single decoder.
<p>Publish Date: 2020-04-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11612>CVE-2020-11612</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://netty.io/news/2020/02/28/4-1-46-Final.html">https://netty.io/news/2020/02/28/4-1-46-Final.html</a></p>
<p>Release Date: 2020-04-07</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.46.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2021-37136</summary>
### Vulnerable Library - <b>netty-codec-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec/4.1.24.Final/netty-codec-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- netty-codec-http-4.1.24.Final.jar
- :x: **netty-codec-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Bzip2 decompression decoder function doesn't allow setting size restrictions on the decompressed output data (which affects the allocation size used during decompression). All users of Bzip2Decoder are affected. The malicious input can trigger an OOME and so a DoS attack
<p>Publish Date: 2021-10-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-37136>CVE-2021-37136</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-grg4-wf29-r9vv">https://github.com/netty/netty/security/advisories/GHSA-grg4-wf29-r9vv</a></p>
<p>Release Date: 2021-10-19</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.68.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2021-37137</summary>
### Vulnerable Library - <b>netty-codec-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec/4.1.24.Final/netty-codec-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- netty-codec-http-4.1.24.Final.jar
- :x: **netty-codec-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Snappy frame decoder function doesn't restrict the chunk length which may lead to excessive memory usage. Beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well. This vulnerability can be triggered by supplying malicious input that decompresses to a very big size (via a network stream or a file) or by sending a huge skippable chunk.
<p>Publish Date: 2021-10-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-37137>CVE-2021-37137</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9vjp-v76f-g363">https://github.com/advisories/GHSA-9vjp-v76f-g363</a></p>
<p>Release Date: 2021-10-19</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.68.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2019-16869</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty before 4.1.42.Final mishandles whitespace before the colon in HTTP headers (such as a "Transfer-Encoding : chunked" line), which leads to HTTP request smuggling.
<p>Publish Date: 2019-09-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-16869>CVE-2019-16869</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16869">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16869</a></p>
<p>Release Date: 2019-09-26</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.42.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-43797</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. Netty prior to version 4.1.71.Final skips control chars when they are present at the beginning / end of the header name. It should instead fail fast as these are not allowed by the spec and could lead to HTTP request smuggling. Failing to do the validation might cause netty to "sanitize" header names before it forward these to another remote system when used as proxy. This remote system can't see the invalid usage anymore, and therefore does not do the validation itself. Users should upgrade to version 4.1.71.Final.
Mend Note: After conducting further research, Mend has determined that all versions of netty up to version 4.1.71.Final are vulnerable to CVE-2021-43797.
<p>Publish Date: 2021-12-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-43797>CVE-2021-43797</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="CVE-2021-43797">CVE-2021-43797</a></p>
<p>Release Date: 2021-12-09</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.71.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-21295</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel's pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`.
<p>Publish Date: 2021-03-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-21295>CVE-2021-21295</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-wm47-8v5p-wjpj">https://github.com/advisories/GHSA-wm47-8v5p-wjpj</a></p>
<p>Release Date: 2021-03-09</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.60.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2022-24823</summary>
### Vulnerable Library - <b>netty-common-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-common/4.1.24.Final/netty-common-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- netty-transport-native-epoll-4.1.24.Final.jar
- :x: **netty-common-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user.
<p>Publish Date: 2022-05-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24823>CVE-2022-24823</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p>
<p>Release Date: 2022-05-06</p>
<p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-21290</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty before version 4.1.59.Final there is a vulnerability on Unix-like systems involving an insecure temp file. When netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. On unix-like systems, the temporary directory is shared between all user. As such, writing to this directory using APIs that do not explicitly set the file/directory permissions can lead to information disclosure. Of note, this does not impact modern MacOS Operating Systems. The method "File.createTempFile" on unix-like systems creates a random file, but, by default will create this file with the permissions "-rw-r--r--". Thus, if sensitive information is written to this file, other local users can read this information. This is the case in netty's "AbstractDiskHttpData" is vulnerable. This has been fixed in version 4.1.59.Final. As a workaround, one may specify your own "java.io.tmpdir" when you start the JVM or use "DefaultHttpDataFactory.setBaseDir(...)" to set the directory to something that is only readable by the current user.
<p>Publish Date: 2021-02-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-21290>CVE-2021-21290</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-5mcr-gq6c-3hq2">https://github.com/netty/netty/security/advisories/GHSA-5mcr-gq6c-3hq2</a></p>
<p>Release Date: 2021-02-08</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.59.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | spring-cloud-starter-openfeign-2.0.0.RC1.jar: 11 vulnerabilities (highest severity is: 9.1) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-cloud-starter-openfeign-2.0.0.RC1.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (spring-cloud-starter-openfeign version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2019-20445](https://www.mend.io/vulnerability-database/CVE-2019-20445) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.1 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2019-20444](https://www.mend.io/vulnerability-database/CVE-2019-20444) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.1 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2020-7238](https://www.mend.io/vulnerability-database/CVE-2020-7238) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2020-11612](https://www.mend.io/vulnerability-database/CVE-2020-11612) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2021-37136](https://www.mend.io/vulnerability-database/CVE-2021-37136) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2021-37137](https://www.mend.io/vulnerability-database/CVE-2021-37137) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2019-16869](https://www.mend.io/vulnerability-database/CVE-2019-16869) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2021-43797](https://www.mend.io/vulnerability-database/CVE-2021-43797) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2021-21295](https://www.mend.io/vulnerability-database/CVE-2021-21295) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.9 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
| [CVE-2022-24823](https://www.mend.io/vulnerability-database/CVE-2022-24823) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.5 | netty-common-4.1.24.Final.jar | Transitive | N/A* | ❌ |
| [CVE-2021-21290](https://www.mend.io/vulnerability-database/CVE-2021-21290) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.5 | netty-codec-http-4.1.24.Final.jar | Transitive | 2.0.0.RELEASE | ✅ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the "Details" section below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2019-20445</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
HttpObjectDecoder.java in Netty before 4.1.44 allows a Content-Length header to be accompanied by a second Content-Length header, or by a Transfer-Encoding header.
<p>Publish Date: 2020-01-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20445>CVE-2019-20445</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-20445">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-20445</a></p>
<p>Release Date: 2020-01-29</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.44.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2019-20444</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
HttpObjectDecoder.java in Netty before 4.1.44 allows an HTTP header that lacks a colon, which might be interpreted as a separate header with an incorrect syntax, or might be interpreted as an "invalid fold."
<p>Publish Date: 2020-01-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-20444>CVE-2019-20444</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-20444">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-20444</a></p>
<p>Release Date: 2020-01-29</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.44.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2020-7238</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty 4.1.43.Final allows HTTP Request Smuggling because it mishandles Transfer-Encoding whitespace (such as a [space]Transfer-Encoding:chunked line) and a later Content-Length header. This issue exists because of an incomplete fix for CVE-2019-16869.
<p>Publish Date: 2020-01-27
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7238>CVE-2020-7238</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-01-27</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.44.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2020-11612</summary>
### Vulnerable Library - <b>netty-codec-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec/4.1.24.Final/netty-codec-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- netty-codec-http-4.1.24.Final.jar
- :x: **netty-codec-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The ZlibDecoders in Netty 4.1.x before 4.1.46 allow for unbounded memory allocation while decoding a ZlibEncoded byte stream. An attacker could send a large ZlibEncoded byte stream to the Netty server, forcing the server to allocate all of its free memory to a single decoder.
<p>Publish Date: 2020-04-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-11612>CVE-2020-11612</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://netty.io/news/2020/02/28/4-1-46-Final.html">https://netty.io/news/2020/02/28/4-1-46-Final.html</a></p>
<p>Release Date: 2020-04-07</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.46.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2021-37136</summary>
### Vulnerable Library - <b>netty-codec-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec/4.1.24.Final/netty-codec-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- netty-codec-http-4.1.24.Final.jar
- :x: **netty-codec-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Bzip2 decompression decoder function doesn't allow setting size restrictions on the decompressed output data (which affects the allocation size used during decompression). All users of Bzip2Decoder are affected. The malicious input can trigger an OOME and so a DoS attack
<p>Publish Date: 2021-10-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-37136>CVE-2021-37136</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-grg4-wf29-r9vv">https://github.com/netty/netty/security/advisories/GHSA-grg4-wf29-r9vv</a></p>
<p>Release Date: 2021-10-19</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.68.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2021-37137</summary>
### Vulnerable Library - <b>netty-codec-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec/4.1.24.Final/netty-codec-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- netty-codec-http-4.1.24.Final.jar
- :x: **netty-codec-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Snappy frame decoder function doesn't restrict the chunk length which may lead to excessive memory usage. Beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well. This vulnerability can be triggered by supplying malicious input that decompresses to a very big size (via a network stream or a file) or by sending a huge skippable chunk.
<p>Publish Date: 2021-10-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-37137>CVE-2021-37137</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9vjp-v76f-g363">https://github.com/advisories/GHSA-9vjp-v76f-g363</a></p>
<p>Release Date: 2021-10-19</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.68.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2019-16869</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty before 4.1.42.Final mishandles whitespace before the colon in HTTP headers (such as a "Transfer-Encoding : chunked" line), which leads to HTTP request smuggling.
<p>Publish Date: 2019-09-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-16869>CVE-2019-16869</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16869">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16869</a></p>
<p>Release Date: 2019-09-26</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.42.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-43797</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. Netty prior to version 4.1.71.Final skips control chars when they are present at the beginning / end of the header name. It should instead fail fast as these are not allowed by the spec and could lead to HTTP request smuggling. Failing to do the validation might cause netty to "sanitize" header names before it forward these to another remote system when used as proxy. This remote system can't see the invalid usage anymore, and therefore does not do the validation itself. Users should upgrade to version 4.1.71.Final.
Mend Note: After conducting further research, Mend has determined that all versions of netty up to version 4.1.71.Final are vulnerable to CVE-2021-43797.
<p>Publish Date: 2021-12-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-43797>CVE-2021-43797</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="CVE-2021-43797">CVE-2021-43797</a></p>
<p>Release Date: 2021-12-09</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.71.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-21295</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty (io.netty:netty-codec-http2) before version 4.1.60.Final there is a vulnerability that enables request smuggling. If a Content-Length header is present in the original HTTP/2 request, the field is not validated by `Http2MultiplexHandler` as it is propagated up. This is fine as long as the request is not proxied through as HTTP/1.1. If the request comes in as an HTTP/2 stream, gets converted into the HTTP/1.1 domain objects (`HttpRequest`, `HttpContent`, etc.) via `Http2StreamFrameToHttpObjectCodec `and then sent up to the child channel's pipeline and proxied through a remote peer as HTTP/1.1 this may result in request smuggling. In a proxy case, users may assume the content-length is validated somehow, which is not the case. If the request is forwarded to a backend channel that is a HTTP/1.1 connection, the Content-Length now has meaning and needs to be checked. An attacker can smuggle requests inside the body as it gets downgraded from HTTP/2 to HTTP/1.1. For an example attack refer to the linked GitHub Advisory. Users are only affected if all of this is true: `HTTP2MultiplexCodec` or `Http2FrameCodec` is used, `Http2StreamFrameToHttpObjectCodec` is used to convert to HTTP/1.1 objects, and these HTTP/1.1 objects are forwarded to another remote peer. This has been patched in 4.1.60.Final As a workaround, the user can do the validation by themselves by implementing a custom `ChannelInboundHandler` that is put in the `ChannelPipeline` behind `Http2StreamFrameToHttpObjectCodec`.
<p>Publish Date: 2021-03-09
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-21295>CVE-2021-21295</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-wm47-8v5p-wjpj">https://github.com/advisories/GHSA-wm47-8v5p-wjpj</a></p>
<p>Release Date: 2021-03-09</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.60.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2022-24823</summary>
### Vulnerable Library - <b>netty-common-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-common/4.1.24.Final/netty-common-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- netty-transport-native-epoll-4.1.24.Final.jar
- :x: **netty-common-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user.
<p>Publish Date: 2022-05-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24823>CVE-2022-24823</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p>
<p>Release Date: 2022-05-06</p>
<p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2021-21290</summary>
### Vulnerable Library - <b>netty-codec-http-4.1.24.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p>
<p>Path to dependency file: /api-pdf-generator/pom.xml</p>
<p>Path to vulnerable library: /home/abrard/.m2/repository/io/netty/netty-codec-http/4.1.24.Final/netty-codec-http-4.1.24.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RC1.jar (Root Library)
- spring-cloud-starter-netflix-ribbon-2.0.0.RC1.jar
- ribbon-2.2.5.jar
- rxnetty-0.4.9.jar
- :x: **netty-codec-http-4.1.24.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty before version 4.1.59.Final there is a vulnerability on Unix-like systems involving an insecure temp file. When netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. On unix-like systems, the temporary directory is shared between all user. As such, writing to this directory using APIs that do not explicitly set the file/directory permissions can lead to information disclosure. Of note, this does not impact modern MacOS Operating Systems. The method "File.createTempFile" on unix-like systems creates a random file, but, by default will create this file with the permissions "-rw-r--r--". Thus, if sensitive information is written to this file, other local users can read this information. This is the case in netty's "AbstractDiskHttpData" is vulnerable. This has been fixed in version 4.1.59.Final. As a workaround, one may specify your own "java.io.tmpdir" when you start the JVM or use "DefaultHttpDataFactory.setBaseDir(...)" to set the directory to something that is only readable by the current user.
<p>Publish Date: 2021-02-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-21290>CVE-2021-21290</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-5mcr-gq6c-3hq2">https://github.com/netty/netty/security/advisories/GHSA-5mcr-gq6c-3hq2</a></p>
<p>Release Date: 2021-02-08</p>
<p>Fix Resolution (io.netty:netty-codec-http): 4.1.59.Final</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 2.0.0.RELEASE</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_priority | spring cloud starter openfeign jar vulnerabilities highest severity is autoclosed vulnerable library spring cloud starter openfeign jar path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec http final netty codec http final jar vulnerabilities cve severity cvss dependency type fixed in spring cloud starter openfeign version remediation available critical netty codec http final jar transitive release critical netty codec http final jar transitive release high netty codec http final jar transitive release high netty codec final jar transitive release high netty codec final jar transitive release high netty codec final jar transitive release high netty codec http final jar transitive release medium netty codec http final jar transitive release medium netty codec http final jar transitive release medium netty common final jar transitive n a medium netty codec http final jar transitive release for some transitive vulnerabilities there is no version of direct dependency with a fix check the details section below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec http final netty codec http final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar x netty codec http final jar vulnerable library found in base branch master vulnerability details httpobjectdecoder java in netty before allows a content length header to be accompanied by a second content length header or by a transfer encoding header publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec http final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec http final netty codec http final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar x netty codec http final jar vulnerable library found in base branch master vulnerability details httpobjectdecoder java in netty before allows an http header that lacks a colon which might be interpreted as a separate header with an incorrect syntax or might be interpreted as an invalid fold publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec http final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec http final netty codec http final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar x netty codec http final jar vulnerable library found in base branch master vulnerability details netty final allows http request smuggling because it mishandles transfer encoding whitespace such as a transfer encoding chunked line and a later content length header this issue exists because of an incomplete fix for cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution io netty netty codec http final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty codec final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec final netty codec final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar netty codec http final jar x netty codec final jar vulnerable library found in base branch master vulnerability details the zlibdecoders in netty x before allow for unbounded memory allocation while decoding a zlibencoded byte stream an attacker could send a large zlibencoded byte stream to the netty server forcing the server to allocate all of its free memory to a single decoder publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty codec final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec final netty codec final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar netty codec http final jar x netty codec final jar vulnerable library found in base branch master vulnerability details the decompression decoder function doesn t allow setting size restrictions on the decompressed output data which affects the allocation size used during decompression all users of are affected the malicious input can trigger an oome and so a dos attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty codec final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec final netty codec final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar netty codec http final jar x netty codec final jar vulnerable library found in base branch master vulnerability details the snappy frame decoder function doesn t restrict the chunk length which may lead to excessive memory usage beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well this vulnerability can be triggered by supplying malicious input that decompresses to a very big size via a network stream or a file or by sending a huge skippable chunk publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec http final netty codec http final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar x netty codec http final jar vulnerable library found in base branch master vulnerability details netty before final mishandles whitespace before the colon in http headers such as a transfer encoding chunked line which leads to http request smuggling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec http final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec http final netty codec http final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar x netty codec http final jar vulnerable library found in base branch master vulnerability details netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients netty prior to version final skips control chars when they are present at the beginning end of the header name it should instead fail fast as these are not allowed by the spec and could lead to http request smuggling failing to do the validation might cause netty to sanitize header names before it forward these to another remote system when used as proxy this remote system can t see the invalid usage anymore and therefore does not do the validation itself users should upgrade to version final mend note after conducting further research mend has determined that all versions of netty up to version final are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin cve release date fix resolution io netty netty codec http final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec http final netty codec http final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar x netty codec http final jar vulnerable library found in base branch master vulnerability details netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty io netty netty codec before version final there is a vulnerability that enables request smuggling if a content length header is present in the original http request the field is not validated by as it is propagated up this is fine as long as the request is not proxied through as http if the request comes in as an http stream gets converted into the http domain objects httprequest httpcontent etc via and then sent up to the child channel s pipeline and proxied through a remote peer as http this may result in request smuggling in a proxy case users may assume the content length is validated somehow which is not the case if the request is forwarded to a backend channel that is a http connection the content length now has meaning and needs to be checked an attacker can smuggle requests inside the body as it gets downgraded from http to http for an example attack refer to the linked github advisory users are only affected if all of this is true or is used is used to convert to http objects and these http objects are forwarded to another remote peer this has been patched in final as a workaround the user can do the validation by themselves by implementing a custom channelinboundhandler that is put in the channelpipeline behind publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec http final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue cve vulnerable library netty common final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty common final netty common final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar netty transport native epoll final jar x netty common final jar vulnerable library found in base branch master vulnerability details netty is an open source asynchronous event driven network application framework the package io netty netty codec http prior to version final contains an insufficient fix for cve when netty s multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled this only impacts applications running on java version and lower additionally this vulnerability impacts code running on unix like systems and very old versions of mac osx and windows as they all share the system temporary directory between all users version final contains a patch for this vulnerability as a workaround specify one s own java io tmpdir when starting the jvm or use defaulthttpdatafactory setbasedir to set the directory to something that is only readable by the current user publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all io netty netty common final cve vulnerable library netty codec http final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to dependency file api pdf generator pom xml path to vulnerable library home abrard repository io netty netty codec http final netty codec http final jar dependency hierarchy spring cloud starter openfeign jar root library spring cloud starter netflix ribbon jar ribbon jar rxnetty jar x netty codec http final jar vulnerable library found in base branch master vulnerability details netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty before version final there is a vulnerability on unix like systems involving an insecure temp file when netty s multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled on unix like systems the temporary directory is shared between all user as such writing to this directory using apis that do not explicitly set the file directory permissions can lead to information disclosure of note this does not impact modern macos operating systems the method file createtempfile on unix like systems creates a random file but by default will create this file with the permissions rw r r thus if sensitive information is written to this file other local users can read this information this is the case in netty s abstractdiskhttpdata is vulnerable this has been fixed in version final as a workaround one may specify your own java io tmpdir when you start the jvm or use defaulthttpdatafactory setbasedir to set the directory to something that is only readable by the current user publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec http final direct dependency fix resolution org springframework cloud spring cloud starter openfeign release rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
379,319 | 26,366,180,645 | IssuesEvent | 2023-01-11 16:44:47 | openreferral/specification | https://api.github.com/repos/openreferral/specification | opened | Update reference and guidance documentation | documentation | This is a holding issue for documenting changes to the structure and content of the HSDS documentation pages following the 3.0 upgrade. This will be broken down into smaller, individual isses. | 1.0 | Update reference and guidance documentation - This is a holding issue for documenting changes to the structure and content of the HSDS documentation pages following the 3.0 upgrade. This will be broken down into smaller, individual isses. | non_priority | update reference and guidance documentation this is a holding issue for documenting changes to the structure and content of the hsds documentation pages following the upgrade this will be broken down into smaller individual isses | 0 |
8,132 | 8,170,275,507 | IssuesEvent | 2018-08-27 07:02:29 | kyma-project/kyma | https://api.github.com/repos/kyma-project/kyma | opened | Enable graphql subscriptions in ui-api-layer for a number of types | area/service-catalog enhancement | Confirm these statements before you submit the issue:
- [x] I have searched open and closed issues for duplicates.
- [x] I have read the contributing guidelines.
---
**Description**
- Make sure `serviceInstanceEvent` works
- Once https://github.com/kyma-project/kyma/pull/180 is completed we need to add one parameter to UI API Virtual Service definition: `websocketUpgrade: true`
- Add 2 new subscriptions for ServiceBinding and ServiceBindingUsage that we will use in views where these resources are listed and their statuses exposed, in the ServiceInstance details view
- Add 2 new subscription for ServiceBrokers that will be used when listing ServiceBrokers and their status, the namespaced ones and the cluster ones.
**Reasons**
Backed must support websockets to enable UI to do realtime updates of changes in the resources, like their status changes
| 1.0 | Enable graphql subscriptions in ui-api-layer for a number of types - Confirm these statements before you submit the issue:
- [x] I have searched open and closed issues for duplicates.
- [x] I have read the contributing guidelines.
---
**Description**
- Make sure `serviceInstanceEvent` works
- Once https://github.com/kyma-project/kyma/pull/180 is completed we need to add one parameter to UI API Virtual Service definition: `websocketUpgrade: true`
- Add 2 new subscriptions for ServiceBinding and ServiceBindingUsage that we will use in views where these resources are listed and their statuses exposed, in the ServiceInstance details view
- Add 2 new subscription for ServiceBrokers that will be used when listing ServiceBrokers and their status, the namespaced ones and the cluster ones.
**Reasons**
Backed must support websockets to enable UI to do realtime updates of changes in the resources, like their status changes
| non_priority | enable graphql subscriptions in ui api layer for a number of types confirm these statements before you submit the issue i have searched open and closed issues for duplicates i have read the contributing guidelines description make sure serviceinstanceevent works once is completed we need to add one parameter to ui api virtual service definition websocketupgrade true add new subscriptions for servicebinding and servicebindingusage that we will use in views where these resources are listed and their statuses exposed in the serviceinstance details view add new subscription for servicebrokers that will be used when listing servicebrokers and their status the namespaced ones and the cluster ones reasons backed must support websockets to enable ui to do realtime updates of changes in the resources like their status changes | 0 |
788,438 | 27,753,082,979 | IssuesEvent | 2023-03-15 22:43:12 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | bt_conn_foreach() includes invalid connection while advertising | bug priority: low area: Bluetooth area: Bluetooth Host | **Describe the bug**
When advertising is active (testing on Nordic nRF52840 Dongle), `bt_conn_foreach()` appears to include a connection that's not yet established with address matching `BT_ADDR_LE_NONE`.
**To Reproduce**
```
static void example_callback(struct bt_conn *conn, void *data) {
const bt_addr_le_t *addr = bt_conn_get_dst(conn);
if (!bt_addr_le_cmp(addr, BT_ADDR_LE_NONE)) {
LOG_ERR("addr none");
} else {
LOG_INF("addr valid");
}
}
...
bt_conn_foreach(BT_CONN_TYPE_LE, example_callback, NULL);
...
```
**Expected behavior**
It's not clear from the documentation whether this is the expected behavior or if there's is a missing check within `bt_conn_foreach()`.
**Impact**
Workaround involves checking the address in the callback.
| 1.0 | bt_conn_foreach() includes invalid connection while advertising - **Describe the bug**
When advertising is active (testing on Nordic nRF52840 Dongle), `bt_conn_foreach()` appears to include a connection that's not yet established with address matching `BT_ADDR_LE_NONE`.
**To Reproduce**
```
static void example_callback(struct bt_conn *conn, void *data) {
const bt_addr_le_t *addr = bt_conn_get_dst(conn);
if (!bt_addr_le_cmp(addr, BT_ADDR_LE_NONE)) {
LOG_ERR("addr none");
} else {
LOG_INF("addr valid");
}
}
...
bt_conn_foreach(BT_CONN_TYPE_LE, example_callback, NULL);
...
```
**Expected behavior**
It's not clear from the documentation whether this is the expected behavior or if there's is a missing check within `bt_conn_foreach()`.
**Impact**
Workaround involves checking the address in the callback.
| priority | bt conn foreach includes invalid connection while advertising describe the bug when advertising is active testing on nordic dongle bt conn foreach appears to include a connection that s not yet established with address matching bt addr le none to reproduce static void example callback struct bt conn conn void data const bt addr le t addr bt conn get dst conn if bt addr le cmp addr bt addr le none log err addr none else log inf addr valid bt conn foreach bt conn type le example callback null expected behavior it s not clear from the documentation whether this is the expected behavior or if there s is a missing check within bt conn foreach impact workaround involves checking the address in the callback | 1 |
528,089 | 15,359,960,553 | IssuesEvent | 2021-03-01 16:24:37 | Plant-for-the-Planet-org/treemapper | https://api.github.com/repos/Plant-for-the-Planet-org/treemapper | closed | Fix Geolocation.getCurrentPosition implementation | Priority bug | The current usage of `Geolocation.getCurrentPosition` has to be check for iOS and Android. On Android explicitly asking for Permission to access the users location is missing as e.g. implemented here for storage permission:
https://github.com/Plant-for-the-Planet-org/treemapper/blob/9da024767c66485739bd15f37e70752804271b2e/app/components/InventoryOverview/index.js#L180-L203
For both Android and iOS all usage of `Geolocation.getCurrentPosition` should be checked for the error handling. Some usage still show debugging alert messages instead of messages suitable for the user to understand:
<img width="416" alt="Bildschirmfoto 2021-02-15 um 16 20 36" src="https://user-images.githubusercontent.com/1532418/107964960-3472e300-6faa-11eb-9ba2-b9d5fd8d0443.png">
Check out the documentation: https://github.com/react-native-geolocation/react-native-geolocation | 1.0 | Fix Geolocation.getCurrentPosition implementation - The current usage of `Geolocation.getCurrentPosition` has to be check for iOS and Android. On Android explicitly asking for Permission to access the users location is missing as e.g. implemented here for storage permission:
https://github.com/Plant-for-the-Planet-org/treemapper/blob/9da024767c66485739bd15f37e70752804271b2e/app/components/InventoryOverview/index.js#L180-L203
For both Android and iOS all usage of `Geolocation.getCurrentPosition` should be checked for the error handling. Some usage still show debugging alert messages instead of messages suitable for the user to understand:
<img width="416" alt="Bildschirmfoto 2021-02-15 um 16 20 36" src="https://user-images.githubusercontent.com/1532418/107964960-3472e300-6faa-11eb-9ba2-b9d5fd8d0443.png">
Check out the documentation: https://github.com/react-native-geolocation/react-native-geolocation | priority | fix geolocation getcurrentposition implementation the current usage of geolocation getcurrentposition has to be check for ios and android on android explicitly asking for permission to access the users location is missing as e g implemented here for storage permission for both android and ios all usage of geolocation getcurrentposition should be checked for the error handling some usage still show debugging alert messages instead of messages suitable for the user to understand img width alt bildschirmfoto um src check out the documentation | 1 |
71,867 | 13,756,558,857 | IssuesEvent | 2020-10-06 20:07:55 | quarkusio/quarkus | https://api.github.com/repos/quarkusio/quarkus | opened | Codestarts: Spring code examples aren't generated | area/codestarts kind/bug | When creating a project, if I select, e.g.,`spring-web` as the only extension then I get a command mode example project with the spring-web extension added in the pom.xml.
If I select the resteasy and spring-web extensions then I get a project with the JAX-RS code example but no Spring Web.
That's when the codestarts are enabled. If the codestarts are disabled then selecting Spring Web will generate the Spring Web example code. | 1.0 | Codestarts: Spring code examples aren't generated - When creating a project, if I select, e.g.,`spring-web` as the only extension then I get a command mode example project with the spring-web extension added in the pom.xml.
If I select the resteasy and spring-web extensions then I get a project with the JAX-RS code example but no Spring Web.
That's when the codestarts are enabled. If the codestarts are disabled then selecting Spring Web will generate the Spring Web example code. | non_priority | codestarts spring code examples aren t generated when creating a project if i select e g spring web as the only extension then i get a command mode example project with the spring web extension added in the pom xml if i select the resteasy and spring web extensions then i get a project with the jax rs code example but no spring web that s when the codestarts are enabled if the codestarts are disabled then selecting spring web will generate the spring web example code | 0 |
829,372 | 31,876,777,905 | IssuesEvent | 2023-09-16 00:15:02 | SparkDevNetwork/Rock | https://api.github.com/repos/SparkDevNetwork/Rock | closed | Internal API Error Issue with Sign Now (Manual Requests Through Document Templates Page) | Type: Bug Status: Confirmed Priority: Low | <!--
Found a security bug? DO NOT file an issue. Email [info@sparkdevnetwork.org](mailto:info@sparkdevnetwork.org) and we'll be in touch shortly.
Before completing your first issue we highly recommend reading this post [https://www.snoyman.com/blog/2017/10/effective-ways-help-from-maintainers](https://www.snoyman.com/blog/2017/10/effective-ways-help-from-maintainers)
-->
### Prerequisites
* [X ] Put an X between the brackets on this line if you have done all of the following:
* Can you reproduce the problem on a fresh install or the [[http://rock.rocksolidchurchdemo.com/](http://rock.rocksolidchurchdemo.com/) demo site ]()? (Demo site doesn't have a Sign Now account to test with -- I reached out to the community for confirmation that this is a widespread bug, but didn't get a response)
* Did you include your Rock version number and [[https://github.com/SparkDevNetwork/Rock/wiki/Environment-and-Diagnostics-Information](https://github.com/SparkDevNetwork/Rock/wiki/Environment-and-Diagnostics-Information) client culture ]() setting?
* Did you [[https://github.com/issues?q=is](https://github.com/issues?q=is) %3Aissue+user%3ASparkDevNetwork+-repo%3ASparkDevNetwork%2FSlack perform a cursory search ]() to see if your bug or enhancement is already reported?
### Description
A clear and concise description of what the bug is, include screenshots if applicable.
### Steps to Reproduce
1. Go to Admin Tools > General Settings > Signature Documents > Select a Template > Click the + button to manually send a new digital signature document / request
![<a href="https://user-images.githubusercontent.com/45953896/55923890-81bae800-5bbc-11e9-9dcf-d7ecbbb170d6.png">https://user-images.githubusercontent.com/45953896/55923890-81bae800-5bbc-11e9-9dcf-d7ecbbb170d6.png</a> image]()
1. Fill in the details
![<a href="https://user-images.githubusercontent.com/45953896/55923943-afa02c80-5bbc-11e9-93ee-d79a68b25e2a.png">https://user-images.githubusercontent.com/45953896/55923943-afa02c80-5bbc-11e9-93ee-d79a68b25e2a.png</a> image]()
1. Click "Save"
1. Find and edit the row of the new document request.
1. Click "Send Invite"
1. See error "Error Sending Signature invite" -- Internal API Error
![<a href="https://user-images.githubusercontent.com/45953896/55923967-cfcfeb80-5bbc-11e9-8ecd-5e53967ee592.png">https://user-images.githubusercontent.com/45953896/55923967-cfcfeb80-5bbc-11e9-8ecd-5e53967ee592.png</a> image]()
**Expected behavior:**
Normally, the signature document sends without any problem -- you get a green success div at the top informing you that the document has been sent, and you click Save. I sent close to 400 documents on 3/5/19 and those were all sent successfully. I have not updated my Rock server between the last time this functionality worked and now.
(Please note that signature documents being sent automatically through Registrations / Groups, etc. ARE working as expected.)
**Actual behavior:**
I first noticed this issue on 3/21/19, so I've already spent a lot of time trying to troubleshoot this with Sign Now between then and now. I have assumed that this issue was related to Sign Now and not Rock, because I got this exact error message when following these same steps at the time we first started using their API (shortly before 3/1/19), and they resolved the issue.
However, one of their API techs today said that the detailed error message they are seeing on their end states that the document is not being generated correctly, because the document ID is missing when the API is called. He said this error is occurring within Rock. Once that is resolved, everything should be golden. The API support tech that I spoke with gave me the call back number of (800) 366-0593 should it be needed.
### Versions
* **Rock Version:** 8.6
* **Client Culture Setting:** en-US
| 1.0 | Internal API Error Issue with Sign Now (Manual Requests Through Document Templates Page) - <!--
Found a security bug? DO NOT file an issue. Email [info@sparkdevnetwork.org](mailto:info@sparkdevnetwork.org) and we'll be in touch shortly.
Before completing your first issue we highly recommend reading this post [https://www.snoyman.com/blog/2017/10/effective-ways-help-from-maintainers](https://www.snoyman.com/blog/2017/10/effective-ways-help-from-maintainers)
-->
### Prerequisites
* [X ] Put an X between the brackets on this line if you have done all of the following:
* Can you reproduce the problem on a fresh install or the [[http://rock.rocksolidchurchdemo.com/](http://rock.rocksolidchurchdemo.com/) demo site ]()? (Demo site doesn't have a Sign Now account to test with -- I reached out to the community for confirmation that this is a widespread bug, but didn't get a response)
* Did you include your Rock version number and [[https://github.com/SparkDevNetwork/Rock/wiki/Environment-and-Diagnostics-Information](https://github.com/SparkDevNetwork/Rock/wiki/Environment-and-Diagnostics-Information) client culture ]() setting?
* Did you [[https://github.com/issues?q=is](https://github.com/issues?q=is) %3Aissue+user%3ASparkDevNetwork+-repo%3ASparkDevNetwork%2FSlack perform a cursory search ]() to see if your bug or enhancement is already reported?
### Description
A clear and concise description of what the bug is, include screenshots if applicable.
### Steps to Reproduce
1. Go to Admin Tools > General Settings > Signature Documents > Select a Template > Click the + button to manually send a new digital signature document / request
![<a href="https://user-images.githubusercontent.com/45953896/55923890-81bae800-5bbc-11e9-9dcf-d7ecbbb170d6.png">https://user-images.githubusercontent.com/45953896/55923890-81bae800-5bbc-11e9-9dcf-d7ecbbb170d6.png</a> image]()
1. Fill in the details
![<a href="https://user-images.githubusercontent.com/45953896/55923943-afa02c80-5bbc-11e9-93ee-d79a68b25e2a.png">https://user-images.githubusercontent.com/45953896/55923943-afa02c80-5bbc-11e9-93ee-d79a68b25e2a.png</a> image]()
1. Click "Save"
1. Find and edit the row of the new document request.
1. Click "Send Invite"
1. See error "Error Sending Signature invite" -- Internal API Error
![<a href="https://user-images.githubusercontent.com/45953896/55923967-cfcfeb80-5bbc-11e9-8ecd-5e53967ee592.png">https://user-images.githubusercontent.com/45953896/55923967-cfcfeb80-5bbc-11e9-8ecd-5e53967ee592.png</a> image]()
**Expected behavior:**
Normally, the signature document sends without any problem -- you get a green success div at the top informing you that the document has been sent, and you click Save. I sent close to 400 documents on 3/5/19 and those were all sent successfully. I have not updated my Rock server between the last time this functionality worked and now.
(Please note that signature documents being sent automatically through Registrations / Groups, etc. ARE working as expected.)
**Actual behavior:**
I first noticed this issue on 3/21/19, so I've already spent a lot of time trying to troubleshoot this with Sign Now between then and now. I have assumed that this issue was related to Sign Now and not Rock, because I got this exact error message when following these same steps at the time we first started using their API (shortly before 3/1/19), and they resolved the issue.
However, one of their API techs today said that the detailed error message they are seeing on their end states that the document is not being generated correctly, because the document ID is missing when the API is called. He said this error is occurring within Rock. Once that is resolved, everything should be golden. The API support tech that I spoke with gave me the call back number of (800) 366-0593 should it be needed.
### Versions
* **Rock Version:** 8.6
* **Client Culture Setting:** en-US
| priority | internal api error issue with sign now manual requests through document templates page found a security bug do not file an issue email mailto info sparkdevnetwork org and we ll be in touch shortly before completing your first issue we highly recommend reading this post prerequisites put an x between the brackets on this line if you have done all of the following can you reproduce the problem on a fresh install or the demo site demo site doesn t have a sign now account to test with i reached out to the community for confirmation that this is a widespread bug but didn t get a response did you include your rock version number and client culture setting did you user repo perform a cursory search to see if your bug or enhancement is already reported description a clear and concise description of what the bug is include screenshots if applicable steps to reproduce go to admin tools general settings signature documents select a template click the button to manually send a new digital signature document request fill in the details click save find and edit the row of the new document request click send invite see error error sending signature invite internal api error expected behavior normally the signature document sends without any problem you get a green success div at the top informing you that the document has been sent and you click save i sent close to documents on and those were all sent successfully i have not updated my rock server between the last time this functionality worked and now please note that signature documents being sent automatically through registrations groups etc are working as expected actual behavior i first noticed this issue on so i ve already spent a lot of time trying to troubleshoot this with sign now between then and now i have assumed that this issue was related to sign now and not rock because i got this exact error message when following these same steps at the time we first started using their api shortly before and they resolved the issue however one of their api techs today said that the detailed error message they are seeing on their end states that the document is not being generated correctly because the document id is missing when the api is called he said this error is occurring within rock once that is resolved everything should be golden the api support tech that i spoke with gave me the call back number of should it be needed versions rock version client culture setting en us | 1 |
366,019 | 10,807,760,018 | IssuesEvent | 2019-11-07 09:09:50 | projectacrn/acrn-hypervisor | https://api.github.com/repos/projectacrn/acrn-hypervisor | closed | hv: remove registration of default port IO and MMIO handlers | priority: P2-High type: bug | The default behaviors of PIO & MMIO handlers are same for all VMs, no need to expose dedicated APIs to register default hanlders. | 1.0 | hv: remove registration of default port IO and MMIO handlers - The default behaviors of PIO & MMIO handlers are same for all VMs, no need to expose dedicated APIs to register default hanlders. | priority | hv remove registration of default port io and mmio handlers the default behaviors of pio mmio handlers are same for all vms no need to expose dedicated apis to register default hanlders | 1 |
810,241 | 30,232,553,875 | IssuesEvent | 2023-07-06 08:01:35 | Team-Ampersand/GKR-Android-Admin | https://api.github.com/repos/Team-Ampersand/GKR-Android-Admin | closed | (#82) 메인 화면 필터 클릭할 때 흰 화면 나오는 이슈 고치기 | 0️⃣ Priority: Critical 🐞 Type: Bug | ### Describe
- 필터 클릭 시 api result가 empty list인 경우 흰 화면이 나옴 | 1.0 | (#82) 메인 화면 필터 클릭할 때 흰 화면 나오는 이슈 고치기 - ### Describe
- 필터 클릭 시 api result가 empty list인 경우 흰 화면이 나옴 | priority | 메인 화면 필터 클릭할 때 흰 화면 나오는 이슈 고치기 describe 필터 클릭 시 api result가 empty list인 경우 흰 화면이 나옴 | 1 |
514,951 | 14,947,300,416 | IssuesEvent | 2021-01-26 08:26:14 | MAIF/daikoku | https://api.github.com/repos/MAIF/daikoku | closed | Tenant admin message issues | bug front-office priority-0 | - [ ] discussion component is displayed for tenant admin
- [ ] tenant admin don't have massage link in top bar
- [ ] tenant admin can't see messages, page is crashing | 1.0 | Tenant admin message issues - - [ ] discussion component is displayed for tenant admin
- [ ] tenant admin don't have massage link in top bar
- [ ] tenant admin can't see messages, page is crashing | priority | tenant admin message issues discussion component is displayed for tenant admin tenant admin don t have massage link in top bar tenant admin can t see messages page is crashing | 1 |
97,984 | 8,673,896,361 | IssuesEvent | 2018-11-30 04:53:45 | humera987/FXLabs-Test-Automation | https://api.github.com/repos/humera987/FXLabs-Test-Automation | closed | FXLabs Testing : ApiV1OrgsSearchGetQueryParamPageInvalidDatatype | FXLabs Testing | Project : FXLabs Testing
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NjBiOTMxMDMtY2IyMi00NzAxLWEyZTUtMzQzNWY1YjMwMzEx; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 30 Nov 2018 04:52:38 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/orgs/search?page=vP7Pkt
Request :
Response :
{
"timestamp" : "2018-11-30T04:52:39.142+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/orgs/search"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | 1.0 | FXLabs Testing : ApiV1OrgsSearchGetQueryParamPageInvalidDatatype - Project : FXLabs Testing
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NjBiOTMxMDMtY2IyMi00NzAxLWEyZTUtMzQzNWY1YjMwMzEx; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Fri, 30 Nov 2018 04:52:38 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/orgs/search?page=vP7Pkt
Request :
Response :
{
"timestamp" : "2018-11-30T04:52:39.142+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/orgs/search"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | non_priority | fxlabs testing project fxlabs testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api orgs search logs assertion resolved to result assertion resolved to result fx bot | 0 |
433,286 | 30,321,522,687 | IssuesEvent | 2023-07-10 19:38:05 | ArkhamX2/StudendaUI | https://api.github.com/repos/ArkhamX2/StudendaUI | opened | QA UX | bug documentation | Составление, планирование и проведение юзабилити тестирования и UX аудита
Составление отчёта о юзабилити-тестировании и UX аудите
Аналитика результатов | 1.0 | QA UX - Составление, планирование и проведение юзабилити тестирования и UX аудита
Составление отчёта о юзабилити-тестировании и UX аудите
Аналитика результатов | non_priority | qa ux составление планирование и проведение юзабилити тестирования и ux аудита составление отчёта о юзабилити тестировании и ux аудите аналитика результатов | 0 |
263,007 | 8,272,953,620 | IssuesEvent | 2018-09-17 01:58:13 | dileep-kishore/beeminder-sync | https://api.github.com/repos/dileep-kishore/beeminder-sync | closed | No module named 'orderedmultidict' | Priority: Critical Status: Review Needed Type: Bug | * Beeminder Sync version: 0.1.0
* Python version: 3.6.5
* Operating System: Manjaro Linux
### Description
Tried to run tests and got an error
### What I Did
```sh
pytest -sv --cov=beeminder_sync
```
Error:
ModuleNotFoundError: No module named 'orderedmultidict'
ERROR: could not load /home/dileep/Documents/Coding/100daysofcode/beeminder-sync/tests/conftest.py
| 1.0 | No module named 'orderedmultidict' - * Beeminder Sync version: 0.1.0
* Python version: 3.6.5
* Operating System: Manjaro Linux
### Description
Tried to run tests and got an error
### What I Did
```sh
pytest -sv --cov=beeminder_sync
```
Error:
ModuleNotFoundError: No module named 'orderedmultidict'
ERROR: could not load /home/dileep/Documents/Coding/100daysofcode/beeminder-sync/tests/conftest.py
| priority | no module named orderedmultidict beeminder sync version python version operating system manjaro linux description tried to run tests and got an error what i did sh pytest sv cov beeminder sync error modulenotfounderror no module named orderedmultidict error could not load home dileep documents coding beeminder sync tests conftest py | 1 |
113,993 | 4,599,866,006 | IssuesEvent | 2016-09-22 01:02:33 | thommoboy/There-are-no-brakes | https://api.github.com/repos/thommoboy/There-are-no-brakes | opened | Spikes are not very visible on the ancient level | Ancients enhancement Priority Low | Spikes are not very visible on the ancient level | 1.0 | Spikes are not very visible on the ancient level - Spikes are not very visible on the ancient level | priority | spikes are not very visible on the ancient level spikes are not very visible on the ancient level | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.