Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 844 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 12 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 248k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
273,087 | 20,769,868,765 | IssuesEvent | 2022-03-16 02:36:45 | Interacao-Humano-Computador/2021.2-Prefeitura-Verdelandia | https://api.github.com/repos/Interacao-Humano-Computador/2021.2-Prefeitura-Verdelandia | opened | Planejamento do storyboard | documentation | ### Descrição:
Issue direcionada a tratar do planejamento do storyboard
### Tarefas:
- [ ] Criar o documento de planejamento do Storyboard
### Critérios de aceitação:
- [ ] Coerência na documentação.
| 1.0 | Planejamento do storyboard - ### Descrição:
Issue direcionada a tratar do planejamento do storyboard
### Tarefas:
- [ ] Criar o documento de planejamento do Storyboard
### Critérios de aceitação:
- [ ] Coerência na documentação.
| non_priority | planejamento do storyboard descrição issue direcionada a tratar do planejamento do storyboard tarefas criar o documento de planejamento do storyboard critérios de aceitação coerência na documentação | 0 |
1,591 | 3,872,703,833 | IssuesEvent | 2016-04-11 14:42:49 | CARLI/vufind-test-repo | https://api.github.com/repos/CARLI/vufind-test-repo | closed | Rename "Storage Retrieval Request" | beta_requirement Major Task | When viewing an item in the catalog, the link to place a request currently says "Place a Storage Retrieval Request". This should be changed to "Request Item".
When viewing Your Account, the menu and heading read "Storage Retrieval Requests". This should be changed to "Requested Items". | 1.0 | Rename "Storage Retrieval Request" - When viewing an item in the catalog, the link to place a request currently says "Place a Storage Retrieval Request". This should be changed to "Request Item".
When viewing Your Account, the menu and heading read "Storage Retrieval Requests". This should be changed to "Requested Items". | non_priority | rename storage retrieval request when viewing an item in the catalog the link to place a request currently says place a storage retrieval request this should be changed to request item when viewing your account the menu and heading read storage retrieval requests this should be changed to requested items | 0 |
177,961 | 14,656,007,381 | IssuesEvent | 2020-12-28 12:24:48 | Schmiddl99/experimenteverwaltung-i2 | https://api.github.com/repos/Schmiddl99/experimenteverwaltung-i2 | closed | Erarbeitung "System wide requirements" nach TS-M2 (A1) DUE: 18.12.20 | documentation requirements | Synthese aus UCs, Vision und Infos aus TS-M1 | 1.0 | Erarbeitung "System wide requirements" nach TS-M2 (A1) DUE: 18.12.20 - Synthese aus UCs, Vision und Infos aus TS-M1 | non_priority | erarbeitung system wide requirements nach ts due synthese aus ucs vision und infos aus ts | 0 |
71,266 | 15,191,770,352 | IssuesEvent | 2021-02-15 20:35:07 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | [Discussion] Adding support/helpers/processors for XML in libbeat | Team:Integrations Team:Security-External Integrations discuss libbeat | This issue is to discuss potential implementations for XML for beats. Looking through different open issues, there is plenty of places in which some sort of XML support would be beneficial.
However there are some pro's and con's to all of them, which is why I wanted to have this open discussions to get peoples viewpoint.
XML in general, using the XML encoder in golang does not support unmarshalling to a interface unlike JSON as a built-in feature, however there are libraries out there that takes care of a lot of that, also in terms of performance, the discussion however does not really need to focus on tooling, as the scope is more important at this stage.
As far as I see it, there is a few places in which we can add this:
**1. Adding it as a new helper in libbeat common, similar to jsontransform and plenty of others.**
**Pro's:**
The reason this is handy is to allow input developers to use the helper instead of either having to rewrite XML handling each time, or implementing different types of functionality.
Compared to a processor, handling the XML on input, before the queue is beneficial in many ways, for example processors does not support splitting of lists, which is a very common usecase when working on similar JSON structures, other usecases would be using the keys or values for any sort of conditional tagging, parsing or other transformations needed during ingest.
**Con's:**
Each input would need to manually add support for this.
**2. Adding it as a new processor in libbeat, that allows any specific beat type to**
**Pro's:**
Anyone can use it, just as with any other processor, makes it easy to cover a much larger scope
**Con's:**
Similar to the Pro's of above, does not make it possible to split or format the data beforehand.
**3. Adding a XML processor for ingest pipeline**
**Pro's:**
Anyone can use it, also outside of beats, similar to how the current Logstash XML filter functions.
**Con's:**
Currently ingest pipelines do not support splitting functionality, and the overhead created by XML is large, transforming it on the beat to JSON before sending would reduce the overhead significantly.
My own opinion on the subject is that all 3 is viable and useful, and could be implemented, but in order of ranking, I would use the same as the order above, especially since the helper created in libbeat could later be used in the processor definition as well.
Any thoughts, or thumbs up/down? | True | [Discussion] Adding support/helpers/processors for XML in libbeat - This issue is to discuss potential implementations for XML for beats. Looking through different open issues, there is plenty of places in which some sort of XML support would be beneficial.
However there are some pro's and con's to all of them, which is why I wanted to have this open discussions to get peoples viewpoint.
XML in general, using the XML encoder in golang does not support unmarshalling to a interface unlike JSON as a built-in feature, however there are libraries out there that takes care of a lot of that, also in terms of performance, the discussion however does not really need to focus on tooling, as the scope is more important at this stage.
As far as I see it, there is a few places in which we can add this:
**1. Adding it as a new helper in libbeat common, similar to jsontransform and plenty of others.**
**Pro's:**
The reason this is handy is to allow input developers to use the helper instead of either having to rewrite XML handling each time, or implementing different types of functionality.
Compared to a processor, handling the XML on input, before the queue is beneficial in many ways, for example processors does not support splitting of lists, which is a very common usecase when working on similar JSON structures, other usecases would be using the keys or values for any sort of conditional tagging, parsing or other transformations needed during ingest.
**Con's:**
Each input would need to manually add support for this.
**2. Adding it as a new processor in libbeat, that allows any specific beat type to**
**Pro's:**
Anyone can use it, just as with any other processor, makes it easy to cover a much larger scope
**Con's:**
Similar to the Pro's of above, does not make it possible to split or format the data beforehand.
**3. Adding a XML processor for ingest pipeline**
**Pro's:**
Anyone can use it, also outside of beats, similar to how the current Logstash XML filter functions.
**Con's:**
Currently ingest pipelines do not support splitting functionality, and the overhead created by XML is large, transforming it on the beat to JSON before sending would reduce the overhead significantly.
My own opinion on the subject is that all 3 is viable and useful, and could be implemented, but in order of ranking, I would use the same as the order above, especially since the helper created in libbeat could later be used in the processor definition as well.
Any thoughts, or thumbs up/down? | non_priority | adding support helpers processors for xml in libbeat this issue is to discuss potential implementations for xml for beats looking through different open issues there is plenty of places in which some sort of xml support would be beneficial however there are some pro s and con s to all of them which is why i wanted to have this open discussions to get peoples viewpoint xml in general using the xml encoder in golang does not support unmarshalling to a interface unlike json as a built in feature however there are libraries out there that takes care of a lot of that also in terms of performance the discussion however does not really need to focus on tooling as the scope is more important at this stage as far as i see it there is a few places in which we can add this adding it as a new helper in libbeat common similar to jsontransform and plenty of others pro s the reason this is handy is to allow input developers to use the helper instead of either having to rewrite xml handling each time or implementing different types of functionality compared to a processor handling the xml on input before the queue is beneficial in many ways for example processors does not support splitting of lists which is a very common usecase when working on similar json structures other usecases would be using the keys or values for any sort of conditional tagging parsing or other transformations needed during ingest con s each input would need to manually add support for this adding it as a new processor in libbeat that allows any specific beat type to pro s anyone can use it just as with any other processor makes it easy to cover a much larger scope con s similar to the pro s of above does not make it possible to split or format the data beforehand adding a xml processor for ingest pipeline pro s anyone can use it also outside of beats similar to how the current logstash xml filter functions con s currently ingest pipelines do not support splitting functionality and the overhead created by xml is large transforming it on the beat to json before sending would reduce the overhead significantly my own opinion on the subject is that all is viable and useful and could be implemented but in order of ranking i would use the same as the order above especially since the helper created in libbeat could later be used in the processor definition as well any thoughts or thumbs up down | 0 |
194,982 | 15,495,861,279 | IssuesEvent | 2021-03-11 01:37:40 | RyanBeiden/cmd-food-client | https://api.github.com/repos/RyanBeiden/cmd-food-client | closed | Wireframe | documentation | # User Story
As a Dev, I should be able to view a Wireframe of _Cmd + Food_ in order to build out this project.
# AC
**WHEN** a Dev needs to build a page for this project
**THEN** they should be able to reference a full Wireframe
# Dev Notes
**[Refer to this video on the full app layout](https://drive.google.com/file/d/15_uVoh27FwwHfTjKru-vKxQRgh2lT_JP/view?usp=sharing). There will be a screenshot of each page on its issue notes.** | 1.0 | Wireframe - # User Story
As a Dev, I should be able to view a Wireframe of _Cmd + Food_ in order to build out this project.
# AC
**WHEN** a Dev needs to build a page for this project
**THEN** they should be able to reference a full Wireframe
# Dev Notes
**[Refer to this video on the full app layout](https://drive.google.com/file/d/15_uVoh27FwwHfTjKru-vKxQRgh2lT_JP/view?usp=sharing). There will be a screenshot of each page on its issue notes.** | non_priority | wireframe user story as a dev i should be able to view a wireframe of cmd food in order to build out this project ac when a dev needs to build a page for this project then they should be able to reference a full wireframe dev notes there will be a screenshot of each page on its issue notes | 0 |
281,314 | 24,382,193,424 | IssuesEvent | 2022-10-04 08:47:06 | Slimefun/Slimefun4 | https://api.github.com/repos/Slimefun/Slimefun4 | closed | Bug SlimeFun Android glitch | 🐞 Bug Report 🎯 Needs testing | ### ❗ Checklist
- [X] I am using the official english version of Slimefun and did not modify the jar.
- [X] I am using an up to date "DEV" (not "RC") version of Slimefun.
- [X] I am aware that issues related to Slimefun addons need to be reported on their bug trackers and not here.
- [X] I searched for similar open issues and could not find an existing bug report on this.
### 📍 Description
Android keeps breaking the mangrove wood endlessly because of the saplin of the wood
server version and addons: https://prnt.sc/lO89wOBSUEBi
### 📑 Reproduction Steps
using android and mangrove tree
### 💡 Expected Behavior
not endlessly generating
### 📷 Screenshots / Videos
Video: https://youtube.com/shorts/nmg8KXrE0ag
### 📜 Server Log
no have
### 📂 `/error-reports/` folder
no have
### 💻 Server Software
Other (please specify in your description)
### 🎮 Minecraft Version
1.19.x
### ⭐ Slimefun version
Slimefun DEV - 1028 (git 1dcc4bdd)
### 🧭 Other plugins
_No response_ | 1.0 | Bug SlimeFun Android glitch - ### ❗ Checklist
- [X] I am using the official english version of Slimefun and did not modify the jar.
- [X] I am using an up to date "DEV" (not "RC") version of Slimefun.
- [X] I am aware that issues related to Slimefun addons need to be reported on their bug trackers and not here.
- [X] I searched for similar open issues and could not find an existing bug report on this.
### 📍 Description
Android keeps breaking the mangrove wood endlessly because of the saplin of the wood
server version and addons: https://prnt.sc/lO89wOBSUEBi
### 📑 Reproduction Steps
using android and mangrove tree
### 💡 Expected Behavior
not endlessly generating
### 📷 Screenshots / Videos
Video: https://youtube.com/shorts/nmg8KXrE0ag
### 📜 Server Log
no have
### 📂 `/error-reports/` folder
no have
### 💻 Server Software
Other (please specify in your description)
### 🎮 Minecraft Version
1.19.x
### ⭐ Slimefun version
Slimefun DEV - 1028 (git 1dcc4bdd)
### 🧭 Other plugins
_No response_ | non_priority | bug slimefun android glitch ❗ checklist i am using the official english version of slimefun and did not modify the jar i am using an up to date dev not rc version of slimefun i am aware that issues related to slimefun addons need to be reported on their bug trackers and not here i searched for similar open issues and could not find an existing bug report on this 📍 description android keeps breaking the mangrove wood endlessly because of the saplin of the wood server version and addons 📑 reproduction steps using android and mangrove tree 💡 expected behavior not endlessly generating 📷 screenshots videos video 📜 server log no have 📂 error reports folder no have 💻 server software other please specify in your description 🎮 minecraft version x ⭐ slimefun version slimefun dev git 🧭 other plugins no response | 0 |
398,031 | 27,183,974,884 | IssuesEvent | 2023-02-19 00:54:14 | hbldh/bleak | https://api.github.com/repos/hbldh/bleak | closed | Docs: what about creating an installation guide? | Documentation | * bleak version:
* Python version:
* Operating System:
* BlueZ version (`bluetoothctl -v`) in case of Linux:
### Description
It would be nice to have an in-depth installation guide for the library for the different platforms. For example, I had a hard time figuring out that `pyobjc` was needed on macOS in order for the library to work properly, and I still have not figured it out the requirements to install it on Windows, since there is no official installation guide on the docs. I think it would be of great value for beginners (like me).
### What I Did
### Logs
| 1.0 | Docs: what about creating an installation guide? - * bleak version:
* Python version:
* Operating System:
* BlueZ version (`bluetoothctl -v`) in case of Linux:
### Description
It would be nice to have an in-depth installation guide for the library for the different platforms. For example, I had a hard time figuring out that `pyobjc` was needed on macOS in order for the library to work properly, and I still have not figured it out the requirements to install it on Windows, since there is no official installation guide on the docs. I think it would be of great value for beginners (like me).
### What I Did
### Logs
| non_priority | docs what about creating an installation guide bleak version python version operating system bluez version bluetoothctl v in case of linux description it would be nice to have an in depth installation guide for the library for the different platforms for example i had a hard time figuring out that pyobjc was needed on macos in order for the library to work properly and i still have not figured it out the requirements to install it on windows since there is no official installation guide on the docs i think it would be of great value for beginners like me what i did logs | 0 |
63,817 | 6,885,155,705 | IssuesEvent | 2017-11-21 15:18:59 | gantry/gantry5 | https://api.github.com/repos/gantry/gantry5 | closed | WordPress Menu Title Attribute Not Displaying | needs testing wordpress | Hi all, my WordPress menu title attribute is not displaying what I input to display.
Instead it is displaying the navigation label.
For instance, my navigation label for a menu item is BYOB.
The Title Attribute is: Build Your Own Bag.
But when you hover your mouse above it, it will show "BYOB" instead of "Build Your OWn Bag."
Other menu items also show their navigation label when you hover your mouse above them,
even when I didn't enter any title attribute.
Note: I have enabled Render Titles in the menu particle. I am using Gantry 5 and Requiem theme.
And my WordPress is up to date.
Any suggestions, please? | 1.0 | WordPress Menu Title Attribute Not Displaying - Hi all, my WordPress menu title attribute is not displaying what I input to display.
Instead it is displaying the navigation label.
For instance, my navigation label for a menu item is BYOB.
The Title Attribute is: Build Your Own Bag.
But when you hover your mouse above it, it will show "BYOB" instead of "Build Your OWn Bag."
Other menu items also show their navigation label when you hover your mouse above them,
even when I didn't enter any title attribute.
Note: I have enabled Render Titles in the menu particle. I am using Gantry 5 and Requiem theme.
And my WordPress is up to date.
Any suggestions, please? | non_priority | wordpress menu title attribute not displaying hi all my wordpress menu title attribute is not displaying what i input to display instead it is displaying the navigation label for instance my navigation label for a menu item is byob the title attribute is build your own bag but when you hover your mouse above it it will show byob instead of build your own bag other menu items also show their navigation label when you hover your mouse above them even when i didn t enter any title attribute note i have enabled render titles in the menu particle i am using gantry and requiem theme and my wordpress is up to date any suggestions please | 0 |
199,773 | 15,782,436,570 | IssuesEvent | 2021-04-01 12:47:55 | ethereum/solidity | https://api.github.com/repos/ethereum/solidity | closed | Document the initial values of elements of newly allocated memory arrays. | documentation :book: | @frangio on gitter: ``are newly allocated memory arrays guaranteed to be zeroed? i couldn't find this in the documentation``
Indeed I couldn't find it being explicitly mentioned anywhere either...
https://docs.soliditylang.org/en/latest/types.html#allocating-memory-arrays seems like the place this should be mentioned. | 1.0 | Document the initial values of elements of newly allocated memory arrays. - @frangio on gitter: ``are newly allocated memory arrays guaranteed to be zeroed? i couldn't find this in the documentation``
Indeed I couldn't find it being explicitly mentioned anywhere either...
https://docs.soliditylang.org/en/latest/types.html#allocating-memory-arrays seems like the place this should be mentioned. | non_priority | document the initial values of elements of newly allocated memory arrays frangio on gitter are newly allocated memory arrays guaranteed to be zeroed i couldn t find this in the documentation indeed i couldn t find it being explicitly mentioned anywhere either seems like the place this should be mentioned | 0 |
110,126 | 23,875,383,157 | IssuesEvent | 2022-09-07 18:31:15 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | closed | Pull Request Template variables | devops/prod devops-code-git/tech needs-more-info cba Pri1 |
Is it possible to access variables about the pull request, specifically the pull request ID, from the pull request template?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a297d54f-48b3-1b68-e828-15d1559d9b2d
* Version Independent ID: fd9dcb35-72f9-f5a9-e0c0-75c3e59ba8d9
* Content: [Improve pull request descriptions with pull request templates - Azure Repos](https://docs.microsoft.com/en-us/azure/devops/repos/git/pull-request-templates?view=azure-devops)
* Content Source: [docs/repos/git/pull-request-templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/repos/git/pull-request-templates.md)
* Product: **devops**
* Technology: **devops-code-git**
* GitHub Login: @vijayma
* Microsoft Alias: **vijayma** | 1.0 | Pull Request Template variables -
Is it possible to access variables about the pull request, specifically the pull request ID, from the pull request template?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: a297d54f-48b3-1b68-e828-15d1559d9b2d
* Version Independent ID: fd9dcb35-72f9-f5a9-e0c0-75c3e59ba8d9
* Content: [Improve pull request descriptions with pull request templates - Azure Repos](https://docs.microsoft.com/en-us/azure/devops/repos/git/pull-request-templates?view=azure-devops)
* Content Source: [docs/repos/git/pull-request-templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/repos/git/pull-request-templates.md)
* Product: **devops**
* Technology: **devops-code-git**
* GitHub Login: @vijayma
* Microsoft Alias: **vijayma** | non_priority | pull request template variables is it possible to access variables about the pull request specifically the pull request id from the pull request template document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops code git github login vijayma microsoft alias vijayma | 0 |
34,221 | 4,534,986,413 | IssuesEvent | 2016-09-08 15:59:54 | coala-analyzer/coala | https://api.github.com/repos/coala-analyzer/coala | opened | Don't prepend non-diff-related characters with '| | |' | status/needs design | ```
test.py
| 2| ••••a=5
| | [NORMAL] SpaceConsistencyBear:
| | Line contains following spacing inconsistencies:
| | - Spaces used instead of tabs.
| | The following actions are applicable to this result:
| | 0: Apply no further actions.
| | 1: Open the affected file(s) in an editor.
| | 2: Apply the patch automatically.
| | 3: Print a diff of the patch that would be applied.
| | Please enter the number of the action you want to execute. 2
| | Patch applied successfully.
| | The following actions are applicable to this result:
| | 0: Apply no further actions.
| | 1: Open the affected file(s) in an editor.
| | Please enter the number of the action you want to execute. 0
```
maybe we should do this? (as it's maybe improving reading)
```
test.py
| 2| ••••a=5
[NORMAL] SpaceConsistencyBear:
Line contains following spacing inconsistencies:
- Spaces used instead of tabs.
The following actions are applicable to this result:
0: Apply no further actions.
1: Open the affected file(s) in an editor.
2: Apply the patch automatically.
3: Print a diff of the patch that would be applied.
Please enter the number of the action you want to execute. 2
Patch applied successfully.
The following actions are applicable to this result:
0: Apply no further actions.
1: Open the affected file(s) in an editor.
Please enter the number of the action you want to execute. 0
``` | 1.0 | Don't prepend non-diff-related characters with '| | |' - ```
test.py
| 2| ••••a=5
| | [NORMAL] SpaceConsistencyBear:
| | Line contains following spacing inconsistencies:
| | - Spaces used instead of tabs.
| | The following actions are applicable to this result:
| | 0: Apply no further actions.
| | 1: Open the affected file(s) in an editor.
| | 2: Apply the patch automatically.
| | 3: Print a diff of the patch that would be applied.
| | Please enter the number of the action you want to execute. 2
| | Patch applied successfully.
| | The following actions are applicable to this result:
| | 0: Apply no further actions.
| | 1: Open the affected file(s) in an editor.
| | Please enter the number of the action you want to execute. 0
```
maybe we should do this? (as it's maybe improving reading)
```
test.py
| 2| ••••a=5
[NORMAL] SpaceConsistencyBear:
Line contains following spacing inconsistencies:
- Spaces used instead of tabs.
The following actions are applicable to this result:
0: Apply no further actions.
1: Open the affected file(s) in an editor.
2: Apply the patch automatically.
3: Print a diff of the patch that would be applied.
Please enter the number of the action you want to execute. 2
Patch applied successfully.
The following actions are applicable to this result:
0: Apply no further actions.
1: Open the affected file(s) in an editor.
Please enter the number of the action you want to execute. 0
``` | non_priority | don t prepend non diff related characters with test py ••••a spaceconsistencybear line contains following spacing inconsistencies spaces used instead of tabs the following actions are applicable to this result apply no further actions open the affected file s in an editor apply the patch automatically print a diff of the patch that would be applied please enter the number of the action you want to execute patch applied successfully the following actions are applicable to this result apply no further actions open the affected file s in an editor please enter the number of the action you want to execute maybe we should do this as it s maybe improving reading test py ••••a spaceconsistencybear line contains following spacing inconsistencies spaces used instead of tabs the following actions are applicable to this result apply no further actions open the affected file s in an editor apply the patch automatically print a diff of the patch that would be applied please enter the number of the action you want to execute patch applied successfully the following actions are applicable to this result apply no further actions open the affected file s in an editor please enter the number of the action you want to execute | 0 |
12,264 | 5,172,366,489 | IssuesEvent | 2017-01-18 13:20:46 | opencv/opencv | https://api.github.com/repos/opencv/opencv | closed | libopencv_core.so.3.2.0: undefined reference to `dpotrf_' | bug category: build/install | I'm installing opencv 3.2 and when I use :+1:
cmake -DWITH_QT=ON -DWITH_OPENGL=ON -DFORCE_VTK=ON -DWITH_TBB=ON -DWITH_GDAL=ON -DWITH_XINE=ON -DBUILD_EXAMPLES=ON ..
and
make -j4 , this error is showing , how can be solved ?
[ 75%] Linking CXX executable ../../bin/opengl-example-opengl_interop
../../lib/libopencv_core.so.3.2.0: undefined reference to `dpotrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgesv_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgels_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgesv_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sposv_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgetrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgetrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgels_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgeqrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `spotrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgeqrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgesdd_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgesdd_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dposv_'
collect2: error: ld returned 1 exit status
make[2]: *** [bin/opencv_version] Error 1
make[1]: *** [apps/version/CMakeFiles/opencv_version.dir/all] Error 2
make[1]: *** Waiting for unfinished jobs....
[ 75%] Built target opencv_annotation
[ 75%] Built target opencv_visualisation
[ 75%] Built target example_opengl_opengl_interop
| 1.0 | libopencv_core.so.3.2.0: undefined reference to `dpotrf_' - I'm installing opencv 3.2 and when I use :+1:
cmake -DWITH_QT=ON -DWITH_OPENGL=ON -DFORCE_VTK=ON -DWITH_TBB=ON -DWITH_GDAL=ON -DWITH_XINE=ON -DBUILD_EXAMPLES=ON ..
and
make -j4 , this error is showing , how can be solved ?
[ 75%] Linking CXX executable ../../bin/opengl-example-opengl_interop
../../lib/libopencv_core.so.3.2.0: undefined reference to `dpotrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgesv_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgels_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgesv_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sposv_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgetrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgetrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgels_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgeqrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `spotrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgeqrf_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `sgesdd_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dgesdd_'
../../lib/libopencv_core.so.3.2.0: undefined reference to `dposv_'
collect2: error: ld returned 1 exit status
make[2]: *** [bin/opencv_version] Error 1
make[1]: *** [apps/version/CMakeFiles/opencv_version.dir/all] Error 2
make[1]: *** Waiting for unfinished jobs....
[ 75%] Built target opencv_annotation
[ 75%] Built target opencv_visualisation
[ 75%] Built target example_opengl_opengl_interop
| non_priority | libopencv core so undefined reference to dpotrf i m installing opencv and when i use cmake dwith qt on dwith opengl on dforce vtk on dwith tbb on dwith gdal on dwith xine on dbuild examples on and make this error is showing how can be solved linking cxx executable bin opengl example opengl interop lib libopencv core so undefined reference to dpotrf lib libopencv core so undefined reference to dgesv lib libopencv core so undefined reference to sgels lib libopencv core so undefined reference to sgesv lib libopencv core so undefined reference to sposv lib libopencv core so undefined reference to dgetrf lib libopencv core so undefined reference to sgetrf lib libopencv core so undefined reference to dgels lib libopencv core so undefined reference to dgeqrf lib libopencv core so undefined reference to spotrf lib libopencv core so undefined reference to sgeqrf lib libopencv core so undefined reference to sgesdd lib libopencv core so undefined reference to dgesdd lib libopencv core so undefined reference to dposv error ld returned exit status make error make error make waiting for unfinished jobs built target opencv annotation built target opencv visualisation built target example opengl opengl interop | 0 |
309,328 | 26,660,648,876 | IssuesEvent | 2023-01-25 20:46:53 | golang/go | https://api.github.com/repos/golang/go | closed | os: TestCloseWithBlockingReadByNewFile failures | Testing NeedsFix | ```
#!watchflakes
post <- pkg == "os" && test == "TestCloseWithBlockingReadByNewFile"
```
Issue created automatically to collect these failures.
Example ([log](https://build.golang.org/log/2f57844f77aa062cab3dbdf7cc31965920c12d73)):
--- FAIL: TestCloseWithBlockingReadByNewFile (1.19s)
pipe_test.go:356: timed out waiting for Read
— [watchflakes](https://go.dev/wiki/Watchflakes)
| 1.0 | os: TestCloseWithBlockingReadByNewFile failures - ```
#!watchflakes
post <- pkg == "os" && test == "TestCloseWithBlockingReadByNewFile"
```
Issue created automatically to collect these failures.
Example ([log](https://build.golang.org/log/2f57844f77aa062cab3dbdf7cc31965920c12d73)):
--- FAIL: TestCloseWithBlockingReadByNewFile (1.19s)
pipe_test.go:356: timed out waiting for Read
— [watchflakes](https://go.dev/wiki/Watchflakes)
| non_priority | os testclosewithblockingreadbynewfile failures watchflakes post pkg os test testclosewithblockingreadbynewfile issue created automatically to collect these failures example fail testclosewithblockingreadbynewfile pipe test go timed out waiting for read — | 0 |
21,137 | 3,686,282,694 | IssuesEvent | 2016-02-25 00:26:04 | www-purple/Mixxy | https://api.github.com/repos/www-purple/Mixxy | closed | Define and develop our use cases | design | First things that come to mind:
- Draw comics
- Edit existing comics
- Fork comics
- Remove comics
- Comments
- Sign in
- Sign out
- Upload a comic
- Feedback | 1.0 | Define and develop our use cases - First things that come to mind:
- Draw comics
- Edit existing comics
- Fork comics
- Remove comics
- Comments
- Sign in
- Sign out
- Upload a comic
- Feedback | non_priority | define and develop our use cases first things that come to mind draw comics edit existing comics fork comics remove comics comments sign in sign out upload a comic feedback | 0 |
40,611 | 16,510,578,998 | IssuesEvent | 2021-05-26 03:13:07 | Azure/azure-powershell | https://api.github.com/repos/Azure/azure-powershell | closed | Get-AzEventHubNamespace should accept Pipeline-Input from Get-AzEventHub | Event Hubs Service Attention customer-reported feature-request | ## Description of the new feature
Currently it doesn't and complains about missing mandatory parameter: Namespace
## Proposed implementation details (optional)
| 1.0 | Get-AzEventHubNamespace should accept Pipeline-Input from Get-AzEventHub - ## Description of the new feature
Currently it doesn't and complains about missing mandatory parameter: Namespace
## Proposed implementation details (optional)
| non_priority | get azeventhubnamespace should accept pipeline input from get azeventhub description of the new feature currently it doesn t and complains about missing mandatory parameter namespace proposed implementation details optional | 0 |
177,115 | 13,683,899,259 | IssuesEvent | 2020-09-30 03:16:16 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Quarantined components tests | area-blazor task test-failure | https://dev.azure.com/dnceng/public/_build/results?buildId=715107&view=ms.vss-test-web.build-test-results-tab&runId=22161926&resultId=100087&paneView=debug
## EventCallbackTest.EventCallback_RerendersOuterComponent
```
OpenQA.Selenium.BrowserAssertFailedException : Xunit.Sdk.NotEmptyException: Assert.NotEmpty() Failure
at Xunit.Assert.NotEmpty(IEnumerable collection) in C:\Dev\xunit\xunit\src\xunit.assert\Asserts\CollectionAsserts.cs:line 331
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass15_0.<Exists>b__0() in /_/src/Shared/E2ETesting/WaitAssert.cs:line 69
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass18_0`1.<WaitAssertCore>b__0(IWebDriver ) in //src/Shared/E2ETesting/WaitAssert.cs:line 106
Screen shot captured at 'F:\workspace\_work\1\s\artifacts\TestResults\Release\Microsoft.AspNetCore.Components.E2ETests\fb5f742b90504482a688b4399d5f5964.png'
Encountered browser errors
[2020-07-02T21:37:10Z] [Debug] http://127.0.0.1:59182/subdir/_framework/dotnet.5.0.0-preview.6.20305.6.js 0:124605 "mono_wasm_runtime_ready" "fe00e07a-5519-4dfe-b35a-f867dbaf2e28"
```
## Microsoft.AspNetCore.Components.E2ETest.ServerExecutionTests.ServerGlobalizationTest.CanSetCultureAndParseCultureInvariantNumbersAndDatesWithInputFields
```
OpenQA.Selenium.BrowserAssertFailedException : Xunit.Sdk.EqualException: Assert.Equal() Failure
Expected: 9000.42
Actual: 0
at Xunit.Assert.Equal[T](T expected, T actual, IEqualityComparer1 comparer) in C:\\Dev\\xunit\\xunit\\src\\xunit.assert\\Asserts\\EqualityAsserts.cs:line 40
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass17_0.<WaitAssertCore>b__0() in /_/src/Shared/E2ETesting/WaitAssert.cs:line 80
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass18_01.<WaitAssertCore>b__0(IWebDriver ) in //src/Shared/E2ETesting/WaitAssert.cs:line 106
Screen shot captured at 'F:\workspace\_work\1\s\artifacts\TestResults\Release\Microsoft.AspNetCore.Components.E2ETests\6b72113071964ccc90c1b32fea26819c.png'
Encountered browser errors
[2020-07-02T21:34:27Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:27.431Z] Information: Normalizing '_blazor' to 'http://127.0.0.1:52168/subdir/_blazor'."
[2020-07-02T21:34:27Z] [Info] http://127.0.0.1:52168/subdir 40:16 "Blazor server-side"
[2020-07-02T21:34:27Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:27.659Z] Information: WebSocket connected to ws://127.0.0.1:52168/subdir/_blazor?id=8g789y_wCQ5B7wMFpe82hA."
[2020-07-02T21:34:29Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:29.453Z] Information: Normalizing '_blazor' to 'http://127.0.0.1:52168/subdir/_blazor'."
[2020-07-02T21:34:29Z] [Info] http://127.0.0.1:52168/subdir 40:16 "Blazor server-side"
[2020-07-02T21:34:29Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:29.489Z] Information: WebSocket connected to ws://127.0.0.1:52168/subdir/_blazor?id=Ffx1jr-XduQoq79VrrTZfA."
[2020-07-02T21:34:32Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:32.725Z] Information: Normalizing '_blazor' to 'http://127.0.0.1:52168/subdir/_blazor'."
[2020-07-02T21:34:32Z] [Info] http://127.0.0.1:52168/subdir 40:16 "Blazor server-side"
[2020-07-02T21:34:32Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:32.736Z] Information: WebSocket connected to ws://127.0.0.1:52168/subdir/_blazor?id=gM7SL5wGyUtcWtrd5eJPiA."
[2020-07-02T21:34:33Z] [Severe] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 18:23161 "[2020-07-02T21:34:33.347Z] Error: Circuit has been shut down due to error."
[2020-07-02T21:34:33Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:33.350Z] Information: Connection disconnected."
[2020-07-02T21:34:33Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:33.699Z] Information: Normalizing '_blazor' to 'http://127.0.0.1:52168/subdir/_blazor'."
[2020-07-02T21:34:33Z] [Info] http://127.0.0.1:52168/subdir 40:16 "Blazor server-side"
[2020-07-02T21:34:33Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:33.709Z] Information: WebSocket connected to ws://127.0.0.1:52168/subdir/_blazor?id=wdwh05klqgpC-1GMz6WZHQ."Page content:
```
## ShowsErrorNotification_OnError_Dismiss
```
OpenQA.Selenium.BrowserAssertFailedException : Xunit.Sdk.NotEmptyException: Assert.NotEmpty() Failure
at Xunit.Assert.NotEmpty(IEnumerable collection) in C:\Dev\xunit\xunit\src\xunit.assert\Asserts\CollectionAsserts.cs:line 331
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass15_0.<Exists>b__0() in /_/src/Shared/E2ETesting/WaitAssert.cs:line 69
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass18_0`1.<WaitAssertCore>b__0(IWebDriver ) in //src/Shared/E2ETesting/WaitAssert.cs:line 106
Screen shot captured at 'F:\workspace\_work\1\s\artifacts\TestResults\Release\Microsoft.AspNetCore.Components.E2ETests\34453aa86e7549729cfddd540d39e011.png'
Encountered browser errors
[2020-07-02T21:37:55Z] [Debug] http://127.0.0.1:62181/subdir/_framework/dotnet.5.0.0-preview.6.20305.6.js 0:124605 "mono_wasm_runtime_ready" "fe00e07a-5519-4dfe-b35a-f867dbaf2e28"Page content:
<head>
<meta charset="utf-8">
<title>Basic test app</title>
<base href="/subdir/">
``` | 1.0 | Quarantined components tests - https://dev.azure.com/dnceng/public/_build/results?buildId=715107&view=ms.vss-test-web.build-test-results-tab&runId=22161926&resultId=100087&paneView=debug
## EventCallbackTest.EventCallback_RerendersOuterComponent
```
OpenQA.Selenium.BrowserAssertFailedException : Xunit.Sdk.NotEmptyException: Assert.NotEmpty() Failure
at Xunit.Assert.NotEmpty(IEnumerable collection) in C:\Dev\xunit\xunit\src\xunit.assert\Asserts\CollectionAsserts.cs:line 331
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass15_0.<Exists>b__0() in /_/src/Shared/E2ETesting/WaitAssert.cs:line 69
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass18_0`1.<WaitAssertCore>b__0(IWebDriver ) in //src/Shared/E2ETesting/WaitAssert.cs:line 106
Screen shot captured at 'F:\workspace\_work\1\s\artifacts\TestResults\Release\Microsoft.AspNetCore.Components.E2ETests\fb5f742b90504482a688b4399d5f5964.png'
Encountered browser errors
[2020-07-02T21:37:10Z] [Debug] http://127.0.0.1:59182/subdir/_framework/dotnet.5.0.0-preview.6.20305.6.js 0:124605 "mono_wasm_runtime_ready" "fe00e07a-5519-4dfe-b35a-f867dbaf2e28"
```
## Microsoft.AspNetCore.Components.E2ETest.ServerExecutionTests.ServerGlobalizationTest.CanSetCultureAndParseCultureInvariantNumbersAndDatesWithInputFields
```
OpenQA.Selenium.BrowserAssertFailedException : Xunit.Sdk.EqualException: Assert.Equal() Failure
Expected: 9000.42
Actual: 0
at Xunit.Assert.Equal[T](T expected, T actual, IEqualityComparer1 comparer) in C:\\Dev\\xunit\\xunit\\src\\xunit.assert\\Asserts\\EqualityAsserts.cs:line 40
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass17_0.<WaitAssertCore>b__0() in /_/src/Shared/E2ETesting/WaitAssert.cs:line 80
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass18_01.<WaitAssertCore>b__0(IWebDriver ) in //src/Shared/E2ETesting/WaitAssert.cs:line 106
Screen shot captured at 'F:\workspace\_work\1\s\artifacts\TestResults\Release\Microsoft.AspNetCore.Components.E2ETests\6b72113071964ccc90c1b32fea26819c.png'
Encountered browser errors
[2020-07-02T21:34:27Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:27.431Z] Information: Normalizing '_blazor' to 'http://127.0.0.1:52168/subdir/_blazor'."
[2020-07-02T21:34:27Z] [Info] http://127.0.0.1:52168/subdir 40:16 "Blazor server-side"
[2020-07-02T21:34:27Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:27.659Z] Information: WebSocket connected to ws://127.0.0.1:52168/subdir/_blazor?id=8g789y_wCQ5B7wMFpe82hA."
[2020-07-02T21:34:29Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:29.453Z] Information: Normalizing '_blazor' to 'http://127.0.0.1:52168/subdir/_blazor'."
[2020-07-02T21:34:29Z] [Info] http://127.0.0.1:52168/subdir 40:16 "Blazor server-side"
[2020-07-02T21:34:29Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:29.489Z] Information: WebSocket connected to ws://127.0.0.1:52168/subdir/_blazor?id=Ffx1jr-XduQoq79VrrTZfA."
[2020-07-02T21:34:32Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:32.725Z] Information: Normalizing '_blazor' to 'http://127.0.0.1:52168/subdir/_blazor'."
[2020-07-02T21:34:32Z] [Info] http://127.0.0.1:52168/subdir 40:16 "Blazor server-side"
[2020-07-02T21:34:32Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:32.736Z] Information: WebSocket connected to ws://127.0.0.1:52168/subdir/_blazor?id=gM7SL5wGyUtcWtrd5eJPiA."
[2020-07-02T21:34:33Z] [Severe] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 18:23161 "[2020-07-02T21:34:33.347Z] Error: Circuit has been shut down due to error."
[2020-07-02T21:34:33Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:33.350Z] Information: Connection disconnected."
[2020-07-02T21:34:33Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:33.699Z] Information: Normalizing '_blazor' to 'http://127.0.0.1:52168/subdir/_blazor'."
[2020-07-02T21:34:33Z] [Info] http://127.0.0.1:52168/subdir 40:16 "Blazor server-side"
[2020-07-02T21:34:33Z] [Info] http://127.0.0.1:52168/subdir/_framework/blazor.server.js 0:5679 "[2020-07-02T21:34:33.709Z] Information: WebSocket connected to ws://127.0.0.1:52168/subdir/_blazor?id=wdwh05klqgpC-1GMz6WZHQ."Page content:
```
## ShowsErrorNotification_OnError_Dismiss
```
OpenQA.Selenium.BrowserAssertFailedException : Xunit.Sdk.NotEmptyException: Assert.NotEmpty() Failure
at Xunit.Assert.NotEmpty(IEnumerable collection) in C:\Dev\xunit\xunit\src\xunit.assert\Asserts\CollectionAsserts.cs:line 331
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass15_0.<Exists>b__0() in /_/src/Shared/E2ETesting/WaitAssert.cs:line 69
at Microsoft.AspNetCore.E2ETesting.WaitAssert.<>c__DisplayClass18_0`1.<WaitAssertCore>b__0(IWebDriver ) in //src/Shared/E2ETesting/WaitAssert.cs:line 106
Screen shot captured at 'F:\workspace\_work\1\s\artifacts\TestResults\Release\Microsoft.AspNetCore.Components.E2ETests\34453aa86e7549729cfddd540d39e011.png'
Encountered browser errors
[2020-07-02T21:37:55Z] [Debug] http://127.0.0.1:62181/subdir/_framework/dotnet.5.0.0-preview.6.20305.6.js 0:124605 "mono_wasm_runtime_ready" "fe00e07a-5519-4dfe-b35a-f867dbaf2e28"Page content:
<head>
<meta charset="utf-8">
<title>Basic test app</title>
<base href="/subdir/">
``` | non_priority | quarantined components tests eventcallbacktest eventcallback rerendersoutercomponent openqa selenium browserassertfailedexception xunit sdk notemptyexception assert notempty failure at xunit assert notempty ienumerable collection in c dev xunit xunit src xunit assert asserts collectionasserts cs line at microsoft aspnetcore waitassert c b in src shared waitassert cs line at microsoft aspnetcore waitassert c b iwebdriver in src shared waitassert cs line screen shot captured at f workspace work s artifacts testresults release microsoft aspnetcore components png encountered browser errors mono wasm runtime ready microsoft aspnetcore components serverexecutiontests serverglobalizationtest cansetcultureandparsecultureinvariantnumbersanddateswithinputfields openqa selenium browserassertfailedexception xunit sdk equalexception assert equal failure expected actual at xunit assert equal t expected t actual comparer in c dev xunit xunit src xunit assert asserts equalityasserts cs line at microsoft aspnetcore waitassert c b in src shared waitassert cs line at microsoft aspnetcore waitassert c b iwebdriver in src shared waitassert cs line screen shot captured at f workspace work s artifacts testresults release microsoft aspnetcore components png encountered browser errors information normalizing blazor to blazor server side information websocket connected to ws subdir blazor id information normalizing blazor to blazor server side information websocket connected to ws subdir blazor id information normalizing blazor to blazor server side information websocket connected to ws subdir blazor id error circuit has been shut down due to error information connection disconnected information normalizing blazor to blazor server side information websocket connected to ws subdir blazor id page content showserrornotification onerror dismiss openqa selenium browserassertfailedexception xunit sdk notemptyexception assert notempty failure at xunit assert notempty ienumerable collection in c dev xunit xunit src xunit assert asserts collectionasserts cs line at microsoft aspnetcore waitassert c b in src shared waitassert cs line at microsoft aspnetcore waitassert c b iwebdriver in src shared waitassert cs line screen shot captured at f workspace work s artifacts testresults release microsoft aspnetcore components png encountered browser errors mono wasm runtime ready page content basic test app | 0 |
7,702 | 6,177,908,174 | IssuesEvent | 2017-07-02 06:54:54 | vmalik-2001/ClubPenguin3D | https://api.github.com/repos/vmalik-2001/ClubPenguin3D | closed | Mail Services | degraded performance investigating Mail | Our mail services are currently facing issue. Mail are being delivered, however, are going to the spam folder. We request you all to report them a 'Not a Spam'. | True | Mail Services - Our mail services are currently facing issue. Mail are being delivered, however, are going to the spam folder. We request you all to report them a 'Not a Spam'. | non_priority | mail services our mail services are currently facing issue mail are being delivered however are going to the spam folder we request you all to report them a not a spam | 0 |
1,012 | 12,892,155,999 | IssuesEvent | 2020-07-13 19:04:45 | esnet/iperf | https://api.github.com/repos/esnet/iperf | closed | iperf3 UDP test functionality depends on SO_REUSEADDR implementation | Help Wanted portability | Hi,
It looks like UDP functionality in part of "connection" management breaks when code is ported to platforms with generic BSD socket API support like ExpressLogic NetX. iperf_udp_accept() is trying to netannounce() the new socket that is attempted to bound to already open UDP port. On NetX platform SO_REUSEADDR socket option is not working the same way for UDP as Linux/Windows implementations, and bind fails. Some research (http://stackoverflow.com/questions/14388706/socket-options-so-reuseaddr-and-so-reuseport-how-do-they-differ-do-they-mean-t/14388707#14388707 and http://hacked10bits.blogspot.com/2014/12/udp-binding-and-port-reuse-in-linux.html) shows that later BSD implementations have SO_REUSEPORT option that represent desired behavior, however this option is not available on most platforms.
I wonder if a more portable solution would be not to reuse port numbers for UDP test.
Thanks.
| True | iperf3 UDP test functionality depends on SO_REUSEADDR implementation - Hi,
It looks like UDP functionality in part of "connection" management breaks when code is ported to platforms with generic BSD socket API support like ExpressLogic NetX. iperf_udp_accept() is trying to netannounce() the new socket that is attempted to bound to already open UDP port. On NetX platform SO_REUSEADDR socket option is not working the same way for UDP as Linux/Windows implementations, and bind fails. Some research (http://stackoverflow.com/questions/14388706/socket-options-so-reuseaddr-and-so-reuseport-how-do-they-differ-do-they-mean-t/14388707#14388707 and http://hacked10bits.blogspot.com/2014/12/udp-binding-and-port-reuse-in-linux.html) shows that later BSD implementations have SO_REUSEPORT option that represent desired behavior, however this option is not available on most platforms.
I wonder if a more portable solution would be not to reuse port numbers for UDP test.
Thanks.
| non_priority | udp test functionality depends on so reuseaddr implementation hi it looks like udp functionality in part of connection management breaks when code is ported to platforms with generic bsd socket api support like expresslogic netx iperf udp accept is trying to netannounce the new socket that is attempted to bound to already open udp port on netx platform so reuseaddr socket option is not working the same way for udp as linux windows implementations and bind fails some research and shows that later bsd implementations have so reuseport option that represent desired behavior however this option is not available on most platforms i wonder if a more portable solution would be not to reuse port numbers for udp test thanks | 0 |
28,512 | 4,415,624,069 | IssuesEvent | 2016-08-14 06:34:19 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | stress: failed test in cockroach/storage/storage.test: TestStoreMetrics | Robot test-failure | Binary: cockroach/static-tests.tar.gz sha: https://github.com/cockroachdb/cockroach/commits/b44ebfb6d0911c8d6e9c6d8fc1099d592791a894
Stress build found a failed test:
```
=== RUN TestStoreMetrics
I160814 04:47:14.456048 storage/engine/rocksdb.go:354 opening in memory rocksdb instance
W160814 04:47:14.456756 gossip/gossip.go:1017 not connected to cluster; use --join to specify a connected node
I160814 04:47:14.458453 storage/store.go:2525 store=1:1: system config not yet available
I160814 04:47:14.459163 storage/replica_trigger.go:308 store=1:1 range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0s [physicalTime=1970-01-01 00:00:00 +0000 UTC]
I160814 04:47:14.459542 storage/engine/rocksdb.go:354 opening in memory rocksdb instance
W160814 04:47:14.460351 gossip/gossip.go:1017 not connected to cluster; use --join to specify a connected node
I160814 04:47:14.461517 storage/store.go:2525 store=2:2: system config not yet available
I160814 04:47:14.461539 gossip/client.go:75 node 2: starting client to 127.0.0.1:35096
I160814 04:47:14.461668 storage/engine/rocksdb.go:354 opening in memory rocksdb instance
I160814 04:47:14.461933 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.461968 gossip/server.go:283 node 1: replying to 2
I160814 04:47:14.462041 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.462095 gossip/server.go:283 node 1: replying to 2
I160814 04:47:14.462231 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.462255 gossip/server.go:283 node 1: replying to 2
I160814 04:47:14.462305 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.462321 gossip/server.go:283 node 1: replying to 2
W160814 04:47:14.462438 gossip/gossip.go:1017 not connected to cluster; use --join to specify a connected node
I160814 04:47:14.463211 storage/store.go:2525 store=3:3: system config not yet available
I160814 04:47:14.463343 gossip/client.go:75 node 3: starting client to 127.0.0.1:35096
I160814 04:47:14.463633 gossip/server.go:185 node 1: received gossip from node 3
I160814 04:47:14.463660 gossip/server.go:283 node 1: replying to 3
I160814 04:47:14.463737 gossip/server.go:185 node 1: received gossip from node 3
I160814 04:47:14.463765 gossip/server.go:283 node 1: replying to 3
I160814 04:47:14.463916 gossip/server.go:185 node 1: received gossip from node 3
I160814 04:47:14.463941 gossip/server.go:283 node 1: replying to 3
I160814 04:47:14.464013 gossip/server.go:185 node 1: received gossip from node 3
I160814 04:47:14.464037 gossip/server.go:283 node 1: replying to 3
I160814 04:47:14.465398 storage/replica_command.go:2216 store=1:1 range=1 [/Min-/Max): initiating a split of this range at key "m"
E160814 04:47:14.467367 storage/queue.go:508 [replicate] (purgatory) on store=1:1 range=1 [/Min-"m"): 0 of 0 stores with an attribute matching []; likely not enough nodes in cluster
E160814 04:47:14.467534 storage/queue.go:508 [replicate] (purgatory) on store=1:1 range=2 ["m"-/Max): 0 of 0 stores with an attribute matching []; likely not enough nodes in cluster
I160814 04:47:14.472228 storage/replica_raftstorage.go:474 generated snapshot for range 2 at index 10 in 110.62µs. encoded size=2407, 23 KV pairs, 0 log entries
I160814 04:47:14.472476 storage/replica_command.go:3059 store=1:1 range=2 ["m"-/Max): change replicas of 2: read existing descriptor range_id:2 start_key:"m" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2
I160814 04:47:14.472874 storage/replica_raftstorage.go:600 store=2:2 range=2 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 2 at index 10 (encoded size=2407, 23 KV pairs, 0 log entries)
I160814 04:47:14.473136 storage/replica.go:2144 store=1:1 range=1 [/Min-"m"): applied part of replica change txn: Noop [/Min,/Min), Noop [/Min,/Min), Put [/Meta2/Max,/Min), pErr=<nil>
I160814 04:47:14.473627 storage/replica.go:2144 store=1:1 range=2 ["m"-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/"m"/RangeDescriptor,/Min), ConditionalPut [/Local/Range/"m"/RangeDescriptor,/Min), Noop [/Min,/Min), pErr=<nil>
I160814 04:47:14.473901 storage/replica.go:1520 store=1:1 range=2 ["m"-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:2} for range 2: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:2 StoreID:2 ReplicaID:2}]
I160814 04:47:14.474091 storage/replica.go:2144 store=1:1 range=2 ["m"-/Max): applied part of replica change txn: EndTransaction [/Local/Range/"m"/RangeDescriptor,/Min), pErr=<nil>
I160814 04:47:14.474699 storage/replica_raftstorage.go:474 generated snapshot for range 2 at index 13 in 121.74µs. encoded size=3231, 25 KV pairs, 3 log entries
W160814 04:47:14.474835 storage/intent_resolver.go:344 could not GC completed transaction: end key /Local/Range/"m"/"rdsc\x00" must be greater than start /Local/Range/"m"/RangeDescriptor
I160814 04:47:14.474904 storage/replica_raftstorage.go:603 store=2:2 range=2 ["m"-/Max): with replicaID [?], applied preemptive snapshot for range 2 in 2.000445ms
I160814 04:47:14.474964 storage/replica_command.go:3059 store=1:1 range=2 ["m"-/Max): change replicas of 2: read existing descriptor range_id:2 start_key:"m" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:2 store_id:2 replica_id:2 > next_replica_id:3
I160814 04:47:14.475276 storage/replica.go:2144 store=1:1 range=1 [/Min-"m"): applied part of replica change txn: Noop [/Min,/Min), Noop [/Min,/Min), Put [/Meta2/Max,/Min), pErr=<nil>
I160814 04:47:14.475399 storage/replica_raftstorage.go:600 store=3:3 range=2 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 2 at index 13 (encoded size=3231, 25 KV pairs, 3 log entries)
I160814 04:47:14.475667 storage/replica_raftstorage.go:603 store=3:3 range=2 ["m"-/Max): with replicaID [?], applied preemptive snapshot for range 2 in 242.983µs
I160814 04:47:14.477021 storage/replica.go:2144 store=2:2 range=2 ["m"-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/"m"/RangeDescriptor,/Min), ConditionalPut [/Local/Range/"m"/RangeDescriptor,/Min), Noop [/Min,/Min), pErr=<nil>
I160814 04:47:14.477215 storage/replica.go:2144 store=2:2 range=2 ["m"-/Max): applied part of replica change txn: EndTransaction [/Local/Range/"m"/RangeDescriptor,/Min), pErr=<nil>
I160814 04:47:14.477338 storage/replica.go:2144 store=1:1 range=2 ["m"-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/"m"/RangeDescriptor,/Min), ConditionalPut [/Local/Range/"m"/RangeDescriptor,/Min), Noop [/Min,/Min), pErr=<nil>
I160814 04:47:14.477632 storage/replica.go:2144 store=2:2 range=2 ["m"-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/"m"/RangeDescriptor,/Min), ConditionalPut [/Local/Range/"m"/RangeDescriptor,/Min), Noop [/Min,/Min), pErr=<nil>
I160814 04:47:14.477701 storage/replica.go:1520 store=1:1 range=2 ["m"-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:3} for range 2: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:2 StoreID:2 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:3}]
I160814 04:47:14.478431 storage/replica.go:2144 store=1:1 range=2 ["m"-/Max): applied part of replica change txn: EndTransaction [/Local/Range/"m"/RangeDescriptor,/Min), pErr=<nil>
I160814 04:47:14.478707 storage/replica.go:2144 store=2:2 range=2 ["m"-/Max): applied part of replica change txn: EndTransaction [/Local/Range/"m"/RangeDescriptor,/Min), pErr=<nil>
I160814 04:47:14.478813 stopper.go:408 quiesceing; tasks left:
1 storage/intent_resolver.go:346
W160814 04:47:14.478905 storage/intent_resolver.go:325 store=1:1 range=2 ["m"-/Max): failed to resolve intents: failed to send RPC: sending to all 1 replicas failed; last error: failed to send RPC: store is stopped
I160814 04:47:14.882525 storage/replica_raftstorage.go:474 generated snapshot for range 2 at index 15 in 157.451µs. encoded size=4094, 27 KV pairs, 5 log entries
I160814 04:47:14.883220 storage/replica_raftstorage.go:600 store=3:3 range=2 [/Min-/Min): with replicaID 3, applying Raft snapshot for range 2 at index 15 (encoded size=4094, 27 KV pairs, 5 log entries)
I160814 04:47:14.883571 storage/replica_raftstorage.go:603 store=3:3 range=2 ["m"-/Max): with replicaID 3, applied Raft snapshot for range 2 in 317.36µs
I160814 04:47:14.883606 storage/replica_trigger.go:308 store=2:2 range=2 ["m"-/Max): new range lease replica {2 2 2} 1970-01-01 00:00:00.9 +0000 UTC 1.8s following replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms [physicalTime=1970-01-01 00:00:01.8 +0000 UTC]
I160814 04:47:14.884006 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.884073 gossip/server.go:283 node 1: replying to 2
E160814 04:47:14.885809 storage/client_test.go:863 engine 0: missing key "z"
E160814 04:47:14.885846 storage/client_test.go:863 engine 2: missing key "z"
E160814 04:47:14.885950 storage/client_test.go:863 engine 0: missing key "z"
E160814 04:47:14.885987 storage/client_test.go:863 engine 2: missing key "z"
I160814 04:47:14.886112 stopper.go:408 quiesceing; tasks left:
1 storage/intent_resolver.go:306
I160814 04:47:14.886133 stopper.go:408 quiesceing; tasks left:
1 storage/client_test.go:425
W160814 04:47:14.886438 storage/replica.go:410 store=2:2 range=2 ["m"-/Max): outgoing raft transport stream to {3 3 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:3 StoreID:3 ReplicaID:3}: no store registered for {NodeID:2 StoreID:2 ReplicaID:2}
W160814 04:47:14.886466 storage/intent_resolver.go:299 store=1:1 range=1 [/Min-"m"): failed to resolve intents: node unavailable; try another peer
W160814 04:47:14.886662 storage/replica.go:410 store=1:1 range=2 ["m"-/Max): outgoing raft transport stream to {2 2 2} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:2}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160814 04:47:14.886977 http2_server.go:276 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:39006->127.0.0.1:57106: use of closed network connection
I160814 04:47:14.886991 http2_server.go:276 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:32884->127.0.0.1:49928: use of closed network connection
--- FAIL: TestStoreMetrics (0.43s)
client_metrics_test.go:94: cluster_id:<00000000-0000-0000-0000-000000000000> node_id:1 store_id:1 : expected intent count to be zero, was 1
```
Run Details:
```
0 runs so far, 0 failures, over 5s
0 runs so far, 0 failures, over 10s
0 runs so far, 0 failures, over 15s
0 runs so far, 0 failures, over 20s
0 runs so far, 0 failures, over 25s
0 runs so far, 0 failures, over 30s
0 runs so far, 0 failures, over 35s
15 runs so far, 0 failures, over 40s
15 runs so far, 0 failures, over 45s
16 runs so far, 0 failures, over 50s
16 runs so far, 0 failures, over 55s
16 runs so far, 0 failures, over 1m0s
16 runs so far, 0 failures, over 1m5s
16 runs so far, 0 failures, over 1m10s
17 runs so far, 0 failures, over 1m15s
19 runs completed, 1 failures, over 1m16s
FAIL
```
Please assign, take a look and update the issue accordingly. | 1.0 | stress: failed test in cockroach/storage/storage.test: TestStoreMetrics - Binary: cockroach/static-tests.tar.gz sha: https://github.com/cockroachdb/cockroach/commits/b44ebfb6d0911c8d6e9c6d8fc1099d592791a894
Stress build found a failed test:
```
=== RUN TestStoreMetrics
I160814 04:47:14.456048 storage/engine/rocksdb.go:354 opening in memory rocksdb instance
W160814 04:47:14.456756 gossip/gossip.go:1017 not connected to cluster; use --join to specify a connected node
I160814 04:47:14.458453 storage/store.go:2525 store=1:1: system config not yet available
I160814 04:47:14.459163 storage/replica_trigger.go:308 store=1:1 range=1 [/Min-/Max): new range lease replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms following replica {0 0 0} 1970-01-01 00:00:00 +0000 UTC 0s [physicalTime=1970-01-01 00:00:00 +0000 UTC]
I160814 04:47:14.459542 storage/engine/rocksdb.go:354 opening in memory rocksdb instance
W160814 04:47:14.460351 gossip/gossip.go:1017 not connected to cluster; use --join to specify a connected node
I160814 04:47:14.461517 storage/store.go:2525 store=2:2: system config not yet available
I160814 04:47:14.461539 gossip/client.go:75 node 2: starting client to 127.0.0.1:35096
I160814 04:47:14.461668 storage/engine/rocksdb.go:354 opening in memory rocksdb instance
I160814 04:47:14.461933 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.461968 gossip/server.go:283 node 1: replying to 2
I160814 04:47:14.462041 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.462095 gossip/server.go:283 node 1: replying to 2
I160814 04:47:14.462231 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.462255 gossip/server.go:283 node 1: replying to 2
I160814 04:47:14.462305 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.462321 gossip/server.go:283 node 1: replying to 2
W160814 04:47:14.462438 gossip/gossip.go:1017 not connected to cluster; use --join to specify a connected node
I160814 04:47:14.463211 storage/store.go:2525 store=3:3: system config not yet available
I160814 04:47:14.463343 gossip/client.go:75 node 3: starting client to 127.0.0.1:35096
I160814 04:47:14.463633 gossip/server.go:185 node 1: received gossip from node 3
I160814 04:47:14.463660 gossip/server.go:283 node 1: replying to 3
I160814 04:47:14.463737 gossip/server.go:185 node 1: received gossip from node 3
I160814 04:47:14.463765 gossip/server.go:283 node 1: replying to 3
I160814 04:47:14.463916 gossip/server.go:185 node 1: received gossip from node 3
I160814 04:47:14.463941 gossip/server.go:283 node 1: replying to 3
I160814 04:47:14.464013 gossip/server.go:185 node 1: received gossip from node 3
I160814 04:47:14.464037 gossip/server.go:283 node 1: replying to 3
I160814 04:47:14.465398 storage/replica_command.go:2216 store=1:1 range=1 [/Min-/Max): initiating a split of this range at key "m"
E160814 04:47:14.467367 storage/queue.go:508 [replicate] (purgatory) on store=1:1 range=1 [/Min-"m"): 0 of 0 stores with an attribute matching []; likely not enough nodes in cluster
E160814 04:47:14.467534 storage/queue.go:508 [replicate] (purgatory) on store=1:1 range=2 ["m"-/Max): 0 of 0 stores with an attribute matching []; likely not enough nodes in cluster
I160814 04:47:14.472228 storage/replica_raftstorage.go:474 generated snapshot for range 2 at index 10 in 110.62µs. encoded size=2407, 23 KV pairs, 0 log entries
I160814 04:47:14.472476 storage/replica_command.go:3059 store=1:1 range=2 ["m"-/Max): change replicas of 2: read existing descriptor range_id:2 start_key:"m" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > next_replica_id:2
I160814 04:47:14.472874 storage/replica_raftstorage.go:600 store=2:2 range=2 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 2 at index 10 (encoded size=2407, 23 KV pairs, 0 log entries)
I160814 04:47:14.473136 storage/replica.go:2144 store=1:1 range=1 [/Min-"m"): applied part of replica change txn: Noop [/Min,/Min), Noop [/Min,/Min), Put [/Meta2/Max,/Min), pErr=<nil>
I160814 04:47:14.473627 storage/replica.go:2144 store=1:1 range=2 ["m"-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/"m"/RangeDescriptor,/Min), ConditionalPut [/Local/Range/"m"/RangeDescriptor,/Min), Noop [/Min,/Min), pErr=<nil>
I160814 04:47:14.473901 storage/replica.go:1520 store=1:1 range=2 ["m"-/Max): proposing ADD_REPLICA {NodeID:2 StoreID:2 ReplicaID:2} for range 2: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:2 StoreID:2 ReplicaID:2}]
I160814 04:47:14.474091 storage/replica.go:2144 store=1:1 range=2 ["m"-/Max): applied part of replica change txn: EndTransaction [/Local/Range/"m"/RangeDescriptor,/Min), pErr=<nil>
I160814 04:47:14.474699 storage/replica_raftstorage.go:474 generated snapshot for range 2 at index 13 in 121.74µs. encoded size=3231, 25 KV pairs, 3 log entries
W160814 04:47:14.474835 storage/intent_resolver.go:344 could not GC completed transaction: end key /Local/Range/"m"/"rdsc\x00" must be greater than start /Local/Range/"m"/RangeDescriptor
I160814 04:47:14.474904 storage/replica_raftstorage.go:603 store=2:2 range=2 ["m"-/Max): with replicaID [?], applied preemptive snapshot for range 2 in 2.000445ms
I160814 04:47:14.474964 storage/replica_command.go:3059 store=1:1 range=2 ["m"-/Max): change replicas of 2: read existing descriptor range_id:2 start_key:"m" end_key:"\377\377" replicas:<node_id:1 store_id:1 replica_id:1 > replicas:<node_id:2 store_id:2 replica_id:2 > next_replica_id:3
I160814 04:47:14.475276 storage/replica.go:2144 store=1:1 range=1 [/Min-"m"): applied part of replica change txn: Noop [/Min,/Min), Noop [/Min,/Min), Put [/Meta2/Max,/Min), pErr=<nil>
I160814 04:47:14.475399 storage/replica_raftstorage.go:600 store=3:3 range=2 [/Min-/Min): with replicaID [?], applying preemptive snapshot for range 2 at index 13 (encoded size=3231, 25 KV pairs, 3 log entries)
I160814 04:47:14.475667 storage/replica_raftstorage.go:603 store=3:3 range=2 ["m"-/Max): with replicaID [?], applied preemptive snapshot for range 2 in 242.983µs
I160814 04:47:14.477021 storage/replica.go:2144 store=2:2 range=2 ["m"-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/"m"/RangeDescriptor,/Min), ConditionalPut [/Local/Range/"m"/RangeDescriptor,/Min), Noop [/Min,/Min), pErr=<nil>
I160814 04:47:14.477215 storage/replica.go:2144 store=2:2 range=2 ["m"-/Max): applied part of replica change txn: EndTransaction [/Local/Range/"m"/RangeDescriptor,/Min), pErr=<nil>
I160814 04:47:14.477338 storage/replica.go:2144 store=1:1 range=2 ["m"-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/"m"/RangeDescriptor,/Min), ConditionalPut [/Local/Range/"m"/RangeDescriptor,/Min), Noop [/Min,/Min), pErr=<nil>
I160814 04:47:14.477632 storage/replica.go:2144 store=2:2 range=2 ["m"-/Max): applied part of replica change txn: BeginTransaction [/Local/Range/"m"/RangeDescriptor,/Min), ConditionalPut [/Local/Range/"m"/RangeDescriptor,/Min), Noop [/Min,/Min), pErr=<nil>
I160814 04:47:14.477701 storage/replica.go:1520 store=1:1 range=2 ["m"-/Max): proposing ADD_REPLICA {NodeID:3 StoreID:3 ReplicaID:3} for range 2: [{NodeID:1 StoreID:1 ReplicaID:1} {NodeID:2 StoreID:2 ReplicaID:2} {NodeID:3 StoreID:3 ReplicaID:3}]
I160814 04:47:14.478431 storage/replica.go:2144 store=1:1 range=2 ["m"-/Max): applied part of replica change txn: EndTransaction [/Local/Range/"m"/RangeDescriptor,/Min), pErr=<nil>
I160814 04:47:14.478707 storage/replica.go:2144 store=2:2 range=2 ["m"-/Max): applied part of replica change txn: EndTransaction [/Local/Range/"m"/RangeDescriptor,/Min), pErr=<nil>
I160814 04:47:14.478813 stopper.go:408 quiesceing; tasks left:
1 storage/intent_resolver.go:346
W160814 04:47:14.478905 storage/intent_resolver.go:325 store=1:1 range=2 ["m"-/Max): failed to resolve intents: failed to send RPC: sending to all 1 replicas failed; last error: failed to send RPC: store is stopped
I160814 04:47:14.882525 storage/replica_raftstorage.go:474 generated snapshot for range 2 at index 15 in 157.451µs. encoded size=4094, 27 KV pairs, 5 log entries
I160814 04:47:14.883220 storage/replica_raftstorage.go:600 store=3:3 range=2 [/Min-/Min): with replicaID 3, applying Raft snapshot for range 2 at index 15 (encoded size=4094, 27 KV pairs, 5 log entries)
I160814 04:47:14.883571 storage/replica_raftstorage.go:603 store=3:3 range=2 ["m"-/Max): with replicaID 3, applied Raft snapshot for range 2 in 317.36µs
I160814 04:47:14.883606 storage/replica_trigger.go:308 store=2:2 range=2 ["m"-/Max): new range lease replica {2 2 2} 1970-01-01 00:00:00.9 +0000 UTC 1.8s following replica {1 1 1} 1970-01-01 00:00:00 +0000 UTC 900ms [physicalTime=1970-01-01 00:00:01.8 +0000 UTC]
I160814 04:47:14.884006 gossip/server.go:185 node 1: received gossip from node 2
I160814 04:47:14.884073 gossip/server.go:283 node 1: replying to 2
E160814 04:47:14.885809 storage/client_test.go:863 engine 0: missing key "z"
E160814 04:47:14.885846 storage/client_test.go:863 engine 2: missing key "z"
E160814 04:47:14.885950 storage/client_test.go:863 engine 0: missing key "z"
E160814 04:47:14.885987 storage/client_test.go:863 engine 2: missing key "z"
I160814 04:47:14.886112 stopper.go:408 quiesceing; tasks left:
1 storage/intent_resolver.go:306
I160814 04:47:14.886133 stopper.go:408 quiesceing; tasks left:
1 storage/client_test.go:425
W160814 04:47:14.886438 storage/replica.go:410 store=2:2 range=2 ["m"-/Max): outgoing raft transport stream to {3 3 3} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:3 StoreID:3 ReplicaID:3}: no store registered for {NodeID:2 StoreID:2 ReplicaID:2}
W160814 04:47:14.886466 storage/intent_resolver.go:299 store=1:1 range=1 [/Min-"m"): failed to resolve intents: node unavailable; try another peer
W160814 04:47:14.886662 storage/replica.go:410 store=1:1 range=2 ["m"-/Max): outgoing raft transport stream to {2 2 2} closed by the remote: rpc error: code = 2 desc = unable to accept Raft message from {NodeID:2 StoreID:2 ReplicaID:2}: no store registered for {NodeID:1 StoreID:1 ReplicaID:1}
I160814 04:47:14.886977 http2_server.go:276 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:39006->127.0.0.1:57106: use of closed network connection
I160814 04:47:14.886991 http2_server.go:276 transport: http2Server.HandleStreams failed to read frame: read tcp 127.0.0.1:32884->127.0.0.1:49928: use of closed network connection
--- FAIL: TestStoreMetrics (0.43s)
client_metrics_test.go:94: cluster_id:<00000000-0000-0000-0000-000000000000> node_id:1 store_id:1 : expected intent count to be zero, was 1
```
Run Details:
```
0 runs so far, 0 failures, over 5s
0 runs so far, 0 failures, over 10s
0 runs so far, 0 failures, over 15s
0 runs so far, 0 failures, over 20s
0 runs so far, 0 failures, over 25s
0 runs so far, 0 failures, over 30s
0 runs so far, 0 failures, over 35s
15 runs so far, 0 failures, over 40s
15 runs so far, 0 failures, over 45s
16 runs so far, 0 failures, over 50s
16 runs so far, 0 failures, over 55s
16 runs so far, 0 failures, over 1m0s
16 runs so far, 0 failures, over 1m5s
16 runs so far, 0 failures, over 1m10s
17 runs so far, 0 failures, over 1m15s
19 runs completed, 1 failures, over 1m16s
FAIL
```
Please assign, take a look and update the issue accordingly. | non_priority | stress failed test in cockroach storage storage test teststoremetrics binary cockroach static tests tar gz sha stress build found a failed test run teststoremetrics storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go store system config not yet available storage replica trigger go store range storage engine rocksdb go opening in memory rocksdb instance gossip gossip go not connected to cluster use join to specify a connected node storage store go store system config not yet available gossip client go node starting client to storage engine rocksdb go opening in memory rocksdb instance gossip server go node received gossip from node gossip server go node replying to gossip server go node received gossip from node gossip server go node replying to gossip server go node received gossip from node gossip server go node replying to gossip server go node received gossip from node gossip server go node replying to gossip gossip go not connected to cluster use join to specify a connected node storage store go store system config not yet available gossip client go node starting client to gossip server go node received gossip from node gossip server go node replying to gossip server go node received gossip from node gossip server go node replying to gossip server go node received gossip from node gossip server go node replying to gossip server go node received gossip from node gossip server go node replying to storage replica command go store range min max initiating a split of this range at key m storage queue go purgatory on store range likely not enough nodes in cluster storage queue go purgatory on store range likely not enough nodes in cluster storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica command go store range m max change replicas of read existing descriptor range id start key m end key replicas next replica id storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica go store range min m applied part of replica change txn noop min min noop min min put max min perr storage replica go store range m max applied part of replica change txn begintransaction local range m rangedescriptor min conditionalput local range m rangedescriptor min noop min min perr storage replica go store range storage replica go store range m max applied part of replica change txn endtransaction local range m rangedescriptor min perr storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage intent resolver go could not gc completed transaction end key local range m rdsc must be greater than start local range m rangedescriptor storage replica raftstorage go store range applied preemptive snapshot for range in storage replica command go store range m max change replicas of read existing descriptor range id start key m end key replicas replicas next replica id storage replica go store range min m applied part of replica change txn noop min min noop min min put max min perr storage replica raftstorage go store range applying preemptive snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range applied preemptive snapshot for range in storage replica go store range m max applied part of replica change txn begintransaction local range m rangedescriptor min conditionalput local range m rangedescriptor min noop min min perr storage replica go store range m max applied part of replica change txn endtransaction local range m rangedescriptor min perr storage replica go store range m max applied part of replica change txn begintransaction local range m rangedescriptor min conditionalput local range m rangedescriptor min noop min min perr storage replica go store range m max applied part of replica change txn begintransaction local range m rangedescriptor min conditionalput local range m rangedescriptor min noop min min perr storage replica go store range storage replica go store range m max applied part of replica change txn endtransaction local range m rangedescriptor min perr storage replica go store range m max applied part of replica change txn endtransaction local range m rangedescriptor min perr stopper go quiesceing tasks left storage intent resolver go storage intent resolver go store range m max failed to resolve intents failed to send rpc sending to all replicas failed last error failed to send rpc store is stopped storage replica raftstorage go generated snapshot for range at index in encoded size kv pairs log entries storage replica raftstorage go store range min min with replicaid applying raft snapshot for range at index encoded size kv pairs log entries storage replica raftstorage go store range m max with replicaid applied raft snapshot for range in storage replica trigger go store range gossip server go node received gossip from node gossip server go node replying to storage client test go engine missing key z storage client test go engine missing key z storage client test go engine missing key z storage client test go engine missing key z stopper go quiesceing tasks left storage intent resolver go stopper go quiesceing tasks left storage client test go storage replica go store range m max outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid storage intent resolver go store range min m failed to resolve intents node unavailable try another peer storage replica go store range m max outgoing raft transport stream to closed by the remote rpc error code desc unable to accept raft message from nodeid storeid replicaid no store registered for nodeid storeid replicaid server go transport handlestreams failed to read frame read tcp use of closed network connection server go transport handlestreams failed to read frame read tcp use of closed network connection fail teststoremetrics client metrics test go cluster id node id store id expected intent count to be zero was run details runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs so far failures over runs completed failures over fail please assign take a look and update the issue accordingly | 0 |
30,082 | 7,160,747,555 | IssuesEvent | 2018-01-28 05:02:43 | zurb/foundation-sites | https://api.github.com/repos/zurb/foundation-sites | closed | Version 6.2.1 priorities | codebase help wanted | We've deliberately waited a little longer for this first patch to 6.2 so we could see what unfolds. Current rough timeline for release is "a few weeks from now".
- [x] Add precompiled JavaScript to `dist/` folder #8406
- [x] Verify foundation-zurb-template will work with Babel's es2015 preset unconfigured #8256 #8324
- [x] Resolve issues with migration to `$foundation-palette` #8281 #8292
- [x] Handle Babel errors in foundation-zurb-template better #8340
- [x] Flexbox mode issues #8325 #8409
- [x] **When 6.2.1 is released,** merge zurb/foundation-zurb-template#34
@andycochran @colin-marshall @JeremyEnglert Have y'alls seen anything critical specific to 6.2 that's worth addressing?
| 1.0 | Version 6.2.1 priorities - We've deliberately waited a little longer for this first patch to 6.2 so we could see what unfolds. Current rough timeline for release is "a few weeks from now".
- [x] Add precompiled JavaScript to `dist/` folder #8406
- [x] Verify foundation-zurb-template will work with Babel's es2015 preset unconfigured #8256 #8324
- [x] Resolve issues with migration to `$foundation-palette` #8281 #8292
- [x] Handle Babel errors in foundation-zurb-template better #8340
- [x] Flexbox mode issues #8325 #8409
- [x] **When 6.2.1 is released,** merge zurb/foundation-zurb-template#34
@andycochran @colin-marshall @JeremyEnglert Have y'alls seen anything critical specific to 6.2 that's worth addressing?
| non_priority | version priorities we ve deliberately waited a little longer for this first patch to so we could see what unfolds current rough timeline for release is a few weeks from now add precompiled javascript to dist folder verify foundation zurb template will work with babel s preset unconfigured resolve issues with migration to foundation palette handle babel errors in foundation zurb template better flexbox mode issues when is released merge zurb foundation zurb template andycochran colin marshall jeremyenglert have y alls seen anything critical specific to that s worth addressing | 0 |
206,647 | 23,396,789,998 | IssuesEvent | 2022-08-12 01:09:21 | rvvergara/react-weather-app | https://api.github.com/repos/rvvergara/react-weather-app | closed | CVE-2021-35065 (High) detected in multiple libraries - autoclosed | security vulnerability | ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-5.1.1.tgz</b>, <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-2.0.0.tgz</b></p></summary>
<p>
<details><summary><b>glob-parent-5.1.1.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/watchpack/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- webpack-4.43.0.tgz (Root Library)
- watchpack-1.7.2.tgz
- chokidar-3.4.1.tgz
- :x: **glob-parent-5.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/watchpack-chokidar2/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- webpack-4.43.0.tgz (Root Library)
- watchpack-1.7.2.tgz
- watchpack-chokidar2-2.0.0.tgz
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-2.0.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- babel-cli-6.26.0.tgz (Root Library)
- chokidar-1.7.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (webpack): 5.0.0</p><p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (webpack): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-35065 (High) detected in multiple libraries - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-5.1.1.tgz</b>, <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-2.0.0.tgz</b></p></summary>
<p>
<details><summary><b>glob-parent-5.1.1.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/watchpack/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- webpack-4.43.0.tgz (Root Library)
- watchpack-1.7.2.tgz
- chokidar-3.4.1.tgz
- :x: **glob-parent-5.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/watchpack-chokidar2/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- webpack-4.43.0.tgz (Root Library)
- watchpack-1.7.2.tgz
- watchpack-chokidar2-2.0.0.tgz
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-2.0.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- babel-cli-6.26.0.tgz (Root Library)
- chokidar-1.7.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (webpack): 5.0.0</p><p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (webpack): 5.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in multiple libraries autoclosed cve high severity vulnerability vulnerable libraries glob parent tgz glob parent tgz glob parent tgz glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file package json path to vulnerable library node modules watchpack node modules glob parent package json dependency hierarchy webpack tgz root library watchpack tgz chokidar tgz x glob parent tgz vulnerable library glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file package json path to vulnerable library node modules watchpack node modules glob parent package json dependency hierarchy webpack tgz root library watchpack tgz watchpack tgz chokidar tgz x glob parent tgz vulnerable library glob parent tgz strips glob magic from a string to provide the parent path library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy babel cli tgz root library chokidar tgz x glob parent tgz vulnerable library found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent direct dependency fix resolution webpack fix resolution glob parent direct dependency fix resolution webpack step up your open source security game with mend | 0 |
165,492 | 26,180,444,615 | IssuesEvent | 2023-01-02 14:56:35 | esi-neuroscience/syncopy | https://api.github.com/repos/esi-neuroscience/syncopy | closed | copy_trial is obsolete | Design Minor Bug | **Describe the bug**
After seven months of testing/running Syncopy, `copy_trial` is still completely unused. We should remove the routine and the corresponding tests to avoid any unintentional side-effects downstream. | 1.0 | copy_trial is obsolete - **Describe the bug**
After seven months of testing/running Syncopy, `copy_trial` is still completely unused. We should remove the routine and the corresponding tests to avoid any unintentional side-effects downstream. | non_priority | copy trial is obsolete describe the bug after seven months of testing running syncopy copy trial is still completely unused we should remove the routine and the corresponding tests to avoid any unintentional side effects downstream | 0 |
93,857 | 27,061,299,629 | IssuesEvent | 2023-02-13 20:00:43 | dotnet/arcade | https://api.github.com/repos/dotnet/arcade | closed | Build failed: Maestro Build Promotion/main #Promoting dotnet-machinelearning build 20230209.1 (166353) to channel(s) '.NET Libraries' # | First Responder Build Failed | Build [#Promoting dotnet-machinelearning build 20230209.1 (166353) to channel(s) '.NET Libraries' #](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2109605) failed
## :x: : internal / Maestro Build Promotion failed
### Summary
**Finished** - Thu, 09 Feb 2023 21:38:23 GMT
**Duration** - 5 minutes
**Requested for** - DotNet Bot
**Reason** - manual
### Details
#### Publishing
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\ccc26692-701d-431b-874b-c3e0d27dff0b\MergedManifest.xml to assets/manifests/dotnet-machinelearning/3.0.0-preview.23109.1/MergedManifest.xml: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44fed9-401e-0045-11ce-3ce5f8000000
Time:2023-02-09T21:36:17.0969836Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44fed9-401e-0045-11ce-3ce5f8000000
Time:2023-02-09T21:36:17.0969836Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 8c44fed9-401e-0045-11ce-3ce5f8000000
x-ms-client-request-id: 72715813-0def-4992-944e-7fd162be1845
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\ddf366f5-ba7f-4b43-8161-e9e45db0f7ed\Microsoft.ML.AutoML.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.AutoML.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:0e6947ed-d01e-004b-52ce-3c09f3000000
Time:2023-02-09T21:36:17.1033689Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:0e6947ed-d01e-004b-52ce-3c09f3000000
Time:2023-02-09T21:36:17.1033689Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 0e6947ed-d01e-004b-52ce-3c09f3000000
x-ms-client-request-id: 03a074e6-80c4-4416-97f7-2116a62f9afe
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\6085ad88-3b4d-4bb8-bca9-caad27c363fe\Microsoft.Data.Analysis.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.Data.Analysis.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:a0e4a0ed-e01e-0048-2ece-3c0af4000000
Time:2023-02-09T21:36:17.0994000Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:a0e4a0ed-e01e-0048-2ece-3c0af4000000
Time:2023-02-09T21:36:17.0994000Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: a0e4a0ed-e01e-0048-2ece-3c0af4000000
x-ms-client-request-id: 3a07a8d7-8bc2-4cd4-8ef6-3994daf4817e
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\4b9396d4-c384-42c3-be3f-7f9943f7f18b\Microsoft.Extensions.ML.3.0.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.Extensions.ML.3.0.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44fef5-401e-0045-2bce-3ce5f8000000
Time:2023-02-09T21:36:17.1169727Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44fef5-401e-0045-2bce-3ce5f8000000
Time:2023-02-09T21:36:17.1169727Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 8c44fef5-401e-0045-2bce-3ce5f8000000
x-ms-client-request-id: f691b990-0694-4782-8c82-29e177f460b2
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\a9ffba31-cb66-40f1-a32e-4da3b5d5d04e\Microsoft.ML.DataView.3.0.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.DataView.3.0.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:3b53d487-701e-002b-5bce-3c4cd1000000
Time:2023-02-09T21:36:17.1166533Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:3b53d487-701e-002b-5bce-3c4cd1000000
Time:2023-02-09T21:36:17.1166533Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 3b53d487-701e-002b-5bce-3c4cd1000000
x-ms-client-request-id: e3b19de3-b3e6-4694-bd0e-f5d13be8e095
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\92277a44-d6c8-4744-8e3f-63b90b23df85\Microsoft.ML.CodeGenerator.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.CodeGenerator.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:a0e4a0fe-e01e-0048-3fce-3c0af4000000
Time:2023-02-09T21:36:17.1323812Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:a0e4a0fe-e01e-0048-3fce-3c0af4000000
Time:2023-02-09T21:36:17.1323812Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: a0e4a0fe-e01e-0048-3fce-3c0af4000000
x-ms-client-request-id: 5b7ad268-6f36-44b8-8547-800bc3064799
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\397a6d2a-2b9e-4f77-a769-246163b70b95\Microsoft.ML.CpuMath.3.0.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.CpuMath.3.0.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:3b53d494-701e-002b-66ce-3c4cd1000000
Time:2023-02-09T21:36:17.1406393Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:3b53d494-701e-002b-66ce-3c4cd1000000
Time:2023-02-09T21:36:17.1406393Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 3b53d494-701e-002b-66ce-3c4cd1000000
x-ms-client-request-id: 57c03d79-c8a7-4b6c-bc8b-4f1768a8503f
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\e2f9a929-52da-4c80-95c1-95f38717c5ad\Microsoft.ML.3.0.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.3.0.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44feff-401e-0045-35ce-3ce5f8000000
Time:2023-02-09T21:36:17.1279663Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44feff-401e-0045-35ce-3ce5f8000000
Time:2023-02-09T21:36:17.1279663Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 8c44feff-401e-0045-35ce-3ce5f8000000
x-ms-client-request-id: 5efae11a-9fe6-440c-b0e5-e4d99ce063ed
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\bda42784-067e-4a08-b581-bfa89290f51d\Microsoft.ML.Ensemble.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.Ensemble.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:88b852f9-101e-009a-03ce-3cb4ac000000
Time:2023-02-09T21:36:42.5872229Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:88b852f9-101e-009a-03ce-3cb4ac000000
Time:2023-02-09T21:36:42.5872229Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 88b852f9-101e-009a-03ce-3cb4ac000000
x-ms-client-request-id: c2f96ed4-b7a2-4737-9322-194bb9ffddcc
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:41 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\07d52e84-b550-494b-8842-cc00af0896c0\Microsoft.ML.Experimental.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.Experimental.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:88b8531e-101e-009a-25ce-3cb4ac000000
Time:2023-02-09T21:36:42.7421335Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:88b8531e-101e-009a-25ce-3cb4ac000000
Time:2023-02-09T21:36:42.7421335Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 88b8531e-101e-009a-25ce-3cb4ac000000
x-ms-client-request-id: 194e78a8-192e-4c2e-981f-a64973dfee7e
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:41 GMT
Content-Length: 302
Content-Type: application/xml
</details>
### Changes
| 1.0 | Build failed: Maestro Build Promotion/main #Promoting dotnet-machinelearning build 20230209.1 (166353) to channel(s) '.NET Libraries' # - Build [#Promoting dotnet-machinelearning build 20230209.1 (166353) to channel(s) '.NET Libraries' #](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2109605) failed
## :x: : internal / Maestro Build Promotion failed
### Summary
**Finished** - Thu, 09 Feb 2023 21:38:23 GMT
**Duration** - 5 minutes
**Requested for** - DotNet Bot
**Reason** - manual
### Details
#### Publishing
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\ccc26692-701d-431b-874b-c3e0d27dff0b\MergedManifest.xml to assets/manifests/dotnet-machinelearning/3.0.0-preview.23109.1/MergedManifest.xml: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44fed9-401e-0045-11ce-3ce5f8000000
Time:2023-02-09T21:36:17.0969836Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44fed9-401e-0045-11ce-3ce5f8000000
Time:2023-02-09T21:36:17.0969836Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 8c44fed9-401e-0045-11ce-3ce5f8000000
x-ms-client-request-id: 72715813-0def-4992-944e-7fd162be1845
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\ddf366f5-ba7f-4b43-8161-e9e45db0f7ed\Microsoft.ML.AutoML.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.AutoML.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:0e6947ed-d01e-004b-52ce-3c09f3000000
Time:2023-02-09T21:36:17.1033689Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:0e6947ed-d01e-004b-52ce-3c09f3000000
Time:2023-02-09T21:36:17.1033689Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 0e6947ed-d01e-004b-52ce-3c09f3000000
x-ms-client-request-id: 03a074e6-80c4-4416-97f7-2116a62f9afe
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\6085ad88-3b4d-4bb8-bca9-caad27c363fe\Microsoft.Data.Analysis.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.Data.Analysis.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:a0e4a0ed-e01e-0048-2ece-3c0af4000000
Time:2023-02-09T21:36:17.0994000Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:a0e4a0ed-e01e-0048-2ece-3c0af4000000
Time:2023-02-09T21:36:17.0994000Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: a0e4a0ed-e01e-0048-2ece-3c0af4000000
x-ms-client-request-id: 3a07a8d7-8bc2-4cd4-8ef6-3994daf4817e
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\4b9396d4-c384-42c3-be3f-7f9943f7f18b\Microsoft.Extensions.ML.3.0.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.Extensions.ML.3.0.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44fef5-401e-0045-2bce-3ce5f8000000
Time:2023-02-09T21:36:17.1169727Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44fef5-401e-0045-2bce-3ce5f8000000
Time:2023-02-09T21:36:17.1169727Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 8c44fef5-401e-0045-2bce-3ce5f8000000
x-ms-client-request-id: f691b990-0694-4782-8c82-29e177f460b2
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\a9ffba31-cb66-40f1-a32e-4da3b5d5d04e\Microsoft.ML.DataView.3.0.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.DataView.3.0.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:3b53d487-701e-002b-5bce-3c4cd1000000
Time:2023-02-09T21:36:17.1166533Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:3b53d487-701e-002b-5bce-3c4cd1000000
Time:2023-02-09T21:36:17.1166533Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 3b53d487-701e-002b-5bce-3c4cd1000000
x-ms-client-request-id: e3b19de3-b3e6-4694-bd0e-f5d13be8e095
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\92277a44-d6c8-4744-8e3f-63b90b23df85\Microsoft.ML.CodeGenerator.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.CodeGenerator.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:a0e4a0fe-e01e-0048-3fce-3c0af4000000
Time:2023-02-09T21:36:17.1323812Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:a0e4a0fe-e01e-0048-3fce-3c0af4000000
Time:2023-02-09T21:36:17.1323812Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: a0e4a0fe-e01e-0048-3fce-3c0af4000000
x-ms-client-request-id: 5b7ad268-6f36-44b8-8547-800bc3064799
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\397a6d2a-2b9e-4f77-a769-246163b70b95\Microsoft.ML.CpuMath.3.0.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.CpuMath.3.0.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:3b53d494-701e-002b-66ce-3c4cd1000000
Time:2023-02-09T21:36:17.1406393Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:3b53d494-701e-002b-66ce-3c4cd1000000
Time:2023-02-09T21:36:17.1406393Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 3b53d494-701e-002b-66ce-3c4cd1000000
x-ms-client-request-id: 57c03d79-c8a7-4b6c-bc8b-4f1768a8503f
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\e2f9a929-52da-4c80-95c1-95f38717c5ad\Microsoft.ML.3.0.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.3.0.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44feff-401e-0045-35ce-3ce5f8000000
Time:2023-02-09T21:36:17.1279663Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:8c44feff-401e-0045-35ce-3ce5f8000000
Time:2023-02-09T21:36:17.1279663Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 8c44feff-401e-0045-35ce-3ce5f8000000
x-ms-client-request-id: 5efae11a-9fe6-440c-b0e5-e4d99ce063ed
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:16 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\bda42784-067e-4a08-b581-bfa89290f51d\Microsoft.ML.Ensemble.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.Ensemble.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:88b852f9-101e-009a-03ce-3cb4ac000000
Time:2023-02-09T21:36:42.5872229Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:88b852f9-101e-009a-03ce-3cb4ac000000
Time:2023-02-09T21:36:42.5872229Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 88b852f9-101e-009a-03ce-3cb4ac000000
x-ms-client-request-id: c2f96ed4-b7a2-4737-9322-194bb9ffddcc
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:41 GMT
Content-Length: 302
Content-Type: application/xml
</details>
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2109605/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23107.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Unexpected exception publishing file D:\a\_work\1\a\07d52e84-b550-494b-8842-cc00af0896c0\Microsoft.ML.Experimental.0.21.0-preview.23109.1.symbols.nupkg to assets/symbols/Microsoft.ML.Experimental.0.21.0-preview.23109.1.symbols.nupkg: Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:88b8531e-101e-009a-25ce-3cb4ac000000
Time:2023-02-09T21:36:42.7421335Z
<details>
Status: 401 (Server failed to authenticate the request. Please refer to the information in the www-authenticate header.)
ErrorCode: NoAuthenticationInformation
Content:
<?xml version="1.0" encoding="utf-8"?><Error><Code>NoAuthenticationInformation</Code><Message>Server failed to authenticate the request. Please refer to the information in the www-authenticate header.
RequestId:88b8531e-101e-009a-25ce-3cb4ac000000
Time:2023-02-09T21:36:42.7421335Z</Message></Error>
Headers:
Server: Windows-Azure-Blob/1.0,Microsoft-HTTPAPI/2.0
x-ms-request-id: 88b8531e-101e-009a-25ce-3cb4ac000000
x-ms-client-request-id: 194e78a8-192e-4c2e-981f-a64973dfee7e
x-ms-version: 2021-08-06
x-ms-error-code: NoAuthenticationInformation
WWW-Authenticate: Bearer authorization_uri=https://login.microsoftonline.com/72f988bf-86f1-41af-91ab-2d7cd011db47/oauth2/authorize resource_id=https://storage.azure.com
Date: Thu, 09 Feb 2023 21:36:41 GMT
Content-Length: 302
Content-Type: application/xml
</details>
### Changes
| non_priority | build failed maestro build promotion main promoting dotnet machinelearning build to channel s net libraries build failed x internal maestro build promotion failed summary finished thu feb gmt duration minutes requested for dotnet bot reason manual details publishing x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a mergedmanifest xml to assets manifests dotnet machinelearning preview mergedmanifest xml server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft ml automl preview symbols nupkg to assets symbols microsoft ml automl preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft data analysis preview symbols nupkg to assets symbols microsoft data analysis preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft extensions ml preview symbols nupkg to assets symbols microsoft extensions ml preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft ml dataview preview symbols nupkg to assets symbols microsoft ml dataview preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft ml codegenerator preview symbols nupkg to assets symbols microsoft ml codegenerator preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft ml cpumath preview symbols nupkg to assets symbols microsoft ml cpumath preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft ml preview symbols nupkg to assets symbols microsoft ml preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft ml ensemble preview symbols nupkg to assets symbols microsoft ml ensemble preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error unexpected exception publishing file d a work a microsoft ml experimental preview symbols nupkg to assets symbols microsoft ml experimental preview symbols nupkg server failed to authenticate the request please refer to the information in the www authenticate header requestid time status server failed to authenticate the request please refer to the information in the www authenticate header errorcode noauthenticationinformation content noauthenticationinformation server failed to authenticate the request please refer to the information in the www authenticate header requestid time headers server windows azure blob microsoft httpapi x ms request id x ms client request id x ms version x ms error code noauthenticationinformation www authenticate bearer authorization uri resource id date thu feb gmt content length content type application xml changes | 0 |
45,242 | 11,614,486,736 | IssuesEvent | 2020-02-26 12:40:58 | spack/spack | https://api.github.com/repos/spack/spack | opened | Installation issue: quantum-espresso does not find fftw3.f | build-error | Installation of quantum-espresso fails, because it can't find fftw3.f.
The issue is kind of known from the mailing list, and it seems like it is more a problem of quantum espresso than of spack.
https://lists.quantum-espresso.org/pipermail/users/2017-September/039159.html
However, it might still be possible to fix this problem in the package.py i hope, but I don't know how to do it.
### Steps to reproduce the issue
```console
$ spack install quantum-espresso~mpi~scalapack target=x86_64 ^/e4wa6u5^/z35g5nd
[+] /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/fftw-3.3.8-e4wa6u53rh6nm2wjaqfz6yuxra232xi7
[+] /shared/spack/opt/spack/linux-fedora29-core2/gcc-8.2.0/openblas-0.3.8-z35g5ndezfyipbxfeydqhmb4nz2nlc75
==> 9345: Installing quantum-espresso
==> Using cached archive: /shared/spack/var/spack/cache/_source-cache/archive/25/258b2a8a6280e86dad779e5c56356d8b35dc96d12ff33dabeee914bc03d6d602.tar.gz
==> Using cached archive: /shared/spack/var/spack/cache/_source-cache/archive/b1/b1aa3179ee1c069964fb9c21f3b832aebeae54947ce8d3cc1a74e7b154c3c10f
==> Staging archive: /shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/q-e-qe-6.5.tar.gz
==> Created stage in /shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io
==> Applied patch https://gitlab.com/QEF/q-e/commit/5fb1195b0844e1052b7601f18ab5c700f9cbe648.diff
==> 9345: quantum-espresso: Building quantum-espresso [Package]
==> 9345: quantum-espresso: Executing phase: 'install'
==> Error: ProcessError: Command exited with status 2:
'make' 'all'
2 errors found in build log:
1318 /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/gcc-8.2.0-ynlewml22i4peay3x4aytn3savkp3atw/bin/gfortran -O3 -g -x f95-cpp-input -D__FFTW3 -I/shared/spack/va
r/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//include -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-q
uantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//FoX/finclude -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5
ljogy2tsqxvqr7bm3io/spack-src//S3DE/iotk/include/ -c fft_scalar.FFTW.f90
1319 /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/gcc-8.2.0-ynlewml22i4peay3x4aytn3savkp3atw/bin/gfortran -O3 -g -x f95-cpp-input -D__FFTW3 -I/shared/spack/va
r/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//include -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-q
uantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//FoX/finclude -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5
ljogy2tsqxvqr7bm3io/spack-src//S3DE/iotk/include/ -c fft_scalar.FFTW3.f90
1320 fft_scalar.FFTW3.f90:37:2:
1321
1322 #endif
1323 1~~~~
>> 1324 Fatal Error: fftw3.f: No such file or directory
1325 compilation terminated.
>> 1326 make[1]: *** [../make.inc:16: fft_scalar.FFTW3.o] Error 1
1327 make[1]: Leaving directory '/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src/FFTXlib'
1328 make: *** [Makefile:207: libfft] Error 1
See build log for details:
/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-build-out.txt
==> Error: Failed to install quantum-espresso due to ChildError: ProcessError: Command exited with status 2:
'make' 'all'
2 errors found in build log:
1318 /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/gcc-8.2.0-ynlewml22i4peay3x4aytn3savkp3atw/bin/gfortran -O3 -g -x f95-cpp-input -D__FFTW3 -I/shared/spack/va
r/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//include -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-q
uantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//FoX/finclude -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5
ljogy2tsqxvqr7bm3io/spack-src//S3DE/iotk/include/ -c fft_scalar.FFTW.f90
1319 /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/gcc-8.2.0-ynlewml22i4peay3x4aytn3savkp3atw/bin/gfortran -O3 -g -x f95-cpp-input -D__FFTW3 -I/shared/spack/va
r/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//include -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-q
uantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//FoX/finclude -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5
ljogy2tsqxvqr7bm3io/spack-src//S3DE/iotk/include/ -c fft_scalar.FFTW3.f90
1320 fft_scalar.FFTW3.f90:37:2:
1321
1322 #endif
1323 1~~~~
>> 1324 Fatal Error: fftw3.f: No such file or directory
1325 compilation terminated.
>> 1326 make[1]: *** [../make.inc:16: fft_scalar.FFTW3.o] Error 1
1327 make[1]: Leaving directory '/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src/FFTXlib'
1328 make: *** [Makefile:207: libfft] Error 1
See build log for details:
/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-build-out.txt
Traceback (most recent call last):
File "/shared/spack/lib/spack/spack/build_environment.py", line 801, in child_process
return_value = function()
File "/shared/spack/lib/spack/spack/installer.py", line 1034, in build_process
phase(pkg.spec, pkg.prefix)
File "/shared/spack/lib/spack/spack/package.py", line 105, in phase_wrapper
phase(spec, prefix)
File "/shared/spack/var/spack/repos/builtin/packages/quantum-espresso/package.py", line 275, in install
File "/shared/spack/lib/spack/spack/build_environment.py", line 131, in __call__
return super(MakeExecutable, self).__call__(*args, **kwargs)
File "/shared/spack/lib/spack/spack/util/executable.py", line 189, in __call__
proc.returncode, long_msg)
spack.util.executable.ProcessError: Command exited with status 2:
'make' 'all'
```
### Platform and user environment
```commandline
$ uname -a
Linux biby33 4.18.18-300.fc29.x86_64 #1 SMP Mon Nov 12 03:12:14 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
$ cat /etc/fedora-release
Fedora release 29 (Twenty Nine)
```
### Additional information
excerpt from spack-build-out.txt
```
--------------------------------------------------------------------
ESPRESSO can take advantage of several optimized numerical libraries
(essl, fftw, mkl...). This configure script attempts to find them,
but may fail if they have been installed in non-standard locations.
If a required library is not found, the local copy will be compiled.
The following libraries have been found:
BLAS_LIBS=-L/shared/spack/opt/spack/linux-fedora29-core2/gcc-8.2.0/openblas-0.3.8-z35g5ndezfyipbxfeydqhmb4nz2nlc75/lib -lopenblas
LAPACK_LIBS= -L/shared/spack/opt/spack/linux-fedora29-core2/gcc-8.2.0/openblas-0.3.8-z35g5ndezfyipbxfeydqhmb4nz2nlc75/lib -lopenblas
FFT_LIBS= -L/shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/fftw-3.3.8-e4wa6u53rh6nm2wjaqfz6yuxra232xi7/lib -lfftw3
Please check if this is what you expect.
If any libraries are missing, you may specify a list of directories
to search and retry, as follows:
./configure LIBDIRS="list of directories, separated by spaces"
Configured for compilation of serial executables.
For more info, read the ESPRESSO User's Guide (Doc/users-guide.tex).
--------------------------------------------------------------------
```
So it finds FFTW but does not include the header files I guess. | 1.0 | Installation issue: quantum-espresso does not find fftw3.f - Installation of quantum-espresso fails, because it can't find fftw3.f.
The issue is kind of known from the mailing list, and it seems like it is more a problem of quantum espresso than of spack.
https://lists.quantum-espresso.org/pipermail/users/2017-September/039159.html
However, it might still be possible to fix this problem in the package.py i hope, but I don't know how to do it.
### Steps to reproduce the issue
```console
$ spack install quantum-espresso~mpi~scalapack target=x86_64 ^/e4wa6u5^/z35g5nd
[+] /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/fftw-3.3.8-e4wa6u53rh6nm2wjaqfz6yuxra232xi7
[+] /shared/spack/opt/spack/linux-fedora29-core2/gcc-8.2.0/openblas-0.3.8-z35g5ndezfyipbxfeydqhmb4nz2nlc75
==> 9345: Installing quantum-espresso
==> Using cached archive: /shared/spack/var/spack/cache/_source-cache/archive/25/258b2a8a6280e86dad779e5c56356d8b35dc96d12ff33dabeee914bc03d6d602.tar.gz
==> Using cached archive: /shared/spack/var/spack/cache/_source-cache/archive/b1/b1aa3179ee1c069964fb9c21f3b832aebeae54947ce8d3cc1a74e7b154c3c10f
==> Staging archive: /shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/q-e-qe-6.5.tar.gz
==> Created stage in /shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io
==> Applied patch https://gitlab.com/QEF/q-e/commit/5fb1195b0844e1052b7601f18ab5c700f9cbe648.diff
==> 9345: quantum-espresso: Building quantum-espresso [Package]
==> 9345: quantum-espresso: Executing phase: 'install'
==> Error: ProcessError: Command exited with status 2:
'make' 'all'
2 errors found in build log:
1318 /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/gcc-8.2.0-ynlewml22i4peay3x4aytn3savkp3atw/bin/gfortran -O3 -g -x f95-cpp-input -D__FFTW3 -I/shared/spack/va
r/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//include -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-q
uantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//FoX/finclude -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5
ljogy2tsqxvqr7bm3io/spack-src//S3DE/iotk/include/ -c fft_scalar.FFTW.f90
1319 /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/gcc-8.2.0-ynlewml22i4peay3x4aytn3savkp3atw/bin/gfortran -O3 -g -x f95-cpp-input -D__FFTW3 -I/shared/spack/va
r/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//include -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-q
uantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//FoX/finclude -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5
ljogy2tsqxvqr7bm3io/spack-src//S3DE/iotk/include/ -c fft_scalar.FFTW3.f90
1320 fft_scalar.FFTW3.f90:37:2:
1321
1322 #endif
1323 1~~~~
>> 1324 Fatal Error: fftw3.f: No such file or directory
1325 compilation terminated.
>> 1326 make[1]: *** [../make.inc:16: fft_scalar.FFTW3.o] Error 1
1327 make[1]: Leaving directory '/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src/FFTXlib'
1328 make: *** [Makefile:207: libfft] Error 1
See build log for details:
/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-build-out.txt
==> Error: Failed to install quantum-espresso due to ChildError: ProcessError: Command exited with status 2:
'make' 'all'
2 errors found in build log:
1318 /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/gcc-8.2.0-ynlewml22i4peay3x4aytn3savkp3atw/bin/gfortran -O3 -g -x f95-cpp-input -D__FFTW3 -I/shared/spack/va
r/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//include -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-q
uantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//FoX/finclude -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5
ljogy2tsqxvqr7bm3io/spack-src//S3DE/iotk/include/ -c fft_scalar.FFTW.f90
1319 /shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/gcc-8.2.0-ynlewml22i4peay3x4aytn3savkp3atw/bin/gfortran -O3 -g -x f95-cpp-input -D__FFTW3 -I/shared/spack/va
r/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//include -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-q
uantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src//FoX/finclude -I/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5
ljogy2tsqxvqr7bm3io/spack-src//S3DE/iotk/include/ -c fft_scalar.FFTW3.f90
1320 fft_scalar.FFTW3.f90:37:2:
1321
1322 #endif
1323 1~~~~
>> 1324 Fatal Error: fftw3.f: No such file or directory
1325 compilation terminated.
>> 1326 make[1]: *** [../make.inc:16: fft_scalar.FFTW3.o] Error 1
1327 make[1]: Leaving directory '/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-src/FFTXlib'
1328 make: *** [Makefile:207: libfft] Error 1
See build log for details:
/shared/spack/var/spack/stage/mbernhardt/spack-stage-quantum-espresso-6.5-3yxmuagwxd2o5ljogy2tsqxvqr7bm3io/spack-build-out.txt
Traceback (most recent call last):
File "/shared/spack/lib/spack/spack/build_environment.py", line 801, in child_process
return_value = function()
File "/shared/spack/lib/spack/spack/installer.py", line 1034, in build_process
phase(pkg.spec, pkg.prefix)
File "/shared/spack/lib/spack/spack/package.py", line 105, in phase_wrapper
phase(spec, prefix)
File "/shared/spack/var/spack/repos/builtin/packages/quantum-espresso/package.py", line 275, in install
File "/shared/spack/lib/spack/spack/build_environment.py", line 131, in __call__
return super(MakeExecutable, self).__call__(*args, **kwargs)
File "/shared/spack/lib/spack/spack/util/executable.py", line 189, in __call__
proc.returncode, long_msg)
spack.util.executable.ProcessError: Command exited with status 2:
'make' 'all'
```
### Platform and user environment
```commandline
$ uname -a
Linux biby33 4.18.18-300.fc29.x86_64 #1 SMP Mon Nov 12 03:12:14 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
$ cat /etc/fedora-release
Fedora release 29 (Twenty Nine)
```
### Additional information
excerpt from spack-build-out.txt
```
--------------------------------------------------------------------
ESPRESSO can take advantage of several optimized numerical libraries
(essl, fftw, mkl...). This configure script attempts to find them,
but may fail if they have been installed in non-standard locations.
If a required library is not found, the local copy will be compiled.
The following libraries have been found:
BLAS_LIBS=-L/shared/spack/opt/spack/linux-fedora29-core2/gcc-8.2.0/openblas-0.3.8-z35g5ndezfyipbxfeydqhmb4nz2nlc75/lib -lopenblas
LAPACK_LIBS= -L/shared/spack/opt/spack/linux-fedora29-core2/gcc-8.2.0/openblas-0.3.8-z35g5ndezfyipbxfeydqhmb4nz2nlc75/lib -lopenblas
FFT_LIBS= -L/shared/spack/opt/spack/linux-fedora29-x86_64/gcc-8.2.0/fftw-3.3.8-e4wa6u53rh6nm2wjaqfz6yuxra232xi7/lib -lfftw3
Please check if this is what you expect.
If any libraries are missing, you may specify a list of directories
to search and retry, as follows:
./configure LIBDIRS="list of directories, separated by spaces"
Configured for compilation of serial executables.
For more info, read the ESPRESSO User's Guide (Doc/users-guide.tex).
--------------------------------------------------------------------
```
So it finds FFTW but does not include the header files I guess. | non_priority | installation issue quantum espresso does not find f installation of quantum espresso fails because it can t find f the issue is kind of known from the mailing list and it seems like it is more a problem of quantum espresso than of spack however it might still be possible to fix this problem in the package py i hope but i don t know how to do it steps to reproduce the issue console spack install quantum espresso mpi scalapack target shared spack opt spack linux gcc fftw shared spack opt spack linux gcc openblas installing quantum espresso using cached archive shared spack var spack cache source cache archive tar gz using cached archive shared spack var spack cache source cache archive staging archive shared spack var spack stage mbernhardt spack stage quantum espresso q e qe tar gz created stage in shared spack var spack stage mbernhardt spack stage quantum espresso applied patch quantum espresso building quantum espresso quantum espresso executing phase install error processerror command exited with status make all errors found in build log shared spack opt spack linux gcc gcc bin gfortran g x cpp input d i shared spack va r spack stage mbernhardt spack stage quantum espresso spack src include i shared spack var spack stage mbernhardt spack stage q uantum espresso spack src fox finclude i shared spack var spack stage mbernhardt spack stage quantum espresso spack src iotk include c fft scalar fftw shared spack opt spack linux gcc gcc bin gfortran g x cpp input d i shared spack va r spack stage mbernhardt spack stage quantum espresso spack src include i shared spack var spack stage mbernhardt spack stage q uantum espresso spack src fox finclude i shared spack var spack stage mbernhardt spack stage quantum espresso spack src iotk include c fft scalar fft scalar endif fatal error f no such file or directory compilation terminated make error make leaving directory shared spack var spack stage mbernhardt spack stage quantum espresso spack src fftxlib make error see build log for details shared spack var spack stage mbernhardt spack stage quantum espresso spack build out txt error failed to install quantum espresso due to childerror processerror command exited with status make all errors found in build log shared spack opt spack linux gcc gcc bin gfortran g x cpp input d i shared spack va r spack stage mbernhardt spack stage quantum espresso spack src include i shared spack var spack stage mbernhardt spack stage q uantum espresso spack src fox finclude i shared spack var spack stage mbernhardt spack stage quantum espresso spack src iotk include c fft scalar fftw shared spack opt spack linux gcc gcc bin gfortran g x cpp input d i shared spack va r spack stage mbernhardt spack stage quantum espresso spack src include i shared spack var spack stage mbernhardt spack stage q uantum espresso spack src fox finclude i shared spack var spack stage mbernhardt spack stage quantum espresso spack src iotk include c fft scalar fft scalar endif fatal error f no such file or directory compilation terminated make error make leaving directory shared spack var spack stage mbernhardt spack stage quantum espresso spack src fftxlib make error see build log for details shared spack var spack stage mbernhardt spack stage quantum espresso spack build out txt traceback most recent call last file shared spack lib spack spack build environment py line in child process return value function file shared spack lib spack spack installer py line in build process phase pkg spec pkg prefix file shared spack lib spack spack package py line in phase wrapper phase spec prefix file shared spack var spack repos builtin packages quantum espresso package py line in install file shared spack lib spack spack build environment py line in call return super makeexecutable self call args kwargs file shared spack lib spack spack util executable py line in call proc returncode long msg spack util executable processerror command exited with status make all platform and user environment commandline uname a linux smp mon nov utc gnu linux cat etc fedora release fedora release twenty nine additional information excerpt from spack build out txt espresso can take advantage of several optimized numerical libraries essl fftw mkl this configure script attempts to find them but may fail if they have been installed in non standard locations if a required library is not found the local copy will be compiled the following libraries have been found blas libs l shared spack opt spack linux gcc openblas lib lopenblas lapack libs l shared spack opt spack linux gcc openblas lib lopenblas fft libs l shared spack opt spack linux gcc fftw lib please check if this is what you expect if any libraries are missing you may specify a list of directories to search and retry as follows configure libdirs list of directories separated by spaces configured for compilation of serial executables for more info read the espresso user s guide doc users guide tex so it finds fftw but does not include the header files i guess | 0 |
28,102 | 8,074,045,629 | IssuesEvent | 2018-08-06 21:27:41 | ethereum/mist | https://api.github.com/repos/ethereum/mist | opened | Better installers | Module: Build process Type: Feature | - [ ] Code sign windows installer (.exe)
- [ ] Smaller packages (asar compression levels)
- [ ] Unify package.json
- [ ] Investigate linux dependency issue
- [ ] Add more metadata to .deb
- [ ] Investigate executable linux package (reported by Ligi) | 1.0 | Better installers - - [ ] Code sign windows installer (.exe)
- [ ] Smaller packages (asar compression levels)
- [ ] Unify package.json
- [ ] Investigate linux dependency issue
- [ ] Add more metadata to .deb
- [ ] Investigate executable linux package (reported by Ligi) | non_priority | better installers code sign windows installer exe smaller packages asar compression levels unify package json investigate linux dependency issue add more metadata to deb investigate executable linux package reported by ligi | 0 |
111,957 | 14,174,679,774 | IssuesEvent | 2020-11-12 20:18:10 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | opened | Update documentation in Storybook that doesn't match Gatsby and design.va.gov | design-system-team | ## Issue Description
Reconcile the differences between existing documentation in Gatsby & Design.va.gov with documentation in Storybook
---
## Tasks
- [ ] Update documentation in Storybook to ensure all components have complete information and is the same as what is on Gatsby & design.va.gov
## Acceptance Criteria
- [ ] Discrepancies with documentation identified during audit are addressed.
- [ ] Component documentation in Storybook matches documentation in Gatsby and design.va.gov
---
## How to configure this issue
- [ ] **Attached to a Milestone** (when will this be completed?)
- [ ] **Attached to an Epic** (what body of work is this a part of?)
- [ ] **Labeled with Team** (`product support`, `analytics-insights`, `operations`, `service-design`, `tools-be`, `tools-fe`)
- [ ] **Labeled with Practice Area** (`backend`, `frontend`, `devops`, `design`, `research`, `product`, `ia`, `qa`, `analytics`, `contact center`, `research`, `accessibility`, `content`)
- [ ] **Labeled with Type** (`bug`, `request`, `discovery`, `documentation`, etc.)
| 1.0 | Update documentation in Storybook that doesn't match Gatsby and design.va.gov - ## Issue Description
Reconcile the differences between existing documentation in Gatsby & Design.va.gov with documentation in Storybook
---
## Tasks
- [ ] Update documentation in Storybook to ensure all components have complete information and is the same as what is on Gatsby & design.va.gov
## Acceptance Criteria
- [ ] Discrepancies with documentation identified during audit are addressed.
- [ ] Component documentation in Storybook matches documentation in Gatsby and design.va.gov
---
## How to configure this issue
- [ ] **Attached to a Milestone** (when will this be completed?)
- [ ] **Attached to an Epic** (what body of work is this a part of?)
- [ ] **Labeled with Team** (`product support`, `analytics-insights`, `operations`, `service-design`, `tools-be`, `tools-fe`)
- [ ] **Labeled with Practice Area** (`backend`, `frontend`, `devops`, `design`, `research`, `product`, `ia`, `qa`, `analytics`, `contact center`, `research`, `accessibility`, `content`)
- [ ] **Labeled with Type** (`bug`, `request`, `discovery`, `documentation`, etc.)
| non_priority | update documentation in storybook that doesn t match gatsby and design va gov issue description reconcile the differences between existing documentation in gatsby design va gov with documentation in storybook tasks update documentation in storybook to ensure all components have complete information and is the same as what is on gatsby design va gov acceptance criteria discrepancies with documentation identified during audit are addressed component documentation in storybook matches documentation in gatsby and design va gov how to configure this issue attached to a milestone when will this be completed attached to an epic what body of work is this a part of labeled with team product support analytics insights operations service design tools be tools fe labeled with practice area backend frontend devops design research product ia qa analytics contact center research accessibility content labeled with type bug request discovery documentation etc | 0 |
213,670 | 24,016,291,941 | IssuesEvent | 2022-09-15 01:15:21 | tt9133github/libraryiotest | https://api.github.com/repos/tt9133github/libraryiotest | opened | WS-2019-0425 (Medium) detected in mocha-2.2.5.js | security vulnerability | ## WS-2019-0425 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mocha-2.2.5.js</b></p></summary>
<p>simple, flexible, fun test framework</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/mocha/2.2.5/mocha.js">https://cdnjs.cloudflare.com/ajax/libs/mocha/2.2.5/mocha.js</a></p>
<p>Path to dependency file: /node_modules/js-base64/test/index.html</p>
<p>Path to vulnerable library: /node_modules/js-base64/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **mocha-2.2.5.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tt9133github/libraryiotest/commit/34ce0f6524391068ef5b880f65575eb3eb499124">34ce0f6524391068ef5b880f65575eb3eb499124</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Mocha is vulnerable to ReDoS attack. If the stack trace in utils.js begins with a large error message, and full-trace is not enabled, utils.stackTraceFilter() will take exponential run time.
<p>Publish Date: 2019-01-24
<p>URL: <a href=https://github.com/mochajs/mocha/commit/1a43d8b11a64e4e85fe2a61aed91c259bbbac559>WS-2019-0425</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-01-24</p>
<p>Fix Resolution: v6.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0425 (Medium) detected in mocha-2.2.5.js - ## WS-2019-0425 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mocha-2.2.5.js</b></p></summary>
<p>simple, flexible, fun test framework</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/mocha/2.2.5/mocha.js">https://cdnjs.cloudflare.com/ajax/libs/mocha/2.2.5/mocha.js</a></p>
<p>Path to dependency file: /node_modules/js-base64/test/index.html</p>
<p>Path to vulnerable library: /node_modules/js-base64/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **mocha-2.2.5.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tt9133github/libraryiotest/commit/34ce0f6524391068ef5b880f65575eb3eb499124">34ce0f6524391068ef5b880f65575eb3eb499124</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Mocha is vulnerable to ReDoS attack. If the stack trace in utils.js begins with a large error message, and full-trace is not enabled, utils.stackTraceFilter() will take exponential run time.
<p>Publish Date: 2019-01-24
<p>URL: <a href=https://github.com/mochajs/mocha/commit/1a43d8b11a64e4e85fe2a61aed91c259bbbac559>WS-2019-0425</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-01-24</p>
<p>Fix Resolution: v6.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in mocha js ws medium severity vulnerability vulnerable library mocha js simple flexible fun test framework library home page a href path to dependency file node modules js test index html path to vulnerable library node modules js test index html dependency hierarchy x mocha js vulnerable library found in head commit a href found in base branch master vulnerability details mocha is vulnerable to redos attack if the stack trace in utils js begins with a large error message and full trace is not enabled utils stacktracefilter will take exponential run time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend | 0 |
303,581 | 26,217,602,227 | IssuesEvent | 2023-01-04 12:20:41 | rancher/cis-operator | https://api.github.com/repos/rancher/cis-operator | closed | Use bitmasks instead of multiple comparisions with "or" for checking file permissions are set to 644 or more restrictive | [zube]: To Test team/area3 team/infracloud | While checking the code, I noticed in some places `op: bitmask` is used (e.g. https://github.com/rancher/security-scan/blob/master/package/cfg/cis-1.6/node.yaml#L11-L19)
However, there are many places that it's not being used and instead I see a long comparison of permissions with `bin_op: or`: https://github.com/rancher/security-scan/blob/master/package/cfg/rke-cis-1.5-permissive/node.yaml#L15-L52
Can we update the places not using `op: bitmask` to use it? This would be more readable. Thanks! | 1.0 | Use bitmasks instead of multiple comparisions with "or" for checking file permissions are set to 644 or more restrictive - While checking the code, I noticed in some places `op: bitmask` is used (e.g. https://github.com/rancher/security-scan/blob/master/package/cfg/cis-1.6/node.yaml#L11-L19)
However, there are many places that it's not being used and instead I see a long comparison of permissions with `bin_op: or`: https://github.com/rancher/security-scan/blob/master/package/cfg/rke-cis-1.5-permissive/node.yaml#L15-L52
Can we update the places not using `op: bitmask` to use it? This would be more readable. Thanks! | non_priority | use bitmasks instead of multiple comparisions with or for checking file permissions are set to or more restrictive while checking the code i noticed in some places op bitmask is used e g however there are many places that it s not being used and instead i see a long comparison of permissions with bin op or can we update the places not using op bitmask to use it this would be more readable thanks | 0 |
233,172 | 18,952,085,868 | IssuesEvent | 2021-11-18 16:07:02 | dagster-io/dagster | https://api.github.com/repos/dagster-io/dagster | closed | dev loop / error discovery is difficult if "dagit -f" file fails to load file | usability flow:dev+test dagit | Right now, we just show a warning, but there's no reason to proceed if the repository doesn't load | 1.0 | dev loop / error discovery is difficult if "dagit -f" file fails to load file - Right now, we just show a warning, but there's no reason to proceed if the repository doesn't load | non_priority | dev loop error discovery is difficult if dagit f file fails to load file right now we just show a warning but there s no reason to proceed if the repository doesn t load | 0 |
50,288 | 10,474,849,911 | IssuesEvent | 2019-09-23 15:13:09 | thats-so-ravenclaw/grace-shopper | https://api.github.com/repos/thats-so-ravenclaw/grace-shopper | closed | API Route: Basic security - don't POST with req.body | code review | From code review, make sure not to POST req.body | 1.0 | API Route: Basic security - don't POST with req.body - From code review, make sure not to POST req.body | non_priority | api route basic security don t post with req body from code review make sure not to post req body | 0 |
329,047 | 24,206,954,293 | IssuesEvent | 2022-09-25 11:19:01 | brotkrueml/typo3-jobrouter-process | https://api.github.com/repos/brotkrueml/typo3-jobrouter-process | opened | Allow attachments for an instance start | documentation feature | - [ ] An attachment can be selected as a process table field type in the process table configuration dialogue
- [ ] The attachment from the form is used when starting an instance.
- [ ] The documentation is adjusted.
- [ ] A changelog entry is added. | 1.0 | Allow attachments for an instance start - - [ ] An attachment can be selected as a process table field type in the process table configuration dialogue
- [ ] The attachment from the form is used when starting an instance.
- [ ] The documentation is adjusted.
- [ ] A changelog entry is added. | non_priority | allow attachments for an instance start an attachment can be selected as a process table field type in the process table configuration dialogue the attachment from the form is used when starting an instance the documentation is adjusted a changelog entry is added | 0 |
311,139 | 26,770,892,888 | IssuesEvent | 2023-01-31 14:03:45 | gear-tech/gear | https://api.github.com/repos/gear-tech/gear | closed | WASM-GEN: check pointer & size passed to syscalls | C1-feature D4-test | ### Problem to Solve
avoid too often OOB errors
<details><summary>Click to expand/collapse</summary>
<p>
```
2023-01-25 14:43:36.021 DEBUG tokio-runtime-worker gear_core_processor::processor: Charged for module instantiation and memory pages. Size: WasmPage(168)
2023-01-25 14:43:36.021 DEBUG tokio-runtime-worker gear_core_processor::executor: Executing program 0x05bad3b550f45b8eb8f8acb172011581c37ca3bee0bd0888b1a47926f805c35f
2023-01-25 14:43:36.021 DEBUG tokio-runtime-worker gear_core_processor::executor: Executing dispatch IncomingDispatch { kind: Init, message: IncomingMessage { id: 0xa663832d1b7bd2f4bdbfe82cf0877a1cc5204f7aee89b5758da917ed0981b08d, source: 0xd43593c715fdd31c61141abd04a99fd6822c8558854ccde39a5684e7a56da27d, payload: 0xbfc0099f8db366762a2028d0d4694ad808d8d4c7723f65dba32f09db6c5c9fc7776b0d53b156dd4836ea0317e5e1697c098f372a73a6ba191075456f21dacd2bba098910e0b65b2187f2b93099b62c201a61e268f14d092897ea4f58f54b657ad30d, gas_limit: 25000000000, value: 0, details: None }, context: None }
2023-01-25 14:43:36.025 TRACE tokio-runtime-worker syscall::gear: debug, args = I32(0x0), I32(0x28),
2023-01-25 14:43:36.025 TRACE tokio-runtime-worker syscall::gear: reply_wgas, args = I32(0x125346d), I32(0xc509defe), I64(0x27f45ba333), I32(0x1092961), I32(0x1bd), I32(0x5312d2),
2023-01-25 14:43:36.057 DEBUG tokio-runtime-worker gear_backend_sandbox::env: SandboxEnvironment::execute res = Err(<wasm:stripped>)
2023-01-25 14:43:36.057 DEBUG tokio-runtime-worker gear_core_processor::executor: Termination reason: Trap(Other(TrimmedString("Memory access out of bounds")))
2023-01-25 14:43:36.057 DEBUG tokio-runtime-worker gear_core_processor::executor: 💥 Trap during execution of 0x05bad3b550f45b8eb8f8acb172011581c37ca3bee0bd0888b1a47926f805c35f
📔 Explanation: Memory access out of bounds
2023-01-25 14:43:36.057 DEBUG tokio-runtime-worker gear_common::gas_provider::internal: Spent 16967757814 of gas
```
```
with seed: 749505056160585358,
```
</p>
</details>
### Possible Solution
check that pointer + size <= memory
### Notes
inspired by #1564 | 1.0 | WASM-GEN: check pointer & size passed to syscalls - ### Problem to Solve
avoid too often OOB errors
<details><summary>Click to expand/collapse</summary>
<p>
```
2023-01-25 14:43:36.021 DEBUG tokio-runtime-worker gear_core_processor::processor: Charged for module instantiation and memory pages. Size: WasmPage(168)
2023-01-25 14:43:36.021 DEBUG tokio-runtime-worker gear_core_processor::executor: Executing program 0x05bad3b550f45b8eb8f8acb172011581c37ca3bee0bd0888b1a47926f805c35f
2023-01-25 14:43:36.021 DEBUG tokio-runtime-worker gear_core_processor::executor: Executing dispatch IncomingDispatch { kind: Init, message: IncomingMessage { id: 0xa663832d1b7bd2f4bdbfe82cf0877a1cc5204f7aee89b5758da917ed0981b08d, source: 0xd43593c715fdd31c61141abd04a99fd6822c8558854ccde39a5684e7a56da27d, payload: 0xbfc0099f8db366762a2028d0d4694ad808d8d4c7723f65dba32f09db6c5c9fc7776b0d53b156dd4836ea0317e5e1697c098f372a73a6ba191075456f21dacd2bba098910e0b65b2187f2b93099b62c201a61e268f14d092897ea4f58f54b657ad30d, gas_limit: 25000000000, value: 0, details: None }, context: None }
2023-01-25 14:43:36.025 TRACE tokio-runtime-worker syscall::gear: debug, args = I32(0x0), I32(0x28),
2023-01-25 14:43:36.025 TRACE tokio-runtime-worker syscall::gear: reply_wgas, args = I32(0x125346d), I32(0xc509defe), I64(0x27f45ba333), I32(0x1092961), I32(0x1bd), I32(0x5312d2),
2023-01-25 14:43:36.057 DEBUG tokio-runtime-worker gear_backend_sandbox::env: SandboxEnvironment::execute res = Err(<wasm:stripped>)
2023-01-25 14:43:36.057 DEBUG tokio-runtime-worker gear_core_processor::executor: Termination reason: Trap(Other(TrimmedString("Memory access out of bounds")))
2023-01-25 14:43:36.057 DEBUG tokio-runtime-worker gear_core_processor::executor: 💥 Trap during execution of 0x05bad3b550f45b8eb8f8acb172011581c37ca3bee0bd0888b1a47926f805c35f
📔 Explanation: Memory access out of bounds
2023-01-25 14:43:36.057 DEBUG tokio-runtime-worker gear_common::gas_provider::internal: Spent 16967757814 of gas
```
```
with seed: 749505056160585358,
```
</p>
</details>
### Possible Solution
check that pointer + size <= memory
### Notes
inspired by #1564 | non_priority | wasm gen check pointer size passed to syscalls problem to solve avoid too often oob errors click to expand collapse debug tokio runtime worker gear core processor processor charged for module instantiation and memory pages size wasmpage debug tokio runtime worker gear core processor executor executing program debug tokio runtime worker gear core processor executor executing dispatch incomingdispatch kind init message incomingmessage id source payload gas limit value details none context none trace tokio runtime worker syscall gear debug args trace tokio runtime worker syscall gear reply wgas args debug tokio runtime worker gear backend sandbox env sandboxenvironment execute res err debug tokio runtime worker gear core processor executor termination reason trap other trimmedstring memory access out of bounds debug tokio runtime worker gear core processor executor 💥 trap during execution of 📔 explanation memory access out of bounds debug tokio runtime worker gear common gas provider internal spent of gas with seed possible solution check that pointer size memory notes inspired by | 0 |
316,505 | 23,635,285,882 | IssuesEvent | 2022-08-25 12:51:28 | andrewchambers/bupstash | https://api.github.com/repos/andrewchambers/bupstash | opened | Document expected system resource usage | documentation enhancement | As users are pushing bupstash harder and harder, we are hitting more limits. We should document limits and what people should expect.
example: https://borgbackup.readthedocs.io/en/2.0.0b1/internals/data-structures.html#indexes-caches-memory-usage
| 1.0 | Document expected system resource usage - As users are pushing bupstash harder and harder, we are hitting more limits. We should document limits and what people should expect.
example: https://borgbackup.readthedocs.io/en/2.0.0b1/internals/data-structures.html#indexes-caches-memory-usage
| non_priority | document expected system resource usage as users are pushing bupstash harder and harder we are hitting more limits we should document limits and what people should expect example | 0 |
131,744 | 12,489,637,072 | IssuesEvent | 2020-05-31 19:47:25 | ash-project/ash | https://api.github.com/repos/ash-project/ash | closed | Initial Thoughts | documentation | **Feature Roadmap**
* Filtering
* Pagination
* Relationships
* Attributes
* JSONAPI
* GraphQL
* Optional Ecto
* Custom Errors
* Escape Hatches
**Niche Issues to Be Aware Of**
* Regular controllers for things such as Twilio callback or pdf generation
* Non-RESTful controller actions that completely break the CRUD pattern
* Custom callbacks before saving/creating such as setting the current user as a property on a model for whodunnit purposes
* Forced sideloading data (this could be more of a function of ember-data being strange)
* Limiting sorting/creating/updating/filtering/attribute serializing based on custom rules (ie: user type)
* Make sure we handle derived IDs (ie on database view) just as nicely as DB generated IDs
* Allow for default values (ie: students are always sorted by last name descending)
* Decorator methods/custom attributes (ie: full name)
* Filter by attribute or create a custom filter
* Multiple associations between the same two tables with different foreign key names (ie: created by user vs updated by user)
* Self associations (ie: users friending users)
**Assumptions:**
* almost all schemas and changesets currently defined within contexts will move to be defined in resources
* all boilerplate autogenerated CRUD code in context files such as School.get_list(id) will be removed
* almost all controllers and views will be removed
* almost all routes in the router file will be removed
**What about...**
* what happens to custom functions defined in a context (ie: import CSV logic)?
* what happens to autogenerated tests for a schema (and would we do this for a resource)?
* what happens to autogenerated tests for a controller (and would we do this for a resource)?
* where would CORS be handled?
* what about Phoenix Channels
* where would error handling/fallback controller happen?
**I'm not clear on**
* The interplay between Ash Resources and Phoenix Contexts - it can't replace 100% of what Contexts do, can it/should it?
* can ecto schemas access database views instead of tables?
* where to define/how to access current user?
* how to determine authenticated vs unauthenticated routes?
**A few specific and somewhat difficult examples we should have a story about**
* Full Text Search on either a db table or multiple db tables
* User registration (non 1-1 db backed endpoint)
* file downloads and uploads
* non CRUD endpoint such as /api/teachers/1/activate
**TODO**
* start building documentation
* start unit testing one isolated piece of the code
* make an integration test for one resource
* CLI
* build out a small demo app for learning purposes
* use Ash to build a substantial API
* get feedback from DY Elixir people, and people like Dan/Larry Gebhardt
* preview this/drum up interest at LoneStar Elixir in Feb 2020?
* present 1.0 stable version at ElixirConf in August 2020 | 1.0 | Initial Thoughts - **Feature Roadmap**
* Filtering
* Pagination
* Relationships
* Attributes
* JSONAPI
* GraphQL
* Optional Ecto
* Custom Errors
* Escape Hatches
**Niche Issues to Be Aware Of**
* Regular controllers for things such as Twilio callback or pdf generation
* Non-RESTful controller actions that completely break the CRUD pattern
* Custom callbacks before saving/creating such as setting the current user as a property on a model for whodunnit purposes
* Forced sideloading data (this could be more of a function of ember-data being strange)
* Limiting sorting/creating/updating/filtering/attribute serializing based on custom rules (ie: user type)
* Make sure we handle derived IDs (ie on database view) just as nicely as DB generated IDs
* Allow for default values (ie: students are always sorted by last name descending)
* Decorator methods/custom attributes (ie: full name)
* Filter by attribute or create a custom filter
* Multiple associations between the same two tables with different foreign key names (ie: created by user vs updated by user)
* Self associations (ie: users friending users)
**Assumptions:**
* almost all schemas and changesets currently defined within contexts will move to be defined in resources
* all boilerplate autogenerated CRUD code in context files such as School.get_list(id) will be removed
* almost all controllers and views will be removed
* almost all routes in the router file will be removed
**What about...**
* what happens to custom functions defined in a context (ie: import CSV logic)?
* what happens to autogenerated tests for a schema (and would we do this for a resource)?
* what happens to autogenerated tests for a controller (and would we do this for a resource)?
* where would CORS be handled?
* what about Phoenix Channels
* where would error handling/fallback controller happen?
**I'm not clear on**
* The interplay between Ash Resources and Phoenix Contexts - it can't replace 100% of what Contexts do, can it/should it?
* can ecto schemas access database views instead of tables?
* where to define/how to access current user?
* how to determine authenticated vs unauthenticated routes?
**A few specific and somewhat difficult examples we should have a story about**
* Full Text Search on either a db table or multiple db tables
* User registration (non 1-1 db backed endpoint)
* file downloads and uploads
* non CRUD endpoint such as /api/teachers/1/activate
**TODO**
* start building documentation
* start unit testing one isolated piece of the code
* make an integration test for one resource
* CLI
* build out a small demo app for learning purposes
* use Ash to build a substantial API
* get feedback from DY Elixir people, and people like Dan/Larry Gebhardt
* preview this/drum up interest at LoneStar Elixir in Feb 2020?
* present 1.0 stable version at ElixirConf in August 2020 | non_priority | initial thoughts feature roadmap filtering pagination relationships attributes jsonapi graphql optional ecto custom errors escape hatches niche issues to be aware of regular controllers for things such as twilio callback or pdf generation non restful controller actions that completely break the crud pattern custom callbacks before saving creating such as setting the current user as a property on a model for whodunnit purposes forced sideloading data this could be more of a function of ember data being strange limiting sorting creating updating filtering attribute serializing based on custom rules ie user type make sure we handle derived ids ie on database view just as nicely as db generated ids allow for default values ie students are always sorted by last name descending decorator methods custom attributes ie full name filter by attribute or create a custom filter multiple associations between the same two tables with different foreign key names ie created by user vs updated by user self associations ie users friending users assumptions almost all schemas and changesets currently defined within contexts will move to be defined in resources all boilerplate autogenerated crud code in context files such as school get list id will be removed almost all controllers and views will be removed almost all routes in the router file will be removed what about what happens to custom functions defined in a context ie import csv logic what happens to autogenerated tests for a schema and would we do this for a resource what happens to autogenerated tests for a controller and would we do this for a resource where would cors be handled what about phoenix channels where would error handling fallback controller happen i m not clear on the interplay between ash resources and phoenix contexts it can t replace of what contexts do can it should it can ecto schemas access database views instead of tables where to define how to access current user how to determine authenticated vs unauthenticated routes a few specific and somewhat difficult examples we should have a story about full text search on either a db table or multiple db tables user registration non db backed endpoint file downloads and uploads non crud endpoint such as api teachers activate todo start building documentation start unit testing one isolated piece of the code make an integration test for one resource cli build out a small demo app for learning purposes use ash to build a substantial api get feedback from dy elixir people and people like dan larry gebhardt preview this drum up interest at lonestar elixir in feb present stable version at elixirconf in august | 0 |
226,143 | 24,946,551,767 | IssuesEvent | 2022-11-01 01:05:51 | rgordon95/sculptureMyra | https://api.github.com/repos/rgordon95/sculptureMyra | opened | CVE-2022-37620 (Medium) detected in html-minifier-3.5.21.tgz | security vulnerability | ## CVE-2022-37620 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>html-minifier-3.5.21.tgz</b></p></summary>
<p>Highly configurable, well-tested, JavaScript-based HTML minifier.</p>
<p>Library home page: <a href="https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.21.tgz">https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.21.tgz</a></p>
<p>Path to dependency file: /sculptureMyra/package.json</p>
<p>Path to vulnerable library: /node_modules/html-minifier/package.json</p>
<p>
Dependency Hierarchy:
- gulp-htmlmin-3.0.0.tgz (Root Library)
- :x: **html-minifier-3.5.21.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular Expression Denial of Service (ReDoS) flaw was found in kangax html-minifier 4.0.0 via the candidate variable in htmlminifier.js.
<p>Publish Date: 2022-10-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37620>CVE-2022-37620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-37620 (Medium) detected in html-minifier-3.5.21.tgz - ## CVE-2022-37620 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>html-minifier-3.5.21.tgz</b></p></summary>
<p>Highly configurable, well-tested, JavaScript-based HTML minifier.</p>
<p>Library home page: <a href="https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.21.tgz">https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.21.tgz</a></p>
<p>Path to dependency file: /sculptureMyra/package.json</p>
<p>Path to vulnerable library: /node_modules/html-minifier/package.json</p>
<p>
Dependency Hierarchy:
- gulp-htmlmin-3.0.0.tgz (Root Library)
- :x: **html-minifier-3.5.21.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Regular Expression Denial of Service (ReDoS) flaw was found in kangax html-minifier 4.0.0 via the candidate variable in htmlminifier.js.
<p>Publish Date: 2022-10-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37620>CVE-2022-37620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in html minifier tgz cve medium severity vulnerability vulnerable library html minifier tgz highly configurable well tested javascript based html minifier library home page a href path to dependency file sculpturemyra package json path to vulnerable library node modules html minifier package json dependency hierarchy gulp htmlmin tgz root library x html minifier tgz vulnerable library vulnerability details a regular expression denial of service redos flaw was found in kangax html minifier via the candidate variable in htmlminifier js publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend | 0 |
82,667 | 10,266,760,714 | IssuesEvent | 2019-08-22 22:30:22 | MozillaReality/FirefoxReality | https://api.github.com/repos/MozillaReality/FirefoxReality | closed | Investigate Send Tab | Needs Design UX enhancement | - Users should be able to find a site they want to visit in a browser outside of VR and then have it displayed in their HMD. This applies to a single URL and not a group of them.
- Sending tabs may be associated with FxA (see #1260), but user should **also** be able to send tabs without making an account or signing in.
- Users should not have another experience (inside FxR or not) disrupted by receipt of the tab.
- Users should not lose data (ie partially filled URL bar or web form) as a result of receiving a tab. | 1.0 | Investigate Send Tab - - Users should be able to find a site they want to visit in a browser outside of VR and then have it displayed in their HMD. This applies to a single URL and not a group of them.
- Sending tabs may be associated with FxA (see #1260), but user should **also** be able to send tabs without making an account or signing in.
- Users should not have another experience (inside FxR or not) disrupted by receipt of the tab.
- Users should not lose data (ie partially filled URL bar or web form) as a result of receiving a tab. | non_priority | investigate send tab users should be able to find a site they want to visit in a browser outside of vr and then have it displayed in their hmd this applies to a single url and not a group of them sending tabs may be associated with fxa see but user should also be able to send tabs without making an account or signing in users should not have another experience inside fxr or not disrupted by receipt of the tab users should not lose data ie partially filled url bar or web form as a result of receiving a tab | 0 |
353,948 | 25,143,097,846 | IssuesEvent | 2022-11-10 01:29:13 | jdrch/Hardware | https://api.github.com/repos/jdrch/Hardware | closed | OS frequently loses network connectivity | bug openindiana dell optiplex 390 mt illumos critical previous experience: no documentation: unknown purchase required: no unix mine: yes | OI frequently loses network connectivity. Per advice from the OI mailing list, do the following the next time it happens:
* `# dladm show-link -s rge0`
* Check for error messages in the /var/adm/messages related to rge0 | 1.0 | OS frequently loses network connectivity - OI frequently loses network connectivity. Per advice from the OI mailing list, do the following the next time it happens:
* `# dladm show-link -s rge0`
* Check for error messages in the /var/adm/messages related to rge0 | non_priority | os frequently loses network connectivity oi frequently loses network connectivity per advice from the oi mailing list do the following the next time it happens dladm show link s check for error messages in the var adm messages related to | 0 |
61,757 | 14,635,284,353 | IssuesEvent | 2020-12-24 07:49:18 | Seagate/cortx-monitor | https://api.github.com/repos/Seagate/cortx-monitor | opened | CVE-2020-14422 (Medium) detected in ipaddress-1.0.23-py2.py3-none-any.whl | security vulnerability | ## CVE-2020-14422 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ipaddress-1.0.23-py2.py3-none-any.whl</b></p></summary>
<p>IPv4/IPv6 manipulation library</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/c2/f8/49697181b1651d8347d24c095ce46c7346c37335ddc7d255833e7cde674d/ipaddress-1.0.23-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/c2/f8/49697181b1651d8347d24c095ce46c7346c37335ddc7d255833e7cde674d/ipaddress-1.0.23-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: cortx-monitor/low-level/requirements.txt</p>
<p>Path to vulnerable library: cortx-monitor/low-level/requirements.txt</p>
<p>
Dependency Hierarchy:
- cryptography-2.8-cp27-cp27mu-manylinux2010_x86_64.whl (Root Library)
- :x: **ipaddress-1.0.23-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-monitor/commit/2d4e585fe7122df981f34c4d821dde3dc5c0d50f">2d4e585fe7122df981f34c4d821dde3dc5c0d50f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lib/ipaddress.py in Python through 3.8.3 improperly computes hash values in the IPv4Interface and IPv6Interface classes, which might allow a remote attacker to cause a denial of service if an application is affected by the performance of a dictionary containing IPv4Interface or IPv6Interface objects, and this attacker can cause many dictionary entries to be created. This is fixed in: v3.5.10, v3.5.10rc1; v3.6.12; v3.7.9; v3.8.4, v3.8.4rc1, v3.8.5, v3.8.6, v3.8.6rc1; v3.9.0, v3.9.0b4, v3.9.0b5, v3.9.0rc1, v3.9.0rc2.
<p>Publish Date: 2020-06-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14422>CVE-2020-14422</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2020-14422">https://security-tracker.debian.org/tracker/CVE-2020-14422</a></p>
<p>Release Date: 2020-06-18</p>
<p>Fix Resolution: 3.5.3-1+deb9u2, 3.7.3-2+deb10u2, 3.8.4~rc1-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-14422 (Medium) detected in ipaddress-1.0.23-py2.py3-none-any.whl - ## CVE-2020-14422 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ipaddress-1.0.23-py2.py3-none-any.whl</b></p></summary>
<p>IPv4/IPv6 manipulation library</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/c2/f8/49697181b1651d8347d24c095ce46c7346c37335ddc7d255833e7cde674d/ipaddress-1.0.23-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/c2/f8/49697181b1651d8347d24c095ce46c7346c37335ddc7d255833e7cde674d/ipaddress-1.0.23-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: cortx-monitor/low-level/requirements.txt</p>
<p>Path to vulnerable library: cortx-monitor/low-level/requirements.txt</p>
<p>
Dependency Hierarchy:
- cryptography-2.8-cp27-cp27mu-manylinux2010_x86_64.whl (Root Library)
- :x: **ipaddress-1.0.23-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Seagate/cortx-monitor/commit/2d4e585fe7122df981f34c4d821dde3dc5c0d50f">2d4e585fe7122df981f34c4d821dde3dc5c0d50f</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lib/ipaddress.py in Python through 3.8.3 improperly computes hash values in the IPv4Interface and IPv6Interface classes, which might allow a remote attacker to cause a denial of service if an application is affected by the performance of a dictionary containing IPv4Interface or IPv6Interface objects, and this attacker can cause many dictionary entries to be created. This is fixed in: v3.5.10, v3.5.10rc1; v3.6.12; v3.7.9; v3.8.4, v3.8.4rc1, v3.8.5, v3.8.6, v3.8.6rc1; v3.9.0, v3.9.0b4, v3.9.0b5, v3.9.0rc1, v3.9.0rc2.
<p>Publish Date: 2020-06-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-14422>CVE-2020-14422</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2020-14422">https://security-tracker.debian.org/tracker/CVE-2020-14422</a></p>
<p>Release Date: 2020-06-18</p>
<p>Fix Resolution: 3.5.3-1+deb9u2, 3.7.3-2+deb10u2, 3.8.4~rc1-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in ipaddress none any whl cve medium severity vulnerability vulnerable library ipaddress none any whl manipulation library library home page a href path to dependency file cortx monitor low level requirements txt path to vulnerable library cortx monitor low level requirements txt dependency hierarchy cryptography whl root library x ipaddress none any whl vulnerable library found in head commit a href found in base branch main vulnerability details lib ipaddress py in python through improperly computes hash values in the and classes which might allow a remote attacker to cause a denial of service if an application is affected by the performance of a dictionary containing or objects and this attacker can cause many dictionary entries to be created this is fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
7,350 | 3,962,830,863 | IssuesEvent | 2016-05-02 18:15:26 | Linuxbrew/homebrew-core | https://api.github.com/repos/Linuxbrew/homebrew-core | opened | txr fails to build on Ubuntu 14.04 | build-error | _From @rwhogg on March 18, 2016 5:27_
`txr` fails to build on Ubuntu 14.04. See https://gist.github.com/rwhogg/87a85424efca56c55efd#file-01-configure-L27
```
Checking what C type we have for integers wider than "long long" ... ./configure: line 1017: [: gcc: integer expression expected
./configure: line 1017: $2: unbound variable
```
- [x] Ran `brew update` and retried your prior step?
- [x] Ran `brew doctor`, fixed as many issues as possible and retried your prior step?
- [x] If you're seeing permission errors tried running `sudo chown -R $(whoami) $(brew --prefix)`?
_Copied from original issue: Linuxbrew/linuxbrew#948_ | 1.0 | txr fails to build on Ubuntu 14.04 - _From @rwhogg on March 18, 2016 5:27_
`txr` fails to build on Ubuntu 14.04. See https://gist.github.com/rwhogg/87a85424efca56c55efd#file-01-configure-L27
```
Checking what C type we have for integers wider than "long long" ... ./configure: line 1017: [: gcc: integer expression expected
./configure: line 1017: $2: unbound variable
```
- [x] Ran `brew update` and retried your prior step?
- [x] Ran `brew doctor`, fixed as many issues as possible and retried your prior step?
- [x] If you're seeing permission errors tried running `sudo chown -R $(whoami) $(brew --prefix)`?
_Copied from original issue: Linuxbrew/linuxbrew#948_ | non_priority | txr fails to build on ubuntu from rwhogg on march txr fails to build on ubuntu see checking what c type we have for integers wider than long long configure line gcc integer expression expected configure line unbound variable ran brew update and retried your prior step ran brew doctor fixed as many issues as possible and retried your prior step if you re seeing permission errors tried running sudo chown r whoami brew prefix copied from original issue linuxbrew linuxbrew | 0 |
156,490 | 12,312,502,538 | IssuesEvent | 2020-05-12 14:01:33 | netsec-ethz/scion-apps | https://api.github.com/repos/netsec-ethz/scion-apps | closed | Add integration tests | testing | Create integration tests for all apps. This can be fairly simple, just checking that the basic functionality works.
The setup can be similar to the integration tests in RAINS. | 1.0 | Add integration tests - Create integration tests for all apps. This can be fairly simple, just checking that the basic functionality works.
The setup can be similar to the integration tests in RAINS. | non_priority | add integration tests create integration tests for all apps this can be fairly simple just checking that the basic functionality works the setup can be similar to the integration tests in rains | 0 |
2,359 | 7,719,218,939 | IssuesEvent | 2018-05-23 18:41:32 | poanetwork/poa-explorer | https://api.github.com/repos/poanetwork/poa-explorer | opened | Collapse receipts into transactions | architecture bug incompatible changes | While having a separate `receipts` table reflects the fact that we use a separate JSONRPC method, `eth_getTransactionReceipt`, the `receipts` table is never show in the UI in isolation, its data is always mixed with data from `transactions`. Joining both the `receipts` and `blocks` table to `transactions` to sort transactions by `blocks.number` and `receipts.transaction_index` has poor performance characteristics as reported by members of Gas Light (@tmecklem, @jimmay5469).
Since `receipts` does not need to exist independently, and it existing independently leads to inefficient join sorts, merge its columns with `transactions` and drop the `receipts` table.
- [ ] Add the `receipts` columns to the `transactions` table
- [ ] Merge `Explorer.Chain.Receipt` into `Explorer.Chain.Transaction`
- [ ] Connect `logs` to `transactions` instead of `receipts`
- [ ] Connect `Explorer.Chain.Log` to `Explorer.Chain.Transaction` instead of `Explorer.Chain.Receipt`
- [ ] Remove the `receipts` migrations
- [ ] Change `Explorer.Indexer` to extract receipt-only parameters and put them in the transactions params before `Explorer.Chain.import_blocks`.
**NOTE: This will require a DB reset and reindex** | 1.0 | Collapse receipts into transactions - While having a separate `receipts` table reflects the fact that we use a separate JSONRPC method, `eth_getTransactionReceipt`, the `receipts` table is never show in the UI in isolation, its data is always mixed with data from `transactions`. Joining both the `receipts` and `blocks` table to `transactions` to sort transactions by `blocks.number` and `receipts.transaction_index` has poor performance characteristics as reported by members of Gas Light (@tmecklem, @jimmay5469).
Since `receipts` does not need to exist independently, and it existing independently leads to inefficient join sorts, merge its columns with `transactions` and drop the `receipts` table.
- [ ] Add the `receipts` columns to the `transactions` table
- [ ] Merge `Explorer.Chain.Receipt` into `Explorer.Chain.Transaction`
- [ ] Connect `logs` to `transactions` instead of `receipts`
- [ ] Connect `Explorer.Chain.Log` to `Explorer.Chain.Transaction` instead of `Explorer.Chain.Receipt`
- [ ] Remove the `receipts` migrations
- [ ] Change `Explorer.Indexer` to extract receipt-only parameters and put them in the transactions params before `Explorer.Chain.import_blocks`.
**NOTE: This will require a DB reset and reindex** | non_priority | collapse receipts into transactions while having a separate receipts table reflects the fact that we use a separate jsonrpc method eth gettransactionreceipt the receipts table is never show in the ui in isolation its data is always mixed with data from transactions joining both the receipts and blocks table to transactions to sort transactions by blocks number and receipts transaction index has poor performance characteristics as reported by members of gas light tmecklem since receipts does not need to exist independently and it existing independently leads to inefficient join sorts merge its columns with transactions and drop the receipts table add the receipts columns to the transactions table merge explorer chain receipt into explorer chain transaction connect logs to transactions instead of receipts connect explorer chain log to explorer chain transaction instead of explorer chain receipt remove the receipts migrations change explorer indexer to extract receipt only parameters and put them in the transactions params before explorer chain import blocks note this will require a db reset and reindex | 0 |
434,417 | 30,452,361,047 | IssuesEvent | 2023-07-16 13:07:07 | ZJUEarthData/geochemistrypi | https://api.github.com/repos/ZJUEarthData/geochemistrypi | reopened | Git -Image Store- Geochemistry π | documentation | # Git
1.
<img width="1174" alt="image" src="https://user-images.githubusercontent.com/47497750/236991014-7c17248c-54cc-41f0-af15-796a67de5307.png">
2.
<img width="920" alt="image" src="https://user-images.githubusercontent.com/47497750/236991322-843c9497-3c32-42d0-a564-1979e093a8d1.png">
3.
<img width="1299" alt="image" src="https://user-images.githubusercontent.com/47497750/236991773-f618dd0f-4503-4688-b29e-c6704dbae3f9.png">
<img width="921" alt="image" src="https://user-images.githubusercontent.com/47497750/237021355-0fff7b74-ebd8-47f6-8c76-8918b11fc8ec.png">
<img width="920" alt="image" src="https://user-images.githubusercontent.com/47497750/237022518-7e0dfa23-7b49-46f3-adf2-ec1f83651f44.png">
4.
<img width="1439" alt="image" src="https://user-images.githubusercontent.com/47497750/237025973-aa9792a3-dbfa-41f3-8adf-3d5824146ddf.png">
<img width="1219" alt="image" src="https://user-images.githubusercontent.com/47497750/237028303-e3edf748-2f76-40a3-8247-3869656dfc88.png">
5.
<img width="1440" alt="image" src="https://user-images.githubusercontent.com/47497750/237012708-79e3e170-dea3-443c-8b5c-6b790b5161c0.png">
<img width="1433" alt="image" src="https://user-images.githubusercontent.com/47497750/237028688-82e628f8-eaac-4cf2-a3a4-c2eb51a85cd8.png">
6.
<img width="1266" alt="image" src="https://user-images.githubusercontent.com/47497750/237016353-3d28a2ea-6c1f-4e2b-aa27-b8a469784a4d.png">
<img width="1274" alt="image" src="https://user-images.githubusercontent.com/47497750/237013451-f7bd73f8-6b3f-4781-8954-5aaf1e171f1f.png">
7.
<img width="1269" alt="image" src="https://user-images.githubusercontent.com/47497750/237016843-8ec4dfc9-74e0-4f59-a4fd-8addcc395eb3.png">
8.
<img width="1433" alt="image" src="https://user-images.githubusercontent.com/47497750/237053125-1b90f6bc-c17f-4840-b6c3-5993913cb4c3.png">
9.
<img width="1438" alt="image" src="https://user-images.githubusercontent.com/47497750/237053725-303b6c15-5939-441e-8bee-2d322c2e8019.png">
10.
<img width="1200" alt="image" src="https://user-images.githubusercontent.com/47497750/237054080-39198eac-e804-41db-b672-aece0417fe38.png">
11.
<img width="1202" alt="image" src="https://user-images.githubusercontent.com/47497750/237054509-e8418115-23c3-49c7-8690-51dc31d2aed1.png">
12.
<img width="1204" alt="image" src="https://user-images.githubusercontent.com/47497750/237055438-38e60ffc-a54d-4a25-82d3-d5e8b33357d6.png">
13.
<img width="1202" alt="image" src="https://user-images.githubusercontent.com/47497750/237055329-cb3935b8-44d7-4205-87f3-4ae4dbcbeca6.png">
<img width="1160" alt="Pasted Graphic 9" src="https://user-images.githubusercontent.com/47497750/237057185-f43b9867-7f42-4505-a90e-0c954923a1da.png">
<img width="627" alt="Pasted Graphic 13" src="https://user-images.githubusercontent.com/47497750/237058359-6616e9aa-ad44-488f-bfea-b290a6c8cb08.png">
<img width="576" alt="Pasted Graphic 14" src="https://user-images.githubusercontent.com/47497750/237058693-95e9b454-92be-4e47-a7ab-9031c7645849.png">
<img width="786" alt="Pasted Graphic 15" src="https://user-images.githubusercontent.com/47497750/237059133-1b2553e0-7d22-4fb8-9cd2-15412a5e2e85.png">
<img width="800" alt="Pasted Graphic 16" src="https://user-images.githubusercontent.com/47497750/237059215-08ef553a-be5e-46bf-9c5c-e76342773d8b.png">
<img width="1109" alt="Pasted Graphic 17" src="https://user-images.githubusercontent.com/47497750/237059295-42b855e9-f70a-4894-8e34-0b406e1d9f53.png">
14.
<img width="1200" alt="image" src="https://user-images.githubusercontent.com/47497750/237062092-0a922f60-4485-4343-93ca-89ed8902acd6.png">
<img width="947" alt="Pasted Graphic 8" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/387b4870-f5b1-4d4d-9865-4564c3b8e5c3">
<img width="1012" alt="Pasted Graphic 3" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/00c628b6-a453-490c-af77-c37929b65a13">
<img width="1005" alt="Pasted Graphic 7" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/82a5fb49-022f-43f9-89b8-862c237de2ba">
<img width="816" alt="Pasted Graphic 5" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/c286eb02-91e4-4927-8b69-981b6aa7a902">
<img width="816" alt="Pasted Graphic 6" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/6472df94-8d65-45aa-b787-1848966fe90d">
15.
<img width="1217" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/ae7de931-dc30-45d5-a29a-0389152075c7">
<img width="1222" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/3ba452c0-c93a-430c-9652-810babd00236">
<img width="1041" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/7cd4a746-2b78-4861-a76d-758c7ab9e21a">
<img width="1193" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/d9ea1bf1-c452-4923-ac76-138a116383fe">
<img width="1185" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/d63e4339-e465-4b79-adb3-ae5a057c7433">
| 1.0 | Git -Image Store- Geochemistry π - # Git
1.
<img width="1174" alt="image" src="https://user-images.githubusercontent.com/47497750/236991014-7c17248c-54cc-41f0-af15-796a67de5307.png">
2.
<img width="920" alt="image" src="https://user-images.githubusercontent.com/47497750/236991322-843c9497-3c32-42d0-a564-1979e093a8d1.png">
3.
<img width="1299" alt="image" src="https://user-images.githubusercontent.com/47497750/236991773-f618dd0f-4503-4688-b29e-c6704dbae3f9.png">
<img width="921" alt="image" src="https://user-images.githubusercontent.com/47497750/237021355-0fff7b74-ebd8-47f6-8c76-8918b11fc8ec.png">
<img width="920" alt="image" src="https://user-images.githubusercontent.com/47497750/237022518-7e0dfa23-7b49-46f3-adf2-ec1f83651f44.png">
4.
<img width="1439" alt="image" src="https://user-images.githubusercontent.com/47497750/237025973-aa9792a3-dbfa-41f3-8adf-3d5824146ddf.png">
<img width="1219" alt="image" src="https://user-images.githubusercontent.com/47497750/237028303-e3edf748-2f76-40a3-8247-3869656dfc88.png">
5.
<img width="1440" alt="image" src="https://user-images.githubusercontent.com/47497750/237012708-79e3e170-dea3-443c-8b5c-6b790b5161c0.png">
<img width="1433" alt="image" src="https://user-images.githubusercontent.com/47497750/237028688-82e628f8-eaac-4cf2-a3a4-c2eb51a85cd8.png">
6.
<img width="1266" alt="image" src="https://user-images.githubusercontent.com/47497750/237016353-3d28a2ea-6c1f-4e2b-aa27-b8a469784a4d.png">
<img width="1274" alt="image" src="https://user-images.githubusercontent.com/47497750/237013451-f7bd73f8-6b3f-4781-8954-5aaf1e171f1f.png">
7.
<img width="1269" alt="image" src="https://user-images.githubusercontent.com/47497750/237016843-8ec4dfc9-74e0-4f59-a4fd-8addcc395eb3.png">
8.
<img width="1433" alt="image" src="https://user-images.githubusercontent.com/47497750/237053125-1b90f6bc-c17f-4840-b6c3-5993913cb4c3.png">
9.
<img width="1438" alt="image" src="https://user-images.githubusercontent.com/47497750/237053725-303b6c15-5939-441e-8bee-2d322c2e8019.png">
10.
<img width="1200" alt="image" src="https://user-images.githubusercontent.com/47497750/237054080-39198eac-e804-41db-b672-aece0417fe38.png">
11.
<img width="1202" alt="image" src="https://user-images.githubusercontent.com/47497750/237054509-e8418115-23c3-49c7-8690-51dc31d2aed1.png">
12.
<img width="1204" alt="image" src="https://user-images.githubusercontent.com/47497750/237055438-38e60ffc-a54d-4a25-82d3-d5e8b33357d6.png">
13.
<img width="1202" alt="image" src="https://user-images.githubusercontent.com/47497750/237055329-cb3935b8-44d7-4205-87f3-4ae4dbcbeca6.png">
<img width="1160" alt="Pasted Graphic 9" src="https://user-images.githubusercontent.com/47497750/237057185-f43b9867-7f42-4505-a90e-0c954923a1da.png">
<img width="627" alt="Pasted Graphic 13" src="https://user-images.githubusercontent.com/47497750/237058359-6616e9aa-ad44-488f-bfea-b290a6c8cb08.png">
<img width="576" alt="Pasted Graphic 14" src="https://user-images.githubusercontent.com/47497750/237058693-95e9b454-92be-4e47-a7ab-9031c7645849.png">
<img width="786" alt="Pasted Graphic 15" src="https://user-images.githubusercontent.com/47497750/237059133-1b2553e0-7d22-4fb8-9cd2-15412a5e2e85.png">
<img width="800" alt="Pasted Graphic 16" src="https://user-images.githubusercontent.com/47497750/237059215-08ef553a-be5e-46bf-9c5c-e76342773d8b.png">
<img width="1109" alt="Pasted Graphic 17" src="https://user-images.githubusercontent.com/47497750/237059295-42b855e9-f70a-4894-8e34-0b406e1d9f53.png">
14.
<img width="1200" alt="image" src="https://user-images.githubusercontent.com/47497750/237062092-0a922f60-4485-4343-93ca-89ed8902acd6.png">
<img width="947" alt="Pasted Graphic 8" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/387b4870-f5b1-4d4d-9865-4564c3b8e5c3">
<img width="1012" alt="Pasted Graphic 3" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/00c628b6-a453-490c-af77-c37929b65a13">
<img width="1005" alt="Pasted Graphic 7" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/82a5fb49-022f-43f9-89b8-862c237de2ba">
<img width="816" alt="Pasted Graphic 5" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/c286eb02-91e4-4927-8b69-981b6aa7a902">
<img width="816" alt="Pasted Graphic 6" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/6472df94-8d65-45aa-b787-1848966fe90d">
15.
<img width="1217" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/ae7de931-dc30-45d5-a29a-0389152075c7">
<img width="1222" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/3ba452c0-c93a-430c-9652-810babd00236">
<img width="1041" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/7cd4a746-2b78-4861-a76d-758c7ab9e21a">
<img width="1193" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/d9ea1bf1-c452-4923-ac76-138a116383fe">
<img width="1185" alt="image" src="https://github.com/ZJUEarthData/geochemistrypi/assets/47497750/d63e4339-e465-4b79-adb3-ae5a057c7433">
| non_priority | git image store geochemistry π git img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src img width alt pasted graphic src img width alt pasted graphic src img width alt pasted graphic src img width alt pasted graphic src img width alt pasted graphic src img width alt pasted graphic src img width alt image src img width alt pasted graphic src img width alt pasted graphic src img width alt pasted graphic src img width alt pasted graphic src img width alt pasted graphic src img width alt image src img width alt image src img width alt image src img width alt image src img width alt image src | 0 |
274,682 | 20,863,450,343 | IssuesEvent | 2022-03-22 02:58:29 | iotaledger/firefly | https://api.github.com/repos/iotaledger/firefly | closed | [Task]: Migrate documentation to handbook | type:documentation | ### Task description
There's lot of documentation that needs to be moved over to the handbook.
### Requirements
Move the following documents over to the handbook:
- [x] Guide: Branching Strategy
- [x] Guide: Coding Conventions
- [x] Guide: Developer Setup
- [x] Guide: Developer Tips
- [x] Guide: Testing
- [x] SOP: Issue Management
- [x] SOP: Release Management
- [x] SOP: Security Incident Response
### Acceptance criteria
Meets the above requirements.
### Creation checklist
- [X] I have assigned this task to the correct people
- [X] I have added the most appropriate labels
- [X] I have linked the correct milestone and/or project | 1.0 | [Task]: Migrate documentation to handbook - ### Task description
There's lot of documentation that needs to be moved over to the handbook.
### Requirements
Move the following documents over to the handbook:
- [x] Guide: Branching Strategy
- [x] Guide: Coding Conventions
- [x] Guide: Developer Setup
- [x] Guide: Developer Tips
- [x] Guide: Testing
- [x] SOP: Issue Management
- [x] SOP: Release Management
- [x] SOP: Security Incident Response
### Acceptance criteria
Meets the above requirements.
### Creation checklist
- [X] I have assigned this task to the correct people
- [X] I have added the most appropriate labels
- [X] I have linked the correct milestone and/or project | non_priority | migrate documentation to handbook task description there s lot of documentation that needs to be moved over to the handbook requirements move the following documents over to the handbook guide branching strategy guide coding conventions guide developer setup guide developer tips guide testing sop issue management sop release management sop security incident response acceptance criteria meets the above requirements creation checklist i have assigned this task to the correct people i have added the most appropriate labels i have linked the correct milestone and or project | 0 |
275,456 | 30,246,285,533 | IssuesEvent | 2023-07-06 16:43:56 | HughC-GH-Demo/Java-Demo-renovate | https://api.github.com/repos/HughC-GH-Demo/Java-Demo-renovate | opened | log4j-1.2.13.jar: 8 vulnerabilities (highest severity is: 9.8) | Mend: dependency security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.13.jar</b></p></summary>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (log4j version) | Remediation Available | Reachability |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | --- |
| [CVE-2022-23305](https://www.mend.io/vulnerability-database/CVE-2022-23305) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.2 | ✅ |
| [CVE-2019-17571](https://www.mend.io/vulnerability-database/CVE-2019-17571) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | log4j-1.2.13.jar | Direct | log4j-manual - 1.2.17-16;log4j-javadoc - 1.2.17-16;log4j - 1.2.17-16,1.2.17-16 | ✅|<p align="center"><a href="#">[<img src='https://whitesource-resources.whitesourcesoftware.com/viaGreen.png' width=19 height=20>](## 'The vulnerability is non-reachable.')</a></p> |
| [CVE-2020-9493](https://www.mend.io/vulnerability-database/CVE-2020-9493) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.1 | ✅|<p align="center"><a href="#">[<img src='https://whitesource-resources.whitesourcesoftware.com/viaGreen.png' width=19 height=20>](## 'The vulnerability is non-reachable.')</a></p> |
| [CVE-2022-23307](https://www.mend.io/vulnerability-database/CVE-2022-23307) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 8.8 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.1 | ✅ |
| [CVE-2022-23302](https://www.mend.io/vulnerability-database/CVE-2022-23302) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 8.8 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.1 | ✅ |
| [CVE-2021-4104](https://www.mend.io/vulnerability-database/CVE-2021-4104) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | log4j-1.2.13.jar | Direct | uom-parent - 1.0.3-3.module,1.0.3-3.module;uom-se-javadoc - 1.0.4-3.module;parfait-examples - 0.5.4-4.module;log4j-manual - 1.2.17-16;si-units-javadoc - 0.6.5-2.module;unit-api - 1.0-5.module,1.0-5.module;unit-api-javadoc - 1.0-5.module;parfait - 0.5.4-4.module,0.5.4-4.module;log4j-javadoc - 1.2.17-16;uom-systems-javadoc - 0.7-1.module;uom-lib-javadoc - 1.0.1-6.module;uom-systems - 0.7-1.module,0.7-1.module;log4j - 1.2.17-16,1.2.17-16;uom-se - 1.0.4-3.module,1.0.4-3.module;uom-lib - 1.0.1-6.module,1.0.1-6.module;parfait-javadoc - 0.5.4-4.module;pcp-parfait-agent - 0.5.4-4.module;si-units - 0.6.5-2.module,0.6.5-2.module | ✅ |
| [CVE-2023-26464](https://www.mend.io/vulnerability-database/CVE-2023-26464) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | log4j-1.2.13.jar | Direct | org.apache.logging.log4j:log4j-core:2.0 | ✅ |
| [CVE-2020-9488](https://www.mend.io/vulnerability-database/CVE-2020-9488) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> Low | 3.7 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.3 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2022-23305</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
By design, the JDBCAppender in Log4j 1.2.x accepts an SQL statement as a configuration parameter where the values to be inserted are converters from PatternLayout. The message converter, %m, is likely to always be included. This allows attackers to manipulate the SQL by entering crafted strings into input fields or headers of an application that are logged allowing unintended SQL queries to be executed. Note this issue only affects Log4j 1.x when specifically configured to use the JDBCAppender, which is not the default. Beginning in version 2.0-beta8, the JDBCAppender was re-introduced with proper support for parameterized SQL queries and further customization over the columns written to in logs. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2022-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23305>CVE-2022-23305</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://reload4j.qos.ch/">https://reload4j.qos.ch/</a></p>
<p>Release Date: 2022-01-18</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.2</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> <img src='https://whitesource-resources.whitesourcesoftware.com/viaGreen.png' width=19 height=20> CVE-2019-17571</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Reachability Analysis
<p>
<p>The vulnerable code is not reachable.</p>
</p>
<p></p>
### Vulnerability Details
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/eea03d504b36e8f870e8321d908e1def1addda16adda04327fe7c125%40%3Cdev.logging.apache.org%3E">https://lists.apache.org/thread.html/eea03d504b36e8f870e8321d908e1def1addda16adda04327fe7c125%40%3Cdev.logging.apache.org%3E</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: log4j-manual - 1.2.17-16;log4j-javadoc - 1.2.17-16;log4j - 1.2.17-16,1.2.17-16</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> <img src='https://whitesource-resources.whitesourcesoftware.com/viaGreen.png' width=19 height=20> CVE-2020-9493</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Reachability Analysis
<p>
<p>The vulnerable code is not reachable.</p>
</p>
<p></p>
### Vulnerability Details
<p>
A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution.
<p>Publish Date: 2021-06-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9493>CVE-2020-9493</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2021/06/16/1">https://www.openwall.com/lists/oss-security/2021/06/16/1</a></p>
<p>Release Date: 2021-06-16</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.1</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2022-23307</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
CVE-2020-9493 identified a deserialization issue that was present in Apache Chainsaw. Prior to Chainsaw V2.0 Chainsaw was a component of Apache Log4j 1.2.x where the same issue exists.
<p>Publish Date: 2022-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23307>CVE-2022-23307</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-01-18</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.1</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2022-23302</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
JMSSink in all versions of Log4j 1.x is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration or if the configuration references an LDAP service the attacker has access to. The attacker can provide a TopicConnectionFactoryBindingName configuration causing JMSSink to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-4104. Note this issue only affects Log4j 1.x when specifically configured to use JMSSink, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2022-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23302>CVE-2022-23302</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://reload4j.qos.ch/">https://reload4j.qos.ch/</a></p>
<p>Release Date: 2022-01-18</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.1</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2021-4104</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2021-12-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-4104>CVE-2021-4104</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-4104">https://nvd.nist.gov/vuln/detail/CVE-2021-4104</a></p>
<p>Release Date: 2021-12-14</p>
<p>Fix Resolution: uom-parent - 1.0.3-3.module,1.0.3-3.module;uom-se-javadoc - 1.0.4-3.module;parfait-examples - 0.5.4-4.module;log4j-manual - 1.2.17-16;si-units-javadoc - 0.6.5-2.module;unit-api - 1.0-5.module,1.0-5.module;unit-api-javadoc - 1.0-5.module;parfait - 0.5.4-4.module,0.5.4-4.module;log4j-javadoc - 1.2.17-16;uom-systems-javadoc - 0.7-1.module;uom-lib-javadoc - 1.0.1-6.module;uom-systems - 0.7-1.module,0.7-1.module;log4j - 1.2.17-16,1.2.17-16;uom-se - 1.0.4-3.module,1.0.4-3.module;uom-lib - 1.0.1-6.module,1.0.1-6.module;parfait-javadoc - 0.5.4-4.module;pcp-parfait-agent - 0.5.4-4.module;si-units - 0.6.5-2.module,0.6.5-2.module</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2023-26464</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
** UNSUPPORTED WHEN ASSIGNED **
When using the Chainsaw or SocketAppender components with Log4j 1.x on JRE less than 1.7, an attacker that manages to cause a logging entry involving a specially-crafted (ie, deeply nested)
hashmap or hashtable (depending on which logging component is in use) to be processed could exhaust the available memory in the virtual machine and achieve Denial of Service when the object is deserialized.
This issue affects Apache Log4j before 2. Affected users are recommended to update to Log4j 2.x.
NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26464>CVE-2023-26464</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-vp98-w2p3-mv35">https://github.com/advisories/GHSA-vp98-w2p3-mv35</a></p>
<p>Release Date: 2023-03-10</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> CVE-2020-9488</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender. Fixed in Apache Log4j 2.12.3 and 2.13.1
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://reload4j.qos.ch/">https://reload4j.qos.ch/</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.3</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details>
***
<p>In order to enable automatic remediation for this issue, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> | True | log4j-1.2.13.jar: 8 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.13.jar</b></p></summary>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (log4j version) | Remediation Available | Reachability |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | --- |
| [CVE-2022-23305](https://www.mend.io/vulnerability-database/CVE-2022-23305) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.2 | ✅ |
| [CVE-2019-17571](https://www.mend.io/vulnerability-database/CVE-2019-17571) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | log4j-1.2.13.jar | Direct | log4j-manual - 1.2.17-16;log4j-javadoc - 1.2.17-16;log4j - 1.2.17-16,1.2.17-16 | ✅|<p align="center"><a href="#">[<img src='https://whitesource-resources.whitesourcesoftware.com/viaGreen.png' width=19 height=20>](## 'The vulnerability is non-reachable.')</a></p> |
| [CVE-2020-9493](https://www.mend.io/vulnerability-database/CVE-2020-9493) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.1 | ✅|<p align="center"><a href="#">[<img src='https://whitesource-resources.whitesourcesoftware.com/viaGreen.png' width=19 height=20>](## 'The vulnerability is non-reachable.')</a></p> |
| [CVE-2022-23307](https://www.mend.io/vulnerability-database/CVE-2022-23307) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 8.8 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.1 | ✅ |
| [CVE-2022-23302](https://www.mend.io/vulnerability-database/CVE-2022-23302) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 8.8 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.1 | ✅ |
| [CVE-2021-4104](https://www.mend.io/vulnerability-database/CVE-2021-4104) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | log4j-1.2.13.jar | Direct | uom-parent - 1.0.3-3.module,1.0.3-3.module;uom-se-javadoc - 1.0.4-3.module;parfait-examples - 0.5.4-4.module;log4j-manual - 1.2.17-16;si-units-javadoc - 0.6.5-2.module;unit-api - 1.0-5.module,1.0-5.module;unit-api-javadoc - 1.0-5.module;parfait - 0.5.4-4.module,0.5.4-4.module;log4j-javadoc - 1.2.17-16;uom-systems-javadoc - 0.7-1.module;uom-lib-javadoc - 1.0.1-6.module;uom-systems - 0.7-1.module,0.7-1.module;log4j - 1.2.17-16,1.2.17-16;uom-se - 1.0.4-3.module,1.0.4-3.module;uom-lib - 1.0.1-6.module,1.0.1-6.module;parfait-javadoc - 0.5.4-4.module;pcp-parfait-agent - 0.5.4-4.module;si-units - 0.6.5-2.module,0.6.5-2.module | ✅ |
| [CVE-2023-26464](https://www.mend.io/vulnerability-database/CVE-2023-26464) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | log4j-1.2.13.jar | Direct | org.apache.logging.log4j:log4j-core:2.0 | ✅ |
| [CVE-2020-9488](https://www.mend.io/vulnerability-database/CVE-2020-9488) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> Low | 3.7 | log4j-1.2.13.jar | Direct | ch.qos.reload4j:reload4j:1.2.18.3 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2022-23305</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
By design, the JDBCAppender in Log4j 1.2.x accepts an SQL statement as a configuration parameter where the values to be inserted are converters from PatternLayout. The message converter, %m, is likely to always be included. This allows attackers to manipulate the SQL by entering crafted strings into input fields or headers of an application that are logged allowing unintended SQL queries to be executed. Note this issue only affects Log4j 1.x when specifically configured to use the JDBCAppender, which is not the default. Beginning in version 2.0-beta8, the JDBCAppender was re-introduced with proper support for parameterized SQL queries and further customization over the columns written to in logs. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2022-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23305>CVE-2022-23305</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://reload4j.qos.ch/">https://reload4j.qos.ch/</a></p>
<p>Release Date: 2022-01-18</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.2</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> <img src='https://whitesource-resources.whitesourcesoftware.com/viaGreen.png' width=19 height=20> CVE-2019-17571</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Reachability Analysis
<p>
<p>The vulnerable code is not reachable.</p>
</p>
<p></p>
### Vulnerability Details
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://lists.apache.org/thread.html/eea03d504b36e8f870e8321d908e1def1addda16adda04327fe7c125%40%3Cdev.logging.apache.org%3E">https://lists.apache.org/thread.html/eea03d504b36e8f870e8321d908e1def1addda16adda04327fe7c125%40%3Cdev.logging.apache.org%3E</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: log4j-manual - 1.2.17-16;log4j-javadoc - 1.2.17-16;log4j - 1.2.17-16,1.2.17-16</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> <img src='https://whitesource-resources.whitesourcesoftware.com/viaGreen.png' width=19 height=20> CVE-2020-9493</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Reachability Analysis
<p>
<p>The vulnerable code is not reachable.</p>
</p>
<p></p>
### Vulnerability Details
<p>
A deserialization flaw was found in Apache Chainsaw versions prior to 2.1.0 which could lead to malicious code execution.
<p>Publish Date: 2021-06-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9493>CVE-2020-9493</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2021/06/16/1">https://www.openwall.com/lists/oss-security/2021/06/16/1</a></p>
<p>Release Date: 2021-06-16</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.1</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2022-23307</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
CVE-2020-9493 identified a deserialization issue that was present in Apache Chainsaw. Prior to Chainsaw V2.0 Chainsaw was a component of Apache Log4j 1.2.x where the same issue exists.
<p>Publish Date: 2022-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23307>CVE-2022-23307</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-01-18</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.1</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2022-23302</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
JMSSink in all versions of Log4j 1.x is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration or if the configuration references an LDAP service the attacker has access to. The attacker can provide a TopicConnectionFactoryBindingName configuration causing JMSSink to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-4104. Note this issue only affects Log4j 1.x when specifically configured to use JMSSink, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2022-01-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23302>CVE-2022-23302</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://reload4j.qos.ch/">https://reload4j.qos.ch/</a></p>
<p>Release Date: 2022-01-18</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.1</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2021-4104</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2021-12-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-4104>CVE-2021-4104</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-4104">https://nvd.nist.gov/vuln/detail/CVE-2021-4104</a></p>
<p>Release Date: 2021-12-14</p>
<p>Fix Resolution: uom-parent - 1.0.3-3.module,1.0.3-3.module;uom-se-javadoc - 1.0.4-3.module;parfait-examples - 0.5.4-4.module;log4j-manual - 1.2.17-16;si-units-javadoc - 0.6.5-2.module;unit-api - 1.0-5.module,1.0-5.module;unit-api-javadoc - 1.0-5.module;parfait - 0.5.4-4.module,0.5.4-4.module;log4j-javadoc - 1.2.17-16;uom-systems-javadoc - 0.7-1.module;uom-lib-javadoc - 1.0.1-6.module;uom-systems - 0.7-1.module,0.7-1.module;log4j - 1.2.17-16,1.2.17-16;uom-se - 1.0.4-3.module,1.0.4-3.module;uom-lib - 1.0.1-6.module,1.0.1-6.module;parfait-javadoc - 0.5.4-4.module;pcp-parfait-agent - 0.5.4-4.module;si-units - 0.6.5-2.module,0.6.5-2.module</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2023-26464</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
** UNSUPPORTED WHEN ASSIGNED **
When using the Chainsaw or SocketAppender components with Log4j 1.x on JRE less than 1.7, an attacker that manages to cause a logging entry involving a specially-crafted (ie, deeply nested)
hashmap or hashtable (depending on which logging component is in use) to be processed could exhaust the available memory in the virtual machine and achieve Denial of Service when the object is deserialized.
This issue affects Apache Log4j before 2. Affected users are recommended to update to Log4j 2.x.
NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26464>CVE-2023-26464</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-vp98-w2p3-mv35">https://github.com/advisories/GHSA-vp98-w2p3-mv35</a></p>
<p>Release Date: 2023-03-10</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> CVE-2020-9488</summary>
### Vulnerable Library - <b>log4j-1.2.13.jar</b></p>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/HughC-GH-Demo/Java-Demo-renovate/commit/7e5c1845e221df395b1901471b8d9b70d106811c">7e5c1845e221df395b1901471b8d9b70d106811c</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender. Fixed in Apache Log4j 2.12.3 and 2.13.1
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://reload4j.qos.ch/">https://reload4j.qos.ch/</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: ch.qos.reload4j:reload4j:1.2.18.3</p>
</p>
<p></p>
<p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
</details>
***
<p>In order to enable automatic remediation for this issue, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> | non_priority | jar vulnerabilities highest severity is vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in version remediation available reachability critical jar direct ch qos critical jar direct manual javadoc the vulnerability is non reachable critical jar direct ch qos the vulnerability is non reachable high jar direct ch qos high jar direct ch qos high jar direct uom parent module module uom se javadoc module parfait examples module manual si units javadoc module unit api module module unit api javadoc module parfait module module javadoc uom systems javadoc module uom lib javadoc module uom systems module module uom se module module uom lib module module parfait javadoc module pcp parfait agent module si units module module high jar direct org apache logging core low jar direct ch qos details cve vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch main vulnerability details by design the jdbcappender in x accepts an sql statement as a configuration parameter where the values to be inserted are converters from patternlayout the message converter m is likely to always be included this allows attackers to manipulate the sql by entering crafted strings into input fields or headers of an application that are logged allowing unintended sql queries to be executed note this issue only affects x when specifically configured to use the jdbcappender which is not the default beginning in version the jdbcappender was re introduced with proper support for parameterized sql queries and further customization over the columns written to in logs apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ch qos in order to enable automatic remediation please create cve vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch main reachability analysis the vulnerable code is not reachable vulnerability details included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution manual javadoc in order to enable automatic remediation please create cve vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch main reachability analysis the vulnerable code is not reachable vulnerability details a deserialization flaw was found in apache chainsaw versions prior to which could lead to malicious code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ch qos in order to enable automatic remediation please create cve vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch main vulnerability details cve identified a deserialization issue that was present in apache chainsaw prior to chainsaw chainsaw was a component of apache x where the same issue exists publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution ch qos in order to enable automatic remediation please create cve vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch main vulnerability details jmssink in all versions of x is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration or if the configuration references an ldap service the attacker has access to the attacker can provide a topicconnectionfactorybindingname configuration causing jmssink to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects x when specifically configured to use jmssink which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ch qos in order to enable automatic remediation please create cve vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch main vulnerability details jmsappender in is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration the attacker can provide topicbindingname and topicconnectionfactorybindingname configurations causing jmsappender to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects when specifically configured to use jmsappender which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution uom parent module module uom se javadoc module parfait examples module manual si units javadoc module unit api module module unit api javadoc module parfait module module javadoc uom systems javadoc module uom lib javadoc module uom systems module module uom se module module uom lib module module parfait javadoc module pcp parfait agent module si units module module in order to enable automatic remediation please create cve vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch main vulnerability details unsupported when assigned when using the chainsaw or socketappender components with x on jre less than an attacker that manages to cause a logging entry involving a specially crafted ie deeply nested hashmap or hashtable depending on which logging component is in use to be processed could exhaust the available memory in the virtual machine and achieve denial of service when the object is deserialized this issue affects apache before affected users are recommended to update to x note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core in order to enable automatic remediation please create cve vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch main vulnerability details improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender fixed in apache and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ch qos in order to enable automatic remediation please create in order to enable automatic remediation for this issue please create | 0 |
172,966 | 14,397,421,845 | IssuesEvent | 2020-12-03 08:08:28 | enthought/traits | https://api.github.com/repos/enthought/traits | opened | Errors on front page of user manual | component: documentation type: bug | From an internal report:
> I noticed two errors on [this page](https://docs.enthought.com/traits/traits_user_manual/intro.html):
> - The two blue buttons at the top both say "user manual" but only the second links to the user manual.
> - At the end of the second paragraph, only the middle of the three bullet points is hyperlinked, and it points to the user manual for traitsui, not traits. It seems like that link should just be removed since linking to the user manual as an additional resource within the user manual itself is odd. | 1.0 | Errors on front page of user manual - From an internal report:
> I noticed two errors on [this page](https://docs.enthought.com/traits/traits_user_manual/intro.html):
> - The two blue buttons at the top both say "user manual" but only the second links to the user manual.
> - At the end of the second paragraph, only the middle of the three bullet points is hyperlinked, and it points to the user manual for traitsui, not traits. It seems like that link should just be removed since linking to the user manual as an additional resource within the user manual itself is odd. | non_priority | errors on front page of user manual from an internal report i noticed two errors on the two blue buttons at the top both say user manual but only the second links to the user manual at the end of the second paragraph only the middle of the three bullet points is hyperlinked and it points to the user manual for traitsui not traits it seems like that link should just be removed since linking to the user manual as an additional resource within the user manual itself is odd | 0 |
274,947 | 30,188,310,022 | IssuesEvent | 2023-07-04 13:36:28 | gabriel-milan/denoising-autoencoder | https://api.github.com/repos/gabriel-milan/denoising-autoencoder | opened | CVE-2021-41204 (Medium) detected in tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl | Mend: dependency security vulnerability | ## CVE-2021-41204 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/aa/fd/993aa1333eb54d9f000863fe8ec61e41d12eb833dea51484c76c038718b5/tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/aa/fd/993aa1333eb54d9f000863fe8ec61e41d12eb833dea51484c76c038718b5/tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /training/requirements.txt</p>
<p>Path to vulnerable library: /training/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gabriel-milan/denoising-autoencoder/commit/22186005a9ff5cf052b53f8bb5aa092b9ea8a670">22186005a9ff5cf052b53f8bb5aa092b9ea8a670</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. In affected versions during TensorFlow's Grappler optimizer phase, constant folding might attempt to deep copy a resource tensor. This results in a segfault, as these tensors are supposed to not change. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-11-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-41204>CVE-2021-41204</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-786j-5qwq-r36x">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-786j-5qwq-r36x</a></p>
<p>Release Date: 2021-11-05</p>
<p>Fix Resolution: 2.5.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-41204 (Medium) detected in tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl - ## CVE-2021-41204 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/aa/fd/993aa1333eb54d9f000863fe8ec61e41d12eb833dea51484c76c038718b5/tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/aa/fd/993aa1333eb54d9f000863fe8ec61e41d12eb833dea51484c76c038718b5/tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /training/requirements.txt</p>
<p>Path to vulnerable library: /training/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-2.5.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gabriel-milan/denoising-autoencoder/commit/22186005a9ff5cf052b53f8bb5aa092b9ea8a670">22186005a9ff5cf052b53f8bb5aa092b9ea8a670</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. In affected versions during TensorFlow's Grappler optimizer phase, constant folding might attempt to deep copy a resource tensor. This results in a segfault, as these tensors are supposed to not change. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-11-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-41204>CVE-2021-41204</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-786j-5qwq-r36x">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-786j-5qwq-r36x</a></p>
<p>Release Date: 2021-11-05</p>
<p>Fix Resolution: 2.5.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file training requirements txt path to vulnerable library training requirements txt dependency hierarchy x tensorflow whl vulnerable library found in head commit a href found in base branch master vulnerability details tensorflow is an open source platform for machine learning in affected versions during tensorflow s grappler optimizer phase constant folding might attempt to deep copy a resource tensor this results in a segfault as these tensors are supposed to not change the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
115,038 | 17,270,607,937 | IssuesEvent | 2021-07-22 19:16:59 | CliffCrerar/ionic-amplify-conference-app | https://api.github.com/repos/CliffCrerar/ionic-amplify-conference-app | opened | CVE-2020-7693 (Medium) detected in sockjs-0.3.19.tgz | security vulnerability | ## CVE-2020-7693 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sockjs-0.3.19.tgz</b></p></summary>
<p>SockJS-node is a server counterpart of SockJS-client a JavaScript library that provides a WebSocket-like object in the browser. SockJS gives you a coherent, cross-browser, Javascript API which creates a low latency, full duplex, cross-domain communication</p>
<p>Library home page: <a href="https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz">https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz</a></p>
<p>Path to dependency file: ionic-amplify-conference-app/package.json</p>
<p>Path to vulnerable library: ionic-amplify-conference-app/node_modules/sockjs/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.803.25.tgz (Root Library)
- webpack-dev-server-3.9.0.tgz
- :x: **sockjs-0.3.19.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/CliffCrerar/ionic-amplify-conference-app/commit/4e9a80774160275a55322301ed17c68e011c3b8f">4e9a80774160275a55322301ed17c68e011c3b8f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Incorrect handling of Upgrade header with the value websocket leads in crashing of containers hosting sockjs apps. This affects the package sockjs before 0.3.20.
<p>Publish Date: 2020-07-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693>CVE-2020-7693</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sockjs/sockjs-node/pull/265">https://github.com/sockjs/sockjs-node/pull/265</a></p>
<p>Release Date: 2020-07-09</p>
<p>Fix Resolution: sockjs - 0.3.20</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7693 (Medium) detected in sockjs-0.3.19.tgz - ## CVE-2020-7693 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sockjs-0.3.19.tgz</b></p></summary>
<p>SockJS-node is a server counterpart of SockJS-client a JavaScript library that provides a WebSocket-like object in the browser. SockJS gives you a coherent, cross-browser, Javascript API which creates a low latency, full duplex, cross-domain communication</p>
<p>Library home page: <a href="https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz">https://registry.npmjs.org/sockjs/-/sockjs-0.3.19.tgz</a></p>
<p>Path to dependency file: ionic-amplify-conference-app/package.json</p>
<p>Path to vulnerable library: ionic-amplify-conference-app/node_modules/sockjs/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.803.25.tgz (Root Library)
- webpack-dev-server-3.9.0.tgz
- :x: **sockjs-0.3.19.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/CliffCrerar/ionic-amplify-conference-app/commit/4e9a80774160275a55322301ed17c68e011c3b8f">4e9a80774160275a55322301ed17c68e011c3b8f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Incorrect handling of Upgrade header with the value websocket leads in crashing of containers hosting sockjs apps. This affects the package sockjs before 0.3.20.
<p>Publish Date: 2020-07-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693>CVE-2020-7693</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sockjs/sockjs-node/pull/265">https://github.com/sockjs/sockjs-node/pull/265</a></p>
<p>Release Date: 2020-07-09</p>
<p>Fix Resolution: sockjs - 0.3.20</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in sockjs tgz cve medium severity vulnerability vulnerable library sockjs tgz sockjs node is a server counterpart of sockjs client a javascript library that provides a websocket like object in the browser sockjs gives you a coherent cross browser javascript api which creates a low latency full duplex cross domain communication library home page a href path to dependency file ionic amplify conference app package json path to vulnerable library ionic amplify conference app node modules sockjs package json dependency hierarchy build angular tgz root library webpack dev server tgz x sockjs tgz vulnerable library found in head commit a href found in base branch master vulnerability details incorrect handling of upgrade header with the value websocket leads in crashing of containers hosting sockjs apps this affects the package sockjs before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution sockjs step up your open source security game with whitesource | 0 |
4,965 | 3,898,675,595 | IssuesEvent | 2016-04-17 08:02:05 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 14291308: Notification can't be discarded | classification:ui/usability reproducible:always status:open | #### Description
Summary:
Notification removal cross is absent from notification centre.
Steps to Reproduce:
1.slide down to show Notification centre.
2.tap "All" to reveal all last notifications
Expected Results:
The should be an "x" in front of each app name to remove its notifications
Actual Results:
No "x" is displayed
Regression:
There was an "x" in iOS 6.1.4
-
Product Version: iOS 7b2
Created: 2013-06-27T17:17:54.864535
Originated: 2013-06-27T19:17:00
Open Radar Link: http://www.openradar.me/14291308 | True | 14291308: Notification can't be discarded - #### Description
Summary:
Notification removal cross is absent from notification centre.
Steps to Reproduce:
1.slide down to show Notification centre.
2.tap "All" to reveal all last notifications
Expected Results:
The should be an "x" in front of each app name to remove its notifications
Actual Results:
No "x" is displayed
Regression:
There was an "x" in iOS 6.1.4
-
Product Version: iOS 7b2
Created: 2013-06-27T17:17:54.864535
Originated: 2013-06-27T19:17:00
Open Radar Link: http://www.openradar.me/14291308 | non_priority | notification can t be discarded description summary notification removal cross is absent from notification centre steps to reproduce slide down to show notification centre tap all to reveal all last notifications expected results the should be an x in front of each app name to remove its notifications actual results no x is displayed regression there was an x in ios product version ios created originated open radar link | 0 |
109,766 | 16,891,979,519 | IssuesEvent | 2021-06-23 10:20:28 | epam/TimeBase | https://api.github.com/repos/epam/TimeBase | closed | CVE-2018-12418 (Medium) detected in junrar-0.7.jar - autoclosed | security vulnerability | ## CVE-2018-12418 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>junrar-0.7.jar</b></p></summary>
<p>rar decompression library in plain java</p>
<p>Library home page: <a href="https://github.com/junrar/junrar">https://github.com/junrar/junrar</a></p>
<p>Path to dependency file: TimeBase/java/installer/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.github.junrar/junrar/0.7/18cc717b85af0b12ba922abf415c2ff4716f8219/junrar-0.7.jar</p>
<p>
Dependency Hierarchy:
- izpack-compiler-5.1.3.jar (Root Library)
- tika-parsers-1.9.jar
- :x: **junrar-0.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/epam/TimeBase/commits/98f6880b361c00a247f77e79a787646e9664cadd">98f6880b361c00a247f77e79a787646e9664cadd</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Archive.java in Junrar before 1.0.1, as used in Apache Tika and other products, is affected by a denial of service vulnerability due to an infinite loop when handling corrupt RAR files.
<p>Publish Date: 2018-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12418>CVE-2018-12418</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12418">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12418</a></p>
<p>Release Date: 2018-06-14</p>
<p>Fix Resolution: 1.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.github.junrar","packageName":"junrar","packageVersion":"0.7","packageFilePaths":["/java/installer/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.codehaus.izpack:izpack-compiler:5.1.3;org.apache.tika:tika-parsers:1.9;com.github.junrar:junrar:0.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.1"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-12418","vulnerabilityDetails":"Archive.java in Junrar before 1.0.1, as used in Apache Tika and other products, is affected by a denial of service vulnerability due to an infinite loop when handling corrupt RAR files.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12418","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-12418 (Medium) detected in junrar-0.7.jar - autoclosed - ## CVE-2018-12418 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>junrar-0.7.jar</b></p></summary>
<p>rar decompression library in plain java</p>
<p>Library home page: <a href="https://github.com/junrar/junrar">https://github.com/junrar/junrar</a></p>
<p>Path to dependency file: TimeBase/java/installer/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/com.github.junrar/junrar/0.7/18cc717b85af0b12ba922abf415c2ff4716f8219/junrar-0.7.jar</p>
<p>
Dependency Hierarchy:
- izpack-compiler-5.1.3.jar (Root Library)
- tika-parsers-1.9.jar
- :x: **junrar-0.7.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/epam/TimeBase/commits/98f6880b361c00a247f77e79a787646e9664cadd">98f6880b361c00a247f77e79a787646e9664cadd</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Archive.java in Junrar before 1.0.1, as used in Apache Tika and other products, is affected by a denial of service vulnerability due to an infinite loop when handling corrupt RAR files.
<p>Publish Date: 2018-06-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12418>CVE-2018-12418</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12418">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12418</a></p>
<p>Release Date: 2018-06-14</p>
<p>Fix Resolution: 1.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.github.junrar","packageName":"junrar","packageVersion":"0.7","packageFilePaths":["/java/installer/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.codehaus.izpack:izpack-compiler:5.1.3;org.apache.tika:tika-parsers:1.9;com.github.junrar:junrar:0.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.0.1"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2018-12418","vulnerabilityDetails":"Archive.java in Junrar before 1.0.1, as used in Apache Tika and other products, is affected by a denial of service vulnerability due to an infinite loop when handling corrupt RAR files.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12418","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in junrar jar autoclosed cve medium severity vulnerability vulnerable library junrar jar rar decompression library in plain java library home page a href path to dependency file timebase java installer build gradle path to vulnerable library home wss scanner gradle caches modules files com github junrar junrar junrar jar dependency hierarchy izpack compiler jar root library tika parsers jar x junrar jar vulnerable library found in head commit a href found in base branch main vulnerability details archive java in junrar before as used in apache tika and other products is affected by a denial of service vulnerability due to an infinite loop when handling corrupt rar files publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org codehaus izpack izpack compiler org apache tika tika parsers com github junrar junrar isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails archive java in junrar before as used in apache tika and other products is affected by a denial of service vulnerability due to an infinite loop when handling corrupt rar files vulnerabilityurl | 0 |
97,233 | 10,996,998,909 | IssuesEvent | 2019-12-03 08:11:29 | emqx/emqx | https://api.github.com/repos/emqx/emqx | closed | zone.external.publish_limit not work! | Documentation | #### Environment
- OS:
- Erlang/OTP:
- EMQ: EMQ X Broker v3.1.1
#### Description
config emqx.conf like this and then reboot emqx but not work
```
## Value: Number,Duration
## Example: 10 messages per minute.
zone.external.publish_limit = 10,1m
``` | 1.0 | zone.external.publish_limit not work! - #### Environment
- OS:
- Erlang/OTP:
- EMQ: EMQ X Broker v3.1.1
#### Description
config emqx.conf like this and then reboot emqx but not work
```
## Value: Number,Duration
## Example: 10 messages per minute.
zone.external.publish_limit = 10,1m
``` | non_priority | zone external publish limit not work environment os erlang otp emq emq x broker description config emqx conf like this and then reboot emqx but not work value number duration example messages per minute zone external publish limit | 0 |
97,412 | 20,254,191,463 | IssuesEvent | 2022-02-14 21:09:02 | learnpack/learnpack | https://api.github.com/repos/learnpack/learnpack | closed | learnpack-reader | good first issue 💻 Code | Create a npm package that can be used with npx to really-fast preview any markdown file. For example:
```bash
$ npx learnpack-reader -f ./path/to/readme.md
```
When the user types that in the console a browser will open (express) and show the HTML website of that markdown.
1. Use the markdown component from this library:
https://github.com/breatheco-de/react-components
2. Learn how to publish a npm package.
3. Learn how to use a npm package with npx.
4. Learn how use express.js to expose a URL pointing to a html. | 1.0 | learnpack-reader - Create a npm package that can be used with npx to really-fast preview any markdown file. For example:
```bash
$ npx learnpack-reader -f ./path/to/readme.md
```
When the user types that in the console a browser will open (express) and show the HTML website of that markdown.
1. Use the markdown component from this library:
https://github.com/breatheco-de/react-components
2. Learn how to publish a npm package.
3. Learn how to use a npm package with npx.
4. Learn how use express.js to expose a URL pointing to a html. | non_priority | learnpack reader create a npm package that can be used with npx to really fast preview any markdown file for example bash npx learnpack reader f path to readme md when the user types that in the console a browser will open express and show the html website of that markdown use the markdown component from this library learn how to publish a npm package learn how to use a npm package with npx learn how use express js to expose a url pointing to a html | 0 |
46,480 | 6,019,263,997 | IssuesEvent | 2017-06-07 14:11:24 | geetsisbac/E55Q3NOULHJ77TLQU57Z4QQM | https://api.github.com/repos/geetsisbac/E55Q3NOULHJ77TLQU57Z4QQM | reopened | y2pVusD2SfZ4RH+gbvQaVKgVcdKDfDQOkzCZSKVWXCmPmyOuLn6lCD4AoiXcE7X8nhgpRuYahDiRQqbTqpYdZVZ0aNsZJLRn7oJFDJHiihMJf5PIrE5/jC1EMYk7LOSGqkPWvcZeXSrzmZJPbwbx2CFAoZlqSVQG4ygEbwhdr48= | design | L6z8wwUwerfz94skZ7HrGzwOrtK2aES99Zl/E4XHz+Tjq+mUAUlTqyMyy14YS3H3Eemd6BGdDMxs2zFqrXktlG2rddAzYnFvTcjqWq9h2AT79iFQpwebZA6zHz24l+NfrkXz5/SNNyi6F4tfhKkVJBECpJ0O/0Fwb36NzTcFnDiYi3QqCX4n9etOhzZzwZjLRIbhuG6BYaDLOcC3QoHA/ESG4bhugWGgyznAt0KBwPxEhuG4boFhoMs5wLdCgcD8RIbhuG6BYaDLOcC3QoHA/AuIBPbtYmv6k7Odj/T4XtQv7o/zLYDg1iTO/3GzVYntYwbxXkQNq8syj8j5PlPOsf6BFpieuV7+vGf1KZ8xFKoKGtajKiA+fdwPomfalnXnLfLVdNq+NS1a2yk7uPl3OOSVzWK5yAspD/7FV8UbN/zY36FogHjZEw2w8KS0mwmm2N+haIB42RMNsPCktJsJpuWvHhqpXRllP4TWvcSadfqt20/GaQn8Fc1khznoENQZA/DZ7bVCSCVX445l4eQef/BbJsgOOpqQZMT6O1qi6XFAKONf0OEZw70RbvCo8j+TghJSVuLOvzZfgvSlegc24ZWR3ss6FZBOJvYy3LY3sKHY36FogHjZEw2w8KS0mwmm2N+haIB42RMNsPCktJsJpsTgLPuhxsJ7BCRFiO2hMkP6C2fiyqBoSiGpwnOeRl6WaQKzeuykpbfShs/f47/7qDEXdAdX1TtXMCduQyiYr19gOHFEsHASEZ6PsLnwVQpH+soF4953vHzrP3nbLdWLJdjfoWiAeNkTDbDwpLSbCabY36FogHjZEw2w8KS0mwmmeV8/WyK9eGf/XH1eU2PWQwDpur9uQjodWNpnByOEqPkMG+LGbkwvP7T3IMVHhJtcXKHVkT5MFADJG8zo6lRiPrx8fer1lRew3r2XUGZ+Julk3++mcV11vwO72DUpZJfFBogu+aMyOkC7u+zwsUtbftjfoWiAeNkTDbDwpLSbCabY36FogHjZEw2w8KS0mwmmDgZM8c2MMPy0MifzkmXvtojPXjVDTqcdio0WB4VWpEbTWSKTkwk/T6p9V8k6RyyVLlU2Rp0eyd6m6efNqMeblscFonOulruFspMUG//IyudFiiAO3yIGju94DzGHKadvxf7LB9mh0rrmaUwd2+WcE9jfoWiAeNkTDbDwpLSbCabY36FogHjZEw2w8KS0mwmmHXcYyLKjP1p4ctGnzmD9w0qE69U8tIiY7zz1Madx9LxVAtz4yeeJc4fbtcxi8urTm3Tg6VrquCvBnjfe1JCz3FnFczxQMpi5n4saq4MhD8JulvIhobJYtRM48vCY2WvF5JXNYrnICykP/sVXxRs3/NjfoWiAeNkTDbDwpLSbCabY36FogHjZEw2w8KS0mwmm5a8eGqldGWU/hNa9xJp1+q3bT8ZpCfwVzWSHOegQ1BkABWoci+KUZW/u29Ex0AKY8FsmyA46mpBkxPo7WqLpcfVLJhSk1O9y1+5RGJb0VeTVL5zZbsT7qiw+QBuizmWh2N+haIB42RMNsPCktJsJptjfoWiAeNkTDbDwpLSbCaZTu4k1Wf6eb3V8XTeFi3TpNmHAGJL8N3cW8zVs7lfi40sSVACFiE078I0d5woQCy6o9Du/MJ+n7hyPUb1Eqnfbxa4cSXsAjLwN6o287+3+e49TJy5x8EL5Y93jOn9qIu+gLIyCCCdqHskAJttnYIYn2N+haIB42RMNsPCktJsJptjfoWiAeNkTDbDwpLSbCaYddxjIsqM/Wnhy0afOYP3DSoTr1Ty0iJjvPPUxp3H0vA6J2yvzMqZvljurtJkhiucL6lBCnBd36VPDO5epzoSFOhKlXiHhNui5M1ao90L+4DjuVRDRyocK5qCSRM7Bvbxl+k62GM8+VJ20KtO18qskttDk5EFQQ4Rt+S1eZ1AcQzFMGT1Y4gQWpvU7Nhl3X8gGmKExPXcny9o0vygRM3OGgQW2Lcwh0KCPurU6GF6irSkUOL65yoD5Th8ScBbfr+e7jdmhBpyj2D9kK8mTbh5yTNvGFBKBhEAHe10zN/t/bQ== | 1.0 | y2pVusD2SfZ4RH+gbvQaVKgVcdKDfDQOkzCZSKVWXCmPmyOuLn6lCD4AoiXcE7X8nhgpRuYahDiRQqbTqpYdZVZ0aNsZJLRn7oJFDJHiihMJf5PIrE5/jC1EMYk7LOSGqkPWvcZeXSrzmZJPbwbx2CFAoZlqSVQG4ygEbwhdr48= - L6z8wwUwerfz94skZ7HrGzwOrtK2aES99Zl/E4XHz+Tjq+mUAUlTqyMyy14YS3H3Eemd6BGdDMxs2zFqrXktlG2rddAzYnFvTcjqWq9h2AT79iFQpwebZA6zHz24l+NfrkXz5/SNNyi6F4tfhKkVJBECpJ0O/0Fwb36NzTcFnDiYi3QqCX4n9etOhzZzwZjLRIbhuG6BYaDLOcC3QoHA/ESG4bhugWGgyznAt0KBwPxEhuG4boFhoMs5wLdCgcD8RIbhuG6BYaDLOcC3QoHA/AuIBPbtYmv6k7Odj/T4XtQv7o/zLYDg1iTO/3GzVYntYwbxXkQNq8syj8j5PlPOsf6BFpieuV7+vGf1KZ8xFKoKGtajKiA+fdwPomfalnXnLfLVdNq+NS1a2yk7uPl3OOSVzWK5yAspD/7FV8UbN/zY36FogHjZEw2w8KS0mwmm2N+haIB42RMNsPCktJsJpuWvHhqpXRllP4TWvcSadfqt20/GaQn8Fc1khznoENQZA/DZ7bVCSCVX445l4eQef/BbJsgOOpqQZMT6O1qi6XFAKONf0OEZw70RbvCo8j+TghJSVuLOvzZfgvSlegc24ZWR3ss6FZBOJvYy3LY3sKHY36FogHjZEw2w8KS0mwmm2N+haIB42RMNsPCktJsJpsTgLPuhxsJ7BCRFiO2hMkP6C2fiyqBoSiGpwnOeRl6WaQKzeuykpbfShs/f47/7qDEXdAdX1TtXMCduQyiYr19gOHFEsHASEZ6PsLnwVQpH+soF4953vHzrP3nbLdWLJdjfoWiAeNkTDbDwpLSbCabY36FogHjZEw2w8KS0mwmmeV8/WyK9eGf/XH1eU2PWQwDpur9uQjodWNpnByOEqPkMG+LGbkwvP7T3IMVHhJtcXKHVkT5MFADJG8zo6lRiPrx8fer1lRew3r2XUGZ+Julk3++mcV11vwO72DUpZJfFBogu+aMyOkC7u+zwsUtbftjfoWiAeNkTDbDwpLSbCabY36FogHjZEw2w8KS0mwmmDgZM8c2MMPy0MifzkmXvtojPXjVDTqcdio0WB4VWpEbTWSKTkwk/T6p9V8k6RyyVLlU2Rp0eyd6m6efNqMeblscFonOulruFspMUG//IyudFiiAO3yIGju94DzGHKadvxf7LB9mh0rrmaUwd2+WcE9jfoWiAeNkTDbDwpLSbCabY36FogHjZEw2w8KS0mwmmHXcYyLKjP1p4ctGnzmD9w0qE69U8tIiY7zz1Madx9LxVAtz4yeeJc4fbtcxi8urTm3Tg6VrquCvBnjfe1JCz3FnFczxQMpi5n4saq4MhD8JulvIhobJYtRM48vCY2WvF5JXNYrnICykP/sVXxRs3/NjfoWiAeNkTDbDwpLSbCabY36FogHjZEw2w8KS0mwmm5a8eGqldGWU/hNa9xJp1+q3bT8ZpCfwVzWSHOegQ1BkABWoci+KUZW/u29Ex0AKY8FsmyA46mpBkxPo7WqLpcfVLJhSk1O9y1+5RGJb0VeTVL5zZbsT7qiw+QBuizmWh2N+haIB42RMNsPCktJsJptjfoWiAeNkTDbDwpLSbCaZTu4k1Wf6eb3V8XTeFi3TpNmHAGJL8N3cW8zVs7lfi40sSVACFiE078I0d5woQCy6o9Du/MJ+n7hyPUb1Eqnfbxa4cSXsAjLwN6o287+3+e49TJy5x8EL5Y93jOn9qIu+gLIyCCCdqHskAJttnYIYn2N+haIB42RMNsPCktJsJptjfoWiAeNkTDbDwpLSbCaYddxjIsqM/Wnhy0afOYP3DSoTr1Ty0iJjvPPUxp3H0vA6J2yvzMqZvljurtJkhiucL6lBCnBd36VPDO5epzoSFOhKlXiHhNui5M1ao90L+4DjuVRDRyocK5qCSRM7Bvbxl+k62GM8+VJ20KtO18qskttDk5EFQQ4Rt+S1eZ1AcQzFMGT1Y4gQWpvU7Nhl3X8gGmKExPXcny9o0vygRM3OGgQW2Lcwh0KCPurU6GF6irSkUOL65yoD5Th8ScBbfr+e7jdmhBpyj2D9kK8mTbh5yTNvGFBKBhEAHe10zN/t/bQ== | non_priority | tjq fdwpomfalnxnlflvdnq kuzw mj t bq | 0 |
150,026 | 23,589,385,768 | IssuesEvent | 2022-08-23 14:07:04 | carbon-design-system/carbon-platform | https://api.github.com/repos/carbon-design-system/carbon-platform | closed | Create landing page illustrations | role: design 🎨 | This sprint we are updating our landing page, which has brought about the need for additional illustration work.
Prototype
https://www.figma.com/file/xjGZCRt3G7S0EXO7Bt4Vtq/v1_Landing_page_revisions?node-id=0%3A1
Illustrations needed:
- [ ] How Carbon Works illustration. This should help people understand that the page will introduce them to the ecosystem and give some definitions. Perhaps imagery of a network?
- [ ] Carbon React feature card. We have an old illustration that uses black. Can we make this lighter (i.e. color scheme of data visualization.)
- [ ] IBM Web Components feature card. This is new. Perhaps it could point towards the editorial/content nature of these components.
- [ ] Roadmap illustration. @jeanservaas will make this image based on the previous roadmap illustrations. | 1.0 | Create landing page illustrations - This sprint we are updating our landing page, which has brought about the need for additional illustration work.
Prototype
https://www.figma.com/file/xjGZCRt3G7S0EXO7Bt4Vtq/v1_Landing_page_revisions?node-id=0%3A1
Illustrations needed:
- [ ] How Carbon Works illustration. This should help people understand that the page will introduce them to the ecosystem and give some definitions. Perhaps imagery of a network?
- [ ] Carbon React feature card. We have an old illustration that uses black. Can we make this lighter (i.e. color scheme of data visualization.)
- [ ] IBM Web Components feature card. This is new. Perhaps it could point towards the editorial/content nature of these components.
- [ ] Roadmap illustration. @jeanservaas will make this image based on the previous roadmap illustrations. | non_priority | create landing page illustrations this sprint we are updating our landing page which has brought about the need for additional illustration work prototype illustrations needed how carbon works illustration this should help people understand that the page will introduce them to the ecosystem and give some definitions perhaps imagery of a network carbon react feature card we have an old illustration that uses black can we make this lighter i e color scheme of data visualization ibm web components feature card this is new perhaps it could point towards the editorial content nature of these components roadmap illustration jeanservaas will make this image based on the previous roadmap illustrations | 0 |
368,497 | 25,796,069,623 | IssuesEvent | 2022-12-10 15:20:05 | pan1caisreal/filmsReviews | https://api.github.com/repos/pan1caisreal/filmsReviews | opened | Техническое задание | documentation | 1. Техническое задание
1.1. Общие сведения:
1.1.1. Полное наименование системы:
Сайт для отслеживания фильмов
1.1.2. Наименования организации-заказчика, разработчика системы:
Заказчик: КФ МГТУ им. Н. Э. Баумана
Разработчик: Студент КФ МГТУ им. Н. Э. Баумана группы ИУК5-52Б Королев Д.И.
1.1.3. Плановые сроки начала и окончания работы по созданию системы:
Плановый срок начала работ – 12 сентября 2022 года.
Плановый срок окончания работ – 1 декабря 2022 года.
1.2. Назначение и цели создания системы
1.2.1. Назначение системы:
Система предназначена любителям киноиндустрии для отслеживания фильмов и оставления отзывов.
1.2.2. Цели создания системы:
Система создается для выбора фильмов по отзывам оставленными другими пользователями, так как не всегда можно разобраться в фильме по его описанию. Также для составления своего списка фильмов, которые можно посмотреть в будущем.
1.3. Требования к системе
1.3.1. Требования к структуре и функционированию системы.
Система представляет из себя несколько модулей:
- клиентская часть, где клиенты смогут ознакомиться со списком фильмов, их отзывами, актерами, зарегистрироваться, оставить свой отзыв и добавить фильм в список желаемого.
- администраторская часть, где администраторы смогут добавлять актеров и фильмы.
Создаваемая система должна реализовывать следующие функции:
Пользователь:
- Система должна предоставлять пользователю страницу со списком фильмов
- Пользователь может регистрироваться в системе (а также авторизовываться после регистрации)
- Система должна давать пользователю возможность добавления фильма в список желаемого
- Система должна давать пользователю возможность оставление, редактирования и удаление оставленного им отзыва
Администратор:
- Добавлять актеров
- Добавлять фильмы
1.3.2. Требования к численности и квалификации персонала системы и режиму его работы.
Система рассчитана на неограниченное количество пользователь со стороны клиентской части, и одного администратора.
1.3.3. Требования по эргономике и технической эстетике:
Интерфейс системы не должен вводить пользователя в заблуждение, содержать элементы, назначение которых неочевидно или элементы, функционал которых отсутствует или не соответствует пояснению в интерфейсе.
Интерфейс должен быть понятен и не перегружен графическими элементами.
1.3.4. Требования к видам обеспечения
1.3.4.1. Требования к лингвистическому обеспечению системы
Средством описания предметной области, также, как и средством взаимодействия пользователя с системой является русский.
1.3.4.2. Требования к программному обеспечению системы
В качестве языка программирования для разработки северной части сайта используется JavaScript Node.js и фреймворк Express, в качестве СУБД используется Postgres, а для создания пользовательского интерфейса используется React.js.
Обязательное требование – наличие браузера.
1.4. Состав и содержание работ по созданию системы
- Анализ требований к системе:
Разработчик должен выполнить анализ области применения разрабатываемой системы с точки зрения определения требований к ней.
- Проектирование программной архитектуры:
Разработчик должен определить программную архитектуру, которая описывает общую структуру объекта и определяет компоненты программного объекта.
- Программирование и тестирование программных средств:
Разработчик должен разработать каждый программный модуль и базу данных, а также процедуры испытаний (тестирования) и данные для тестирования каждого программного модуля и базы данных.
1.5. Порядок контроля и приемки системы
Система разрабатывается с применением тестирования работоспособности функционала. При добавлении новых функций предыдущие тесты должны сохранять работоспособность.
Приемка работы осуществляется комиссией КФ МГТУ им. Н. Э. Баумана кафедры ИУК5.
1.6. Требования к составу и содержанию работ по подготовке объекта автоматизации к вводу системы в действие.
Для корректной работы системы, на ПК должен быть доступ в интернет, установлена Windows 10 и выше, а также установлен любой браузер.
1.7. Требования к документированию
Требуется предоставить:
- Техническое задание в соответствии с ГОСТ 34.602-89
- Расчетно-пояснительную записку, включающую исследовательскую часть, проектно-конструкторскую часть и проектно-технологическую часть, включающую в себя руководство пользователя и руководство программиста (администратора). Расчетно-пояснительная записка выполняется с учетом требований, предусмотренных ГОСТ 7.32-2001 и 2.105-95. | 1.0 | Техническое задание - 1. Техническое задание
1.1. Общие сведения:
1.1.1. Полное наименование системы:
Сайт для отслеживания фильмов
1.1.2. Наименования организации-заказчика, разработчика системы:
Заказчик: КФ МГТУ им. Н. Э. Баумана
Разработчик: Студент КФ МГТУ им. Н. Э. Баумана группы ИУК5-52Б Королев Д.И.
1.1.3. Плановые сроки начала и окончания работы по созданию системы:
Плановый срок начала работ – 12 сентября 2022 года.
Плановый срок окончания работ – 1 декабря 2022 года.
1.2. Назначение и цели создания системы
1.2.1. Назначение системы:
Система предназначена любителям киноиндустрии для отслеживания фильмов и оставления отзывов.
1.2.2. Цели создания системы:
Система создается для выбора фильмов по отзывам оставленными другими пользователями, так как не всегда можно разобраться в фильме по его описанию. Также для составления своего списка фильмов, которые можно посмотреть в будущем.
1.3. Требования к системе
1.3.1. Требования к структуре и функционированию системы.
Система представляет из себя несколько модулей:
- клиентская часть, где клиенты смогут ознакомиться со списком фильмов, их отзывами, актерами, зарегистрироваться, оставить свой отзыв и добавить фильм в список желаемого.
- администраторская часть, где администраторы смогут добавлять актеров и фильмы.
Создаваемая система должна реализовывать следующие функции:
Пользователь:
- Система должна предоставлять пользователю страницу со списком фильмов
- Пользователь может регистрироваться в системе (а также авторизовываться после регистрации)
- Система должна давать пользователю возможность добавления фильма в список желаемого
- Система должна давать пользователю возможность оставление, редактирования и удаление оставленного им отзыва
Администратор:
- Добавлять актеров
- Добавлять фильмы
1.3.2. Требования к численности и квалификации персонала системы и режиму его работы.
Система рассчитана на неограниченное количество пользователь со стороны клиентской части, и одного администратора.
1.3.3. Требования по эргономике и технической эстетике:
Интерфейс системы не должен вводить пользователя в заблуждение, содержать элементы, назначение которых неочевидно или элементы, функционал которых отсутствует или не соответствует пояснению в интерфейсе.
Интерфейс должен быть понятен и не перегружен графическими элементами.
1.3.4. Требования к видам обеспечения
1.3.4.1. Требования к лингвистическому обеспечению системы
Средством описания предметной области, также, как и средством взаимодействия пользователя с системой является русский.
1.3.4.2. Требования к программному обеспечению системы
В качестве языка программирования для разработки северной части сайта используется JavaScript Node.js и фреймворк Express, в качестве СУБД используется Postgres, а для создания пользовательского интерфейса используется React.js.
Обязательное требование – наличие браузера.
1.4. Состав и содержание работ по созданию системы
- Анализ требований к системе:
Разработчик должен выполнить анализ области применения разрабатываемой системы с точки зрения определения требований к ней.
- Проектирование программной архитектуры:
Разработчик должен определить программную архитектуру, которая описывает общую структуру объекта и определяет компоненты программного объекта.
- Программирование и тестирование программных средств:
Разработчик должен разработать каждый программный модуль и базу данных, а также процедуры испытаний (тестирования) и данные для тестирования каждого программного модуля и базы данных.
1.5. Порядок контроля и приемки системы
Система разрабатывается с применением тестирования работоспособности функционала. При добавлении новых функций предыдущие тесты должны сохранять работоспособность.
Приемка работы осуществляется комиссией КФ МГТУ им. Н. Э. Баумана кафедры ИУК5.
1.6. Требования к составу и содержанию работ по подготовке объекта автоматизации к вводу системы в действие.
Для корректной работы системы, на ПК должен быть доступ в интернет, установлена Windows 10 и выше, а также установлен любой браузер.
1.7. Требования к документированию
Требуется предоставить:
- Техническое задание в соответствии с ГОСТ 34.602-89
- Расчетно-пояснительную записку, включающую исследовательскую часть, проектно-конструкторскую часть и проектно-технологическую часть, включающую в себя руководство пользователя и руководство программиста (администратора). Расчетно-пояснительная записка выполняется с учетом требований, предусмотренных ГОСТ 7.32-2001 и 2.105-95. | non_priority | техническое задание техническое задание общие сведения полное наименование системы сайт для отслеживания фильмов наименования организации заказчика разработчика системы заказчик кф мгту им н э баумана разработчик студент кф мгту им н э баумана группы королев д и плановые сроки начала и окончания работы по созданию системы плановый срок начала работ – сентября года плановый срок окончания работ – декабря года назначение и цели создания системы назначение системы система предназначена любителям киноиндустрии для отслеживания фильмов и оставления отзывов цели создания системы система создается для выбора фильмов по отзывам оставленными другими пользователями так как не всегда можно разобраться в фильме по его описанию также для составления своего списка фильмов которые можно посмотреть в будущем требования к системе требования к структуре и функционированию системы система представляет из себя несколько модулей клиентская часть где клиенты смогут ознакомиться со списком фильмов их отзывами актерами зарегистрироваться оставить свой отзыв и добавить фильм в список желаемого администраторская часть где администраторы смогут добавлять актеров и фильмы создаваемая система должна реализовывать следующие функции пользователь система должна предоставлять пользователю страницу со списком фильмов пользователь может регистрироваться в системе а также авторизовываться после регистрации система должна давать пользователю возможность добавления фильма в список желаемого система должна давать пользователю возможность оставление редактирования и удаление оставленного им отзыва администратор добавлять актеров добавлять фильмы требования к численности и квалификации персонала системы и режиму его работы система рассчитана на неограниченное количество пользователь со стороны клиентской части и одного администратора требования по эргономике и технической эстетике интерфейс системы не должен вводить пользователя в заблуждение содержать элементы назначение которых неочевидно или элементы функционал которых отсутствует или не соответствует пояснению в интерфейсе интерфейс должен быть понятен и не перегружен графическими элементами требования к видам обеспечения требования к лингвистическому обеспечению системы средством описания предметной области также как и средством взаимодействия пользователя с системой является русский требования к программному обеспечению системы в качестве языка программирования для разработки северной части сайта используется javascript node js и фреймворк express в качестве субд используется postgres а для создания пользовательского интерфейса используется react js обязательное требование – наличие браузера состав и содержание работ по созданию системы анализ требований к системе разработчик должен выполнить анализ области применения разрабатываемой системы с точки зрения определения требований к ней проектирование программной архитектуры разработчик должен определить программную архитектуру которая описывает общую структуру объекта и определяет компоненты программного объекта программирование и тестирование программных средств разработчик должен разработать каждый программный модуль и базу данных а также процедуры испытаний тестирования и данные для тестирования каждого программного модуля и базы данных порядок контроля и приемки системы система разрабатывается с применением тестирования работоспособности функционала при добавлении новых функций предыдущие тесты должны сохранять работоспособность приемка работы осуществляется комиссией кф мгту им н э баумана кафедры требования к составу и содержанию работ по подготовке объекта автоматизации к вводу системы в действие для корректной работы системы на пк должен быть доступ в интернет установлена windows и выше а также установлен любой браузер требования к документированию требуется предоставить техническое задание в соответствии с гост расчетно пояснительную записку включающую исследовательскую часть проектно конструкторскую часть и проектно технологическую часть включающую в себя руководство пользователя и руководство программиста администратора расчетно пояснительная записка выполняется с учетом требований предусмотренных гост и | 0 |
76,510 | 26,471,282,493 | IssuesEvent | 2023-01-17 07:34:42 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Metrics collector throws NullPointerException when metrics are disabled on client | Type: Defect Team: Client Source: Community Module: Metrics | **Describe the bug**
When using a hazelcast client, setting the following property:
```
<metrics enabled="false">
<jmx enabled="false"/>
</metrics>
```
causes a null pointer exception to be thrown during any metric collection:
`java.lang.NullPointerException: Cannot invoke "com.hazelcast.internal.metrics.collectors.MetricsCollector.collectException(com.hazelcast.internal.metrics.MetricDescriptor, java.lang.Exception)" because "collector" is null
at com.hazelcast.internal.metrics.impl.CompositeMetricsCollector.collectException(CompositeMetricsCollector.java:47) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.internal.metrics.impl.MetricsCollectionCycle.collectLong(MetricsCollectionCycle.java:177) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.internal.metrics.impl.MetricsCollectionCycle.collectStaticMetrics(MetricsCollectionCycle.java:82) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.internal.metrics.impl.MetricsRegistryImpl.collect(MetricsRegistryImpl.java:319) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.client.impl.statistics.ClientStatisticsService.collectAndSendStats(ClientStatisticsService.java:144) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) ~[?:?]
at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?]
at java.lang.Thread.run(Thread.java:833) ~[?:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]`
it looks like the ClientStatisticsService property -> publisherMetricsCollector is null when it is passed into the CompositeMetricsCollector constructor, which causes the issue.
**Expected behavior**
I can disable metrics on the client, and not see hazelcast exceptions during normal execution of the service.
**To Reproduce**
Steps to reproduce the behavior:
1. Use version 5.2.1 of hazelcast client.
2. Disable jmx metrics with the mentioned config
3. Run the client for a period of time.
**Additional context**
I can see in ClientStatisticsService it returns early on the start method, not setting the publisherMetricsCollector if its not enabled. Possibly, there would be no attempt to collect any metrics if no metrics are enabled, or the collect and send stats method should return early also, if its not enabled. For context, we disable all client metrics due to the performance implications on our GC, as we run Java in a low-latency environment with real-time deadlines. | 1.0 | Metrics collector throws NullPointerException when metrics are disabled on client - **Describe the bug**
When using a hazelcast client, setting the following property:
```
<metrics enabled="false">
<jmx enabled="false"/>
</metrics>
```
causes a null pointer exception to be thrown during any metric collection:
`java.lang.NullPointerException: Cannot invoke "com.hazelcast.internal.metrics.collectors.MetricsCollector.collectException(com.hazelcast.internal.metrics.MetricDescriptor, java.lang.Exception)" because "collector" is null
at com.hazelcast.internal.metrics.impl.CompositeMetricsCollector.collectException(CompositeMetricsCollector.java:47) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.internal.metrics.impl.MetricsCollectionCycle.collectLong(MetricsCollectionCycle.java:177) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.internal.metrics.impl.MetricsCollectionCycle.collectStaticMetrics(MetricsCollectionCycle.java:82) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.internal.metrics.impl.MetricsRegistryImpl.collect(MetricsRegistryImpl.java:319) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.client.impl.statistics.ClientStatisticsService.collectAndSendStats(ClientStatisticsService.java:144) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:539) ~[?:?]
at java.util.concurrent.FutureTask.run(FutureTask.java:264) ~[?:?]
at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:304) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1136) ~[?:?]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:635) ~[?:?]
at java.lang.Thread.run(Thread.java:833) ~[?:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.executeRun(HazelcastManagedThread.java:76) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]
at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) ~[streaming-server-nio-0.0.1-SNAPSHOT-0.0.1.jar:?]`
it looks like the ClientStatisticsService property -> publisherMetricsCollector is null when it is passed into the CompositeMetricsCollector constructor, which causes the issue.
**Expected behavior**
I can disable metrics on the client, and not see hazelcast exceptions during normal execution of the service.
**To Reproduce**
Steps to reproduce the behavior:
1. Use version 5.2.1 of hazelcast client.
2. Disable jmx metrics with the mentioned config
3. Run the client for a period of time.
**Additional context**
I can see in ClientStatisticsService it returns early on the start method, not setting the publisherMetricsCollector if its not enabled. Possibly, there would be no attempt to collect any metrics if no metrics are enabled, or the collect and send stats method should return early also, if its not enabled. For context, we disable all client metrics due to the performance implications on our GC, as we run Java in a low-latency environment with real-time deadlines. | non_priority | metrics collector throws nullpointerexception when metrics are disabled on client describe the bug when using a hazelcast client setting the following property causes a null pointer exception to be thrown during any metric collection java lang nullpointerexception cannot invoke com hazelcast internal metrics collectors metricscollector collectexception com hazelcast internal metrics metricdescriptor java lang exception because collector is null at com hazelcast internal metrics impl compositemetricscollector collectexception compositemetricscollector java at com hazelcast internal metrics impl metricscollectioncycle collectlong metricscollectioncycle java at com hazelcast internal metrics impl metricscollectioncycle collectstaticmetrics metricscollectioncycle java at com hazelcast internal metrics impl metricsregistryimpl collect metricsregistryimpl java at com hazelcast client impl statistics clientstatisticsservice collectandsendstats clientstatisticsservice java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at java util concurrent scheduledthreadpoolexecutor scheduledfuturetask run scheduledthreadpoolexecutor java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java at com hazelcast internal util executor hazelcastmanagedthread executerun hazelcastmanagedthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java it looks like the clientstatisticsservice property publishermetricscollector is null when it is passed into the compositemetricscollector constructor which causes the issue expected behavior i can disable metrics on the client and not see hazelcast exceptions during normal execution of the service to reproduce steps to reproduce the behavior use version of hazelcast client disable jmx metrics with the mentioned config run the client for a period of time additional context i can see in clientstatisticsservice it returns early on the start method not setting the publishermetricscollector if its not enabled possibly there would be no attempt to collect any metrics if no metrics are enabled or the collect and send stats method should return early also if its not enabled for context we disable all client metrics due to the performance implications on our gc as we run java in a low latency environment with real time deadlines | 0 |
7,981 | 10,145,015,689 | IssuesEvent | 2019-08-05 02:00:41 | Wabbit0101/mods_hoardercraft | https://api.github.com/repos/Wabbit0101/mods_hoardercraft | closed | SmartHoppers pouches incompatible with SpongeForge | incompatibility mod_smarthoppers | Pouches work fine in singleplayer, but on a SpongeForge server, when trying to open any pouch other than the Ender Chest Pouch, the following message appears in console, and nothing is shown to the client:
`[Server thread/WARN] [net.minecraft.entity.EntityTrackerEntry]: Fetching addPacket for removed entity` | True | SmartHoppers pouches incompatible with SpongeForge - Pouches work fine in singleplayer, but on a SpongeForge server, when trying to open any pouch other than the Ender Chest Pouch, the following message appears in console, and nothing is shown to the client:
`[Server thread/WARN] [net.minecraft.entity.EntityTrackerEntry]: Fetching addPacket for removed entity` | non_priority | smarthoppers pouches incompatible with spongeforge pouches work fine in singleplayer but on a spongeforge server when trying to open any pouch other than the ender chest pouch the following message appears in console and nothing is shown to the client fetching addpacket for removed entity | 0 |
25,077 | 18,098,381,511 | IssuesEvent | 2021-09-22 11:42:27 | Mudlet/Mudlet | https://api.github.com/repos/Mudlet/Mudlet | opened | Auto-tag regression issues to the next milestone | help wanted Public infrastructure | #### Brief summary of issue / Description of requested feature:
Auto-tag regression issues to the next milestone so that we know what issues are blocking a release.
#### Steps to reproduce the issue / Reasons for adding feature:
https://github.com/Mudlet/Mudlet/issues/5370 was a regression that slipped the horizon in 4.13 release because it wasn't tagged to the milestone.
#### Error output / Expected result of feature
A github actions workflow would suffice for this. Help wanted! | 1.0 | Auto-tag regression issues to the next milestone - #### Brief summary of issue / Description of requested feature:
Auto-tag regression issues to the next milestone so that we know what issues are blocking a release.
#### Steps to reproduce the issue / Reasons for adding feature:
https://github.com/Mudlet/Mudlet/issues/5370 was a regression that slipped the horizon in 4.13 release because it wasn't tagged to the milestone.
#### Error output / Expected result of feature
A github actions workflow would suffice for this. Help wanted! | non_priority | auto tag regression issues to the next milestone brief summary of issue description of requested feature auto tag regression issues to the next milestone so that we know what issues are blocking a release steps to reproduce the issue reasons for adding feature was a regression that slipped the horizon in release because it wasn t tagged to the milestone error output expected result of feature a github actions workflow would suffice for this help wanted | 0 |
39,825 | 16,101,096,208 | IssuesEvent | 2021-04-27 09:23:18 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Omit other bucket request if it's empty | :KibanaApp/fix-it-week Team:AppServices Team:KibanaApp | Other bucket is always executed, even if the first response indicates there will be no data.
Consider the following response:
```
{
"doc_count_error_upper_bound": 0,
"sum_other_doc_count": 0,
"buckets": [
{
"key": "CN",
"doc_count": 2
},
{
"key": "FR",
"doc_count": 2
},
{
"key": "ET",
"doc_count": 1
},
{
"key": "ID",
"doc_count": 1
},
{
"key": "IN",
"doc_count": 1
},
{
"key": "US",
"doc_count": 1
}
]
}
```
With `"sum_other_doc_count": 0`, the second request becomes unnecessary:
```
{
"other-filter": {
"filters": {
"filters": {
"": {
"bool": {
"must": [],
"filter": [
{
"exists": {
"field": "geo.src"
}
}
],
"should": [],
"must_not": [
{
"match_phrase": {
"geo.src": "CN"
}
},
{
"match_phrase": {
"geo.src": "FR"
}
},
{
"match_phrase": {
"geo.src": "ET"
}
},
{
"match_phrase": {
"geo.src": "ID"
}
},
{
"match_phrase": {
"geo.src": "IN"
}
},
{
"match_phrase": {
"geo.src": "US"
}
}
]
}
}
}
}
}
}
```
Tagging this with Kibana app because it's important for the Lens default behavior. | 1.0 | Omit other bucket request if it's empty - Other bucket is always executed, even if the first response indicates there will be no data.
Consider the following response:
```
{
"doc_count_error_upper_bound": 0,
"sum_other_doc_count": 0,
"buckets": [
{
"key": "CN",
"doc_count": 2
},
{
"key": "FR",
"doc_count": 2
},
{
"key": "ET",
"doc_count": 1
},
{
"key": "ID",
"doc_count": 1
},
{
"key": "IN",
"doc_count": 1
},
{
"key": "US",
"doc_count": 1
}
]
}
```
With `"sum_other_doc_count": 0`, the second request becomes unnecessary:
```
{
"other-filter": {
"filters": {
"filters": {
"": {
"bool": {
"must": [],
"filter": [
{
"exists": {
"field": "geo.src"
}
}
],
"should": [],
"must_not": [
{
"match_phrase": {
"geo.src": "CN"
}
},
{
"match_phrase": {
"geo.src": "FR"
}
},
{
"match_phrase": {
"geo.src": "ET"
}
},
{
"match_phrase": {
"geo.src": "ID"
}
},
{
"match_phrase": {
"geo.src": "IN"
}
},
{
"match_phrase": {
"geo.src": "US"
}
}
]
}
}
}
}
}
}
```
Tagging this with Kibana app because it's important for the Lens default behavior. | non_priority | omit other bucket request if it s empty other bucket is always executed even if the first response indicates there will be no data consider the following response doc count error upper bound sum other doc count buckets key cn doc count key fr doc count key et doc count key id doc count key in doc count key us doc count with sum other doc count the second request becomes unnecessary other filter filters filters bool must filter exists field geo src should must not match phrase geo src cn match phrase geo src fr match phrase geo src et match phrase geo src id match phrase geo src in match phrase geo src us tagging this with kibana app because it s important for the lens default behavior | 0 |
131,020 | 18,170,647,488 | IssuesEvent | 2021-09-27 19:35:27 | byorgey/swarm | https://api.github.com/repos/byorgey/swarm | opened | Inter-robot communication | Z-Feature C-Project S-Nice to have L-Language design G-Design | This may be still a ways off, but eventually we certainly want to be able to have inter-robot communication. This would enable lots of cool things such as remotely controlling or reprogramming robots, creating teams of robots that can dynamically coordinate among one another, etc. There is some discussion of this in https://github.com/byorgey/swarm/issues/17#issuecomment-927669308 .
There is a lot of language design work to be done here. My initial vague thought to look for inspiration to things like pi-calculus and session types. Probably full-on session types are too heavy-handed, though. Maybe something simple like typed broadcast channels: a value of type `chan t` is a channel over which values of type `t` can be broadcast, supporting operations like
```
newChannel : cmd (chan t)
broadcast : forall t. chan t -> t -> cmd ()
read : forall t. chan t -> cmd t
readNonBlocking : forall t. chan t -> cmd (maybe t)
```
Channels will be private in the sense that only robots with a reference to a channel can use it. There would have to be some special support for `newChannel` similar to `def`, so a robot can declare a new channel and then have it available globally to all robots it built after that.
There are also game design issues to think about; e.g. what sort(s) of device(s) should be needed to use these features? | 2.0 | Inter-robot communication - This may be still a ways off, but eventually we certainly want to be able to have inter-robot communication. This would enable lots of cool things such as remotely controlling or reprogramming robots, creating teams of robots that can dynamically coordinate among one another, etc. There is some discussion of this in https://github.com/byorgey/swarm/issues/17#issuecomment-927669308 .
There is a lot of language design work to be done here. My initial vague thought to look for inspiration to things like pi-calculus and session types. Probably full-on session types are too heavy-handed, though. Maybe something simple like typed broadcast channels: a value of type `chan t` is a channel over which values of type `t` can be broadcast, supporting operations like
```
newChannel : cmd (chan t)
broadcast : forall t. chan t -> t -> cmd ()
read : forall t. chan t -> cmd t
readNonBlocking : forall t. chan t -> cmd (maybe t)
```
Channels will be private in the sense that only robots with a reference to a channel can use it. There would have to be some special support for `newChannel` similar to `def`, so a robot can declare a new channel and then have it available globally to all robots it built after that.
There are also game design issues to think about; e.g. what sort(s) of device(s) should be needed to use these features? | non_priority | inter robot communication this may be still a ways off but eventually we certainly want to be able to have inter robot communication this would enable lots of cool things such as remotely controlling or reprogramming robots creating teams of robots that can dynamically coordinate among one another etc there is some discussion of this in there is a lot of language design work to be done here my initial vague thought to look for inspiration to things like pi calculus and session types probably full on session types are too heavy handed though maybe something simple like typed broadcast channels a value of type chan t is a channel over which values of type t can be broadcast supporting operations like newchannel cmd chan t broadcast forall t chan t t cmd read forall t chan t cmd t readnonblocking forall t chan t cmd maybe t channels will be private in the sense that only robots with a reference to a channel can use it there would have to be some special support for newchannel similar to def so a robot can declare a new channel and then have it available globally to all robots it built after that there are also game design issues to think about e g what sort s of device s should be needed to use these features | 0 |
959 | 2,789,872,504 | IssuesEvent | 2015-05-08 22:04:18 | biocore/scikit-bio | https://api.github.com/repos/biocore/scikit-bio | closed | test use of numpy array instead of str for storing BiologicalSequence._sequence | chapter enhancement performance | Suggested by @wasade. I did some initial testing, and it shouldn't be hard to make this change. See discussion on #53, but in particular:
https://github.com/biocore/bipy/pull/53#discussion-diff-9160149
https://github.com/biocore/bipy/pull/53#discussion-diff-9171472
| True | test use of numpy array instead of str for storing BiologicalSequence._sequence - Suggested by @wasade. I did some initial testing, and it shouldn't be hard to make this change. See discussion on #53, but in particular:
https://github.com/biocore/bipy/pull/53#discussion-diff-9160149
https://github.com/biocore/bipy/pull/53#discussion-diff-9171472
| non_priority | test use of numpy array instead of str for storing biologicalsequence sequence suggested by wasade i did some initial testing and it shouldn t be hard to make this change see discussion on but in particular | 0 |
48,736 | 13,184,726,766 | IssuesEvent | 2020-08-12 19:59:01 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | omkey inherits from i3Frame object (Trac #117) | IceTray Incomplete Migration Migrated from Trac defect | <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/117
, reported by troy and owned by troy_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "where did this come from?",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1416713876900096",
"component": "IceTray",
"summary": "omkey inherits from i3Frame object",
"priority": "major",
"keywords": "",
"time": "2008-08-30T16:06:14",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| 1.0 | omkey inherits from i3Frame object (Trac #117) - <details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/117
, reported by troy and owned by troy_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-11-23T03:37:56",
"description": "where did this come from?",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1416713876900096",
"component": "IceTray",
"summary": "omkey inherits from i3Frame object",
"priority": "major",
"keywords": "",
"time": "2008-08-30T16:06:14",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
</p>
</details>
| non_priority | omkey inherits from object trac migrated from reported by troy and owned by troy json status closed changetime description where did this come from reporter troy cc resolution wont or cant fix ts component icetray summary omkey inherits from object priority major keywords time milestone owner troy type defect | 0 |
182,858 | 21,678,001,240 | IssuesEvent | 2022-05-09 01:05:15 | renfei/start | https://api.github.com/repos/renfei/start | opened | CVE-2022-24823 (Medium) detected in netty-common-4.1.63.Final.jar | security vulnerability | ## CVE-2022-24823 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-common-4.1.63.Final.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-common/4.1.63.Final/netty-common-4.1.63.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-redis-reactive-2.4.5.jar (Root Library)
- spring-boot-starter-data-redis-2.4.5.jar
- lettuce-core-6.0.4.RELEASE.jar
- :x: **netty-common-4.1.63.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user.
<p>Publish Date: 2022-05-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823>CVE-2022-24823</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p>
<p>Release Date: 2022-05-06</p>
<p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-24823 (Medium) detected in netty-common-4.1.63.Final.jar - ## CVE-2022-24823 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-common-4.1.63.Final.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-common/4.1.63.Final/netty-common-4.1.63.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-redis-reactive-2.4.5.jar (Root Library)
- spring-boot-starter-data-redis-2.4.5.jar
- lettuce-core-6.0.4.RELEASE.jar
- :x: **netty-common-4.1.63.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user.
<p>Publish Date: 2022-05-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823>CVE-2022-24823</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p>
<p>Release Date: 2022-05-06</p>
<p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in netty common final jar cve medium severity vulnerability vulnerable library netty common final jar library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository io netty netty common final netty common final jar dependency hierarchy spring boot starter data redis reactive jar root library spring boot starter data redis jar lettuce core release jar x netty common final jar vulnerable library found in base branch master vulnerability details netty is an open source asynchronous event driven network application framework the package io netty netty codec http prior to version final contains an insufficient fix for cve when netty s multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled this only impacts applications running on java version and lower additionally this vulnerability impacts code running on unix like systems and very old versions of mac osx and windows as they all share the system temporary directory between all users version final contains a patch for this vulnerability as a workaround specify one s own java io tmpdir when starting the jvm or use defaulthttpdatafactory setbasedir to set the directory to something that is only readable by the current user publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all io netty netty common final step up your open source security game with whitesource | 0 |
279,134 | 30,702,453,477 | IssuesEvent | 2023-07-27 01:31:28 | panasalap/linux-4.1.15 | https://api.github.com/repos/panasalap/linux-4.1.15 | closed | CVE-2018-1000004 (Medium) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246, linux179e72b561d3d331c850e1a5779688d7a7de5246 - autoclosed | Mend: dependency security vulnerability | ## CVE-2018-1000004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b>, <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel 4.12, 3.10, 2.6 and possibly earlier versions a race condition vulnerability exists in the sound system, this can lead to a deadlock and denial of service condition.
<p>Publish Date: 2018-01-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-1000004>CVE-2018-1000004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000004">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000004</a></p>
<p>Release Date: 2018-01-16</p>
<p>Fix Resolution: v4.15-rc9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-1000004 (Medium) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246, linux179e72b561d3d331c850e1a5779688d7a7de5246 - autoclosed - ## CVE-2018-1000004 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b>, <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel 4.12, 3.10, 2.6 and possibly earlier versions a race condition vulnerability exists in the sound system, this can lead to a deadlock and denial of service condition.
<p>Publish Date: 2018-01-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-1000004>CVE-2018-1000004</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000004">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1000004</a></p>
<p>Release Date: 2018-01-16</p>
<p>Fix Resolution: v4.15-rc9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in autoclosed cve medium severity vulnerability vulnerable libraries vulnerability details in the linux kernel and possibly earlier versions a race condition vulnerability exists in the sound system this can lead to a deadlock and denial of service condition publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
40,312 | 8,773,214,587 | IssuesEvent | 2018-12-18 16:19:05 | eventespresso/event-espresso-core | https://api.github.com/repos/eventespresso/event-espresso-core | opened | Filter expired events from next/previous post linkslinks | category:frontend-ui-&-ux category:templates-shortcodes-and-widgets | Started on Codebase where Dean wrote:
```
http://eventespresso.com/topic/removing-out-of-date-events-from-front-end-of-website/
Basically, if you are on a single event and the theme supports next and previous post links, expired events can still show up and be clicked.
I had a discussion with Darren in Hipchat about this as I didn't see this as a bug per se, but I couldn't get the next_post_links/prev_post_links to play nicely in excluding an event category (my ideas was to attach an event category to the expired events and exclude it in the next_post_links function).
Darren considers this a bug as no links should point to an expired event except the direct URL.
```
I put together a gist that basically re-orders the next/previous post links for EE events so they sort by the last date time. It's not the perfect solution (maybe they want the next/previous post links to order by 1st date time if not expired) but for most folks this is probably sufficient.
https://gist.github.com/joshfeck/a8e28b9cfecd89308fbea4c152f546c9
| 1.0 | Filter expired events from next/previous post linkslinks - Started on Codebase where Dean wrote:
```
http://eventespresso.com/topic/removing-out-of-date-events-from-front-end-of-website/
Basically, if you are on a single event and the theme supports next and previous post links, expired events can still show up and be clicked.
I had a discussion with Darren in Hipchat about this as I didn't see this as a bug per se, but I couldn't get the next_post_links/prev_post_links to play nicely in excluding an event category (my ideas was to attach an event category to the expired events and exclude it in the next_post_links function).
Darren considers this a bug as no links should point to an expired event except the direct URL.
```
I put together a gist that basically re-orders the next/previous post links for EE events so they sort by the last date time. It's not the perfect solution (maybe they want the next/previous post links to order by 1st date time if not expired) but for most folks this is probably sufficient.
https://gist.github.com/joshfeck/a8e28b9cfecd89308fbea4c152f546c9
| non_priority | filter expired events from next previous post linkslinks started on codebase where dean wrote basically if you are on a single event and the theme supports next and previous post links expired events can still show up and be clicked i had a discussion with darren in hipchat about this as i didn t see this as a bug per se but i couldn t get the next post links prev post links to play nicely in excluding an event category my ideas was to attach an event category to the expired events and exclude it in the next post links function darren considers this a bug as no links should point to an expired event except the direct url i put together a gist that basically re orders the next previous post links for ee events so they sort by the last date time it s not the perfect solution maybe they want the next previous post links to order by date time if not expired but for most folks this is probably sufficient | 0 |
218,423 | 16,990,426,539 | IssuesEvent | 2021-06-30 19:38:52 | microsoft/BotFramework-Composer | https://api.github.com/repos/microsoft/BotFramework-Composer | closed | Add "Watch" as new tab to debugging pane | P0 Team: Authoring Team: Platform Testing & Debugging feature-request | ## Summary
Tracking state throughout a conversation in Web Chat (or Emulator) is possible, but difficult. The user has to find the "Bot State" log item, select it, and then dig into the inspector pane to find the property they are looking for. Elevating bot state to its own tab and allowing users to select the custom properties thy want to follow will
1. reduce the clicking needed to find a property
2. enable a persistent view into the properties of interest
## Features
### P0s
- [x] Add tab that filters the Web Chat logs to Bot State (follow same layout: logs on left, inspector in the right side)
- [x] Add filter for properties to watch **_- update: instead of filter, using a button to add a new row to the table that opens the property picker._**
- [x] Visual presentation of the user-defined properties only
- [ ] Abillty to autoexpand to the properties/scope selected in the filter
## P1
- [ ] ~~Ability to move to previous/next state~~ - **update: _removed per discussion june 9_. may reconsider after R14 based on user feedback.**
## P2
- [ ] ~~Ability to click on messages in Web Chat and select the state for that turn~~ **update: _removed. per discussion june 9, w are making an assumption that users are only interested in the latest state, not in the history of what's in memory.**
## Design mockups
(to come)
| 1.0 | Add "Watch" as new tab to debugging pane - ## Summary
Tracking state throughout a conversation in Web Chat (or Emulator) is possible, but difficult. The user has to find the "Bot State" log item, select it, and then dig into the inspector pane to find the property they are looking for. Elevating bot state to its own tab and allowing users to select the custom properties thy want to follow will
1. reduce the clicking needed to find a property
2. enable a persistent view into the properties of interest
## Features
### P0s
- [x] Add tab that filters the Web Chat logs to Bot State (follow same layout: logs on left, inspector in the right side)
- [x] Add filter for properties to watch **_- update: instead of filter, using a button to add a new row to the table that opens the property picker._**
- [x] Visual presentation of the user-defined properties only
- [ ] Abillty to autoexpand to the properties/scope selected in the filter
## P1
- [ ] ~~Ability to move to previous/next state~~ - **update: _removed per discussion june 9_. may reconsider after R14 based on user feedback.**
## P2
- [ ] ~~Ability to click on messages in Web Chat and select the state for that turn~~ **update: _removed. per discussion june 9, w are making an assumption that users are only interested in the latest state, not in the history of what's in memory.**
## Design mockups
(to come)
| non_priority | add watch as new tab to debugging pane summary tracking state throughout a conversation in web chat or emulator is possible but difficult the user has to find the bot state log item select it and then dig into the inspector pane to find the property they are looking for elevating bot state to its own tab and allowing users to select the custom properties thy want to follow will reduce the clicking needed to find a property enable a persistent view into the properties of interest features add tab that filters the web chat logs to bot state follow same layout logs on left inspector in the right side add filter for properties to watch update instead of filter using a button to add a new row to the table that opens the property picker visual presentation of the user defined properties only abillty to autoexpand to the properties scope selected in the filter ability to move to previous next state update removed per discussion june may reconsider after based on user feedback ability to click on messages in web chat and select the state for that turn update removed per discussion june w are making an assumption that users are only interested in the latest state not in the history of what s in memory design mockups to come | 0 |
23,307 | 11,866,614,547 | IssuesEvent | 2020-03-26 04:19:52 | cBioPortal/cbioportal | https://api.github.com/repos/cBioPortal/cbioportal | closed | Improve Session Service Configuration | enhancement session-service wontfix | - [ ] `session.service.url` doesn't work without trailing slash `session.service.url=the/url/to/session-service`. Prolly good to add this automatically if it's missing.
- [ ] Maybe do some check on startup whether session service actually works. It takes quite a lot of steps to figure out whether it's actually operational | 1.0 | Improve Session Service Configuration - - [ ] `session.service.url` doesn't work without trailing slash `session.service.url=the/url/to/session-service`. Prolly good to add this automatically if it's missing.
- [ ] Maybe do some check on startup whether session service actually works. It takes quite a lot of steps to figure out whether it's actually operational | non_priority | improve session service configuration session service url doesn t work without trailing slash session service url the url to session service prolly good to add this automatically if it s missing maybe do some check on startup whether session service actually works it takes quite a lot of steps to figure out whether it s actually operational | 0 |
117,390 | 9,933,714,737 | IssuesEvent | 2019-07-02 12:59:41 | xcat2/xcat-core | https://api.github.com/repos/xcat2/xcat-core | closed | [OpenBMC] When BMC IP is DHCP, NTP servers cannot be set, by design, block command. | component:openbmc status:pending test:testcase_requested type:usability | In this scenario, xCAT should prevent this from being set since it is user error and will cause admins to go down a path to try to understand why ntpservers cannot be specified.
Here's what happens:
```
[root@stratton01 ~]# rspconfig f5u14 ntpservers=10.6.29.1
[stratton01]: f5u14: BMC Setting NTPServers...
[stratton01]: f5u14: BMC NTP Servers: 10.6.29.1
[stratton01]: f5u14: Warning: time will not be synchronized until the host is powered off.
```
After few seconds, querying shows None:
```
root@stratton01 ~]# rspconfig f5u14 ntpservers
f5u14: BMC NTP Servers: None
```
This is because, by design, openbmc does not set this if IPSRC is DHCP.
```
[root@stratton01 ~]# rspconfig f5u14 ipsrc
f5u14: BMC IP Source: DHCP
```
So I think xCAT should block the set of this value if IPSRC is DHCP ... | 2.0 | [OpenBMC] When BMC IP is DHCP, NTP servers cannot be set, by design, block command. - In this scenario, xCAT should prevent this from being set since it is user error and will cause admins to go down a path to try to understand why ntpservers cannot be specified.
Here's what happens:
```
[root@stratton01 ~]# rspconfig f5u14 ntpservers=10.6.29.1
[stratton01]: f5u14: BMC Setting NTPServers...
[stratton01]: f5u14: BMC NTP Servers: 10.6.29.1
[stratton01]: f5u14: Warning: time will not be synchronized until the host is powered off.
```
After few seconds, querying shows None:
```
root@stratton01 ~]# rspconfig f5u14 ntpservers
f5u14: BMC NTP Servers: None
```
This is because, by design, openbmc does not set this if IPSRC is DHCP.
```
[root@stratton01 ~]# rspconfig f5u14 ipsrc
f5u14: BMC IP Source: DHCP
```
So I think xCAT should block the set of this value if IPSRC is DHCP ... | non_priority | when bmc ip is dhcp ntp servers cannot be set by design block command in this scenario xcat should prevent this from being set since it is user error and will cause admins to go down a path to try to understand why ntpservers cannot be specified here s what happens rspconfig ntpservers bmc setting ntpservers bmc ntp servers warning time will not be synchronized until the host is powered off after few seconds querying shows none root rspconfig ntpservers bmc ntp servers none this is because by design openbmc does not set this if ipsrc is dhcp rspconfig ipsrc bmc ip source dhcp so i think xcat should block the set of this value if ipsrc is dhcp | 0 |
73,867 | 9,739,156,657 | IssuesEvent | 2019-06-01 08:41:12 | hedgerh/hiring-with-records | https://api.github.com/repos/hedgerh/hiring-with-records | opened | How should companies be listed/organized? | discussion documentation help wanted | How should companies be organized on this project? Will brainstorm a bit tomorrow. | 1.0 | How should companies be listed/organized? - How should companies be organized on this project? Will brainstorm a bit tomorrow. | non_priority | how should companies be listed organized how should companies be organized on this project will brainstorm a bit tomorrow | 0 |
34,626 | 16,618,176,411 | IssuesEvent | 2021-06-02 19:41:45 | magento/pwa-studio | https://api.github.com/repos/magento/pwa-studio | closed | [feature]: Implement graphQL query caching at Service Worker Level. | Progress: done enhancement performance | **Is your feature request related to a problem? Please describe.**
Related to #1673
**Describe the solution you'd like**
As of today, we are not caching graphQL query results. This will be useful when it comes to offline experience and also quick instant responses when it comes to data that might not change often like `storeConfig`.
1. When a graphQL call is made, analyze if it is a GET or POST equivalent. If GET, use the network first approach and store the data before returning to the client for future use.
2. If the network is taking too long, return the data in the cache and update the cache when the network comes back with the data.
**Please let us know what packages this feature is in regards to:**
- [x] `venia-concept`
- [x] `venia-ui`
- [ ] `pwa-buildpack`
- [ ] `peregrine`
- [ ] `pwa-devdocs`
- [ ] `upward-js`
- [ ] `upward-spec`
| True | [feature]: Implement graphQL query caching at Service Worker Level. - **Is your feature request related to a problem? Please describe.**
Related to #1673
**Describe the solution you'd like**
As of today, we are not caching graphQL query results. This will be useful when it comes to offline experience and also quick instant responses when it comes to data that might not change often like `storeConfig`.
1. When a graphQL call is made, analyze if it is a GET or POST equivalent. If GET, use the network first approach and store the data before returning to the client for future use.
2. If the network is taking too long, return the data in the cache and update the cache when the network comes back with the data.
**Please let us know what packages this feature is in regards to:**
- [x] `venia-concept`
- [x] `venia-ui`
- [ ] `pwa-buildpack`
- [ ] `peregrine`
- [ ] `pwa-devdocs`
- [ ] `upward-js`
- [ ] `upward-spec`
| non_priority | implement graphql query caching at service worker level is your feature request related to a problem please describe related to describe the solution you d like as of today we are not caching graphql query results this will be useful when it comes to offline experience and also quick instant responses when it comes to data that might not change often like storeconfig when a graphql call is made analyze if it is a get or post equivalent if get use the network first approach and store the data before returning to the client for future use if the network is taking too long return the data in the cache and update the cache when the network comes back with the data please let us know what packages this feature is in regards to venia concept venia ui pwa buildpack peregrine pwa devdocs upward js upward spec | 0 |
101,862 | 16,529,727,173 | IssuesEvent | 2021-05-27 03:11:11 | alpersonalwebsite/postME | https://api.github.com/repos/alpersonalwebsite/postME | opened | CVE-2021-23337 (High) detected in lodash-4.17.11.tgz | security vulnerability | ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p>
<p>Path to dependency file: postME/frontend/package.json</p>
<p>Path to vulnerable library: postME/frontend/node_modules/lodash</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- jest-circus-26.6.0.tgz
- jest-snapshot-26.6.2.tgz
- types-7.4.4.tgz
- :x: **lodash-4.17.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/postME/commit/3d3c22bfea6f240b03d6274d257a6fdb2102f310">3d3c22bfea6f240b03d6274d257a6fdb2102f310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23337 (High) detected in lodash-4.17.11.tgz - ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p>
<p>Path to dependency file: postME/frontend/package.json</p>
<p>Path to vulnerable library: postME/frontend/node_modules/lodash</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- jest-circus-26.6.0.tgz
- jest-snapshot-26.6.2.tgz
- types-7.4.4.tgz
- :x: **lodash-4.17.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/alpersonalwebsite/postME/commit/3d3c22bfea6f240b03d6274d257a6fdb2102f310">3d3c22bfea6f240b03d6274d257a6fdb2102f310</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file postme frontend package json path to vulnerable library postme frontend node modules lodash dependency hierarchy react scripts tgz root library jest circus tgz jest snapshot tgz types tgz x lodash tgz vulnerable library found in head commit a href vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource | 0 |
42,892 | 11,098,167,233 | IssuesEvent | 2019-12-16 14:43:49 | NEZNAMY/TAB | https://api.github.com/repos/NEZNAMY/TAB | closed | Do not use marker tag for ArmorStands for 1.8.x clients | Check new build Problem Unlimited nametags Version specific | Due do a visual bug in 1.8.x and actual lack of marker tag in 1.8.0, it will be better to not use marker tag for 1.8.x clients and use the old way (full hitbox with packet redirecting).
Problem 1: 1.8.x clients see slight flicker when players sneak.
Problem 2: 1.8.0 clients see the armor stand 2 blocks higher. | 1.0 | Do not use marker tag for ArmorStands for 1.8.x clients - Due do a visual bug in 1.8.x and actual lack of marker tag in 1.8.0, it will be better to not use marker tag for 1.8.x clients and use the old way (full hitbox with packet redirecting).
Problem 1: 1.8.x clients see slight flicker when players sneak.
Problem 2: 1.8.0 clients see the armor stand 2 blocks higher. | non_priority | do not use marker tag for armorstands for x clients due do a visual bug in x and actual lack of marker tag in it will be better to not use marker tag for x clients and use the old way full hitbox with packet redirecting problem x clients see slight flicker when players sneak problem clients see the armor stand blocks higher | 0 |
1,400 | 15,830,652,178 | IssuesEvent | 2021-04-06 12:46:23 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | [QUIC] Update MsQuic GCHandles | area-System.Net.Quic enhancement tenet-reliability | Some of the classes in MsQuic implementation are allocating a `GCHandle` for `this` and is depending on `Dispose` or finalizer to deallocate `this`. If a user forgets to call `Dispose`, the GC will never collect the object due to the `GCHandle`. | True | [QUIC] Update MsQuic GCHandles - Some of the classes in MsQuic implementation are allocating a `GCHandle` for `this` and is depending on `Dispose` or finalizer to deallocate `this`. If a user forgets to call `Dispose`, the GC will never collect the object due to the `GCHandle`. | non_priority | update msquic gchandles some of the classes in msquic implementation are allocating a gchandle for this and is depending on dispose or finalizer to deallocate this if a user forgets to call dispose the gc will never collect the object due to the gchandle | 0 |
203,862 | 15,392,703,277 | IssuesEvent | 2021-03-03 15:54:58 | smartdevicelink/sdl_ios | https://api.github.com/repos/smartdevicelink/sdl_ios | opened | Fix Sometimes Failing CI Tests | testing | ### Bug Report
Timing issues sometimes cause CI tests to fail. | 1.0 | Fix Sometimes Failing CI Tests - ### Bug Report
Timing issues sometimes cause CI tests to fail. | non_priority | fix sometimes failing ci tests bug report timing issues sometimes cause ci tests to fail | 0 |
68,686 | 17,380,500,929 | IssuesEvent | 2021-07-31 15:59:06 | luvit/luvi | https://api.github.com/repos/luvit/luvi | closed | miss luv_ctx_t and luv_context in winsvc.c | ? build | static svc_baton* svc_create_baton(lua_State* L, const char* name, int main_ref, int cb_ref) {
luv_ctx_t* ctx = luv_context(L); **** error ****
...
}
what is luv_ctx_t and luv_context? | 1.0 | miss luv_ctx_t and luv_context in winsvc.c - static svc_baton* svc_create_baton(lua_State* L, const char* name, int main_ref, int cb_ref) {
luv_ctx_t* ctx = luv_context(L); **** error ****
...
}
what is luv_ctx_t and luv_context? | non_priority | miss luv ctx t and luv context in winsvc c static svc baton svc create baton lua state l const char name int main ref int cb ref luv ctx t ctx luv context l error what is luv ctx t and luv context | 0 |
324,642 | 24,010,716,447 | IssuesEvent | 2022-09-14 18:32:57 | Pradumnasaraf/open-source-with-pradumna | https://api.github.com/repos/Pradumnasaraf/open-source-with-pradumna | closed | [DOCS] Add process/guide - "How to archive a repo" | documentation good first issue EddieHub:good-first-issue how-to OSWP | ### Description
Hey! Contributor,
This issue is for adding the guide/process for **How to archive a repo**. The step needs to be well defined and include the Screenshot/Screen recording for better explanation. To get an overview of how the docs should look, check the available docs [`here`](https://github.com/Pradumnasaraf/open-source-with-pradumna/blob/main/pages/How-to/guide)
- Path to the dedicate Markdown for this issue - [`open-source-with-pradumna/pages/How-to/guide/archive-repo.md`](https://github.com/Pradumnasaraf/open-source-with-pradumna/blob/main/pages/How-to/guide/archive-repo.md)
---
If you have any suggestions feel free to Open an [Issue](https://github.com/Pradumnasaraf/open-source-with-pradumna/issues)
**Also if you need any kind of help, feel free to ping!**
| 1.0 | [DOCS] Add process/guide - "How to archive a repo" - ### Description
Hey! Contributor,
This issue is for adding the guide/process for **How to archive a repo**. The step needs to be well defined and include the Screenshot/Screen recording for better explanation. To get an overview of how the docs should look, check the available docs [`here`](https://github.com/Pradumnasaraf/open-source-with-pradumna/blob/main/pages/How-to/guide)
- Path to the dedicate Markdown for this issue - [`open-source-with-pradumna/pages/How-to/guide/archive-repo.md`](https://github.com/Pradumnasaraf/open-source-with-pradumna/blob/main/pages/How-to/guide/archive-repo.md)
---
If you have any suggestions feel free to Open an [Issue](https://github.com/Pradumnasaraf/open-source-with-pradumna/issues)
**Also if you need any kind of help, feel free to ping!**
| non_priority | add process guide how to archive a repo description hey contributor this issue is for adding the guide process for how to archive a repo the step needs to be well defined and include the screenshot screen recording for better explanation to get an overview of how the docs should look check the available docs path to the dedicate markdown for this issue if you have any suggestions feel free to open an also if you need any kind of help feel free to ping | 0 |
10,034 | 7,059,816,872 | IssuesEvent | 2018-01-05 04:26:50 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | tf.scatter_add causes error in loop | type:bug/performance | ### System information
- **Have I written custom code (as opposed to using a stock example script provided in TensorFlow)**: Yes, it's below.
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: MacOSX
- **TensorFlow installed from (source or binary)**: binary
- **TensorFlow version (use command below)**: v1.3.0-rc2-20-g0787eee 1.3.0
- **Python version**: 3.5
- **Bazel version (if compiling from source)**:
- **CUDA/cuDNN version**: no CUDA
- **GPU model and memory**: no GPU
- **Exact command to reproduce**:
### Describe the problem
I found very strange behavior of tf.scatter_add: I created a tf.while_loop that creates a Tensor wrapped inside a tf.Variable. If I don't add something to the Variable outside the loop, tensorflow causes an error telling me that the Variable is not mutable.
I asked the on StackOverflow and was told to create a bug report.
https://stackoverflow.com/questions/46935216/tf-scatter-add-causes-error-in-loop?noredirect=1#comment80914069_46935216
Uncommenting the commented line removes the error. But I don't think this is intended behavior.
### Source code / logs
import tensorflow as tf
m = 25
batch_num = 32
num_bus = 50
C = tf.zeros((m, batch_num, num_bus, m),tf.float64)
C = tf.Variable(C)
c = tf.ones((batch_num, num_bus, m), tf.float64)
#C = tf.scatter_add(C,0,c)
k = tf.constant(1)
stop_cond = lambda k,C: k<m
def construct_C(k, C):
upd_c = c+1
C = tf.scatter_add(C,k,upd_c)
return k+1,C
k,C = tf.while_loop(stop_cond,construct_C, (k,C))
sess = tf.Session()
sess.run(tf.global_variables_initializer())
C1 = sess.run(C)
| True | tf.scatter_add causes error in loop - ### System information
- **Have I written custom code (as opposed to using a stock example script provided in TensorFlow)**: Yes, it's below.
- **OS Platform and Distribution (e.g., Linux Ubuntu 16.04)**: MacOSX
- **TensorFlow installed from (source or binary)**: binary
- **TensorFlow version (use command below)**: v1.3.0-rc2-20-g0787eee 1.3.0
- **Python version**: 3.5
- **Bazel version (if compiling from source)**:
- **CUDA/cuDNN version**: no CUDA
- **GPU model and memory**: no GPU
- **Exact command to reproduce**:
### Describe the problem
I found very strange behavior of tf.scatter_add: I created a tf.while_loop that creates a Tensor wrapped inside a tf.Variable. If I don't add something to the Variable outside the loop, tensorflow causes an error telling me that the Variable is not mutable.
I asked the on StackOverflow and was told to create a bug report.
https://stackoverflow.com/questions/46935216/tf-scatter-add-causes-error-in-loop?noredirect=1#comment80914069_46935216
Uncommenting the commented line removes the error. But I don't think this is intended behavior.
### Source code / logs
import tensorflow as tf
m = 25
batch_num = 32
num_bus = 50
C = tf.zeros((m, batch_num, num_bus, m),tf.float64)
C = tf.Variable(C)
c = tf.ones((batch_num, num_bus, m), tf.float64)
#C = tf.scatter_add(C,0,c)
k = tf.constant(1)
stop_cond = lambda k,C: k<m
def construct_C(k, C):
upd_c = c+1
C = tf.scatter_add(C,k,upd_c)
return k+1,C
k,C = tf.while_loop(stop_cond,construct_C, (k,C))
sess = tf.Session()
sess.run(tf.global_variables_initializer())
C1 = sess.run(C)
| non_priority | tf scatter add causes error in loop system information have i written custom code as opposed to using a stock example script provided in tensorflow yes it s below os platform and distribution e g linux ubuntu macosx tensorflow installed from source or binary binary tensorflow version use command below python version bazel version if compiling from source cuda cudnn version no cuda gpu model and memory no gpu exact command to reproduce describe the problem i found very strange behavior of tf scatter add i created a tf while loop that creates a tensor wrapped inside a tf variable if i don t add something to the variable outside the loop tensorflow causes an error telling me that the variable is not mutable i asked the on stackoverflow and was told to create a bug report uncommenting the commented line removes the error but i don t think this is intended behavior source code logs import tensorflow as tf m batch num num bus c tf zeros m batch num num bus m tf c tf variable c c tf ones batch num num bus m tf c tf scatter add c c k tf constant stop cond lambda k c k m def construct c k c upd c c c tf scatter add c k upd c return k c k c tf while loop stop cond construct c k c sess tf session sess run tf global variables initializer sess run c | 0 |
183,079 | 21,691,361,696 | IssuesEvent | 2022-05-09 15:38:37 | turkdevops/vuex | https://api.github.com/repos/turkdevops/vuex | closed | CVE-2020-7720 (High) detected in node-forge-0.9.0.tgz - autoclosed | security vulnerability | ## CVE-2020-7720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- vuepress-1.8.0.tgz (Root Library)
- core-1.8.0.tgz
- webpack-dev-server-3.11.0.tgz
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/vuex/commit/65977d233acef63b783ff0a2a37000fc4eaaf247">65977d233acef63b783ff0a2a37000fc4eaaf247</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package node-forge before 0.10.0 is vulnerable to Prototype Pollution via the util.setPath function. Note: Version 0.10.0 is a breaking change removing the vulnerable functions.
<p>Publish Date: 2020-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720>CVE-2020-7720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/digitalbazaar/forge/blob/master/CHANGELOG.md">https://github.com/digitalbazaar/forge/blob/master/CHANGELOG.md</a></p>
<p>Release Date: 2020-09-01</p>
<p>Fix Resolution (node-forge): 0.10.0</p>
<p>Direct dependency fix Resolution (vuepress): 1.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7720 (High) detected in node-forge-0.9.0.tgz - autoclosed - ## CVE-2020-7720 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- vuepress-1.8.0.tgz (Root Library)
- core-1.8.0.tgz
- webpack-dev-server-3.11.0.tgz
- selfsigned-1.10.7.tgz
- :x: **node-forge-0.9.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/vuex/commit/65977d233acef63b783ff0a2a37000fc4eaaf247">65977d233acef63b783ff0a2a37000fc4eaaf247</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package node-forge before 0.10.0 is vulnerable to Prototype Pollution via the util.setPath function. Note: Version 0.10.0 is a breaking change removing the vulnerable functions.
<p>Publish Date: 2020-09-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720>CVE-2020-7720</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/digitalbazaar/forge/blob/master/CHANGELOG.md">https://github.com/digitalbazaar/forge/blob/master/CHANGELOG.md</a></p>
<p>Release Date: 2020-09-01</p>
<p>Fix Resolution (node-forge): 0.10.0</p>
<p>Direct dependency fix Resolution (vuepress): 1.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in node forge tgz autoclosed cve high severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules node forge package json dependency hierarchy vuepress tgz root library core tgz webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library found in head commit a href found in base branch dev vulnerability details the package node forge before is vulnerable to prototype pollution via the util setpath function note version is a breaking change removing the vulnerable functions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution vuepress step up your open source security game with whitesource | 0 |
58,071 | 11,837,967,860 | IssuesEvent | 2020-03-23 14:58:22 | microsoft/Azure-Kinect-Sensor-SDK | https://api.github.com/repos/microsoft/Azure-Kinect-Sensor-SDK | closed | playback memory leak help | Bug Code Sample Triage Needed | I`m having hard time figure out why i`m getting memory leak in
this function which run every frame.
before this function I create captures only once.
after the function i release the images ( every frame).
any help will be appreciated.
**Update :
I made mistake and placed the release in wrong place.**
resolved.
```
CPUMemoryTOP::bundleCaptures()
{
if (result == K4A_RESULT_SUCCEEDED)
{
//printf("%-32s %12s %12s %12s\n", "Source file", "COLOR", "DEPTH", "IR");
//printf("==========================================================================\n");
// Print 8 captures ( frame ) at the time in order of timestamp across all the recordings.
//
for (int frame = 0; frame < file_count ; ++frame)
{
uint64_t min_timestamp = (uint64_t)-1; // very big number trick
recording_t* min_file = NULL;
size_t min_index = -1;
// Find the lowest timestamp out of each of the current captures.
for (size_t i = 0; i < file_count; ++i)
{
if (files[i].capture != NULL)
{
uint64_t timestamp = first_capture_timestamp(files[i].capture);
if (timestamp < min_timestamp)
{
min_timestamp = timestamp;
min_file = &files[i];
min_index = i;
}
}
}
if (min_file == nullptr)
{
std::cout << "no capture available" << std::endl;
return;
}
BGRAImageK[min_index] = k4a_capture_get_color_image(min_file->capture);
depthImageK[min_index] = k4a_capture_get_depth_image(min_file->capture);
if (k4a_image_get_size(BGRAImageK[min_index]) != 0 && k4a_image_get_size(depthImageK[min_index]) != 0 ) // check if there is image, if not paint it black.
{
bgraIn[min_index] = k4a_image_get_buffer(BGRAImageK[min_index]);
depthIn[min_index] = (unsigned short*)k4a_image_get_buffer(depthImageK[min_index]);
}
else
{
bgraIn[min_index] = (unsigned char*)&bgraBlack;
depthIn[min_index] = (unsigned short*)&depthBlack;
}
//test
k4a_capture_release(min_file->capture);
min_file->capture = NULL;
// Advance the recording with the lowest current timestamp forward.
k4a_stream_result_t stream_result = k4a_playback_get_next_capture(min_file->handle, &min_file->capture);
if (stream_result == K4A_STREAM_RESULT_FAILED)
{
printf("ERROR: Failed to read next capture from file: %s\n", min_file->filename);
result = K4A_RESULT_FAILED;
break;
}
}
}
}
```
Update :
I made mistake and placed the release in wrong place.
```
for (uint32_t i = 0; i < file_count; ++i)
{
k4a_image_release(BGRAImageK[i]);
k4a_image_release(depthImageK[i]);
}
``` | 1.0 | playback memory leak help - I`m having hard time figure out why i`m getting memory leak in
this function which run every frame.
before this function I create captures only once.
after the function i release the images ( every frame).
any help will be appreciated.
**Update :
I made mistake and placed the release in wrong place.**
resolved.
```
CPUMemoryTOP::bundleCaptures()
{
if (result == K4A_RESULT_SUCCEEDED)
{
//printf("%-32s %12s %12s %12s\n", "Source file", "COLOR", "DEPTH", "IR");
//printf("==========================================================================\n");
// Print 8 captures ( frame ) at the time in order of timestamp across all the recordings.
//
for (int frame = 0; frame < file_count ; ++frame)
{
uint64_t min_timestamp = (uint64_t)-1; // very big number trick
recording_t* min_file = NULL;
size_t min_index = -1;
// Find the lowest timestamp out of each of the current captures.
for (size_t i = 0; i < file_count; ++i)
{
if (files[i].capture != NULL)
{
uint64_t timestamp = first_capture_timestamp(files[i].capture);
if (timestamp < min_timestamp)
{
min_timestamp = timestamp;
min_file = &files[i];
min_index = i;
}
}
}
if (min_file == nullptr)
{
std::cout << "no capture available" << std::endl;
return;
}
BGRAImageK[min_index] = k4a_capture_get_color_image(min_file->capture);
depthImageK[min_index] = k4a_capture_get_depth_image(min_file->capture);
if (k4a_image_get_size(BGRAImageK[min_index]) != 0 && k4a_image_get_size(depthImageK[min_index]) != 0 ) // check if there is image, if not paint it black.
{
bgraIn[min_index] = k4a_image_get_buffer(BGRAImageK[min_index]);
depthIn[min_index] = (unsigned short*)k4a_image_get_buffer(depthImageK[min_index]);
}
else
{
bgraIn[min_index] = (unsigned char*)&bgraBlack;
depthIn[min_index] = (unsigned short*)&depthBlack;
}
//test
k4a_capture_release(min_file->capture);
min_file->capture = NULL;
// Advance the recording with the lowest current timestamp forward.
k4a_stream_result_t stream_result = k4a_playback_get_next_capture(min_file->handle, &min_file->capture);
if (stream_result == K4A_STREAM_RESULT_FAILED)
{
printf("ERROR: Failed to read next capture from file: %s\n", min_file->filename);
result = K4A_RESULT_FAILED;
break;
}
}
}
}
```
Update :
I made mistake and placed the release in wrong place.
```
for (uint32_t i = 0; i < file_count; ++i)
{
k4a_image_release(BGRAImageK[i]);
k4a_image_release(depthImageK[i]);
}
``` | non_priority | playback memory leak help i m having hard time figure out why i m getting memory leak in this function which run every frame before this function i create captures only once after the function i release the images every frame any help will be appreciated update i made mistake and placed the release in wrong place resolved cpumemorytop bundlecaptures if result result succeeded printf n source file color depth ir printf n print captures frame at the time in order of timestamp across all the recordings for int frame frame file count frame t min timestamp t very big number trick recording t min file null size t min index find the lowest timestamp out of each of the current captures for size t i i file count i if files capture null t timestamp first capture timestamp files capture if timestamp min timestamp min timestamp timestamp min file files min index i if min file nullptr std cout no capture available std endl return bgraimagek capture get color image min file capture depthimagek capture get depth image min file capture if image get size bgraimagek image get size depthimagek check if there is image if not paint it black bgrain image get buffer bgraimagek depthin unsigned short image get buffer depthimagek else bgrain unsigned char bgrablack depthin unsigned short depthblack test capture release min file capture min file capture null advance the recording with the lowest current timestamp forward stream result t stream result playback get next capture min file handle min file capture if stream result stream result failed printf error failed to read next capture from file s n min file filename result result failed break update i made mistake and placed the release in wrong place for t i i file count i image release bgraimagek image release depthimagek | 0 |
427,758 | 29,834,031,329 | IssuesEvent | 2023-06-18 16:09:25 | JakeStanger/corn | https://api.github.com/repos/JakeStanger/corn | closed | crates.io pages missing readme | bug documentation | Moving to workspace with readme in top level now means crates don't have readme info. | 1.0 | crates.io pages missing readme - Moving to workspace with readme in top level now means crates don't have readme info. | non_priority | crates io pages missing readme moving to workspace with readme in top level now means crates don t have readme info | 0 |
172,615 | 27,306,569,549 | IssuesEvent | 2023-02-24 08:46:27 | OdyseeTeam/odysee-frontend | https://api.github.com/repos/OdyseeTeam/odysee-frontend | closed | Mini sidebar sometimes blank | design | Sometimes I get a blank mini sidebar, and requires several refresh and/or window resize to fix.
Hard to reproduce ... I think happens on specific window sizes.
<img width="659" alt="image" src="https://user-images.githubusercontent.com/64950861/155685644-63276477-f550-4ce5-b430-a853671c6df8.png">
| 1.0 | Mini sidebar sometimes blank - Sometimes I get a blank mini sidebar, and requires several refresh and/or window resize to fix.
Hard to reproduce ... I think happens on specific window sizes.
<img width="659" alt="image" src="https://user-images.githubusercontent.com/64950861/155685644-63276477-f550-4ce5-b430-a853671c6df8.png">
| non_priority | mini sidebar sometimes blank sometimes i get a blank mini sidebar and requires several refresh and or window resize to fix hard to reproduce i think happens on specific window sizes img width alt image src | 0 |
280,184 | 30,805,139,826 | IssuesEvent | 2023-08-01 06:32:01 | Satheesh575555/linux-4.1.15 | https://api.github.com/repos/Satheesh575555/linux-4.1.15 | reopened | CVE-2020-36386 (High) detected in linuxlinux-4.6 | Mend: dependency security vulnerability | ## CVE-2020-36386 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_event.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_event.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.8.1. net/bluetooth/hci_event.c has a slab out-of-bounds read in hci_extended_inquiry_result_evt, aka CID-51c19bf3d5cf.
<p>Publish Date: 2021-06-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36386>CVE-2020-36386</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-36386">https://www.linuxkernelcves.com/cves/CVE-2020-36386</a></p>
<p>Release Date: 2021-06-07</p>
<p>Fix Resolution: v4.4.233, v4.9.233, v4.14.194, v4.19.139, v5.4.58, v5.7.15, v5.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-36386 (High) detected in linuxlinux-4.6 - ## CVE-2020-36386 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_event.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/bluetooth/hci_event.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.8.1. net/bluetooth/hci_event.c has a slab out-of-bounds read in hci_extended_inquiry_result_evt, aka CID-51c19bf3d5cf.
<p>Publish Date: 2021-06-07
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-36386>CVE-2020-36386</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-36386">https://www.linuxkernelcves.com/cves/CVE-2020-36386</a></p>
<p>Release Date: 2021-06-07</p>
<p>Fix Resolution: v4.4.233, v4.9.233, v4.14.194, v4.19.139, v5.4.58, v5.7.15, v5.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files net bluetooth hci event c net bluetooth hci event c vulnerability details an issue was discovered in the linux kernel before net bluetooth hci event c has a slab out of bounds read in hci extended inquiry result evt aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
61,619 | 14,631,859,797 | IssuesEvent | 2020-12-23 20:53:28 | idonthaveafifaaddiction/initializr | https://api.github.com/repos/idonthaveafifaaddiction/initializr | opened | CVE-2020-1945 (Medium) detected in ant-1.10.2.jar | security vulnerability | ## CVE-2020-1945 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.10.2.jar</b></p></summary>
<p>master POM</p>
<p>Path to dependency file: initializr/initializr-web/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/org/apache/ant/ant/1.10.2/ant-1.10.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **ant-1.10.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/initializr/commit/2d68126d61075760436ec9c8a26ab19ba32ae5fa">2d68126d61075760436ec9c8a26ab19ba32ae5fa</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Ant 1.1 to 1.9.14 and 1.10.0 to 1.10.7 uses the default temporary directory identified by the Java system property java.io.tmpdir for several tasks and may thus leak sensitive information. The fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process.
<p>Publish Date: 2020-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1945>CVE-2020-1945</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://ant.apache.org/security.html">https://ant.apache.org/security.html</a></p>
<p>Release Date: 2020-05-14</p>
<p>Fix Resolution: org.apache.ant:ant-junitlauncher:1.10.8;org.apache.ant:ant:1.9.15,1.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.ant","packageName":"ant","packageVersion":"1.10.2","isTransitiveDependency":false,"dependencyTree":"org.apache.ant:ant:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.ant:ant-junitlauncher:1.10.8;org.apache.ant:ant:1.9.15,1.10.8"}],"vulnerabilityIdentifier":"CVE-2020-1945","vulnerabilityDetails":"Apache Ant 1.1 to 1.9.14 and 1.10.0 to 1.10.7 uses the default temporary directory identified by the Java system property java.io.tmpdir for several tasks and may thus leak sensitive information. The fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1945","cvss3Severity":"medium","cvss3Score":"6.3","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-1945 (Medium) detected in ant-1.10.2.jar - ## CVE-2020-1945 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ant-1.10.2.jar</b></p></summary>
<p>master POM</p>
<p>Path to dependency file: initializr/initializr-web/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/org/apache/ant/ant/1.10.2/ant-1.10.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **ant-1.10.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/idonthaveafifaaddiction/initializr/commit/2d68126d61075760436ec9c8a26ab19ba32ae5fa">2d68126d61075760436ec9c8a26ab19ba32ae5fa</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Ant 1.1 to 1.9.14 and 1.10.0 to 1.10.7 uses the default temporary directory identified by the Java system property java.io.tmpdir for several tasks and may thus leak sensitive information. The fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process.
<p>Publish Date: 2020-05-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1945>CVE-2020-1945</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://ant.apache.org/security.html">https://ant.apache.org/security.html</a></p>
<p>Release Date: 2020-05-14</p>
<p>Fix Resolution: org.apache.ant:ant-junitlauncher:1.10.8;org.apache.ant:ant:1.9.15,1.10.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.ant","packageName":"ant","packageVersion":"1.10.2","isTransitiveDependency":false,"dependencyTree":"org.apache.ant:ant:1.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.ant:ant-junitlauncher:1.10.8;org.apache.ant:ant:1.9.15,1.10.8"}],"vulnerabilityIdentifier":"CVE-2020-1945","vulnerabilityDetails":"Apache Ant 1.1 to 1.9.14 and 1.10.0 to 1.10.7 uses the default temporary directory identified by the Java system property java.io.tmpdir for several tasks and may thus leak sensitive information. The fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1945","cvss3Severity":"medium","cvss3Score":"6.3","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in ant jar cve medium severity vulnerability vulnerable library ant jar master pom path to dependency file initializr initializr web pom xml path to vulnerable library canner repository org apache ant ant ant jar dependency hierarchy x ant jar vulnerable library found in head commit a href found in base branch master vulnerability details apache ant to and to uses the default temporary directory identified by the java system property java io tmpdir for several tasks and may thus leak sensitive information the fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache ant ant junitlauncher org apache ant ant isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails apache ant to and to uses the default temporary directory identified by the java system property java io tmpdir for several tasks and may thus leak sensitive information the fixcrlf and replaceregexp tasks also copy files from the temporary directory back into the build tree allowing an attacker to inject modified source files into the build process vulnerabilityurl | 0 |
1,108 | 2,856,001,064 | IssuesEvent | 2015-06-02 13:00:32 | zaphoyd/websocketpp | https://api.github.com/repos/zaphoyd/websocketpp | closed | Two SSL records for one message. | iostream Transport Performance Issue | Hello,
I wrote a programm using Websocket++, what exchange mesages between client and server.
I didn't use any embeded transports, but use library in iostream manner, handling send and receive by myself.
Whenewer I call client.send(), I got two writes in my write handle, set by set_write_handler().
First for hrader, and second - for message body.
As a result, I got two SSL records in channel.
What produces unnecessary overhead. Two bytes of header results in may bytes of SSL record.
I, as a library user, would prefer to get in write handle complete message, with header, body and footer (if any), so I can wrap it into single SSL record. | True | Two SSL records for one message. - Hello,
I wrote a programm using Websocket++, what exchange mesages between client and server.
I didn't use any embeded transports, but use library in iostream manner, handling send and receive by myself.
Whenewer I call client.send(), I got two writes in my write handle, set by set_write_handler().
First for hrader, and second - for message body.
As a result, I got two SSL records in channel.
What produces unnecessary overhead. Two bytes of header results in may bytes of SSL record.
I, as a library user, would prefer to get in write handle complete message, with header, body and footer (if any), so I can wrap it into single SSL record. | non_priority | two ssl records for one message hello i wrote a programm using websocket what exchange mesages between client and server i didn t use any embeded transports but use library in iostream manner handling send and receive by myself whenewer i call client send i got two writes in my write handle set by set write handler first for hrader and second for message body as a result i got two ssl records in channel what produces unnecessary overhead two bytes of header results in may bytes of ssl record i as a library user would prefer to get in write handle complete message with header body and footer if any so i can wrap it into single ssl record | 0 |
176,794 | 14,596,431,187 | IssuesEvent | 2020-12-20 15:49:29 | paul-buerkner/brms | https://api.github.com/repos/paul-buerkner/brms | closed | Clarify sigma/sd prior defaults in the reference manual | documentation | At the moment, the [reference manual (version 2.14.4)](https://cran.r-project.org/web/packages/brms/brms.pdf) appears to give the incorrect information on the default setting for priors of `class = sigma`.
In the *Parameters for specific families* subsection of the `set_prior` section, we read: “By default, `sigma` has a half student-t prior that scales in the same way as the group-level standard deviations.” Further up in the *Standard deviations of group-level (’random’) effects* subsection, we read: “These parameters are restricted to be non-negative and, by default, have a half student-t prior with 3 degrees of freedom and a scale parameter that depends on the standard deviation of the response after applying the link function. Minimally, the scale parameter is 10.”
However, based on responses to the [Default student_t priors in brms](https://discourse.mc-stan.org/t/default-student-t-priors-in-brms/17197) thread on the Stan Forums, it appears **brms** currently uses the formula `max(2.5, mad(data$y))` to set the scale parameter, where `y` is the response variable in the `data`. Based on the model I'm currently working with, I can confirm **brms** seems to use the formula `max(2.5, mad(data$y))` to calculate the scale parameter for the default.
It'd be helpful if this was clarified in the reference manual.
| 1.0 | Clarify sigma/sd prior defaults in the reference manual - At the moment, the [reference manual (version 2.14.4)](https://cran.r-project.org/web/packages/brms/brms.pdf) appears to give the incorrect information on the default setting for priors of `class = sigma`.
In the *Parameters for specific families* subsection of the `set_prior` section, we read: “By default, `sigma` has a half student-t prior that scales in the same way as the group-level standard deviations.” Further up in the *Standard deviations of group-level (’random’) effects* subsection, we read: “These parameters are restricted to be non-negative and, by default, have a half student-t prior with 3 degrees of freedom and a scale parameter that depends on the standard deviation of the response after applying the link function. Minimally, the scale parameter is 10.”
However, based on responses to the [Default student_t priors in brms](https://discourse.mc-stan.org/t/default-student-t-priors-in-brms/17197) thread on the Stan Forums, it appears **brms** currently uses the formula `max(2.5, mad(data$y))` to set the scale parameter, where `y` is the response variable in the `data`. Based on the model I'm currently working with, I can confirm **brms** seems to use the formula `max(2.5, mad(data$y))` to calculate the scale parameter for the default.
It'd be helpful if this was clarified in the reference manual.
| non_priority | clarify sigma sd prior defaults in the reference manual at the moment the appears to give the incorrect information on the default setting for priors of class sigma in the parameters for specific families subsection of the set prior section we read “by default sigma has a half student t prior that scales in the same way as the group level standard deviations ” further up in the standard deviations of group level ’random’ effects subsection we read “these parameters are restricted to be non negative and by default have a half student t prior with degrees of freedom and a scale parameter that depends on the standard deviation of the response after applying the link function minimally the scale parameter is ” however based on responses to the thread on the stan forums it appears brms currently uses the formula max mad data y to set the scale parameter where y is the response variable in the data based on the model i m currently working with i can confirm brms seems to use the formula max mad data y to calculate the scale parameter for the default it d be helpful if this was clarified in the reference manual | 0 |
2,214 | 7,478,272,506 | IssuesEvent | 2018-04-04 11:03:36 | fabric8-services/fabric8-wit | https://api.github.com/repos/fabric8-services/fabric8-wit | closed | Problems with the /search endpoint and search-by-name | :fire: tech-dept :grey_exclamation: important :house: architecture :mag: search :question: question :recycle: refactoring kind/enhancement | ## Current situation
We have a generic search endpoint at `/search`.
You can pass essentially any query to that endpoint. It doesn't need to contain a space ID.
Currently we don't support search by iteration or area name but it shall come at some point.
## Problem
An area name or iteration name isn't unique. That is why we need a space ID in order to translate from any name (e.g. iteration or area) to an ID on the backend.
Getting the relevant space ID out of a query can be very complex due to all the logical nesting that is possible within a search query.
## Possible solutions
### Pass space ID as Query Option
If we dedicate a field for the space ID inside the query options we have an easy time on the backend because we can easily lookup thinks in the query option. There's no logic there but just key-value pairs. This is my favored solution to this problem.
As a consequence we must raise an error every time somebody tries to query by name of some entity but didn't provide a space ID.
### Create a /spaces/UUID/search endpoint
If we create a search endpoint under the spaces endpoint we would be given the space ID as part of the controller context in GOA. I must say that I dislike this idea because we don't have access to the controller in the search package. And to pass the space ID in to the search package would require a placeholder anyway. That placeholder could easily be the aforementioned space ID key in the query options.
I would really appreciate feedback on this topic or any of the proposed solution | 1.0 | Problems with the /search endpoint and search-by-name - ## Current situation
We have a generic search endpoint at `/search`.
You can pass essentially any query to that endpoint. It doesn't need to contain a space ID.
Currently we don't support search by iteration or area name but it shall come at some point.
## Problem
An area name or iteration name isn't unique. That is why we need a space ID in order to translate from any name (e.g. iteration or area) to an ID on the backend.
Getting the relevant space ID out of a query can be very complex due to all the logical nesting that is possible within a search query.
## Possible solutions
### Pass space ID as Query Option
If we dedicate a field for the space ID inside the query options we have an easy time on the backend because we can easily lookup thinks in the query option. There's no logic there but just key-value pairs. This is my favored solution to this problem.
As a consequence we must raise an error every time somebody tries to query by name of some entity but didn't provide a space ID.
### Create a /spaces/UUID/search endpoint
If we create a search endpoint under the spaces endpoint we would be given the space ID as part of the controller context in GOA. I must say that I dislike this idea because we don't have access to the controller in the search package. And to pass the space ID in to the search package would require a placeholder anyway. That placeholder could easily be the aforementioned space ID key in the query options.
I would really appreciate feedback on this topic or any of the proposed solution | non_priority | problems with the search endpoint and search by name current situation we have a generic search endpoint at search you can pass essentially any query to that endpoint it doesn t need to contain a space id currently we don t support search by iteration or area name but it shall come at some point problem an area name or iteration name isn t unique that is why we need a space id in order to translate from any name e g iteration or area to an id on the backend getting the relevant space id out of a query can be very complex due to all the logical nesting that is possible within a search query possible solutions pass space id as query option if we dedicate a field for the space id inside the query options we have an easy time on the backend because we can easily lookup thinks in the query option there s no logic there but just key value pairs this is my favored solution to this problem as a consequence we must raise an error every time somebody tries to query by name of some entity but didn t provide a space id create a spaces uuid search endpoint if we create a search endpoint under the spaces endpoint we would be given the space id as part of the controller context in goa i must say that i dislike this idea because we don t have access to the controller in the search package and to pass the space id in to the search package would require a placeholder anyway that placeholder could easily be the aforementioned space id key in the query options i would really appreciate feedback on this topic or any of the proposed solution | 0 |
102,398 | 12,794,770,518 | IssuesEvent | 2020-07-02 07:34:37 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | This is the best ms product ever | *as-designed | <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Also please test using the latest insiders build to make sure your issue has not already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- Use Help > Report Issue to prefill these. -->
- VSCode Version:
- OS Version:
Steps to Reproduce:
1.
2.
<!-- Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes/No
| 1.0 | This is the best ms product ever - <!-- ⚠️⚠️ Do Not Delete This! bug_report_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Also please test using the latest insiders build to make sure your issue has not already been fixed: https://code.visualstudio.com/insiders/ -->
<!-- Use Help > Report Issue to prefill these. -->
- VSCode Version:
- OS Version:
Steps to Reproduce:
1.
2.
<!-- Launch with `code --disable-extensions` to check. -->
Does this issue occur when all extensions are disabled?: Yes/No
| non_priority | this is the best ms product ever report issue to prefill these vscode version os version steps to reproduce does this issue occur when all extensions are disabled yes no | 0 |
159,006 | 20,036,582,816 | IssuesEvent | 2022-02-02 12:34:22 | kapseliboi/sttm-web | https://api.github.com/repos/kapseliboi/sttm-web | opened | CVE-2020-7733 (High) detected in ua-parser-js-0.7.21.tgz | security vulnerability | ## CVE-2020-7733 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.21.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.21.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.21.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- react-smartbanner-5.1.4.tgz (Root Library)
- :x: **ua-parser-js-0.7.21.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/sttm-web/commit/7afd13366ed3b25b964f044b38a5a0595b7d4e82">7afd13366ed3b25b964f044b38a5a0595b7d4e82</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA.
<p>Publish Date: 2020-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7733>CVE-2020-7733</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p>
<p>Release Date: 2020-09-16</p>
<p>Fix Resolution: 0.7.22</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7733 (High) detected in ua-parser-js-0.7.21.tgz - ## CVE-2020-7733 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.21.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.21.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.21.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- react-smartbanner-5.1.4.tgz (Root Library)
- :x: **ua-parser-js-0.7.21.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/sttm-web/commit/7afd13366ed3b25b964f044b38a5a0595b7d4e82">7afd13366ed3b25b964f044b38a5a0595b7d4e82</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA.
<p>Publish Date: 2020-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7733>CVE-2020-7733</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p>
<p>Release Date: 2020-09-16</p>
<p>Fix Resolution: 0.7.22</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in ua parser js tgz cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file package json path to vulnerable library node modules ua parser js package json dependency hierarchy react smartbanner tgz root library x ua parser js tgz vulnerable library found in head commit a href found in base branch dev vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos via the regex for redmi phones and mi pad tablets ua publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
128,720 | 18,070,099,598 | IssuesEvent | 2021-09-21 01:11:12 | RG4421/node-clair-client | https://api.github.com/repos/RG4421/node-clair-client | opened | CVE-2021-3807 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2021-3807 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-4.1.0.tgz</b>, <b>ansi-regex-2.1.1.tgz</b>, <b>ansi-regex-3.0.0.tgz</b></p></summary>
<p>
<details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p>
<p>Path to dependency file: node-clair-client/package.json</p>
<p>Path to vulnerable library: node-clair-client/node_modules/yargs/node_modules/ansi-regex/package.json,node-clair-client/node_modules/table/node_modules/ansi-regex/package.json,node-clair-client/node_modules/inquirer/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- eslint-5.16.0.tgz (Root Library)
- inquirer-6.3.1.tgz
- strip-ansi-5.2.0.tgz
- :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>ansi-regex-2.1.1.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.1.1.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.1.1.tgz</a></p>
<p>Path to dependency file: node-clair-client/package.json</p>
<p>Path to vulnerable library: node-clair-client/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- mocha-6.1.4.tgz (Root Library)
- yargs-13.2.2.tgz
- cliui-4.1.0.tgz
- wrap-ansi-2.1.0.tgz
- strip-ansi-3.0.1.tgz
- :x: **ansi-regex-2.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>ansi-regex-3.0.0.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p>
<p>Path to dependency file: node-clair-client/package.json</p>
<p>Path to vulnerable library: node-clair-client/node_modules/strip-ansi/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- eslint-5.16.0.tgz (Root Library)
- strip-ansi-4.0.0.tgz
- :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ansi-regex is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"4.1.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"eslint:5.16.0;inquirer:6.3.1;strip-ansi:5.2.0;ansi-regex:4.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"2.1.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"mocha:6.1.4;yargs:13.2.2;cliui:4.1.0;wrap-ansi:2.1.0;strip-ansi:3.0.1;ansi-regex:2.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"3.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"eslint:5.16.0;strip-ansi:4.0.0;ansi-regex:3.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3807","vulnerabilityDetails":"ansi-regex is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-3807 (Medium) detected in multiple libraries - ## CVE-2021-3807 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-4.1.0.tgz</b>, <b>ansi-regex-2.1.1.tgz</b>, <b>ansi-regex-3.0.0.tgz</b></p></summary>
<p>
<details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p>
<p>Path to dependency file: node-clair-client/package.json</p>
<p>Path to vulnerable library: node-clair-client/node_modules/yargs/node_modules/ansi-regex/package.json,node-clair-client/node_modules/table/node_modules/ansi-regex/package.json,node-clair-client/node_modules/inquirer/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- eslint-5.16.0.tgz (Root Library)
- inquirer-6.3.1.tgz
- strip-ansi-5.2.0.tgz
- :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>ansi-regex-2.1.1.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.1.1.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-2.1.1.tgz</a></p>
<p>Path to dependency file: node-clair-client/package.json</p>
<p>Path to vulnerable library: node-clair-client/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- mocha-6.1.4.tgz (Root Library)
- yargs-13.2.2.tgz
- cliui-4.1.0.tgz
- wrap-ansi-2.1.0.tgz
- strip-ansi-3.0.1.tgz
- :x: **ansi-regex-2.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>ansi-regex-3.0.0.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p>
<p>Path to dependency file: node-clair-client/package.json</p>
<p>Path to vulnerable library: node-clair-client/node_modules/strip-ansi/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- eslint-5.16.0.tgz (Root Library)
- strip-ansi-4.0.0.tgz
- :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ansi-regex is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"4.1.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"eslint:5.16.0;inquirer:6.3.1;strip-ansi:5.2.0;ansi-regex:4.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"2.1.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"mocha:6.1.4;yargs:13.2.2;cliui:4.1.0;wrap-ansi:2.1.0;strip-ansi:3.0.1;ansi-regex:2.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"},{"packageType":"javascript/Node.js","packageName":"ansi-regex","packageVersion":"3.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"eslint:5.16.0;strip-ansi:4.0.0;ansi-regex:3.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"ansi-regex - 5.0.1,6.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-3807","vulnerabilityDetails":"ansi-regex is vulnerable to Inefficient Regular Expression Complexity","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries ansi regex tgz ansi regex tgz ansi regex tgz ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file node clair client package json path to vulnerable library node clair client node modules yargs node modules ansi regex package json node clair client node modules table node modules ansi regex package json node clair client node modules inquirer node modules ansi regex package json dependency hierarchy eslint tgz root library inquirer tgz strip ansi tgz x ansi regex tgz vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file node clair client package json path to vulnerable library node clair client node modules ansi regex package json dependency hierarchy mocha tgz root library yargs tgz cliui tgz wrap ansi tgz strip ansi tgz x ansi regex tgz vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file node clair client package json path to vulnerable library node clair client node modules strip ansi node modules ansi regex package json dependency hierarchy eslint tgz root library strip ansi tgz x ansi regex tgz vulnerable library found in base branch master vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree eslint inquirer strip ansi ansi regex isminimumfixversionavailable true minimumfixversion ansi regex packagetype javascript node js packagename ansi regex packageversion packagefilepaths istransitivedependency true dependencytree mocha yargs cliui wrap ansi strip ansi ansi regex isminimumfixversionavailable true minimumfixversion ansi regex packagetype javascript node js packagename ansi regex packageversion packagefilepaths istransitivedependency true dependencytree eslint strip ansi ansi regex isminimumfixversionavailable true minimumfixversion ansi regex basebranches vulnerabilityidentifier cve vulnerabilitydetails ansi regex is vulnerable to inefficient regular expression complexity vulnerabilityurl | 0 |
749 | 2,541,283,833 | IssuesEvent | 2015-01-28 06:52:38 | http4s/http4s | https://api.github.com/repos/http4s/http4s | closed | Need a template for getting started | documentation enhancement freezer | I'm willing to entertain arguments why it should be activator. But there should be something to get people up and running quickly. | 1.0 | Need a template for getting started - I'm willing to entertain arguments why it should be activator. But there should be something to get people up and running quickly. | non_priority | need a template for getting started i m willing to entertain arguments why it should be activator but there should be something to get people up and running quickly | 0 |
2,738 | 3,619,122,865 | IssuesEvent | 2016-02-08 14:53:09 | mozilla/pdf.js | https://api.github.com/repos/mozilla/pdf.js | opened | Major performance issues for a simple PDF file | 2-performance | I have created a simple PDF file using Scribus 1.5.0svn. Notice that the file size is large for a two-page PDF file, so I can only suspect that Scribus is doing something really unefficient when exporting the PDF file. Nevertheless, the PDF file below renders instantly with both Adobe Acrobat Reader DC and Foxit Reader (within 0.5 seconds), however PDF.js takes *27 seconds* to render the first page of this file. I have no idea why PDF.js is taking such an excessive amount of time, but there must be something we can do to optimize this, given that other viewers do not have any problems with this file.
Below is the PDF file. I made this myself, so anyone is free to use this as a test case in a PR that addresses this issue:
[test.pdf](https://github.com/mozilla/pdf.js/files/121712/test.pdf)
| True | Major performance issues for a simple PDF file - I have created a simple PDF file using Scribus 1.5.0svn. Notice that the file size is large for a two-page PDF file, so I can only suspect that Scribus is doing something really unefficient when exporting the PDF file. Nevertheless, the PDF file below renders instantly with both Adobe Acrobat Reader DC and Foxit Reader (within 0.5 seconds), however PDF.js takes *27 seconds* to render the first page of this file. I have no idea why PDF.js is taking such an excessive amount of time, but there must be something we can do to optimize this, given that other viewers do not have any problems with this file.
Below is the PDF file. I made this myself, so anyone is free to use this as a test case in a PR that addresses this issue:
[test.pdf](https://github.com/mozilla/pdf.js/files/121712/test.pdf)
| non_priority | major performance issues for a simple pdf file i have created a simple pdf file using scribus notice that the file size is large for a two page pdf file so i can only suspect that scribus is doing something really unefficient when exporting the pdf file nevertheless the pdf file below renders instantly with both adobe acrobat reader dc and foxit reader within seconds however pdf js takes seconds to render the first page of this file i have no idea why pdf js is taking such an excessive amount of time but there must be something we can do to optimize this given that other viewers do not have any problems with this file below is the pdf file i made this myself so anyone is free to use this as a test case in a pr that addresses this issue | 0 |
135,400 | 12,683,509,922 | IssuesEvent | 2020-06-19 19:56:11 | AllYarnsAreBeautiful/ayab-desktop | https://api.github.com/repos/AllYarnsAreBeautiful/ayab-desktop | closed | Include Help | documentation | Check if and how to add a Help menu to the software, to help users getting started and provide support for common issues.
Questions:
* Content:
* Explanation of GUI elements
* How to configure
* How to knit (using the different modes)
* FAQ
* Languages?
* English (for now)
* Technology
* Output: HTML tomake documentation available online (ayab-knitting.com, readthedocs.org) and ship an offline version with each release.
* Tool
* https://www.mkdocs.org/
* Sphinx
* gitbook.com
* Separate repository for documentation | 1.0 | Include Help - Check if and how to add a Help menu to the software, to help users getting started and provide support for common issues.
Questions:
* Content:
* Explanation of GUI elements
* How to configure
* How to knit (using the different modes)
* FAQ
* Languages?
* English (for now)
* Technology
* Output: HTML tomake documentation available online (ayab-knitting.com, readthedocs.org) and ship an offline version with each release.
* Tool
* https://www.mkdocs.org/
* Sphinx
* gitbook.com
* Separate repository for documentation | non_priority | include help check if and how to add a help menu to the software to help users getting started and provide support for common issues questions content explanation of gui elements how to configure how to knit using the different modes faq languages english for now technology output html tomake documentation available online ayab knitting com readthedocs org and ship an offline version with each release tool sphinx gitbook com separate repository for documentation | 0 |
334,536 | 24,423,161,924 | IssuesEvent | 2022-10-05 22:38:38 | jonathanloganmoran/ND0013-Self-Driving-Car-Engineer | https://api.github.com/repos/jonathanloganmoran/ND0013-Self-Driving-Car-Engineer | closed | Accessing Waymo OD attribute data in Exercise 1.1.3 does not work | bug documentation | The access chains given in Sect. 1.1 of [Exercise 1.1.3](https://github.com/jonathanloganmoran/ND0013-Self-Driving-Car-Engineer/blob/main/1-Object-Detection-in-Urban-Environments/Exercises/1-1-3-Creating-TF-Records/2022-08-03-Creating-TF-Records.ipynb) is not correct for Waymo OD v1.2.
#### More info
* `frame.camera_labels` retrieves the bounding boxes for all cameras/classes;
* Each camera can be accessed individually with e.g., `frame.camera_labels[0]` (here that pulls all data for `FRONT` camera);
* Note that the `CameraName` enum definitions from [`dataset.proto`](https://github.com/waymo-research/waymo-open-dataset/blob/master/waymo_open_dataset/dataset_proto#L48) aren't entirely correct (no `UNKNOWN` label, retrieving by index starts at 0).
* `frame.camera_labels[0].labels` retrieves a list of all labels for camera `FRONT`, indexing `labels` only indexes the list;
* Indexing `.labels[..]` with the `Type` enum will not produce desired result mentioned;
* e.g., `frame.camera_labels[0].labels[1]` would "get all `FRONT` camera boxes of the `TYPE_VEHICLE` class (not true, only returns the first label in the `labels` list. | 1.0 | Accessing Waymo OD attribute data in Exercise 1.1.3 does not work - The access chains given in Sect. 1.1 of [Exercise 1.1.3](https://github.com/jonathanloganmoran/ND0013-Self-Driving-Car-Engineer/blob/main/1-Object-Detection-in-Urban-Environments/Exercises/1-1-3-Creating-TF-Records/2022-08-03-Creating-TF-Records.ipynb) is not correct for Waymo OD v1.2.
#### More info
* `frame.camera_labels` retrieves the bounding boxes for all cameras/classes;
* Each camera can be accessed individually with e.g., `frame.camera_labels[0]` (here that pulls all data for `FRONT` camera);
* Note that the `CameraName` enum definitions from [`dataset.proto`](https://github.com/waymo-research/waymo-open-dataset/blob/master/waymo_open_dataset/dataset_proto#L48) aren't entirely correct (no `UNKNOWN` label, retrieving by index starts at 0).
* `frame.camera_labels[0].labels` retrieves a list of all labels for camera `FRONT`, indexing `labels` only indexes the list;
* Indexing `.labels[..]` with the `Type` enum will not produce desired result mentioned;
* e.g., `frame.camera_labels[0].labels[1]` would "get all `FRONT` camera boxes of the `TYPE_VEHICLE` class (not true, only returns the first label in the `labels` list. | non_priority | accessing waymo od attribute data in exercise does not work the access chains given in sect of is not correct for waymo od more info frame camera labels retrieves the bounding boxes for all cameras classes each camera can be accessed individually with e g frame camera labels here that pulls all data for front camera note that the cameraname enum definitions from aren t entirely correct no unknown label retrieving by index starts at frame camera labels labels retrieves a list of all labels for camera front indexing labels only indexes the list indexing labels with the type enum will not produce desired result mentioned e g frame camera labels labels would get all front camera boxes of the type vehicle class not true only returns the first label in the labels list | 0 |
69,110 | 17,574,395,266 | IssuesEvent | 2021-08-15 10:14:53 | badsaarow/quasar2-storybook-boilerplate | https://api.github.com/repos/badsaarow/quasar2-storybook-boilerplate | opened | CircleCI build error | build | * The job has no executor type specified. The job should have one of the following keys specified: "machine", "docker", "macos" | 1.0 | CircleCI build error - * The job has no executor type specified. The job should have one of the following keys specified: "machine", "docker", "macos" | non_priority | circleci build error the job has no executor type specified the job should have one of the following keys specified machine docker macos | 0 |
273,713 | 23,780,143,372 | IssuesEvent | 2022-09-02 03:12:39 | openmsupply/conforma-web-app | https://api.github.com/repos/openmsupply/conforma-web-app | opened | Loading stuck on edit user details submission | Bugs during tests | After submitting edit user details form, submission fails and form is stuck on loading.
No console logs. | 1.0 | Loading stuck on edit user details submission - After submitting edit user details form, submission fails and form is stuck on loading.
No console logs. | non_priority | loading stuck on edit user details submission after submitting edit user details form submission fails and form is stuck on loading no console logs | 0 |
185,604 | 15,026,866,185 | IssuesEvent | 2021-02-01 23:30:27 | openshift/okd | https://api.github.com/repos/openshift/okd | closed | OKD4 oVirt IPI doc refers to sudo; causes openshift-install FATAL error | kind/documentation platform/ovirt | **Describe the bug**
OKD4 oVirt IPI doc refers to sudo; causes openshift-install FATAL error
**Version**
4.5.0
**How reproducible**
every time. Removing sudo lets installer injest SSH and create $HOME/.ovirt directory & ovirt-config.yaml
**Log bundle**
```
sudo ./openshift-install create cluster --dir=wont-work --log-level=info
[sudo] password for jupittma:
? Platform ovirt
FATAL failed to fetch Metadata: failed to fetch dependency of "Metadata": failed to fetch dependency of "Cluster ID": failed to fetch dependency of "Install Config": failed to fetch dependency of "Base Domain": failed to generate asset "Platform": Error during SSO authentication access_denied : Cannot authenticate user 'admin@internal': Unable to log in. Verify your login information or contact the system administrator..
[jupittma@jupittma live-okd4-install]$ ./openshift-install create cluster --dir=will-work --log-level=info
? SSH Public Key /home/jupittma/.ssh/okd4_rsa.pub
? Platform ovirt
? oVirt cluster HomeLab_Cluster
? oVirt storage domain [Use arrows to move, enter to select, type to filter, ? for more help]
> hosted_storage
FATAL failed to fetch Metadata: failed to fetch dependency of "Metadata": failed to fetch dependency of "Cluster ID": failed to fetch dependency of "Install Config": failed to fetch dependency of "Base Domain": failed to generate asset "Platform": interrupt
```
| 1.0 | OKD4 oVirt IPI doc refers to sudo; causes openshift-install FATAL error - **Describe the bug**
OKD4 oVirt IPI doc refers to sudo; causes openshift-install FATAL error
**Version**
4.5.0
**How reproducible**
every time. Removing sudo lets installer injest SSH and create $HOME/.ovirt directory & ovirt-config.yaml
**Log bundle**
```
sudo ./openshift-install create cluster --dir=wont-work --log-level=info
[sudo] password for jupittma:
? Platform ovirt
FATAL failed to fetch Metadata: failed to fetch dependency of "Metadata": failed to fetch dependency of "Cluster ID": failed to fetch dependency of "Install Config": failed to fetch dependency of "Base Domain": failed to generate asset "Platform": Error during SSO authentication access_denied : Cannot authenticate user 'admin@internal': Unable to log in. Verify your login information or contact the system administrator..
[jupittma@jupittma live-okd4-install]$ ./openshift-install create cluster --dir=will-work --log-level=info
? SSH Public Key /home/jupittma/.ssh/okd4_rsa.pub
? Platform ovirt
? oVirt cluster HomeLab_Cluster
? oVirt storage domain [Use arrows to move, enter to select, type to filter, ? for more help]
> hosted_storage
FATAL failed to fetch Metadata: failed to fetch dependency of "Metadata": failed to fetch dependency of "Cluster ID": failed to fetch dependency of "Install Config": failed to fetch dependency of "Base Domain": failed to generate asset "Platform": interrupt
```
| non_priority | ovirt ipi doc refers to sudo causes openshift install fatal error describe the bug ovirt ipi doc refers to sudo causes openshift install fatal error version how reproducible every time removing sudo lets installer injest ssh and create home ovirt directory ovirt config yaml log bundle sudo openshift install create cluster dir wont work log level info password for jupittma platform ovirt fatal failed to fetch metadata failed to fetch dependency of metadata failed to fetch dependency of cluster id failed to fetch dependency of install config failed to fetch dependency of base domain failed to generate asset platform error during sso authentication access denied cannot authenticate user admin internal unable to log in verify your login information or contact the system administrator openshift install create cluster dir will work log level info ssh public key home jupittma ssh rsa pub platform ovirt ovirt cluster homelab cluster ovirt storage domain hosted storage fatal failed to fetch metadata failed to fetch dependency of metadata failed to fetch dependency of cluster id failed to fetch dependency of install config failed to fetch dependency of base domain failed to generate asset platform interrupt | 0 |
6,604 | 5,542,495,248 | IssuesEvent | 2017-03-22 15:07:52 | hadley/dplyr | https://api.github.com/repos/hadley/dplyr | closed | Order-of-magnitude performance drop, possibly due to rlang | data frame performance | @lionel-: I see a substantial performance drop for several verbs somewhere around the introduction of rlang, but I can't measure intermediate versions because we didn't think about pinning the rlang revision back then, and older code doesn't work with current rlang. The benchmark code is here: https://github.com/krlmlr/dplyr.benchmark/blob/a27d672a19b80a168a0bffc626d8050a41ae18e2/tests/testthat/code/microbenchmark.R
To rerun the benchmark to detect which particular revision introduced the slowdown, I could use a list of rlang revisions for each dplyr revision between f39db50 and 4c5a978. Or we could just profile and see where the time is spent.
One of the things that have changed is the use of `overscope_eval()`. Could this be the reason? Do we really need to build the overscope for each call?





| True | Order-of-magnitude performance drop, possibly due to rlang - @lionel-: I see a substantial performance drop for several verbs somewhere around the introduction of rlang, but I can't measure intermediate versions because we didn't think about pinning the rlang revision back then, and older code doesn't work with current rlang. The benchmark code is here: https://github.com/krlmlr/dplyr.benchmark/blob/a27d672a19b80a168a0bffc626d8050a41ae18e2/tests/testthat/code/microbenchmark.R
To rerun the benchmark to detect which particular revision introduced the slowdown, I could use a list of rlang revisions for each dplyr revision between f39db50 and 4c5a978. Or we could just profile and see where the time is spent.
One of the things that have changed is the use of `overscope_eval()`. Could this be the reason? Do we really need to build the overscope for each call?





| non_priority | order of magnitude performance drop possibly due to rlang lionel i see a substantial performance drop for several verbs somewhere around the introduction of rlang but i can t measure intermediate versions because we didn t think about pinning the rlang revision back then and older code doesn t work with current rlang the benchmark code is here to rerun the benchmark to detect which particular revision introduced the slowdown i could use a list of rlang revisions for each dplyr revision between and or we could just profile and see where the time is spent one of the things that have changed is the use of overscope eval could this be the reason do we really need to build the overscope for each call | 0 |
287,238 | 24,817,714,362 | IssuesEvent | 2022-10-25 14:16:36 | netlify/build | https://api.github.com/repos/netlify/build | closed | Improve testing of network requests | test type: chore area: api area: testing stale | Netlify Build performs few network requests to:
- Bitballoon API, specifically the `getSite`, `cancelDeploy` and `createPluginRun` endpoints
- Telemetry
In integration tests, we currently [mock those](https://github.com/netlify/build/blob/836b84a30e951942394f45922c59b6b6d7c6d561/packages/build/tests/helpers/server.js#L7). This makes the test faster and more resilient to network failures.
However, this is also based on assumptions on the response of those servers. For example, if the network response was to change shape, or we made the initial assumption wrong, those tests would not catch this.
This created one instance of a production bug, where we assumed that a specific property of the `getSite` endpoint was `undefined` instead of `null` when empty.
We should run some tests against the real servers. | 2.0 | Improve testing of network requests - Netlify Build performs few network requests to:
- Bitballoon API, specifically the `getSite`, `cancelDeploy` and `createPluginRun` endpoints
- Telemetry
In integration tests, we currently [mock those](https://github.com/netlify/build/blob/836b84a30e951942394f45922c59b6b6d7c6d561/packages/build/tests/helpers/server.js#L7). This makes the test faster and more resilient to network failures.
However, this is also based on assumptions on the response of those servers. For example, if the network response was to change shape, or we made the initial assumption wrong, those tests would not catch this.
This created one instance of a production bug, where we assumed that a specific property of the `getSite` endpoint was `undefined` instead of `null` when empty.
We should run some tests against the real servers. | non_priority | improve testing of network requests netlify build performs few network requests to bitballoon api specifically the getsite canceldeploy and createpluginrun endpoints telemetry in integration tests we currently this makes the test faster and more resilient to network failures however this is also based on assumptions on the response of those servers for example if the network response was to change shape or we made the initial assumption wrong those tests would not catch this this created one instance of a production bug where we assumed that a specific property of the getsite endpoint was undefined instead of null when empty we should run some tests against the real servers | 0 |
114,638 | 11,852,759,609 | IssuesEvent | 2020-03-24 20:34:44 | fthomas/scala-steward | https://api.github.com/repos/fthomas/scala-steward | closed | Firejail doesn't work in Docker | documentation | Hi @fthomas, I've discovered that firejail doesn't work in Docker:
```bash
$ firejail echo hello
Warning: an existing sandbox was detected. echo will run without any additional sandboxing features
```
Running it with the `--force` flag, which is supposed to disable the PID namespace checking, results in the following error:
```bash
$ firejail --force echo hello
Error clone: main.c:2519 main: Operation not permitted
```
Additionally, the `--force` flag was removed in firejail 0.9.54 and apparently it is not possible to run firejail in Docker anymore (I asked about this recently https://github.com/netblue30/firejail/issues/2579).
I think we should remove firejail from scala-steward docker image and force it to run without the sandbox when the command isn't found or make it explicit in the docs that the sandboxing features don't work when running scala-steward in Docker. | 1.0 | Firejail doesn't work in Docker - Hi @fthomas, I've discovered that firejail doesn't work in Docker:
```bash
$ firejail echo hello
Warning: an existing sandbox was detected. echo will run without any additional sandboxing features
```
Running it with the `--force` flag, which is supposed to disable the PID namespace checking, results in the following error:
```bash
$ firejail --force echo hello
Error clone: main.c:2519 main: Operation not permitted
```
Additionally, the `--force` flag was removed in firejail 0.9.54 and apparently it is not possible to run firejail in Docker anymore (I asked about this recently https://github.com/netblue30/firejail/issues/2579).
I think we should remove firejail from scala-steward docker image and force it to run without the sandbox when the command isn't found or make it explicit in the docs that the sandboxing features don't work when running scala-steward in Docker. | non_priority | firejail doesn t work in docker hi fthomas i ve discovered that firejail doesn t work in docker bash firejail echo hello warning an existing sandbox was detected echo will run without any additional sandboxing features running it with the force flag which is supposed to disable the pid namespace checking results in the following error bash firejail force echo hello error clone main c main operation not permitted additionally the force flag was removed in firejail and apparently it is not possible to run firejail in docker anymore i asked about this recently i think we should remove firejail from scala steward docker image and force it to run without the sandbox when the command isn t found or make it explicit in the docs that the sandboxing features don t work when running scala steward in docker | 0 |
12,743 | 7,969,056,431 | IssuesEvent | 2018-07-16 07:34:02 | shopsys/shopsys | https://api.github.com/repos/shopsys/shopsys | reopened | coding standards: speed up ecs check | Bug :bug: Performance :gear: | <!--- Title should contain short general summary what is the issue about -->
On linux OS it was noticed that next build-demo-dev phing targets are slower with ecs for packages than with the checkers there were before (php_cs....).
### What is happening
<!--- What are preconditions and your setting e.g. Shopsys version or in case of Docker issues your operating system -->
Ubuntu 18, i7-8550, 16GB RAM, SSD
<!--- Best is to include steps to reproduce this issue if relevant-->
#### Steps
```sh
version=v7.0.0-alpha2 # v7.0.0-alpha3
git checkout
composer update
./phing clean db-create test-db-create
for i in 1 2 3
do
./phing build-demo-dev
done
```
<!--- Tell us what happens -->
Ecs is using symfony cache and if `composer install` is triggered the cache is invalidated.
### Expected result
<!--- Tell us what should happened instead -->
Optimalization of spped of ecs target and solved the issue with cache would be great if possible. | True | coding standards: speed up ecs check - <!--- Title should contain short general summary what is the issue about -->
On linux OS it was noticed that next build-demo-dev phing targets are slower with ecs for packages than with the checkers there were before (php_cs....).
### What is happening
<!--- What are preconditions and your setting e.g. Shopsys version or in case of Docker issues your operating system -->
Ubuntu 18, i7-8550, 16GB RAM, SSD
<!--- Best is to include steps to reproduce this issue if relevant-->
#### Steps
```sh
version=v7.0.0-alpha2 # v7.0.0-alpha3
git checkout
composer update
./phing clean db-create test-db-create
for i in 1 2 3
do
./phing build-demo-dev
done
```
<!--- Tell us what happens -->
Ecs is using symfony cache and if `composer install` is triggered the cache is invalidated.
### Expected result
<!--- Tell us what should happened instead -->
Optimalization of spped of ecs target and solved the issue with cache would be great if possible. | non_priority | coding standards speed up ecs check on linux os it was noticed that next build demo dev phing targets are slower with ecs for packages than with the checkers there were before php cs what is happening ubuntu ram ssd steps sh version git checkout composer update phing clean db create test db create for i in do phing build demo dev done ecs is using symfony cache and if composer install is triggered the cache is invalidated expected result optimalization of spped of ecs target and solved the issue with cache would be great if possible | 0 |
89,808 | 8,214,489,129 | IssuesEvent | 2018-09-04 23:40:20 | mautic/mautic | https://api.github.com/repos/mautic/mautic | closed | SugarCRM - Companies Sync problems | Bug Ready To Test | **Please DO NOT report security vulnerabilities here. Send them to security@mautic.com instead.**
What type of report is this:
| Q | A
| ---| ---
| Bug report? | Y
| Feature request? | N
| Enhancement? | N
## Description:
Companies created in SugarCRM are pulled empty in Mautic and empty line created multiple times
## If a bug:
| Q | A
| --- | ---
| Mautic version | 2.12.1
| PHP version | 7.0.4
### Steps to reproduce:
1. Make sure a SugarCRM 6.x integration is properly setup and syncing
2. Create a company in Sugar with all needed fields.
3. Wait 2min
4. Trigger the fetchleads console/cron job php app/console mautic:integration:fetchleads --integration=Sugarcrm --time-interval="5 minute"
5. Wait for the sync to complete and then check the corresponding record in the Mautic UI. Observe that empty company is created
### Log errors:
On the second console run, you have error below :
`PHP Notice: Undefined offset: 1 in /app/bundles/LeadBundle/Helper/IdentifyCompanyHelper.php on line 132`
| 1.0 | SugarCRM - Companies Sync problems - **Please DO NOT report security vulnerabilities here. Send them to security@mautic.com instead.**
What type of report is this:
| Q | A
| ---| ---
| Bug report? | Y
| Feature request? | N
| Enhancement? | N
## Description:
Companies created in SugarCRM are pulled empty in Mautic and empty line created multiple times
## If a bug:
| Q | A
| --- | ---
| Mautic version | 2.12.1
| PHP version | 7.0.4
### Steps to reproduce:
1. Make sure a SugarCRM 6.x integration is properly setup and syncing
2. Create a company in Sugar with all needed fields.
3. Wait 2min
4. Trigger the fetchleads console/cron job php app/console mautic:integration:fetchleads --integration=Sugarcrm --time-interval="5 minute"
5. Wait for the sync to complete and then check the corresponding record in the Mautic UI. Observe that empty company is created
### Log errors:
On the second console run, you have error below :
`PHP Notice: Undefined offset: 1 in /app/bundles/LeadBundle/Helper/IdentifyCompanyHelper.php on line 132`
| non_priority | sugarcrm companies sync problems please do not report security vulnerabilities here send them to security mautic com instead what type of report is this q a bug report y feature request n enhancement n description companies created in sugarcrm are pulled empty in mautic and empty line created multiple times if a bug q a mautic version php version steps to reproduce make sure a sugarcrm x integration is properly setup and syncing create a company in sugar with all needed fields wait trigger the fetchleads console cron job php app console mautic integration fetchleads integration sugarcrm time interval minute wait for the sync to complete and then check the corresponding record in the mautic ui observe that empty company is created log errors on the second console run you have error below php notice undefined offset in app bundles leadbundle helper identifycompanyhelper php on line | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.