Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
32,790 | 26,998,420,948 | IssuesEvent | 2023-02-10 04:39:02 | ossf/package-analysis | https://api.github.com/repos/ossf/package-analysis | closed | Update cloudbuild.yaml to use Makefile / build static analysis sandbox image | infrastructure | Currently, cloudbuild.yaml uses the old build_docker.sh script to build production images. However, this does not support building the static analysis image.
Before static analysis is enabled in production, cloudbuild.yaml needs to be updated to also build the static analysis sandbox image, which is done via the Makefile
In fact, the entire build can be done using the Makefile, which may make the change easier to implement | 1.0 | Update cloudbuild.yaml to use Makefile / build static analysis sandbox image - Currently, cloudbuild.yaml uses the old build_docker.sh script to build production images. However, this does not support building the static analysis image.
Before static analysis is enabled in production, cloudbuild.yaml needs to be updated to also build the static analysis sandbox image, which is done via the Makefile
In fact, the entire build can be done using the Makefile, which may make the change easier to implement | infrastructure | update cloudbuild yaml to use makefile build static analysis sandbox image currently cloudbuild yaml uses the old build docker sh script to build production images however this does not support building the static analysis image before static analysis is enabled in production cloudbuild yaml needs to be updated to also build the static analysis sandbox image which is done via the makefile in fact the entire build can be done using the makefile which may make the change easier to implement | 1 |
6,866 | 15,682,331,111 | IssuesEvent | 2021-03-25 07:07:56 | bithyve/hexa | https://api.github.com/repos/bithyve/hexa | closed | PoC - Use SQLite with persist to compare peformance | 1.5.1 Architecture DevTask NFR | Use sqlite with redux persist instead of AsyncStorage and compare performance on a low end android and ios device | 1.0 | PoC - Use SQLite with persist to compare peformance - Use sqlite with redux persist instead of AsyncStorage and compare performance on a low end android and ios device | non_infrastructure | poc use sqlite with persist to compare peformance use sqlite with redux persist instead of asyncstorage and compare performance on a low end android and ios device | 0 |
12,568 | 9,852,832,241 | IssuesEvent | 2019-06-19 13:40:24 | crystal-lang/crystal | https://api.github.com/repos/crystal-lang/crystal | closed | Using nightly builds on Travis CI is broken | topic:infrastructure | Using nightly builds on Travis CI has been advertised in [docs for Crystal language](https://docs.travis-ci.com/user/languages/crystal) for some time now, yet ATM it doesn't work.
Refs: https://github.com/travis-ci/travis-build/blob/b2a89dad8e7020c4c28b396639b36d87e5b1d4f6/lib/travis/build/script/crystal.rb#L80
---
Without nightly releases users caring about having their codebase tested on Travis CI against current Crystal are left with 3 less-than-optimal-options:
1. Wait for next Crystal release—which might happen... who knows when?
2. Use docker with nightly tag—~~but! `nightly` is 11 days old as of now, and new builds are tagged with `nightly-YYYYMMDD` scheme, making it hard to point at just that last release~~—fixed in #5837
3. Build it manually—ouch
Since nightly releases are being build as of #5182, the only thing missing would be hooking this up to apt package index, which is used by Travis CI.
With that in place we'd be able to use nice'n'simple `.travis.yml`:
```yml
language: crystal
crystal:
- latest
- nightly
```
Howgh! Thanks for reading! | 1.0 | Using nightly builds on Travis CI is broken - Using nightly builds on Travis CI has been advertised in [docs for Crystal language](https://docs.travis-ci.com/user/languages/crystal) for some time now, yet ATM it doesn't work.
Refs: https://github.com/travis-ci/travis-build/blob/b2a89dad8e7020c4c28b396639b36d87e5b1d4f6/lib/travis/build/script/crystal.rb#L80
---
Without nightly releases users caring about having their codebase tested on Travis CI against current Crystal are left with 3 less-than-optimal-options:
1. Wait for next Crystal release—which might happen... who knows when?
2. Use docker with nightly tag—~~but! `nightly` is 11 days old as of now, and new builds are tagged with `nightly-YYYYMMDD` scheme, making it hard to point at just that last release~~—fixed in #5837
3. Build it manually—ouch
Since nightly releases are being build as of #5182, the only thing missing would be hooking this up to apt package index, which is used by Travis CI.
With that in place we'd be able to use nice'n'simple `.travis.yml`:
```yml
language: crystal
crystal:
- latest
- nightly
```
Howgh! Thanks for reading! | infrastructure | using nightly builds on travis ci is broken using nightly builds on travis ci has been advertised in for some time now yet atm it doesn t work refs without nightly releases users caring about having their codebase tested on travis ci against current crystal are left with less than optimal options wait for next crystal release—which might happen who knows when use docker with nightly tag— but nightly is days old as of now and new builds are tagged with nightly yyyymmdd scheme making it hard to point at just that last release —fixed in build it manually—ouch since nightly releases are being build as of the only thing missing would be hooking this up to apt package index which is used by travis ci with that in place we d be able to use nice n simple travis yml yml language crystal crystal latest nightly howgh thanks for reading | 1 |
12,545 | 9,819,977,379 | IssuesEvent | 2019-06-14 00:16:17 | RITlug/TigerOS | https://api.github.com/repos/RITlug/TigerOS | opened | Move to RITlug GitLab Roadmap | infrastructure priority:critical | # What This Is
This issue is where we are going to create our roadmap for what is needed to be done in order to successfully migrate TigerOS over to git.ritlug.com
| 1.0 | Move to RITlug GitLab Roadmap - # What This Is
This issue is where we are going to create our roadmap for what is needed to be done in order to successfully migrate TigerOS over to git.ritlug.com
| infrastructure | move to ritlug gitlab roadmap what this is this issue is where we are going to create our roadmap for what is needed to be done in order to successfully migrate tigeros over to git ritlug com | 1 |
77,940 | 22,047,859,851 | IssuesEvent | 2022-05-30 05:13:47 | google/mediapipe | https://api.github.com/repos/google/mediapipe | closed | <Solved> from mediapipe.python._framework_bindings import | type:build/install stat:awaiting response platform:python stalled | **Versions:**
python version - **3.8** (I did not use anaconda)
OS - **Windows** system terminal
the rest are all based on installation **tutorial**
Suppose you are not using 'pip3 install mediapipe' to install mediapipe. Instead, you use python package, and the last step is '**python setup.py install --link-opencv**'.
**Errors:**
from mediapipe.python._framework_bindings import resource_util ImportError: DLL load failed: The specified module could not be found.
**Solutions:**
If #1839 #1405 doesn't work for you, please check if you run 'import mediapipe' in the mediapipe repository.
(I do not need Visual C++ redistributable packages OR msvc-runtime)
**Reminder:** DO NOT run test ‘import mediapipe’ in the mediapipe repository
Firstly, you should run '**pip3 install absl-py attrs matplotlib opencv-contrib-python protobuf --user**' in your terminal since requirements.txt lacks some libraries.
Then, for me, I run 'import mediapipe' in **IDLE (Python 3.8 64-bit)**.
It should work!
**Last check step** ->
>> pip3 list
```
Package Version
--------------------- ---------
matplotlib 3.5.1
mediapipe dev
numpy 1.22.2
opencv-contrib-python 4.5.5.62
opencv-python 4.5.5.62
...
```
There should not be version number here. Enjoy editing and creating your own mediapipe :)
I made this post since there are still many people asking the same question. I am glad to share my solution which is unqiue but effective.
| 1.0 | <Solved> from mediapipe.python._framework_bindings import - **Versions:**
python version - **3.8** (I did not use anaconda)
OS - **Windows** system terminal
the rest are all based on installation **tutorial**
Suppose you are not using 'pip3 install mediapipe' to install mediapipe. Instead, you use python package, and the last step is '**python setup.py install --link-opencv**'.
**Errors:**
from mediapipe.python._framework_bindings import resource_util ImportError: DLL load failed: The specified module could not be found.
**Solutions:**
If #1839 #1405 doesn't work for you, please check if you run 'import mediapipe' in the mediapipe repository.
(I do not need Visual C++ redistributable packages OR msvc-runtime)
**Reminder:** DO NOT run test ‘import mediapipe’ in the mediapipe repository
Firstly, you should run '**pip3 install absl-py attrs matplotlib opencv-contrib-python protobuf --user**' in your terminal since requirements.txt lacks some libraries.
Then, for me, I run 'import mediapipe' in **IDLE (Python 3.8 64-bit)**.
It should work!
**Last check step** ->
>> pip3 list
```
Package Version
--------------------- ---------
matplotlib 3.5.1
mediapipe dev
numpy 1.22.2
opencv-contrib-python 4.5.5.62
opencv-python 4.5.5.62
...
```
There should not be version number here. Enjoy editing and creating your own mediapipe :)
I made this post since there are still many people asking the same question. I am glad to share my solution which is unqiue but effective.
| non_infrastructure | from mediapipe python framework bindings import versions python version i did not use anaconda os windows system terminal the rest are all based on installation tutorial suppose you are not using install mediapipe to install mediapipe instead you use python package and the last step is python setup py install link opencv errors from mediapipe python framework bindings import resource util importerror dll load failed the specified module could not be found solutions if doesn t work for you please check if you run import mediapipe in the mediapipe repository i do not need visual c redistributable packages or msvc runtime reminder do not run test ‘import mediapipe’ in the mediapipe repository firstly you should run install absl py attrs matplotlib opencv contrib python protobuf user in your terminal since requirements txt lacks some libraries then for me i run import mediapipe in idle python bit it should work last check step list package version matplotlib mediapipe dev numpy opencv contrib python opencv python there should not be version number here enjoy editing and creating your own mediapipe i made this post since there are still many people asking the same question i am glad to share my solution which is unqiue but effective | 0 |
76,340 | 3,487,299,391 | IssuesEvent | 2016-01-01 19:07:47 | Apollo-Community/ApolloStation | https://api.github.com/repos/Apollo-Community/ApolloStation | closed | suggest adding a little code bit to check for table-type before attempting to combine tables. | easy to fix priority: low suggestion | It's really kind of attrociously ugly when you see an R table and a wood table trying to combine into one. Probbably easy to fix.
On another note, the chef has an attrocious combination of an R table and a glass table right next to eachother. | 1.0 | suggest adding a little code bit to check for table-type before attempting to combine tables. - It's really kind of attrociously ugly when you see an R table and a wood table trying to combine into one. Probbably easy to fix.
On another note, the chef has an attrocious combination of an R table and a glass table right next to eachother. | non_infrastructure | suggest adding a little code bit to check for table type before attempting to combine tables it s really kind of attrociously ugly when you see an r table and a wood table trying to combine into one probbably easy to fix on another note the chef has an attrocious combination of an r table and a glass table right next to eachother | 0 |
295,517 | 9,087,609,164 | IssuesEvent | 2019-02-18 14:10:59 | status-im/status-react | https://api.github.com/repos/status-im/status-react | closed | UX updates on Recovery flows | feature high-priority stale wallet | ### User Story
As a user, I want to have a right perception on what kind of password do I need to provide during the recovery flow that I'm not trying to recall my old password.
(Needed to ETHBerlin)
### Description
*Type*: Feature
*Summary*: Current recovery flow confuses users, makes them think they need to provide their old password which is not correct. I propose to add separate 2 steps for password creation and confirmation. Also, I renamed (Hester proposed this many times) action on the main screen from `I already have an account` to `Access account`.
#### Expected behavior

#### Actual behavior
<img width="459" alt="screenshot 2018-08-18 11 54 35" src="https://user-images.githubusercontent.com/6559881/44298585-9ec54280-a2dd-11e8-9f02-47cbf84c2fa6.png"> | 1.0 | UX updates on Recovery flows - ### User Story
As a user, I want to have a right perception on what kind of password do I need to provide during the recovery flow that I'm not trying to recall my old password.
(Needed to ETHBerlin)
### Description
*Type*: Feature
*Summary*: Current recovery flow confuses users, makes them think they need to provide their old password which is not correct. I propose to add separate 2 steps for password creation and confirmation. Also, I renamed (Hester proposed this many times) action on the main screen from `I already have an account` to `Access account`.
#### Expected behavior

#### Actual behavior
<img width="459" alt="screenshot 2018-08-18 11 54 35" src="https://user-images.githubusercontent.com/6559881/44298585-9ec54280-a2dd-11e8-9f02-47cbf84c2fa6.png"> | non_infrastructure | ux updates on recovery flows user story as a user i want to have a right perception on what kind of password do i need to provide during the recovery flow that i m not trying to recall my old password needed to ethberlin description type feature summary current recovery flow confuses users makes them think they need to provide their old password which is not correct i propose to add separate steps for password creation and confirmation also i renamed hester proposed this many times action on the main screen from i already have an account to access account expected behavior actual behavior img width alt screenshot src | 0 |
642,625 | 20,908,575,907 | IssuesEvent | 2022-03-24 06:43:03 | apache/dolphinscheduler | https://api.github.com/repos/apache/dolphinscheduler | closed | [Bug] [UI Next][V1.0.0-Alpha] Batch delete not working while click the border of the button | bug UI priority:middle | ### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues.
### What happened

### What you expected to happen
above.
### How to reproduce
above.
### Anything else
_No response_
### Version
dev
### Are you willing to submit PR?
- [ ] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
| 1.0 | [Bug] [UI Next][V1.0.0-Alpha] Batch delete not working while click the border of the button - ### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues.
### What happened

### What you expected to happen
above.
### How to reproduce
above.
### Anything else
_No response_
### Version
dev
### Are you willing to submit PR?
- [ ] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
| non_infrastructure | batch delete not working while click the border of the button search before asking i had searched in the and found no similar issues what happened what you expected to happen above how to reproduce above anything else no response version dev are you willing to submit pr yes i am willing to submit a pr code of conduct i agree to follow this project s | 0 |
22,257 | 3,947,116,486 | IssuesEvent | 2016-04-28 08:34:46 | backbee/backbee-standard | https://api.github.com/repos/backbee/backbee-standard | closed | [TREE] If user chose to close tree popin, don't reload the tree popin when page is refreshed | enhancement To test | If user chose to close tree popin, don't reload the tree popin when page is refreshed | 1.0 | [TREE] If user chose to close tree popin, don't reload the tree popin when page is refreshed - If user chose to close tree popin, don't reload the tree popin when page is refreshed | non_infrastructure | if user chose to close tree popin don t reload the tree popin when page is refreshed if user chose to close tree popin don t reload the tree popin when page is refreshed | 0 |
367,576 | 25,750,457,305 | IssuesEvent | 2022-12-08 12:53:22 | sidebase/nuxt-auth | https://api.github.com/repos/sidebase/nuxt-auth | closed | Setup deployment of the docs page | documentation | **Goal**
In order to improve our DX, we want to publish are own extensive docs page. The docs have already been created in https://github.com/sidebase/docs and need to be transferred into the respective module repositories.
Afterwards we hope to host each documentation with GitHub Pages, allowing them to be redeployed after every merge to the main branch.
In order to do this we first need to check how we can correctly assign subdomains to a specific github page. In this case we want to use the subdomain https://nuxt-auth.sidebase.io to publish our docs.
As far as I see, we just need to add a CNAME record for all subdomains of sidebase.io to point to the GitHub servers. Once done, we can verify the subdomain through the GitHub verification process to provide proof of ownership.
**Links**
https://docs.github.com/en/pages/configuring-a-custom-domain-for-your-github-pages-site/about-custom-domains-and-github-pages
https://sean-coughlin.medium.com/how-to-host-a-site-with-a-subdomain-on-github-pages-33202341f711
**Tasks**
- [ ] Finish documentation (@zoey-kaiser for finishing, @BracketJohn for review)
- [ ] Setup CNAME record for the domain (@IronicUsername)
- [ ] Add documentation folder to `nuxt-auth` repo (@zoey-kaiser)
- [ ] Setup Github Actions to rebuild docs so we do not need to do it manually (optional, depending on how much extra work it presents, @IronicUsername) | 1.0 | Setup deployment of the docs page - **Goal**
In order to improve our DX, we want to publish are own extensive docs page. The docs have already been created in https://github.com/sidebase/docs and need to be transferred into the respective module repositories.
Afterwards we hope to host each documentation with GitHub Pages, allowing them to be redeployed after every merge to the main branch.
In order to do this we first need to check how we can correctly assign subdomains to a specific github page. In this case we want to use the subdomain https://nuxt-auth.sidebase.io to publish our docs.
As far as I see, we just need to add a CNAME record for all subdomains of sidebase.io to point to the GitHub servers. Once done, we can verify the subdomain through the GitHub verification process to provide proof of ownership.
**Links**
https://docs.github.com/en/pages/configuring-a-custom-domain-for-your-github-pages-site/about-custom-domains-and-github-pages
https://sean-coughlin.medium.com/how-to-host-a-site-with-a-subdomain-on-github-pages-33202341f711
**Tasks**
- [ ] Finish documentation (@zoey-kaiser for finishing, @BracketJohn for review)
- [ ] Setup CNAME record for the domain (@IronicUsername)
- [ ] Add documentation folder to `nuxt-auth` repo (@zoey-kaiser)
- [ ] Setup Github Actions to rebuild docs so we do not need to do it manually (optional, depending on how much extra work it presents, @IronicUsername) | non_infrastructure | setup deployment of the docs page goal in order to improve our dx we want to publish are own extensive docs page the docs have already been created in and need to be transferred into the respective module repositories afterwards we hope to host each documentation with github pages allowing them to be redeployed after every merge to the main branch in order to do this we first need to check how we can correctly assign subdomains to a specific github page in this case we want to use the subdomain to publish our docs as far as i see we just need to add a cname record for all subdomains of sidebase io to point to the github servers once done we can verify the subdomain through the github verification process to provide proof of ownership links tasks finish documentation zoey kaiser for finishing bracketjohn for review setup cname record for the domain ironicusername add documentation folder to nuxt auth repo zoey kaiser setup github actions to rebuild docs so we do not need to do it manually optional depending on how much extra work it presents ironicusername | 0 |
21,915 | 14,934,226,603 | IssuesEvent | 2021-01-25 10:15:21 | qutebrowser/qutebrowser | https://api.github.com/repos/qutebrowser/qutebrowser | opened | Switch pytest import mode to importlib | component: infrastructure priority: 2 - low | With https://github.com/pytest-dev/pytest-bdd/pull/381 plus various improvements inside pytest, I should give this a try again. `__init__.py` files in `tests/` are needed for various utils, see e.g. #5249. | 1.0 | Switch pytest import mode to importlib - With https://github.com/pytest-dev/pytest-bdd/pull/381 plus various improvements inside pytest, I should give this a try again. `__init__.py` files in `tests/` are needed for various utils, see e.g. #5249. | infrastructure | switch pytest import mode to importlib with plus various improvements inside pytest i should give this a try again init py files in tests are needed for various utils see e g | 1 |
268,546 | 28,566,220,166 | IssuesEvent | 2023-04-21 02:30:11 | Thanraj/packages_apps_Settings | https://api.github.com/repos/Thanraj/packages_apps_Settings | closed | CVE-2023-20955 (High) detected in Settingsandroid-10.0.0_r44 - autoclosed | Mend: dependency security vulnerability | ## CVE-2023-20955 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r44</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/packages_apps_Settings/commit/cd4be0a8e3f7aad8748a372235eac74f3a3bcaaa">cd4be0a8e3f7aad8748a372235eac74f3a3bcaaa</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/applications/appinfo/AppInfoDashboardFragment.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In onPrepareOptionsMenu of AppInfoDashboardFragment.java, there is a possible way to bypass admin restrictions and uninstall applications for all users due to a missing permission check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-11 Android-12 Android-12L Android-13Android ID: A-258653813
<p>Publish Date: 2023-03-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-20955>CVE-2023-20955</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://android.googlesource.com/platform/packages/apps/Settings/+/f3b323e378ee5d98875711216cbd92f4fa795fc0">https://android.googlesource.com/platform/packages/apps/Settings/+/f3b323e378ee5d98875711216cbd92f4fa795fc0</a></p>
<p>Release Date: 2023-03-24</p>
<p>Fix Resolution: android-13.0.0_r32</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-20955 (High) detected in Settingsandroid-10.0.0_r44 - autoclosed - ## CVE-2023-20955 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Settingsandroid-10.0.0_r44</b></p></summary>
<p>
<p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Settings>https://android.googlesource.com/platform/packages/apps/Settings</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Thanraj/packages_apps_Settings/commit/cd4be0a8e3f7aad8748a372235eac74f3a3bcaaa">cd4be0a8e3f7aad8748a372235eac74f3a3bcaaa</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/com/android/settings/applications/appinfo/AppInfoDashboardFragment.java</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In onPrepareOptionsMenu of AppInfoDashboardFragment.java, there is a possible way to bypass admin restrictions and uninstall applications for all users due to a missing permission check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-11 Android-12 Android-12L Android-13Android ID: A-258653813
<p>Publish Date: 2023-03-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-20955>CVE-2023-20955</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://android.googlesource.com/platform/packages/apps/Settings/+/f3b323e378ee5d98875711216cbd92f4fa795fc0">https://android.googlesource.com/platform/packages/apps/Settings/+/f3b323e378ee5d98875711216cbd92f4fa795fc0</a></p>
<p>Release Date: 2023-03-24</p>
<p>Fix Resolution: android-13.0.0_r32</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in settingsandroid autoclosed cve high severity vulnerability vulnerable library settingsandroid library home page a href found in head commit a href found in base branch master vulnerable source files src com android settings applications appinfo appinfodashboardfragment java vulnerability details in onprepareoptionsmenu of appinfodashboardfragment java there is a possible way to bypass admin restrictions and uninstall applications for all users due to a missing permission check this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android step up your open source security game with mend | 0 |
2,672 | 3,799,734,820 | IssuesEvent | 2016-03-23 16:48:42 | codeforamerica/brigade-staff | https://api.github.com/repos/codeforamerica/brigade-staff | closed | Advise the Community of Practice site | content focus areas infrastructure network growth | All the focus areas want help on building communities of practice. I think infrastructure is a good way to do it.
A good CoP site could include:
* The general overview - GitHub ReadMe
* Discussion forums - GitHub Issues
* Active projects - CfAPI projects tagged 'health' or 'S&J'
* Asks for help - Civic Issues
* People - PeopleDB and Badges | 1.0 | Advise the Community of Practice site - All the focus areas want help on building communities of practice. I think infrastructure is a good way to do it.
A good CoP site could include:
* The general overview - GitHub ReadMe
* Discussion forums - GitHub Issues
* Active projects - CfAPI projects tagged 'health' or 'S&J'
* Asks for help - Civic Issues
* People - PeopleDB and Badges | infrastructure | advise the community of practice site all the focus areas want help on building communities of practice i think infrastructure is a good way to do it a good cop site could include the general overview github readme discussion forums github issues active projects cfapi projects tagged health or s j asks for help civic issues people peopledb and badges | 1 |
34,467 | 29,999,851,319 | IssuesEvent | 2023-06-26 08:35:30 | dotnet/extensions | https://api.github.com/repos/dotnet/extensions | closed | start-vs.cmd does not work | work in progress :construction: area-infrastructure | ### Description
The batch file `start-vs.cmd` in the root of the repository references a solution file that does not exist in the repository:
https://github.com/dotnet/extensions/blob/14917b87e8fc81f10d44ceea52d9b24e50e26550/start-vs.cmd#L26
### Reproduction Steps
Run the following commands on a Windows machine:
1. `restore.cmd`
2. `start-vs.cmd`
### Expected behavior
A solution file containing the repository's projects opens in Visual Studio ready to work on the code.
### Actual behavior
The batch file displays the following error and exits:
```
The system cannot find the file .\SDK.sln.
```
Visual Studio is not launched.
### Regression?
I guess there used to be a solution file in an old repo before the code was made public. If that was the case, then yes.
### Known Workarounds
Manually create a solution file with that name in the repository and then run the batch file.
### Configuration
N/A
### Other information
_No response_ | 1.0 | start-vs.cmd does not work - ### Description
The batch file `start-vs.cmd` in the root of the repository references a solution file that does not exist in the repository:
https://github.com/dotnet/extensions/blob/14917b87e8fc81f10d44ceea52d9b24e50e26550/start-vs.cmd#L26
### Reproduction Steps
Run the following commands on a Windows machine:
1. `restore.cmd`
2. `start-vs.cmd`
### Expected behavior
A solution file containing the repository's projects opens in Visual Studio ready to work on the code.
### Actual behavior
The batch file displays the following error and exits:
```
The system cannot find the file .\SDK.sln.
```
Visual Studio is not launched.
### Regression?
I guess there used to be a solution file in an old repo before the code was made public. If that was the case, then yes.
### Known Workarounds
Manually create a solution file with that name in the repository and then run the batch file.
### Configuration
N/A
### Other information
_No response_ | infrastructure | start vs cmd does not work description the batch file start vs cmd in the root of the repository references a solution file that does not exist in the repository reproduction steps run the following commands on a windows machine restore cmd start vs cmd expected behavior a solution file containing the repository s projects opens in visual studio ready to work on the code actual behavior the batch file displays the following error and exits the system cannot find the file sdk sln visual studio is not launched regression i guess there used to be a solution file in an old repo before the code was made public if that was the case then yes known workarounds manually create a solution file with that name in the repository and then run the batch file configuration n a other information no response | 1 |
461,124 | 13,223,931,405 | IssuesEvent | 2020-08-17 18:09:31 | epam/cloud-pipeline | https://api.github.com/repos/epam/cloud-pipeline | closed | Set "bucket-owner-full-control" ACL for any S3 upload operation | goal/dorado :sparkles: kind/enhancement priority/high state/verify sys/core sys/gui | **Background**
Once the #1036 was introduced - it is quite common to have the S3 buckets attached from other accounts.
In such cases, the bucket owners can't access the object, uploaded by the Cloud Pipeline.
This is caused by AWS ACLs: the https://aws.amazon.com/premiumsupport/knowledge-center/s3-bucket-owner-access/
**Approach**
* We shall set the ACL `bucket-owner-full-control` for any object `PUT` operation, which is initiated by the `Cloud Pipeline`:
* CLI cp/mv
* pipe mount
* s3fs/goofys mount
* GUI upload
* This ACL shall not harm anyone, so we can set it as a default | 1.0 | Set "bucket-owner-full-control" ACL for any S3 upload operation - **Background**
Once the #1036 was introduced - it is quite common to have the S3 buckets attached from other accounts.
In such cases, the bucket owners can't access the object, uploaded by the Cloud Pipeline.
This is caused by AWS ACLs: the https://aws.amazon.com/premiumsupport/knowledge-center/s3-bucket-owner-access/
**Approach**
* We shall set the ACL `bucket-owner-full-control` for any object `PUT` operation, which is initiated by the `Cloud Pipeline`:
* CLI cp/mv
* pipe mount
* s3fs/goofys mount
* GUI upload
* This ACL shall not harm anyone, so we can set it as a default | non_infrastructure | set bucket owner full control acl for any upload operation background once the was introduced it is quite common to have the buckets attached from other accounts in such cases the bucket owners can t access the object uploaded by the cloud pipeline this is caused by aws acls the approach we shall set the acl bucket owner full control for any object put operation which is initiated by the cloud pipeline cli cp mv pipe mount goofys mount gui upload this acl shall not harm anyone so we can set it as a default | 0 |
12,706 | 9,923,934,905 | IssuesEvent | 2019-07-01 08:31:27 | libero/libero | https://api.github.com/repos/libero/libero | closed | Create unstable-jats-ingester-logs bucket | infrastructure jats-ingester | Same for other environments.
Application (with current credentials) must be able to read and write to it. | 1.0 | Create unstable-jats-ingester-logs bucket - Same for other environments.
Application (with current credentials) must be able to read and write to it. | infrastructure | create unstable jats ingester logs bucket same for other environments application with current credentials must be able to read and write to it | 1 |
291,920 | 25,185,530,737 | IssuesEvent | 2022-11-11 17:33:18 | multipath-tcp/mptcp_net-next | https://api.github.com/repos/multipath-tcp/mptcp_net-next | closed | selftests: `simult_flows`: transfer slower than expected! | bug selftests | It is only failing occasionally on the public CI, e.g.
```
# balanced bwidth 7389 max 7561 [ OK ]
# balanced bwidth - reverse direction 7420 max 7561 [ OK ]
# balanced bwidth with unbalanced delay 7448 max 7561 [ OK ]
# balanced bwidth with unbalanced delay - reverse direction 7386 max 7561 [ OK ]
# unbalanced bwidth 3826 max 4005 [ OK ]
# unbalanced bwidth - reverse direction 3818 max 4005 [ OK ]
# unbalanced bwidth with unbalanced delay 3949 max 4005 [ OK ]
# unbalanced bwidth with unbalanced delay - reverse direction 3794 max 4005 [ OK ]
# unbalanced bwidth with opposed, unbalanced delay 3862 max 4005 [ OK ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4143 ms, expected 4005 ms max 4005 [ fail ]
# client exit code 0, server 1
#
# netns ns3-0-NkCPyi socket stat for 10010:
# State Recv-Q Send-Q Local Address:Port Peer Address:Port Process
# TIME-WAIT 0 0 10.0.3.3:10010 10.0.1.1:45520 timer:(timewait,59sec,0)
#
# TIME-WAIT 0 0 10.0.3.3:10010 10.0.2.1:46451 timer:(timewait,59sec,0)
#
#
# netns ns1-0-NkCPyi socket stat for 10010:
# State Recv-Q Send-Q Local Address:Port Peer Address:Port Process
# TIME-WAIT 0 0 10.0.1.1:45520 10.0.3.3:10010 timer:(timewait,59sec,0)
#
# TIME-WAIT 0 0 10.0.2.1%ns1eth2:46451 10.0.3.3:10010 timer:(timewait,59sec,0)
#
# -rw------- 1 root root 81920 Oct 10 07:10 /tmp/tmp.HPaPdz7J6C
# -rw------- 1 root root 81920 Oct 10 07:09 /tmp/tmp.ovbsZnLTYN
# -rw------- 1 root root 16777216 Oct 10 07:09 /tmp/tmp.U4M3NESqYO
# -rw------- 1 root root 16777216 Oct 10 07:10 /tmp/tmp.v7DK6RJCYy
```
Only in non debug kernel config (without additional tolerance).
From:
* https://api.cirrus-ci.com/v1/artifact/task/4778608395812864/summary/summary.txt (https://cirrus-ci.com/task/4778608395812864)
If I look at the last 75 builds from the Tessares CI, we can see a few occurrences, always without a debug kernel config:
```
# unbalanced bwidth with unbalanced delay - reverse direction transfer slower than expected! runtime 4123 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4070 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4035 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4076 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4028 ms, expected 4005 ms max 4005 [ fail ]
# balanced bwidth with unbalanced delay transfer slower than expected! runtime 7616 ms, expected 7561 ms max 7561 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4008 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4185 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4152 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4187 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with unbalanced delay transfer slower than expected! runtime 4025 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4093 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with unbalanced delay transfer slower than expected! runtime 4006 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth - reverse direction transfer slower than expected! runtime 4053 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with unbalanced delay - reverse direction transfer slower than expected! runtime 4086 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth - reverse direction transfer slower than expected! runtime 4083 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4129 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4074 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4059 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4020 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4083 ms, expected 4005 ms max 4005 [ fail ]
```
It is not regularly over the limit, most of the time under 100ms, only a few times over 150ms.
Should we increase the tolerance? Or do we need to take something else into account? A bit more time for the creation of a connection + ramp-up? | 1.0 | selftests: `simult_flows`: transfer slower than expected! - It is only failing occasionally on the public CI, e.g.
```
# balanced bwidth 7389 max 7561 [ OK ]
# balanced bwidth - reverse direction 7420 max 7561 [ OK ]
# balanced bwidth with unbalanced delay 7448 max 7561 [ OK ]
# balanced bwidth with unbalanced delay - reverse direction 7386 max 7561 [ OK ]
# unbalanced bwidth 3826 max 4005 [ OK ]
# unbalanced bwidth - reverse direction 3818 max 4005 [ OK ]
# unbalanced bwidth with unbalanced delay 3949 max 4005 [ OK ]
# unbalanced bwidth with unbalanced delay - reverse direction 3794 max 4005 [ OK ]
# unbalanced bwidth with opposed, unbalanced delay 3862 max 4005 [ OK ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4143 ms, expected 4005 ms max 4005 [ fail ]
# client exit code 0, server 1
#
# netns ns3-0-NkCPyi socket stat for 10010:
# State Recv-Q Send-Q Local Address:Port Peer Address:Port Process
# TIME-WAIT 0 0 10.0.3.3:10010 10.0.1.1:45520 timer:(timewait,59sec,0)
#
# TIME-WAIT 0 0 10.0.3.3:10010 10.0.2.1:46451 timer:(timewait,59sec,0)
#
#
# netns ns1-0-NkCPyi socket stat for 10010:
# State Recv-Q Send-Q Local Address:Port Peer Address:Port Process
# TIME-WAIT 0 0 10.0.1.1:45520 10.0.3.3:10010 timer:(timewait,59sec,0)
#
# TIME-WAIT 0 0 10.0.2.1%ns1eth2:46451 10.0.3.3:10010 timer:(timewait,59sec,0)
#
# -rw------- 1 root root 81920 Oct 10 07:10 /tmp/tmp.HPaPdz7J6C
# -rw------- 1 root root 81920 Oct 10 07:09 /tmp/tmp.ovbsZnLTYN
# -rw------- 1 root root 16777216 Oct 10 07:09 /tmp/tmp.U4M3NESqYO
# -rw------- 1 root root 16777216 Oct 10 07:10 /tmp/tmp.v7DK6RJCYy
```
Only in non debug kernel config (without additional tolerance).
From:
* https://api.cirrus-ci.com/v1/artifact/task/4778608395812864/summary/summary.txt (https://cirrus-ci.com/task/4778608395812864)
If I look at the last 75 builds from the Tessares CI, we can see a few occurrences, always without a debug kernel config:
```
# unbalanced bwidth with unbalanced delay - reverse direction transfer slower than expected! runtime 4123 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4070 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4035 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4076 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4028 ms, expected 4005 ms max 4005 [ fail ]
# balanced bwidth with unbalanced delay transfer slower than expected! runtime 7616 ms, expected 7561 ms max 7561 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4008 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4185 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4152 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4187 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with unbalanced delay transfer slower than expected! runtime 4025 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4093 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with unbalanced delay transfer slower than expected! runtime 4006 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth - reverse direction transfer slower than expected! runtime 4053 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with unbalanced delay - reverse direction transfer slower than expected! runtime 4086 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth - reverse direction transfer slower than expected! runtime 4083 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4129 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4074 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4059 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay - reverse directiontransfer slower than expected! runtime 4020 ms, expected 4005 ms max 4005 [ fail ]
# unbalanced bwidth with opposed, unbalanced delay transfer slower than expected! runtime 4083 ms, expected 4005 ms max 4005 [ fail ]
```
It is not regularly over the limit, most of the time under 100ms, only a few times over 150ms.
Should we increase the tolerance? Or do we need to take something else into account? A bit more time for the creation of a connection + ramp-up? | non_infrastructure | selftests simult flows transfer slower than expected it is only failing occasionally on the public ci e g balanced bwidth max balanced bwidth reverse direction max balanced bwidth with unbalanced delay max balanced bwidth with unbalanced delay reverse direction max unbalanced bwidth max unbalanced bwidth reverse direction max unbalanced bwidth with unbalanced delay max unbalanced bwidth with unbalanced delay reverse direction max unbalanced bwidth with opposed unbalanced delay max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max client exit code server netns nkcpyi socket stat for state recv q send q local address port peer address port process time wait timer timewait time wait timer timewait netns nkcpyi socket stat for state recv q send q local address port peer address port process time wait timer timewait time wait timer timewait rw root root oct tmp tmp rw root root oct tmp tmp ovbsznltyn rw root root oct tmp tmp rw root root oct tmp tmp only in non debug kernel config without additional tolerance from if i look at the last builds from the tessares ci we can see a few occurrences always without a debug kernel config unbalanced bwidth with unbalanced delay reverse direction transfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay transfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay transfer slower than expected runtime ms expected ms max balanced bwidth with unbalanced delay transfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay transfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max unbalanced bwidth with unbalanced delay transfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay transfer slower than expected runtime ms expected ms max unbalanced bwidth with unbalanced delay transfer slower than expected runtime ms expected ms max unbalanced bwidth reverse direction transfer slower than expected runtime ms expected ms max unbalanced bwidth with unbalanced delay reverse direction transfer slower than expected runtime ms expected ms max unbalanced bwidth reverse direction transfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay transfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay reverse directiontransfer slower than expected runtime ms expected ms max unbalanced bwidth with opposed unbalanced delay transfer slower than expected runtime ms expected ms max it is not regularly over the limit most of the time under only a few times over should we increase the tolerance or do we need to take something else into account a bit more time for the creation of a connection ramp up | 0 |
8,523 | 2,996,769,999 | IssuesEvent | 2015-07-23 00:17:45 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Acl rules are not being programmed for one of the listeners when the LB has multiple listeners. | kind/bug status/blocker status/to-test | Server version - V0.29.0-rc1
Steps to reproduce the problem:
Created a LB rule with following rules having 2 listeners - 80:80 and 81:81
<img width="983" alt="screen shot 2015-07-22 at 1 14 55 pm" src="https://cloud.githubusercontent.com/assets/4266958/8838586/7aa2358c-3082-11e5-895c-771ea7479474.png">
There are no acl rules for frontend - 81 programmed in haproxy and all only 1 backend with entries from all the services are present.
<img width="676" alt="screen shot 2015-07-22 at 2 57 43 pm" src="https://cloud.githubusercontent.com/assets/4266958/8838642/fee947fe-3082-11e5-9189-812560613ac4.png">
LB redirects traffic to all services when accessing - wget --header="host: abc" http://104.154.66.248:81/service3.html
| 1.0 | Acl rules are not being programmed for one of the listeners when the LB has multiple listeners. - Server version - V0.29.0-rc1
Steps to reproduce the problem:
Created a LB rule with following rules having 2 listeners - 80:80 and 81:81
<img width="983" alt="screen shot 2015-07-22 at 1 14 55 pm" src="https://cloud.githubusercontent.com/assets/4266958/8838586/7aa2358c-3082-11e5-895c-771ea7479474.png">
There are no acl rules for frontend - 81 programmed in haproxy and all only 1 backend with entries from all the services are present.
<img width="676" alt="screen shot 2015-07-22 at 2 57 43 pm" src="https://cloud.githubusercontent.com/assets/4266958/8838642/fee947fe-3082-11e5-9189-812560613ac4.png">
LB redirects traffic to all services when accessing - wget --header="host: abc" http://104.154.66.248:81/service3.html
| non_infrastructure | acl rules are not being programmed for one of the listeners when the lb has multiple listeners server version steps to reproduce the problem created a lb rule with following rules having listeners and img width alt screen shot at pm src there are no acl rules for frontend programmed in haproxy and all only backend with entries from all the services are present img width alt screen shot at pm src lb redirects traffic to all services when accessing wget header host abc | 0 |
121,441 | 4,816,619,970 | IssuesEvent | 2016-11-04 10:43:32 | fetlife/android | https://api.github.com/repos/fetlife/android | closed | About page force closes app | bug priority high | I just got the newest version of the app and when I try to go to the about page I get a force close issue with the app. | 1.0 | About page force closes app - I just got the newest version of the app and when I try to go to the about page I get a force close issue with the app. | non_infrastructure | about page force closes app i just got the newest version of the app and when i try to go to the about page i get a force close issue with the app | 0 |
32,388 | 26,662,301,741 | IssuesEvent | 2023-01-25 22:22:48 | ForNeVeR/wpf-math | https://api.github.com/repos/ForNeVeR/wpf-math | opened | Separate NuGet package for WpfMath.Shared | kind:infrastructure | WpfMath.Shared assembly has been created in #327.
When publishing a new major release with Avalonia, we'll need to move it into a separate NuGet package.
Look for `TODO[#xxx]` in the repository to find some relevant code. | 1.0 | Separate NuGet package for WpfMath.Shared - WpfMath.Shared assembly has been created in #327.
When publishing a new major release with Avalonia, we'll need to move it into a separate NuGet package.
Look for `TODO[#xxx]` in the repository to find some relevant code. | infrastructure | separate nuget package for wpfmath shared wpfmath shared assembly has been created in when publishing a new major release with avalonia we ll need to move it into a separate nuget package look for todo in the repository to find some relevant code | 1 |
239,972 | 18,289,388,070 | IssuesEvent | 2021-10-05 13:47:34 | web3community/blockchain-dev-path | https://api.github.com/repos/web3community/blockchain-dev-path | opened | docs: Fill some content for the advance level #19 | documentation hacktoberfest | > This is the second level before [medium](https://github.com/web3community/blockchain-dev-path/blob/main/learning-guide/medium.md)
We need content for Blockchain Dev Path (advance level)
Link: https://github.com/web3community/blockchain-dev-path/blob/main/learning-guide/advance.md
You can suggest below | 1.0 | docs: Fill some content for the advance level #19 - > This is the second level before [medium](https://github.com/web3community/blockchain-dev-path/blob/main/learning-guide/medium.md)
We need content for Blockchain Dev Path (advance level)
Link: https://github.com/web3community/blockchain-dev-path/blob/main/learning-guide/advance.md
You can suggest below | non_infrastructure | docs fill some content for the advance level this is the second level before we need content for blockchain dev path advance level link you can suggest below | 0 |
22,205 | 15,044,503,310 | IssuesEvent | 2021-02-03 03:08:23 | ignitionrobotics/docs | https://api.github.com/repos/ignitionrobotics/docs | opened | Add issue and pull request templates to all code repositories | infrastructure | Templates help contributors ticket more straightforward issues and pull requests. We should add them to all our code repositories.
## How
Here's some GitHub documentation about it:
https://docs.github.com/en/github/building-a-strong-community/manually-creating-a-single-issue-template-for-your-repository
I added some rough examples to the testing repository, see all the `TEMPLATE`s:
https://github.com/ignitionrobotics/testing/tree/main/.github
The two options show up when ticketing an issue:
https://github.com/ignitionrobotics/testing/issues/new/choose
Opening a PR uses the default pull request template (the one outside the folder):
https://github.com/ignitionrobotics/testing/compare/main...another_branch
And the other PR templates can be accessed using the `?template=` query (unfortunately, [not through the UI yet](https://github.community/t/multiple-pull-request-templates/1850)):
* https://github.com/ignitionrobotics/testing/compare/main...another_branch?template=bug_fix.md
* https://github.com/ignitionrobotics/testing/compare/main...another_branch?template=new_feature.md
* https://github.com/ignitionrobotics/testing/compare/main...another_branch?template=port.md
Unfortunately, it looks like the automated title and labels aren't working for PR templates :confused:
## Content
We should have 2 issue templates: bugs and features, and 3 PR templates: bugs, features and ports.
There are lots of nice guides online for what to include in these templates. I'll brainstorm some ideas below and mention some Ignition-specific things.
### Issues
#### Bug
* User environment (versions, installation method, OS...)
* Reproduction steps
* Screenshots, logs, backtraces...
* ...
### Feature
* Desired behaviour
* Implementation suggestion
* Alternatives considered
### PRs
#### Bug
* Issue that is fixed
* How to reproduce bug before and after
* DCO
* Added tests?
* Message to squash-merge
#### Feature
* Issue that is addressed
* How to try out the feature manually
* DCO
* Added tests?
* Added example world? Tutorial?
* Message to squash-merge
### Ports
* From branch
* To branch
* Comparison link
* Message to merge with commit (no squash or rebase)
| 1.0 | Add issue and pull request templates to all code repositories - Templates help contributors ticket more straightforward issues and pull requests. We should add them to all our code repositories.
## How
Here's some GitHub documentation about it:
https://docs.github.com/en/github/building-a-strong-community/manually-creating-a-single-issue-template-for-your-repository
I added some rough examples to the testing repository, see all the `TEMPLATE`s:
https://github.com/ignitionrobotics/testing/tree/main/.github
The two options show up when ticketing an issue:
https://github.com/ignitionrobotics/testing/issues/new/choose
Opening a PR uses the default pull request template (the one outside the folder):
https://github.com/ignitionrobotics/testing/compare/main...another_branch
And the other PR templates can be accessed using the `?template=` query (unfortunately, [not through the UI yet](https://github.community/t/multiple-pull-request-templates/1850)):
* https://github.com/ignitionrobotics/testing/compare/main...another_branch?template=bug_fix.md
* https://github.com/ignitionrobotics/testing/compare/main...another_branch?template=new_feature.md
* https://github.com/ignitionrobotics/testing/compare/main...another_branch?template=port.md
Unfortunately, it looks like the automated title and labels aren't working for PR templates :confused:
## Content
We should have 2 issue templates: bugs and features, and 3 PR templates: bugs, features and ports.
There are lots of nice guides online for what to include in these templates. I'll brainstorm some ideas below and mention some Ignition-specific things.
### Issues
#### Bug
* User environment (versions, installation method, OS...)
* Reproduction steps
* Screenshots, logs, backtraces...
* ...
### Feature
* Desired behaviour
* Implementation suggestion
* Alternatives considered
### PRs
#### Bug
* Issue that is fixed
* How to reproduce bug before and after
* DCO
* Added tests?
* Message to squash-merge
#### Feature
* Issue that is addressed
* How to try out the feature manually
* DCO
* Added tests?
* Added example world? Tutorial?
* Message to squash-merge
### Ports
* From branch
* To branch
* Comparison link
* Message to merge with commit (no squash or rebase)
| infrastructure | add issue and pull request templates to all code repositories templates help contributors ticket more straightforward issues and pull requests we should add them to all our code repositories how here s some github documentation about it i added some rough examples to the testing repository see all the template s the two options show up when ticketing an issue opening a pr uses the default pull request template the one outside the folder and the other pr templates can be accessed using the template query unfortunately unfortunately it looks like the automated title and labels aren t working for pr templates confused content we should have issue templates bugs and features and pr templates bugs features and ports there are lots of nice guides online for what to include in these templates i ll brainstorm some ideas below and mention some ignition specific things issues bug user environment versions installation method os reproduction steps screenshots logs backtraces feature desired behaviour implementation suggestion alternatives considered prs bug issue that is fixed how to reproduce bug before and after dco added tests message to squash merge feature issue that is addressed how to try out the feature manually dco added tests added example world tutorial message to squash merge ports from branch to branch comparison link message to merge with commit no squash or rebase | 1 |
215,820 | 16,618,196,382 | IssuesEvent | 2021-06-02 19:43:27 | NRCan/CanFlood | https://api.github.com/repos/NRCan/CanFlood | closed | [CLOSED] How to install with the OSGeo4W Network Installer | documentation migrate | <a href="https://github.com/cefect"><img src="https://avatars.githubusercontent.com/u/47092034?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [cefect](https://github.com/cefect)**
_Wednesday Aug 19, 2020 at 00:21 GMT # Sample: Friday Sep 13, 2013 at 22:58 GMT_
_Originally opened as https://github.com/IBIGroupCanWest/CanFlood/issues/47_
----
download the [OSGeo4W Network Installer](https://qgis.org/en/site/forusers/download.html)
select 'Advanced Install option'
On the 'Select Packages' page, type 'qgis-ltr-full' into the search bar, expand 'Desktop' and ensure the 'New' (or 'Current') version shows 3.10-2 or higher.
OPTIONAL: check the dependency versions listed [here ](https://github.com/IBIGroupCanWest/CanFlood/tree/master/requirements) (e.g. type 'python3-pandas' into the search bar, and ensure 'New' shows 1.0.1-1 or higher.)
Click through and this should install Qgis 3.10 nicely
| 1.0 | [CLOSED] How to install with the OSGeo4W Network Installer - <a href="https://github.com/cefect"><img src="https://avatars.githubusercontent.com/u/47092034?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [cefect](https://github.com/cefect)**
_Wednesday Aug 19, 2020 at 00:21 GMT # Sample: Friday Sep 13, 2013 at 22:58 GMT_
_Originally opened as https://github.com/IBIGroupCanWest/CanFlood/issues/47_
----
download the [OSGeo4W Network Installer](https://qgis.org/en/site/forusers/download.html)
select 'Advanced Install option'
On the 'Select Packages' page, type 'qgis-ltr-full' into the search bar, expand 'Desktop' and ensure the 'New' (or 'Current') version shows 3.10-2 or higher.
OPTIONAL: check the dependency versions listed [here ](https://github.com/IBIGroupCanWest/CanFlood/tree/master/requirements) (e.g. type 'python3-pandas' into the search bar, and ensure 'New' shows 1.0.1-1 or higher.)
Click through and this should install Qgis 3.10 nicely
| non_infrastructure | how to install with the network installer issue by wednesday aug at gmt sample friday sep at gmt originally opened as download the select advanced install option on the select packages page type qgis ltr full into the search bar expand desktop and ensure the new or current version shows or higher optional check the dependency versions listed e g type pandas into the search bar and ensure new shows or higher click through and this should install qgis nicely | 0 |
259,836 | 22,554,486,906 | IssuesEvent | 2022-06-27 09:01:50 | wazuh/wazuh | https://api.github.com/repos/wazuh/wazuh | closed | Release 4.3.5 - Release Candidate 1 - Register | release test/4.3.5 | ### Packages tests metrics information
| | |
|---------------------------------|--------------------------------------------|
| **Main release candidate issue** | #13966 |
| **Main packages metrics issue** | #13971 |
| **Version** | 4.3.5 |
| **Release candidate #** | RC1 |
| **Tag** | https://github.com/wazuh/wazuh/tree/v4.3.5-rc1 |
### Packages used
- Repository: `packages-dev.wazuh.com`
- Package path: `pre-release`
- Package revision: `1`
---
| AMD64 Build |
| -- |
| https://ci.wazuh.info/view/Tests/job/Test_registration_tier/36/ |
| https://ci.wazuh.info/view/Tests/job/Test_registration/7856/ |
---
| System | Test Result |
| -- | -- |
| CentOS 8 | 🟢 |
| CentOS 7 | 🟢 |
| CentOS 6 | 🟢 |
| CentOS 5 | 🟢 |
| Debian 10 | 🟢 |
| Debian 9 | 🟢 |
| Debian 8 | 🟢 |
| Debian 7 | 🟢 |
| Fedora 32 | 🟢 |
| Fedora 31 | 🟢 |
| Ubuntu Focal | 🟢 |
| Ubuntu Bionic | 🟢 |
| Ubuntu Xenial | 🟢 |
| Ubuntu Precise | 🟢 |
| Ubuntu Trusty | 🟢 |
| Amazon Linux 1 | 🟢 |
| Amazon Linux 2 | 🟢 |
| Oracle Linux 6 | 🟢 |
| Oracle Linux 7 | 🟢 |
| Oracle Linux 8 | 🟢 |
| OpenSuse Tumbleweed | 🟢 |
| Redhat 6 | 🟢 |
| Redhat 7 | 🟢 |
| Redhat 8 | 🟢 |
Legend:
Pending/In progress: 🕐
Test not launched: ⚫
Test failed: 🔴
Test success: 🟢
Need review: 🟡 | 1.0 | Release 4.3.5 - Release Candidate 1 - Register - ### Packages tests metrics information
| | |
|---------------------------------|--------------------------------------------|
| **Main release candidate issue** | #13966 |
| **Main packages metrics issue** | #13971 |
| **Version** | 4.3.5 |
| **Release candidate #** | RC1 |
| **Tag** | https://github.com/wazuh/wazuh/tree/v4.3.5-rc1 |
### Packages used
- Repository: `packages-dev.wazuh.com`
- Package path: `pre-release`
- Package revision: `1`
---
| AMD64 Build |
| -- |
| https://ci.wazuh.info/view/Tests/job/Test_registration_tier/36/ |
| https://ci.wazuh.info/view/Tests/job/Test_registration/7856/ |
---
| System | Test Result |
| -- | -- |
| CentOS 8 | 🟢 |
| CentOS 7 | 🟢 |
| CentOS 6 | 🟢 |
| CentOS 5 | 🟢 |
| Debian 10 | 🟢 |
| Debian 9 | 🟢 |
| Debian 8 | 🟢 |
| Debian 7 | 🟢 |
| Fedora 32 | 🟢 |
| Fedora 31 | 🟢 |
| Ubuntu Focal | 🟢 |
| Ubuntu Bionic | 🟢 |
| Ubuntu Xenial | 🟢 |
| Ubuntu Precise | 🟢 |
| Ubuntu Trusty | 🟢 |
| Amazon Linux 1 | 🟢 |
| Amazon Linux 2 | 🟢 |
| Oracle Linux 6 | 🟢 |
| Oracle Linux 7 | 🟢 |
| Oracle Linux 8 | 🟢 |
| OpenSuse Tumbleweed | 🟢 |
| Redhat 6 | 🟢 |
| Redhat 7 | 🟢 |
| Redhat 8 | 🟢 |
Legend:
Pending/In progress: 🕐
Test not launched: ⚫
Test failed: 🔴
Test success: 🟢
Need review: 🟡 | non_infrastructure | release release candidate register packages tests metrics information main release candidate issue main packages metrics issue version release candidate tag packages used repository packages dev wazuh com package path pre release package revision build system test result centos 🟢 centos 🟢 centos 🟢 centos 🟢 debian 🟢 debian 🟢 debian 🟢 debian 🟢 fedora 🟢 fedora 🟢 ubuntu focal 🟢 ubuntu bionic 🟢 ubuntu xenial 🟢 ubuntu precise 🟢 ubuntu trusty 🟢 amazon linux 🟢 amazon linux 🟢 oracle linux 🟢 oracle linux 🟢 oracle linux 🟢 opensuse tumbleweed 🟢 redhat 🟢 redhat 🟢 redhat 🟢 legend pending in progress 🕐 test not launched ⚫ test failed 🔴 test success 🟢 need review 🟡 | 0 |
10,515 | 8,603,237,913 | IssuesEvent | 2018-11-16 16:17:08 | uclibs/ucrate | https://api.github.com/repos/uclibs/ucrate | closed | Make the needed changes to deploy to servers | status: review type: infrastructure | ### Descriptive summary
(We've already started a `scholar-deploy` branch for this)
Make the needed changes to this repo and the UCIT servers to support deploying the application via Bamboo. A deploy to scholar-dev.uc.edu has already been done, but more tweaks are needed.
We will merge the `scholar-deploy` branch once everything is working. | 1.0 | Make the needed changes to deploy to servers - ### Descriptive summary
(We've already started a `scholar-deploy` branch for this)
Make the needed changes to this repo and the UCIT servers to support deploying the application via Bamboo. A deploy to scholar-dev.uc.edu has already been done, but more tweaks are needed.
We will merge the `scholar-deploy` branch once everything is working. | infrastructure | make the needed changes to deploy to servers descriptive summary we ve already started a scholar deploy branch for this make the needed changes to this repo and the ucit servers to support deploying the application via bamboo a deploy to scholar dev uc edu has already been done but more tweaks are needed we will merge the scholar deploy branch once everything is working | 1 |
149,463 | 13,281,767,267 | IssuesEvent | 2020-08-23 19:01:17 | sebschlicht/ansible-nas | https://api.github.com/repos/sebschlicht/ansible-nas | closed | Document configuration options | documentation feature | * dynv6: hostname + token
* minidlna: friendly_name, list of media_dirs
* openssh: SSH key paths per user (determines AllowUsers)
* nas: unattended_upgrades_mail_recipient, send mail setup, mounts (base, mount point, UUID), users (name, SSH key, initial password, groups, private shares), public shares, mirroring, hostname
* name + ask-sudo-pw/pass SSH key for remote admin user | 1.0 | Document configuration options - * dynv6: hostname + token
* minidlna: friendly_name, list of media_dirs
* openssh: SSH key paths per user (determines AllowUsers)
* nas: unattended_upgrades_mail_recipient, send mail setup, mounts (base, mount point, UUID), users (name, SSH key, initial password, groups, private shares), public shares, mirroring, hostname
* name + ask-sudo-pw/pass SSH key for remote admin user | non_infrastructure | document configuration options hostname token minidlna friendly name list of media dirs openssh ssh key paths per user determines allowusers nas unattended upgrades mail recipient send mail setup mounts base mount point uuid users name ssh key initial password groups private shares public shares mirroring hostname name ask sudo pw pass ssh key for remote admin user | 0 |
32,958 | 27,109,137,886 | IssuesEvent | 2023-02-15 14:13:38 | Anthony-Nolan/Atlas | https://api.github.com/repos/Anthony-Nolan/Atlas | reopened | Test Pipeline should be runnable outside of AN environment | bug automated-tests infrastructure | **Describe the bug**
Test Pipeline uses unique resource names that are already deployed in AN's environment. This results in test pipeline failures during the Terraform and Matching Algorithm Validation stages.
**To Reproduce**
Steps to reproduce the behavior:
1. Use test-pipeline.yml to create a new Pipeline in Azure DevOps
2. Create a Variable Group named 'Terraform', and update with Client Id / Secret / Access keys necessary for Terraform to connect to Azure.
3. Run pipeline
4. Pipeline fails at the Terraform (unique resource names) and Matching Algorithm Valdation (no access to AN nuget feed).
**Expected behaviour**
Test pipeline should successfully deploy resource to the test resource group and be able to fetch all packages needed for tests to run successfully.
**Screenshots**
N/A
**Inputs/Outputs**
Set the following values in the Terraform variable group:
ARM_ACCESS_KEY: [access key to storage account]
ARM_CLIENT_ID: [client id used by terraform to connect to resources]
ARM_CLIENT_SECRET: [client secret used by terraform to connect resources]
ARM_TENANT_ID: [azure tenant id test resources are deployed to[
BACKEND_RESOURCE_GROUP_NAME: [Name of the back-end resource group]
BACKEND_STORAGE_ACCOUNT_NAME: [Name of the storage account used by terraform]
BACKEND_STORAGE_CONTAINER_NAME: [Name of the storage container used by terraform]
**Atlas Build & Runtime Info (please complete the following information):**
- Runtime Environment: Azure Cloud
- Atlas Version: 1.4.2
- GitHub commit ID: e8bf95de65ce445f52231a8185fe607f8f1e0e7d
**Additional context**
This issue was discovered while running test pipelines in WMDA's DevOps instance
| 1.0 | Test Pipeline should be runnable outside of AN environment - **Describe the bug**
Test Pipeline uses unique resource names that are already deployed in AN's environment. This results in test pipeline failures during the Terraform and Matching Algorithm Validation stages.
**To Reproduce**
Steps to reproduce the behavior:
1. Use test-pipeline.yml to create a new Pipeline in Azure DevOps
2. Create a Variable Group named 'Terraform', and update with Client Id / Secret / Access keys necessary for Terraform to connect to Azure.
3. Run pipeline
4. Pipeline fails at the Terraform (unique resource names) and Matching Algorithm Valdation (no access to AN nuget feed).
**Expected behaviour**
Test pipeline should successfully deploy resource to the test resource group and be able to fetch all packages needed for tests to run successfully.
**Screenshots**
N/A
**Inputs/Outputs**
Set the following values in the Terraform variable group:
ARM_ACCESS_KEY: [access key to storage account]
ARM_CLIENT_ID: [client id used by terraform to connect to resources]
ARM_CLIENT_SECRET: [client secret used by terraform to connect resources]
ARM_TENANT_ID: [azure tenant id test resources are deployed to[
BACKEND_RESOURCE_GROUP_NAME: [Name of the back-end resource group]
BACKEND_STORAGE_ACCOUNT_NAME: [Name of the storage account used by terraform]
BACKEND_STORAGE_CONTAINER_NAME: [Name of the storage container used by terraform]
**Atlas Build & Runtime Info (please complete the following information):**
- Runtime Environment: Azure Cloud
- Atlas Version: 1.4.2
- GitHub commit ID: e8bf95de65ce445f52231a8185fe607f8f1e0e7d
**Additional context**
This issue was discovered while running test pipelines in WMDA's DevOps instance
| infrastructure | test pipeline should be runnable outside of an environment describe the bug test pipeline uses unique resource names that are already deployed in an s environment this results in test pipeline failures during the terraform and matching algorithm validation stages to reproduce steps to reproduce the behavior use test pipeline yml to create a new pipeline in azure devops create a variable group named terraform and update with client id secret access keys necessary for terraform to connect to azure run pipeline pipeline fails at the terraform unique resource names and matching algorithm valdation no access to an nuget feed expected behaviour test pipeline should successfully deploy resource to the test resource group and be able to fetch all packages needed for tests to run successfully screenshots n a inputs outputs set the following values in the terraform variable group arm access key arm client id arm client secret arm tenant id azure tenant id test resources are deployed to backend resource group name backend storage account name backend storage container name atlas build runtime info please complete the following information runtime environment azure cloud atlas version github commit id additional context this issue was discovered while running test pipelines in wmda s devops instance | 1 |
12,367 | 9,747,660,215 | IssuesEvent | 2019-06-03 14:49:45 | cockroachdb/docs | https://api.github.com/repos/cockroachdb/docs | opened | Translation research | A-docs-infrastructure A-general C-docs-project O-internal | Check into translation services that support Markdown and Github integration and check on initial and continuing translation costs for select languages. | 1.0 | Translation research - Check into translation services that support Markdown and Github integration and check on initial and continuing translation costs for select languages. | infrastructure | translation research check into translation services that support markdown and github integration and check on initial and continuing translation costs for select languages | 1 |
22,573 | 15,285,463,676 | IssuesEvent | 2021-02-23 13:35:36 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | NuGet restore failure reading feed information | area-Infrastructure-libraries untriaged | <!-- runfo report start -->
Runfo Tracking Issue: [NuGet restore failure reading feed information](https://runfo.azurewebsites.net/tracking/issue/69)
|Definition|Build|Kind|Job Name|
|---|---|---|---|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Android x64 Release AllSubsets_Mono_RuntimeTests|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Browser wasm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Browser wasm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Browser wasm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Browser wasm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887684](https://dev.azure.com/dnceng/public/_build/results?buildId=887684)|[PR 44675](https://github.com/dotnet/runtime/pull/44675)|CoreCLR Product Build windows arm release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887676](https://dev.azure.com/dnceng/public/_build/results?buildId=887676)|[PR 44649](https://github.com/dotnet/runtime/pull/44649)|Build Android arm64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887613](https://dev.azure.com/dnceng/public/_build/results?buildId=887613)|[PR 44706](https://github.com/dotnet/runtime/pull/44706)|CoreCLR Product Build Linux arm64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887230](https://dev.azure.com/dnceng/public/_build/results?buildId=887230)|Rolling|Libraries Build windows allConfigurations x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Mono Product Build OSX x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Build tvOS x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Libraries Build OSX x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Mono Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Build tvOS arm64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Build iOS x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|CoreCLR Product Build OSX arm64 checked|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Build OSX x64 Release AllSubsets_Mono_LLVMJIT|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|CoreCLR Product Build OSX arm64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887152](https://dev.azure.com/dnceng/public/_build/results?buildId=887152)|[PR 42972](https://github.com/dotnet/runtime/pull/42972)|Libraries Build windows x86 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887146](https://dev.azure.com/dnceng/public/_build/results?buildId=887146)|[PR 44709](https://github.com/dotnet/runtime/pull/44709)|Libraries Build windows arm Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887102](https://dev.azure.com/dnceng/public/_build/results?buildId=887102)|[PR 44691](https://github.com/dotnet/runtime/pull/44691)|Build tvOS x64 Debug AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887072](https://dev.azure.com/dnceng/public/_build/results?buildId=887072)|[PR 44671](https://github.com/dotnet/runtime/pull/44671)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887072](https://dev.azure.com/dnceng/public/_build/results?buildId=887072)|[PR 44671](https://github.com/dotnet/runtime/pull/44671)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887069](https://dev.azure.com/dnceng/public/_build/results?buildId=887069)|[PR 44581](https://github.com/dotnet/runtime/pull/44581)|Libraries Build windows allConfigurations x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887052](https://dev.azure.com/dnceng/public/_build/results?buildId=887052)|[PR 44465](https://github.com/dotnet/runtime/pull/44465)|Build Browser wasm Release AllSubsets_Mono_RuntimeTests|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887052](https://dev.azure.com/dnceng/public/_build/results?buildId=887052)|[PR 44465](https://github.com/dotnet/runtime/pull/44465)|Build Browser wasm Release AllSubsets_Mono_RuntimeTests|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887052](https://dev.azure.com/dnceng/public/_build/results?buildId=887052)|[PR 44465](https://github.com/dotnet/runtime/pull/44465)|Mono Product Build windows x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887052](https://dev.azure.com/dnceng/public/_build/results?buildId=887052)|[PR 44465](https://github.com/dotnet/runtime/pull/44465)|Build Android arm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887005](https://dev.azure.com/dnceng/public/_build/results?buildId=887005)|[PR 43024](https://github.com/dotnet/runtime/pull/43024)|Installer Build and Test coreclr windows_x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[886999](https://dev.azure.com/dnceng/public/_build/results?buildId=886999)|[PR 44505](https://github.com/dotnet/runtime/pull/44505)|Build Android x64 Release AllSubsets_Mono_RuntimeTests|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[886999](https://dev.azure.com/dnceng/public/_build/results?buildId=886999)|[PR 44505](https://github.com/dotnet/runtime/pull/44505)|Libraries Build windows x86 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[886999](https://dev.azure.com/dnceng/public/_build/results?buildId=886999)|[PR 44505](https://github.com/dotnet/runtime/pull/44505)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[874440](https://dev.azure.com/dnceng/public/_build/results?buildId=874440)|[PR 39294](https://github.com/dotnet/runtime/pull/39294)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[874440](https://dev.azure.com/dnceng/public/_build/results?buildId=874440)|[PR 39294](https://github.com/dotnet/runtime/pull/39294)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[860374](https://dev.azure.com/dnceng/public/_build/results?buildId=860374)|[PR 42097](https://github.com/dotnet/runtime/pull/42097)|Libraries Test Run release coreclr OSX x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[860335](https://dev.azure.com/dnceng/public/_build/results?buildId=860335)|[PR 43601](https://github.com/dotnet/runtime/pull/43601)|CoreCLR Product Build OSX x64 checked|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[860291](https://dev.azure.com/dnceng/public/_build/results?buildId=860291)|[PR 43650](https://github.com/dotnet/runtime/pull/43650)|Build Android arm64 Debug AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[855865](https://dev.azure.com/dnceng/public/_build/results?buildId=855865)|[PR 43397](https://github.com/dotnet/runtime/pull/43397)|Installer Build and Test coreclr Windows_NT_x64 Release|
Build Result Summary
|Day Hit Count|Week Hit Count|Month Hit Count|
|---|---|---|
|0|0|0|
<!-- runfo report end -->
| 1.0 | NuGet restore failure reading feed information - <!-- runfo report start -->
Runfo Tracking Issue: [NuGet restore failure reading feed information](https://runfo.azurewebsites.net/tracking/issue/69)
|Definition|Build|Kind|Job Name|
|---|---|---|---|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Android x64 Release AllSubsets_Mono_RuntimeTests|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Browser wasm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Browser wasm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Browser wasm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[888041](https://dev.azure.com/dnceng/public/_build/results?buildId=888041)|[PR 44629](https://github.com/dotnet/runtime/pull/44629)|Build Browser wasm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887684](https://dev.azure.com/dnceng/public/_build/results?buildId=887684)|[PR 44675](https://github.com/dotnet/runtime/pull/44675)|CoreCLR Product Build windows arm release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887676](https://dev.azure.com/dnceng/public/_build/results?buildId=887676)|[PR 44649](https://github.com/dotnet/runtime/pull/44649)|Build Android arm64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887613](https://dev.azure.com/dnceng/public/_build/results?buildId=887613)|[PR 44706](https://github.com/dotnet/runtime/pull/44706)|CoreCLR Product Build Linux arm64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887230](https://dev.azure.com/dnceng/public/_build/results?buildId=887230)|Rolling|Libraries Build windows allConfigurations x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Mono Product Build OSX x64 debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Build tvOS x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Libraries Build OSX x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Mono Product Build OSX x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Build tvOS arm64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Build iOS x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|CoreCLR Product Build OSX arm64 checked|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|Build OSX x64 Release AllSubsets_Mono_LLVMJIT|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887161](https://dev.azure.com/dnceng/public/_build/results?buildId=887161)|Rolling|CoreCLR Product Build OSX arm64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887152](https://dev.azure.com/dnceng/public/_build/results?buildId=887152)|[PR 42972](https://github.com/dotnet/runtime/pull/42972)|Libraries Build windows x86 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887146](https://dev.azure.com/dnceng/public/_build/results?buildId=887146)|[PR 44709](https://github.com/dotnet/runtime/pull/44709)|Libraries Build windows arm Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887102](https://dev.azure.com/dnceng/public/_build/results?buildId=887102)|[PR 44691](https://github.com/dotnet/runtime/pull/44691)|Build tvOS x64 Debug AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887072](https://dev.azure.com/dnceng/public/_build/results?buildId=887072)|[PR 44671](https://github.com/dotnet/runtime/pull/44671)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887072](https://dev.azure.com/dnceng/public/_build/results?buildId=887072)|[PR 44671](https://github.com/dotnet/runtime/pull/44671)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887069](https://dev.azure.com/dnceng/public/_build/results?buildId=887069)|[PR 44581](https://github.com/dotnet/runtime/pull/44581)|Libraries Build windows allConfigurations x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887052](https://dev.azure.com/dnceng/public/_build/results?buildId=887052)|[PR 44465](https://github.com/dotnet/runtime/pull/44465)|Build Browser wasm Release AllSubsets_Mono_RuntimeTests|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887052](https://dev.azure.com/dnceng/public/_build/results?buildId=887052)|[PR 44465](https://github.com/dotnet/runtime/pull/44465)|Build Browser wasm Release AllSubsets_Mono_RuntimeTests|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887052](https://dev.azure.com/dnceng/public/_build/results?buildId=887052)|[PR 44465](https://github.com/dotnet/runtime/pull/44465)|Mono Product Build windows x64 release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887052](https://dev.azure.com/dnceng/public/_build/results?buildId=887052)|[PR 44465](https://github.com/dotnet/runtime/pull/44465)|Build Android arm Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[887005](https://dev.azure.com/dnceng/public/_build/results?buildId=887005)|[PR 43024](https://github.com/dotnet/runtime/pull/43024)|Installer Build and Test coreclr windows_x64 Release|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[886999](https://dev.azure.com/dnceng/public/_build/results?buildId=886999)|[PR 44505](https://github.com/dotnet/runtime/pull/44505)|Build Android x64 Release AllSubsets_Mono_RuntimeTests|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[886999](https://dev.azure.com/dnceng/public/_build/results?buildId=886999)|[PR 44505](https://github.com/dotnet/runtime/pull/44505)|Libraries Build windows x86 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[886999](https://dev.azure.com/dnceng/public/_build/results?buildId=886999)|[PR 44505](https://github.com/dotnet/runtime/pull/44505)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[874440](https://dev.azure.com/dnceng/public/_build/results?buildId=874440)|[PR 39294](https://github.com/dotnet/runtime/pull/39294)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[874440](https://dev.azure.com/dnceng/public/_build/results?buildId=874440)|[PR 39294](https://github.com/dotnet/runtime/pull/39294)|Build Android x64 Release AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[860374](https://dev.azure.com/dnceng/public/_build/results?buildId=860374)|[PR 42097](https://github.com/dotnet/runtime/pull/42097)|Libraries Test Run release coreclr OSX x64 Debug|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[860335](https://dev.azure.com/dnceng/public/_build/results?buildId=860335)|[PR 43601](https://github.com/dotnet/runtime/pull/43601)|CoreCLR Product Build OSX x64 checked|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[860291](https://dev.azure.com/dnceng/public/_build/results?buildId=860291)|[PR 43650](https://github.com/dotnet/runtime/pull/43650)|Build Android arm64 Debug AllSubsets_Mono|
|[runtime](https://dnceng.visualstudio.com/public/_build?definitionId=686)|[855865](https://dev.azure.com/dnceng/public/_build/results?buildId=855865)|[PR 43397](https://github.com/dotnet/runtime/pull/43397)|Installer Build and Test coreclr Windows_NT_x64 Release|
Build Result Summary
|Day Hit Count|Week Hit Count|Month Hit Count|
|---|---|---|
|0|0|0|
<!-- runfo report end -->
| infrastructure | nuget restore failure reading feed information runfo tracking issue definition build kind job name android release allsubsets mono runtimetests browser wasm release allsubsets mono browser wasm release allsubsets mono browser wasm release allsubsets mono browser wasm release allsubsets mono product build windows arm release android release allsubsets mono product build linux release build windows allconfigurations release product build osx debug tvos release allsubsets mono build osx release product build osx release tvos release allsubsets mono ios release allsubsets mono product build osx checked osx release allsubsets mono llvmjit product build osx release build windows release build windows arm release tvos debug allsubsets mono android release allsubsets mono android release allsubsets mono build windows allconfigurations debug browser wasm release allsubsets mono runtimetests browser wasm release allsubsets mono runtimetests product build windows release android arm release allsubsets mono build and test coreclr windows release android release allsubsets mono runtimetests build windows debug android release allsubsets mono android release allsubsets mono android release allsubsets mono test run release coreclr osx debug product build osx checked android debug allsubsets mono build and test coreclr windows nt release build result summary day hit count week hit count month hit count | 1 |
43,047 | 11,141,171,025 | IssuesEvent | 2019-12-21 20:19:12 | neuronsimulator/nrn | https://api.github.com/repos/neuronsimulator/nrn | closed | Build issues with GCC v9 | bug building | While building NEURON with CMake using GCC 9 (or 9.2), I am seeing below errors:
```
$ cmake .. -DNRN_ENABLE_CORENEURON=OFF -DNRN_ENABLE_INTERVIEWS=OFF
....
-- The C compiler identification is GNU 9.2.0
-- The CXX compiler identification is GNU 9.2.0
-- Check for working C compiler: /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/gcc
-- Check for working C compiler: /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/gcc -- works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Detecting C compile features
...
--
-- Configured NEURON 7.8
--
-- Some things you can do now:
-- --------------+--------------------------------------------------------------
-- Command | Description
-- --------------+--------------------------------------------------------------
-- make install | Will install NEURON to: /usr/local
-- | Change the install location of NEURON using:
-- | cmake <src_path> -DCMAKE_INSTALL_PREFIX=<install_path>
-- make docs | Build the API documentation
-- make uninstall| Removes files installed by make install (todo)
-- --------------+--------------------------------------------------------------
-- Build option | Status
-- --------------+--------------------------------------------------------------
-- Shared | ON
-- Legacy FR | ON
-- MPI | ON
-- INC | /opt/hpe/hpc/mpt/mpt-2.16/include
-- LIB | /lib64/libpthread.so
-- Python | ON
-- EXE | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/bin/python
-- INC | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/include/python3.6m
-- LIB | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/lib/libpython3.6m.so
-- DYNAMIC | OFF
-- RX3D | ON
-- OptLevel | 0
-- Interviews | OFF
-- CoreNEURON | OFF
-- --------------+--------------------------------------------------------------
-- See documentation : https://www.neuron.yale.edu/neuron/
-- --------------+--------------------------------------------------------------
--
-- Configuring done
-- Generating done
-- Build files have been written to: /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build
...
$ make VERBOSE=1
/gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/tools/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/cmake-3.13.0-ivumccos42/bin/cmake -S/gpfs/bbp.cscs.ch/home/kumbhar/tmp/
....
[ 6%] Building CXX object src/nrniv/CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o
cd /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrniv && /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/g++ -DHAVE_CONFIG_H -Dnrniv_lib_EXPORTS -DCABLE=1 -DOOP=1 -DUSEBBS=1 -DUSECVODE=1 -DUSEMATRIX=1 -I/opt/hpe/hpc/mpt/mpt-2.16/include -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/oc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrncvode -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnjava -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnpython -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/oc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/parallel -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/sundials -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/sundials/shared -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/gnu -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/memacs -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/mesch -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrncvode -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnmpi -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnpython -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/parallel -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sparse -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sparse13 -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/cvodes -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/ida -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/shared -I/gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/include/python3.6m -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/nrniv_lib -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivos -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/ivos -fPIC -std=c++98 -o CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o -c /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp
In file included from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected template-name before ‘<’ token
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected ‘{’ before ‘<’ token
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected unqualified-id before ‘<’ token
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In constructor ‘BBSDirect::BBSDirect()’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:48:27: error: invalid use of incomplete type ‘class KeepArgs’
48 | keepargs_ = new KeepArgs();
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In destructor ‘virtual BBSDirect::~BBSDirect()’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: warning: possible problem detected in invocation of delete operator: [-Wdelete-incomplete]
56 | delete keepargs_;
| ^~~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: warning: invalid use of incomplete type ‘class KeepArgs’
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: note: neither the destructor nor the class-specific operator delete will be called, even if they are declared when the class is defined
56 | delete keepargs_;
| ^~~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In member function ‘virtual void BBSDirect::save_args(int)’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:269:11: error: invalid use of incomplete type ‘class KeepArgs’
269 | keepargs_->insert(
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:3: error: ‘pair’ was not declared in this scope; did you mean ‘std::pair’?
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^~~~
| std::pair
In file included from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/stl_algobase.h:64,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/char_traits.h:39,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/ios:40,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/ostream:38,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/iostream:39,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivos/ivstream.h:4,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrncvode/../../nrnconf.h:370,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:1,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/stl_pair.h:208:12: note: ‘std::pair’ declared here
208 | struct pair
| ^~~~
In file included from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:8: error: expected primary-expression before ‘const’
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:28: error: expected primary-expression before ‘*’ token
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:29: error: expected primary-expression before ‘>’ token
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In member function ‘virtual void BBSDirect::return_args(int)’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:279:12: error: incomplete type ‘KeepArgs’ used in nested name specifier
279 | KeepArgs::iterator i = keepargs_->find(userid);
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:282:6: error: ‘i’ was not declared in this scope
282 | if (i != keepargs_->end()) {
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:282:20: error: invalid use of incomplete type ‘class KeepArgs’
282 | if (i != keepargs_->end()) {
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:284:12: error: invalid use of incomplete type ‘class KeepArgs’
284 | keepargs_->erase(i);
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
make[2]: *** [src/nrniv/CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o] Error 1
make[2]: Leaving directory `/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build'
make[1]: *** [src/nrniv/CMakeFiles/nrniv_lib.dir/all] Error 2
make[1]: Leaving directory `/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build'
make: *** [all] Error 2
```
This seems to be related to #358 but I am surprised that this is only appearing with GCC v9.
I am shorty creating PR with the fix.
cc: @anilbey
| 1.0 | Build issues with GCC v9 - While building NEURON with CMake using GCC 9 (or 9.2), I am seeing below errors:
```
$ cmake .. -DNRN_ENABLE_CORENEURON=OFF -DNRN_ENABLE_INTERVIEWS=OFF
....
-- The C compiler identification is GNU 9.2.0
-- The CXX compiler identification is GNU 9.2.0
-- Check for working C compiler: /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/gcc
-- Check for working C compiler: /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/gcc -- works
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Detecting C compile features
...
--
-- Configured NEURON 7.8
--
-- Some things you can do now:
-- --------------+--------------------------------------------------------------
-- Command | Description
-- --------------+--------------------------------------------------------------
-- make install | Will install NEURON to: /usr/local
-- | Change the install location of NEURON using:
-- | cmake <src_path> -DCMAKE_INSTALL_PREFIX=<install_path>
-- make docs | Build the API documentation
-- make uninstall| Removes files installed by make install (todo)
-- --------------+--------------------------------------------------------------
-- Build option | Status
-- --------------+--------------------------------------------------------------
-- Shared | ON
-- Legacy FR | ON
-- MPI | ON
-- INC | /opt/hpe/hpc/mpt/mpt-2.16/include
-- LIB | /lib64/libpthread.so
-- Python | ON
-- EXE | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/bin/python
-- INC | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/include/python3.6m
-- LIB | /gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/lib/libpython3.6m.so
-- DYNAMIC | OFF
-- RX3D | ON
-- OptLevel | 0
-- Interviews | OFF
-- CoreNEURON | OFF
-- --------------+--------------------------------------------------------------
-- See documentation : https://www.neuron.yale.edu/neuron/
-- --------------+--------------------------------------------------------------
--
-- Configuring done
-- Generating done
-- Build files have been written to: /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build
...
$ make VERBOSE=1
/gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/tools/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/cmake-3.13.0-ivumccos42/bin/cmake -S/gpfs/bbp.cscs.ch/home/kumbhar/tmp/
....
[ 6%] Building CXX object src/nrniv/CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o
cd /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrniv && /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/bin/g++ -DHAVE_CONFIG_H -Dnrniv_lib_EXPORTS -DCABLE=1 -DOOP=1 -DUSEBBS=1 -DUSECVODE=1 -DUSEMATRIX=1 -I/opt/hpe/hpc/mpt/mpt-2.16/include -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/oc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrncvode -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnjava -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnoc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrnpython -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/oc -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/parallel -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/sundials -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/sundials/shared -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/gnu -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/memacs -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/mesch -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrncvode -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnmpi -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrnpython -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/parallel -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sparse -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sparse13 -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/cvodes -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/ida -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/sundials/shared -I/gpfs/bbp.cscs.ch/apps/hpc/jenkins/deploy/external-libraries/2018-12-19/linux-rhel7-x86_64/gcc-6.4.0/python-3.6.5-ukuow6gd2f/include/python3.6m -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/nrniv_lib -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivos -I/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/ivos -fPIC -std=c++98 -o CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o -c /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp
In file included from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected template-name before ‘<’ token
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected ‘{’ before ‘<’ token
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:28: error: expected unqualified-id before ‘<’ token
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In constructor ‘BBSDirect::BBSDirect()’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:48:27: error: invalid use of incomplete type ‘class KeepArgs’
48 | keepargs_ = new KeepArgs();
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In destructor ‘virtual BBSDirect::~BBSDirect()’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: warning: possible problem detected in invocation of delete operator: [-Wdelete-incomplete]
56 | delete keepargs_;
| ^~~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: warning: invalid use of incomplete type ‘class KeepArgs’
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:56:9: note: neither the destructor nor the class-specific operator delete will be called, even if they are declared when the class is defined
56 | delete keepargs_;
| ^~~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In member function ‘virtual void BBSDirect::save_args(int)’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:269:11: error: invalid use of incomplete type ‘class KeepArgs’
269 | keepargs_->insert(
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:3: error: ‘pair’ was not declared in this scope; did you mean ‘std::pair’?
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^~~~
| std::pair
In file included from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/stl_algobase.h:64,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/char_traits.h:39,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/ios:40,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/ostream:38,
from /gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/iostream:39,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/ivos/ivstream.h:4,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build/src/nrncvode/../../nrnconf.h:370,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:1,
from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/apps/hpc/jenkins/i_would_like_to_be_deleted---merge-2019-10/deploy/compilers/2019-10-02/linux-rhel7-x86_64/gcc-4.8.5/gcc-9.2.0-tuxi4uvc6t/include/c++/9.2.0/bits/stl_pair.h:208:12: note: ‘std::pair’ declared here
208 | struct pair
| ^~~~
In file included from /gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/bbsdirect.cpp:3:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:8: error: expected primary-expression before ‘const’
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:28: error: expected primary-expression before ‘*’ token
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:270:29: error: expected primary-expression before ‘>’ token
270 | pair<const int, bbsmpibuf*>(userid, sendbuf_)
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp: In member function ‘virtual void BBSDirect::return_args(int)’:
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:279:12: error: incomplete type ‘KeepArgs’ used in nested name specifier
279 | KeepArgs::iterator i = keepargs_->find(userid);
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:282:6: error: ‘i’ was not declared in this scope
282 | if (i != keepargs_->end()) {
| ^
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:282:20: error: invalid use of incomplete type ‘class KeepArgs’
282 | if (i != keepargs_->end()) {
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:284:12: error: invalid use of incomplete type ‘class KeepArgs’
284 | keepargs_->erase(i);
| ^~
/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/src/nrniv/../parallel/bbsdirectmpi.cpp:35:7: note: forward declaration of ‘class KeepArgs’
35 | class KeepArgs : public map<int, bbsmpibuf*, ltint>{};
| ^~~~~~~~
make[2]: *** [src/nrniv/CMakeFiles/nrniv_lib.dir/bbsdirect.cpp.o] Error 1
make[2]: Leaving directory `/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build'
make[1]: *** [src/nrniv/CMakeFiles/nrniv_lib.dir/all] Error 2
make[1]: Leaving directory `/gpfs/bbp.cscs.ch/home/kumbhar/tmp/nrn/build'
make: *** [all] Error 2
```
This seems to be related to #358 but I am surprised that this is only appearing with GCC v9.
I am shorty creating PR with the fix.
cc: @anilbey
| non_infrastructure | build issues with gcc while building neuron with cmake using gcc or i am seeing below errors cmake dnrn enable coreneuron off dnrn enable interviews off the c compiler identification is gnu the cxx compiler identification is gnu check for working c compiler gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc bin gcc check for working c compiler gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc bin gcc works detecting c compiler abi info detecting c compiler abi info done detecting c compile features configured neuron some things you can do now command description make install will install neuron to usr local change the install location of neuron using cmake dcmake install prefix make docs build the api documentation make uninstall removes files installed by make install todo build option status shared on legacy fr on mpi on inc opt hpe hpc mpt mpt include lib libpthread so python on exe gpfs bbp cscs ch apps hpc jenkins deploy external libraries linux gcc python bin python inc gpfs bbp cscs ch apps hpc jenkins deploy external libraries linux gcc python include lib gpfs bbp cscs ch apps hpc jenkins deploy external libraries linux gcc python lib so dynamic off on optlevel interviews off coreneuron off see documentation configuring done generating done build files have been written to gpfs bbp cscs ch home kumbhar tmp nrn build make verbose gpfs bbp cscs ch apps hpc jenkins deploy tools linux gcc cmake bin cmake s gpfs bbp cscs ch home kumbhar tmp building cxx object src nrniv cmakefiles nrniv lib dir bbsdirect cpp o cd gpfs bbp cscs ch home kumbhar tmp nrn build src nrniv gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc bin g dhave config h dnrniv lib exports dcable doop dusebbs dusecvode dusematrix i opt hpe hpc mpt mpt include i gpfs bbp cscs ch home kumbhar tmp nrn src ivoc i gpfs bbp cscs ch home kumbhar tmp nrn src nrniv i gpfs bbp cscs ch home kumbhar tmp nrn src nrnoc i gpfs bbp cscs ch home kumbhar tmp nrn src oc i gpfs bbp cscs ch home kumbhar tmp nrn build i gpfs bbp cscs ch home kumbhar tmp nrn build src nrncvode i gpfs bbp cscs ch home kumbhar tmp nrn build src nrnjava i gpfs bbp cscs ch home kumbhar tmp nrn build src nrnoc i gpfs bbp cscs ch home kumbhar tmp nrn build src nrnpython i gpfs bbp cscs ch home kumbhar tmp nrn build src oc i gpfs bbp cscs ch home kumbhar tmp nrn build src parallel i gpfs bbp cscs ch home kumbhar tmp nrn build src sundials i gpfs bbp cscs ch home kumbhar tmp nrn build src sundials shared i gpfs bbp cscs ch home kumbhar tmp nrn src i gpfs bbp cscs ch home kumbhar tmp nrn src gnu i gpfs bbp cscs ch home kumbhar tmp nrn src memacs i gpfs bbp cscs ch home kumbhar tmp nrn src mesch i gpfs bbp cscs ch home kumbhar tmp nrn src nrncvode i gpfs bbp cscs ch home kumbhar tmp nrn src nrnmpi i gpfs bbp cscs ch home kumbhar tmp nrn src nrnpython i gpfs bbp cscs ch home kumbhar tmp nrn src parallel i gpfs bbp cscs ch home kumbhar tmp nrn src sparse i gpfs bbp cscs ch home kumbhar tmp nrn src i gpfs bbp cscs ch home kumbhar tmp nrn src sundials i gpfs bbp cscs ch home kumbhar tmp nrn src sundials cvodes i gpfs bbp cscs ch home kumbhar tmp nrn src sundials ida i gpfs bbp cscs ch home kumbhar tmp nrn src sundials shared i gpfs bbp cscs ch apps hpc jenkins deploy external libraries linux gcc python include i gpfs bbp cscs ch home kumbhar tmp nrn src nrniv nrniv lib i gpfs bbp cscs ch home kumbhar tmp nrn src ivos i gpfs bbp cscs ch home kumbhar tmp nrn build src ivos fpic std c o cmakefiles nrniv lib dir bbsdirect cpp o c gpfs bbp cscs ch home kumbhar tmp nrn src nrniv bbsdirect cpp in file included from gpfs bbp cscs ch home kumbhar tmp nrn src nrniv bbsdirect cpp gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected template name before ‘ ’ token class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected ‘ ’ before ‘ ’ token gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected unqualified id before ‘ ’ token gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp in constructor ‘bbsdirect bbsdirect ’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error invalid use of incomplete type ‘class keepargs’ keepargs new keepargs gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp in destructor ‘virtual bbsdirect bbsdirect ’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp warning possible problem detected in invocation of delete operator delete keepargs gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp warning invalid use of incomplete type ‘class keepargs’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note neither the destructor nor the class specific operator delete will be called even if they are declared when the class is defined delete keepargs gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp in member function ‘virtual void bbsdirect save args int ’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error invalid use of incomplete type ‘class keepargs’ keepargs insert gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error ‘pair’ was not declared in this scope did you mean ‘std pair’ pair userid sendbuf std pair in file included from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c bits stl algobase h from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c bits char traits h from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c ios from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c ostream from gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c iostream from gpfs bbp cscs ch home kumbhar tmp nrn src ivos ivstream h from gpfs bbp cscs ch home kumbhar tmp nrn build src nrncvode nrnconf h from gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp from gpfs bbp cscs ch home kumbhar tmp nrn src nrniv bbsdirect cpp gpfs bbp cscs ch apps hpc jenkins i would like to be deleted merge deploy compilers linux gcc gcc include c bits stl pair h note ‘std pair’ declared here struct pair in file included from gpfs bbp cscs ch home kumbhar tmp nrn src nrniv bbsdirect cpp gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected primary expression before ‘const’ pair userid sendbuf gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected primary expression before ‘ ’ token pair userid sendbuf gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error expected primary expression before ‘ ’ token pair userid sendbuf gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp in member function ‘virtual void bbsdirect return args int ’ gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error incomplete type ‘keepargs’ used in nested name specifier keepargs iterator i keepargs find userid gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error ‘i’ was not declared in this scope if i keepargs end gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error invalid use of incomplete type ‘class keepargs’ if i keepargs end gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp error invalid use of incomplete type ‘class keepargs’ keepargs erase i gpfs bbp cscs ch home kumbhar tmp nrn src nrniv parallel bbsdirectmpi cpp note forward declaration of ‘class keepargs’ class keepargs public map make error make leaving directory gpfs bbp cscs ch home kumbhar tmp nrn build make error make leaving directory gpfs bbp cscs ch home kumbhar tmp nrn build make error this seems to be related to but i am surprised that this is only appearing with gcc i am shorty creating pr with the fix cc anilbey | 0 |
6,789 | 7,754,237,276 | IssuesEvent | 2018-05-31 05:38:57 | terraform-providers/terraform-provider-aws | https://api.github.com/repos/terraform-providers/terraform-provider-aws | closed | AWS_KMS_GRANT missing operations | question service/kms waiting-response | Hi,
just wanted to use aws_kms_grant for allowing an ASG to start instances with an encrypted root volume.
Therefore I need to add the permissions "kms:CreateGrant", "kms:ListGrants" and "kms:RevokeGrant" for the ASG service-role to the kms policy. Unfortunately TF does not allow the operations "ListGrants" and "RevokeGrant".
Please fix this small issue.
Thanks
### Terraform Version
v0.11.3
### AWS Provider version
v1.12.0
### Affected Resource(s)
- aws_kms_grant
### Terraform Configuration Files
```hcl
resource "aws_kms_key" "ecs" {
description = "ECS instance key"
deletion_window_in_days = 10
key_usage = "ENCRYPT_DECRYPT"
is_enabled = true
enable_key_rotation = true
}
resource "aws_kms_grant" "ecs_asg" {
name = "ECS_ASG"
key_id = "${aws_kms_key.ecs.key_id}"
grantee_principal = "arn:aws:iam::${data.aws_caller_identity.current.account_id}:role/aws-service-role/autoscaling.amazonaws.com/AWSServiceRoleForAutoScaling"
operations = ["CreateGrant", "ListGrants", "RevokeGrant"]
}
```
### Debug Output
```
Error: aws_kms_grant.ecs_asg: expected operations.1 to be one of [CreateGrant Decrypt DescribeKey Encrypt GenerateDataKey GenerateDataKeyWithoutPlaintext ReEncryptFrom ReEncryptTo RetireGrant], got ListGrants
Error: aws_kms_grant.ecs_asg: expected operations.2 to be one of [CreateGrant Decrypt DescribeKey Encrypt GenerateDataKey GenerateDataKeyWithoutPlaintext ReEncryptFrom ReEncryptTo RetireGrant], got RevokeGrant
```
### Expected Behavior
Well, just create the damn policy :-D
### Actual Behavior
Error message above
### Steps to Reproduce
Please list the steps required to reproduce the issue, for example:
1. `terraform plan`
| 1.0 | AWS_KMS_GRANT missing operations - Hi,
just wanted to use aws_kms_grant for allowing an ASG to start instances with an encrypted root volume.
Therefore I need to add the permissions "kms:CreateGrant", "kms:ListGrants" and "kms:RevokeGrant" for the ASG service-role to the kms policy. Unfortunately TF does not allow the operations "ListGrants" and "RevokeGrant".
Please fix this small issue.
Thanks
### Terraform Version
v0.11.3
### AWS Provider version
v1.12.0
### Affected Resource(s)
- aws_kms_grant
### Terraform Configuration Files
```hcl
resource "aws_kms_key" "ecs" {
description = "ECS instance key"
deletion_window_in_days = 10
key_usage = "ENCRYPT_DECRYPT"
is_enabled = true
enable_key_rotation = true
}
resource "aws_kms_grant" "ecs_asg" {
name = "ECS_ASG"
key_id = "${aws_kms_key.ecs.key_id}"
grantee_principal = "arn:aws:iam::${data.aws_caller_identity.current.account_id}:role/aws-service-role/autoscaling.amazonaws.com/AWSServiceRoleForAutoScaling"
operations = ["CreateGrant", "ListGrants", "RevokeGrant"]
}
```
### Debug Output
```
Error: aws_kms_grant.ecs_asg: expected operations.1 to be one of [CreateGrant Decrypt DescribeKey Encrypt GenerateDataKey GenerateDataKeyWithoutPlaintext ReEncryptFrom ReEncryptTo RetireGrant], got ListGrants
Error: aws_kms_grant.ecs_asg: expected operations.2 to be one of [CreateGrant Decrypt DescribeKey Encrypt GenerateDataKey GenerateDataKeyWithoutPlaintext ReEncryptFrom ReEncryptTo RetireGrant], got RevokeGrant
```
### Expected Behavior
Well, just create the damn policy :-D
### Actual Behavior
Error message above
### Steps to Reproduce
Please list the steps required to reproduce the issue, for example:
1. `terraform plan`
| non_infrastructure | aws kms grant missing operations hi just wanted to use aws kms grant for allowing an asg to start instances with an encrypted root volume therefore i need to add the permissions kms creategrant kms listgrants and kms revokegrant for the asg service role to the kms policy unfortunately tf does not allow the operations listgrants and revokegrant please fix this small issue thanks terraform version aws provider version affected resource s aws kms grant terraform configuration files hcl resource aws kms key ecs description ecs instance key deletion window in days key usage encrypt decrypt is enabled true enable key rotation true resource aws kms grant ecs asg name ecs asg key id aws kms key ecs key id grantee principal arn aws iam data aws caller identity current account id role aws service role autoscaling amazonaws com awsserviceroleforautoscaling operations debug output error aws kms grant ecs asg expected operations to be one of got listgrants error aws kms grant ecs asg expected operations to be one of got revokegrant expected behavior well just create the damn policy d actual behavior error message above steps to reproduce please list the steps required to reproduce the issue for example terraform plan | 0 |
9,026 | 7,779,469,964 | IssuesEvent | 2018-06-05 16:55:24 | dotnet/wcf | https://api.github.com/repos/dotnet/wcf | closed | branch 2.1.0 servicing package validation error. | Infrastructure Servicing-Consider-2.1 | With https://github.com/StephenBonikowsky/wcf/tree/stebon/2.1.0/UpdatePkgAndAssembliesForServicing I am moving our release/2.1.0 branch to servicing.
After making the necessary changes I am getting a package validation error only for System.ServiceModel.Security.
The errors are...
```
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1/win-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1/win-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1/win7-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1/win7-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2/win-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2/win-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2/win7-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2/win7-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7/win-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7/win-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7/win7-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7/win7-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
``` | 1.0 | branch 2.1.0 servicing package validation error. - With https://github.com/StephenBonikowsky/wcf/tree/stebon/2.1.0/UpdatePkgAndAssembliesForServicing I am moving our release/2.1.0 branch to servicing.
After making the necessary changes I am getting a package validation error only for System.ServiceModel.Security.
The errors are...
```
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1/win-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1/win-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1/win7-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.1/win7-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net461" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2/win-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2/win-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2/win7-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.6.2/win7-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net462" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7/win-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7/win-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7/win7-x86 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
C:\OSS\wcf\Tools\Packaging.targets(1131,5): error : System.ServiceModel.Security should be supported on .NETFramework,Version=v4.7/win7-x64 but has a compile placeholder. You may need to remove InboxOnTargetFramework Include="net47" /> from your project. [C:\OSS\wcf\src\System.ServiceModel.Security\pkg\System.ServiceModel.Security.pkgproj]
``` | infrastructure | branch servicing package validation error with i am moving our release branch to servicing after making the necessary changes i am getting a package validation error only for system servicemodel security the errors are c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version win but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version win but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version win but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version win but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version win but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version win but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project c oss wcf tools packaging targets error system servicemodel security should be supported on netframework version but has a compile placeholder you may need to remove inboxontargetframework include from your project | 1 |
15,566 | 11,592,068,587 | IssuesEvent | 2020-02-24 10:42:21 | flutter/website | https://api.github.com/repos/flutter/website | closed | The search function on the API site is inefficient | infrastructure | Consider the situation where you want to search `setState` on api.flutter.dev
Logically, it should bring `State.setState` as the first result, since this is the only class that implements the method.
The others just inherit from it.
The reality is different:
<img width="488" alt="Screenshot 2020-02-21 at 14 42 23" src="https://user-images.githubusercontent.com/20165741/75043692-609f7500-54b8-11ea-9d0c-f3f3ee74b956.png">
The search result is polluted by all the stateful widgets in the framework.
In the end, we don't even see `State.setState` because there are too many results. | 1.0 | The search function on the API site is inefficient - Consider the situation where you want to search `setState` on api.flutter.dev
Logically, it should bring `State.setState` as the first result, since this is the only class that implements the method.
The others just inherit from it.
The reality is different:
<img width="488" alt="Screenshot 2020-02-21 at 14 42 23" src="https://user-images.githubusercontent.com/20165741/75043692-609f7500-54b8-11ea-9d0c-f3f3ee74b956.png">
The search result is polluted by all the stateful widgets in the framework.
In the end, we don't even see `State.setState` because there are too many results. | infrastructure | the search function on the api site is inefficient consider the situation where you want to search setstate on api flutter dev logically it should bring state setstate as the first result since this is the only class that implements the method the others just inherit from it the reality is different img width alt screenshot at src the search result is polluted by all the stateful widgets in the framework in the end we don t even see state setstate because there are too many results | 1 |
26,461 | 20,142,441,606 | IssuesEvent | 2022-02-09 01:34:49 | OpenHistoricalMap/issues | https://api.github.com/repos/OpenHistoricalMap/issues | closed | Dealing with coastlines | infrastructure high_pri | **What's your idea for a cool feature that would help you use OHM better.**
It would be great to be able to render islands and / or wetlands more reliably in OHM.
This is important b/c many coasts / shores are full of marshes before they get built up with seawalls, etc.
Perhaps I'm missing something, but it's not clear what's up.
**Current workarounds**
None that I can figure out:
**Additional info**
OHM: https://www.openhistoricalmap.org/way/198314785#map=16/47.5644/-122.3383&layers=OD
<img width="761" alt="image" src="https://user-images.githubusercontent.com/1494546/84733903-5ff18980-af54-11ea-817d-30502c802104.png">
OSM:https://www.openstreetmap.org/way/116923338#map=16/37.9331/-122.5089&layers=D
<img width="729" alt="image" src="https://user-images.githubusercontent.com/1494546/84734061-d2626980-af54-11ea-838e-1b8c6fdfb241.png">
OSM: https://www.openstreetmap.org/way/318323319#map=18/37.92417/-122.50433&layers=D
<img width="518" alt="image" src="https://user-images.githubusercontent.com/1494546/84736391-8e726300-af5a-11ea-85af-44210271e554.png">
| 1.0 | Dealing with coastlines - **What's your idea for a cool feature that would help you use OHM better.**
It would be great to be able to render islands and / or wetlands more reliably in OHM.
This is important b/c many coasts / shores are full of marshes before they get built up with seawalls, etc.
Perhaps I'm missing something, but it's not clear what's up.
**Current workarounds**
None that I can figure out:
**Additional info**
OHM: https://www.openhistoricalmap.org/way/198314785#map=16/47.5644/-122.3383&layers=OD
<img width="761" alt="image" src="https://user-images.githubusercontent.com/1494546/84733903-5ff18980-af54-11ea-817d-30502c802104.png">
OSM:https://www.openstreetmap.org/way/116923338#map=16/37.9331/-122.5089&layers=D
<img width="729" alt="image" src="https://user-images.githubusercontent.com/1494546/84734061-d2626980-af54-11ea-838e-1b8c6fdfb241.png">
OSM: https://www.openstreetmap.org/way/318323319#map=18/37.92417/-122.50433&layers=D
<img width="518" alt="image" src="https://user-images.githubusercontent.com/1494546/84736391-8e726300-af5a-11ea-85af-44210271e554.png">
| infrastructure | dealing with coastlines what s your idea for a cool feature that would help you use ohm better it would be great to be able to render islands and or wetlands more reliably in ohm this is important b c many coasts shores are full of marshes before they get built up with seawalls etc perhaps i m missing something but it s not clear what s up current workarounds none that i can figure out additional info ohm img width alt image src osm img width alt image src osm img width alt image src | 1 |
3,313 | 4,214,883,177 | IssuesEvent | 2016-06-30 00:27:21 | Microsoft/TypeScript | https://api.github.com/repos/Microsoft/TypeScript | opened | Phase TypeScript to use --strictNullChecks | Infrastructure | This is important for consumers of our API and for testing the feature in general | 1.0 | Phase TypeScript to use --strictNullChecks - This is important for consumers of our API and for testing the feature in general | infrastructure | phase typescript to use strictnullchecks this is important for consumers of our api and for testing the feature in general | 1 |
333,830 | 29,811,867,667 | IssuesEvent | 2023-06-16 15:39:58 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | DISABLED test_backward_multiple_output_tensors (__main__.TensorPipeDistAutogradTest) | oncall: distributed module: flaky-tests skipped | Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_backward_multiple_output_tensors%2C%20TensorPipeDistAutogradTest) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/6443982488).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 red and 6 green.
cc @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @SciPioneer @H-Huang | 1.0 | DISABLED test_backward_multiple_output_tensors (__main__.TensorPipeDistAutogradTest) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_backward_multiple_output_tensors%2C%20TensorPipeDistAutogradTest) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/6443982488).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 red and 6 green.
cc @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @osalpekar @jiayisuse @SciPioneer @H-Huang | non_infrastructure | disabled test backward multiple output tensors main tensorpipedistautogradtest platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with red and green cc pietern mrshenli zhaojuanmao satgera rohan varma gqchen aazzolini osalpekar jiayisuse scipioneer h huang | 0 |
107,250 | 11,524,516,535 | IssuesEvent | 2020-02-15 01:00:55 | UBC-LA-Hackathon/student-dashboard | https://api.github.com/repos/UBC-LA-Hackathon/student-dashboard | opened | Add list of volunteers under README.md#Volunteers | documentation | There are 7 confirmed external volunteers.
Ask for:
* Github/Linkedin link?
* Photo?
* Company name/link? | 1.0 | Add list of volunteers under README.md#Volunteers - There are 7 confirmed external volunteers.
Ask for:
* Github/Linkedin link?
* Photo?
* Company name/link? | non_infrastructure | add list of volunteers under readme md volunteers there are confirmed external volunteers ask for github linkedin link photo company name link | 0 |
25,783 | 19,180,039,617 | IssuesEvent | 2021-12-04 08:01:16 | KurnakovMaksim/eHomeDesigner.API | https://api.github.com/repos/KurnakovMaksim/eHomeDesigner.API | closed | Data layer | infrastructure data | Now we need to add data layer. It will have:
* Entities
* Contexts
* Repositories | 1.0 | Data layer - Now we need to add data layer. It will have:
* Entities
* Contexts
* Repositories | infrastructure | data layer now we need to add data layer it will have entities contexts repositories | 1 |
28,614 | 23,391,974,134 | IssuesEvent | 2022-08-11 18:46:42 | accessibility-exchange/platform | https://api.github.com/repos/accessibility-exchange/platform | closed | Configure Twilio (Sendgrid) for transactional email delivery | help wanted infrastructure | We need to add Twilio Sendgrid configuration to the staging and production environments. | 1.0 | Configure Twilio (Sendgrid) for transactional email delivery - We need to add Twilio Sendgrid configuration to the staging and production environments. | infrastructure | configure twilio sendgrid for transactional email delivery we need to add twilio sendgrid configuration to the staging and production environments | 1 |
15,731 | 11,687,125,822 | IssuesEvent | 2020-03-05 12:13:06 | hotosm/tasking-manager | https://api.github.com/repos/hotosm/tasking-manager | opened | User data moved over to another profile | Component: Backend Component: Infrastructure Priority: High Status: Needs implementation Type: Bug | During the initial mapathon testing one bug was discovered, that looks similar to #1715:
> For a single user, when logged in, the name, gender, email address showed on his profile was the one of a different user. But the image was correct. No other users reported this issue. After logging out and in again, the user still had the same problem. | 1.0 | User data moved over to another profile - During the initial mapathon testing one bug was discovered, that looks similar to #1715:
> For a single user, when logged in, the name, gender, email address showed on his profile was the one of a different user. But the image was correct. No other users reported this issue. After logging out and in again, the user still had the same problem. | infrastructure | user data moved over to another profile during the initial mapathon testing one bug was discovered that looks similar to for a single user when logged in the name gender email address showed on his profile was the one of a different user but the image was correct no other users reported this issue after logging out and in again the user still had the same problem | 1 |
2,442 | 3,686,415,203 | IssuesEvent | 2016-02-25 01:09:57 | aspnet/Performance | https://api.github.com/repos/aspnet/Performance | closed | Build using a CI and add a build health badge to README.md | 2 - Working Infrastructure task | Consider using Travis (like [Entity Framework](https://github.com/aspnet/EntityFramework)) or Jenkins (like [cli](https://github.com/dotnet/cli)). We should keep a very high quality bar on the health status of this repository. | 1.0 | Build using a CI and add a build health badge to README.md - Consider using Travis (like [Entity Framework](https://github.com/aspnet/EntityFramework)) or Jenkins (like [cli](https://github.com/dotnet/cli)). We should keep a very high quality bar on the health status of this repository. | infrastructure | build using a ci and add a build health badge to readme md consider using travis like or jenkins like we should keep a very high quality bar on the health status of this repository | 1 |
110,837 | 24,015,635,194 | IssuesEvent | 2022-09-15 00:10:46 | qhy040404/Library-One-Tap-Android | https://api.github.com/repos/qhy040404/Library-One-Tap-Android | closed | Rewrite AboutActivity to use partial chrome | enhancement large code low priority UI / UX external | ### Enhancement propose
Better UX
### Solution

### Additional info
_No response_ | 1.0 | Rewrite AboutActivity to use partial chrome - ### Enhancement propose
Better UX
### Solution

### Additional info
_No response_ | non_infrastructure | rewrite aboutactivity to use partial chrome enhancement propose better ux solution additional info no response | 0 |
28,199 | 23,078,970,077 | IssuesEvent | 2022-07-26 04:33:13 | zer0Kerbal/Fuselage | https://api.github.com/repos/zer0Kerbal/Fuselage | closed | Asset Updates | issue: model issue: texture type: infrastructure | <!--<FileName> v1.0.0.0
Fuselage (DMF)
created: 01 May 2022
updated:
CC BY-ND 4.0 by zer0Kerbal
-->
### Asset Updates
* [x] create Assets/ folder
* [x] convert from mesh to MODEL
* [x] rename
* [x] models to unique names
* [x] textures to unique names
* [ ] update
* [x] model pointers (.png et al to .dds)
* [ ] model texture pointers to new names
* [x] relocate assets to Assets/
* [ ] eliminate
* [ ] duplicate textures
* [ ] duplicate models
* [x] relocate part.cfg to Parts/ | 1.0 | Asset Updates - <!--<FileName> v1.0.0.0
Fuselage (DMF)
created: 01 May 2022
updated:
CC BY-ND 4.0 by zer0Kerbal
-->
### Asset Updates
* [x] create Assets/ folder
* [x] convert from mesh to MODEL
* [x] rename
* [x] models to unique names
* [x] textures to unique names
* [ ] update
* [x] model pointers (.png et al to .dds)
* [ ] model texture pointers to new names
* [x] relocate assets to Assets/
* [ ] eliminate
* [ ] duplicate textures
* [ ] duplicate models
* [x] relocate part.cfg to Parts/ | infrastructure | asset updates fuselage dmf created may updated cc by nd by asset updates create assets folder convert from mesh to model rename models to unique names textures to unique names update model pointers png et al to dds model texture pointers to new names relocate assets to assets eliminate duplicate textures duplicate models relocate part cfg to parts | 1 |
219,175 | 7,333,744,397 | IssuesEvent | 2018-03-05 20:24:12 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | SERVER ISSUE: System.InvalidOperationException: Collection was modified; enumeration operation may not execute. | High Priority | Dedicated server on local network to me: my client was receiving notifications that people were disconnecting and reconnecting immediately. They did not realize this was happening but it showed up in my chat log.
I could continue performing actions and so could other users but they could not connect at some point. I checked the server and then saw the error with dump below:
`--BEGIN DUMP--
Dump Time
02/19/2018 16:47:18
Exception
System.InvalidOperationException: Collection was modified; enumeration operation may not execute.
at System.ThrowHelper.ThrowInvalidOperationException(ExceptionResource resource)
at System.Collections.Generic.Dictionary`2.Enumerator.MoveNext()
at System.Linq.Enumerable.WhereEnumerableIterator`1.MoveNext()
at Eco.Shared.Utils.EnumerableExtensions.ForEach[T](IEnumerable`1 enumeration, Action`1 action)
at Eco.Gameplay.Minimap.MinimapManager.GetChangesSince(ChunkMap chunkMap, Double time, String valueTypeName)
at Eco.Gameplay.Minimap.MinimapManager.SendInitialState(BSONObject bsonObj, INetObjectViewer viewer)
at Eco.Shared.Networking.NetObjectManager.GetInitialObjectStates(List`1 newObjects, INetObjectViewer viewer)
at Eco.Plugins.Networking.Clients.Client.Update()
at Eco.Plugins.Networking.Clients.Client.<.ctor>b__55_1()
--END DUMP--`
Had to force stop EcoServer.exe and everyone rolled back about 3 minutes. Unfortunately, no one knows what actions they were performing as none were out of the ordinary might have caused this. | 1.0 | SERVER ISSUE: System.InvalidOperationException: Collection was modified; enumeration operation may not execute. - Dedicated server on local network to me: my client was receiving notifications that people were disconnecting and reconnecting immediately. They did not realize this was happening but it showed up in my chat log.
I could continue performing actions and so could other users but they could not connect at some point. I checked the server and then saw the error with dump below:
`--BEGIN DUMP--
Dump Time
02/19/2018 16:47:18
Exception
System.InvalidOperationException: Collection was modified; enumeration operation may not execute.
at System.ThrowHelper.ThrowInvalidOperationException(ExceptionResource resource)
at System.Collections.Generic.Dictionary`2.Enumerator.MoveNext()
at System.Linq.Enumerable.WhereEnumerableIterator`1.MoveNext()
at Eco.Shared.Utils.EnumerableExtensions.ForEach[T](IEnumerable`1 enumeration, Action`1 action)
at Eco.Gameplay.Minimap.MinimapManager.GetChangesSince(ChunkMap chunkMap, Double time, String valueTypeName)
at Eco.Gameplay.Minimap.MinimapManager.SendInitialState(BSONObject bsonObj, INetObjectViewer viewer)
at Eco.Shared.Networking.NetObjectManager.GetInitialObjectStates(List`1 newObjects, INetObjectViewer viewer)
at Eco.Plugins.Networking.Clients.Client.Update()
at Eco.Plugins.Networking.Clients.Client.<.ctor>b__55_1()
--END DUMP--`
Had to force stop EcoServer.exe and everyone rolled back about 3 minutes. Unfortunately, no one knows what actions they were performing as none were out of the ordinary might have caused this. | non_infrastructure | server issue system invalidoperationexception collection was modified enumeration operation may not execute dedicated server on local network to me my client was receiving notifications that people were disconnecting and reconnecting immediately they did not realize this was happening but it showed up in my chat log i could continue performing actions and so could other users but they could not connect at some point i checked the server and then saw the error with dump below begin dump dump time exception system invalidoperationexception collection was modified enumeration operation may not execute at system throwhelper throwinvalidoperationexception exceptionresource resource at system collections generic dictionary enumerator movenext at system linq enumerable whereenumerableiterator movenext at eco shared utils enumerableextensions foreach ienumerable enumeration action action at eco gameplay minimap minimapmanager getchangessince chunkmap chunkmap double time string valuetypename at eco gameplay minimap minimapmanager sendinitialstate bsonobject bsonobj inetobjectviewer viewer at eco shared networking netobjectmanager getinitialobjectstates list newobjects inetobjectviewer viewer at eco plugins networking clients client update at eco plugins networking clients client b end dump had to force stop ecoserver exe and everyone rolled back about minutes unfortunately no one knows what actions they were performing as none were out of the ordinary might have caused this | 0 |
234,207 | 25,806,042,041 | IssuesEvent | 2022-12-11 12:20:28 | SmartBear/readyapi-swagger-assertion-plugin | https://api.github.com/repos/SmartBear/readyapi-swagger-assertion-plugin | closed | CVE-2020-10969 (High) detected in jackson-databind-2.4.2.jar - autoclosed | security vulnerability | ## CVE-2020-10969 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.4.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.4.2/jackson-databind-2.4.2.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.7.0.jar (Root Library)
- jackson-datatype-joda-2.4.2.jar
- :x: **jackson-databind-2.4.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to javax.swing.JEditorPane.
<p>Publish Date: 2020-03-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-10969>CVE-2020-10969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10969">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10969</a></p>
<p>Release Date: 2020-03-26</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.6;com.fasterxml.jackson.core:jackson-databind:2.7.9.7</p>
</p>
</details>
<p></p>
| True | CVE-2020-10969 (High) detected in jackson-databind-2.4.2.jar - autoclosed - ## CVE-2020-10969 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.4.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.4.2/jackson-databind-2.4.2.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.7.0.jar (Root Library)
- jackson-datatype-joda-2.4.2.jar
- :x: **jackson-databind-2.4.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to javax.swing.JEditorPane.
<p>Publish Date: 2020-03-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-10969>CVE-2020-10969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10969">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10969</a></p>
<p>Release Date: 2020-03-26</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.6;com.fasterxml.jackson.core:jackson-databind:2.7.9.7</p>
</p>
</details>
<p></p>
| non_infrastructure | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy ready api soapui pro jar root library jackson datatype joda jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to javax swing jeditorpane publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind com fasterxml jackson core jackson databind | 0 |
274,223 | 29,934,254,517 | IssuesEvent | 2023-06-22 11:40:45 | tomkerkhove/promitor | https://api.github.com/repos/tomkerkhove/promitor | opened | CVE-2023-29331 in both Scraper and Resource Discovery security | security | ### Report
The vulnerable [System.Security.Cryptography.Pkcs](https://www.nuget.org/packages/System.Security.Cryptography.Pkcs) v7.0.0 is used by both agents.
### Vulnerability Information
https://github.com/dotnet/runtime/issues/87498
### Affected Component(s)
Resource Discovery, Scraper
### Affected Version(s)
latest (preview)
### Vulnerability Migitation
_No response_
### Vulnerability Fix
Upgrade System.Security.Cryptography.Pkcs to v7.0.2. I couldn't find a better way to do that than pinning the version in the csproj files.
### Contact Details
_No response_ | True | CVE-2023-29331 in both Scraper and Resource Discovery security - ### Report
The vulnerable [System.Security.Cryptography.Pkcs](https://www.nuget.org/packages/System.Security.Cryptography.Pkcs) v7.0.0 is used by both agents.
### Vulnerability Information
https://github.com/dotnet/runtime/issues/87498
### Affected Component(s)
Resource Discovery, Scraper
### Affected Version(s)
latest (preview)
### Vulnerability Migitation
_No response_
### Vulnerability Fix
Upgrade System.Security.Cryptography.Pkcs to v7.0.2. I couldn't find a better way to do that than pinning the version in the csproj files.
### Contact Details
_No response_ | non_infrastructure | cve in both scraper and resource discovery security report the vulnerable is used by both agents vulnerability information affected component s resource discovery scraper affected version s latest preview vulnerability migitation no response vulnerability fix upgrade system security cryptography pkcs to i couldn t find a better way to do that than pinning the version in the csproj files contact details no response | 0 |
28,758 | 23,474,034,950 | IssuesEvent | 2022-08-17 02:51:53 | woowacourse-teams/2022-kkogkkog | https://api.github.com/repos/woowacourse-teams/2022-kkogkkog | closed | [FE] 사파리 브라우저 이슈 해결 | 🦄 frontend 🌐 infrastructure | ## 배경
사파리, Mac Chrome, Microsoft Edge 등 다양한 브라우저 환경에서도 사용자 경험을 잃지 않아야 한다.
IOS, Android OS 등 다양한 디바이스 환경에서도 사용자 경험을 잃지 않아야 한다.
## 작업
- input date 이슈
- input zoom
- 줌이 과하게 되는 이슈
- 100vh 이슈
- 화면을 넘어가던 이슈
## 진행사항
- Safari, Chrome, Microsoft Edge 대응
- IOS, Android OS 대응
<!--
## 공유사항
해당 작업을 수행함에 있어 주의해야 할 사항에 대한 설명
-->
<!--
아래 작업들을 완료 후 주석은 전부 제거
1. Assignees에 해당 작업과 관련된 팀원들만 배정되도록 수정
2. labels 목록 수정
3. Projects에 현재 진행 중인 스프린트에 해당되는 칸반보드 등록
4. 개별 작업을 진행하면서 진행사항의 체크리스트들을 칸반보드에서 하나씩 체크
-->
| 1.0 | [FE] 사파리 브라우저 이슈 해결 - ## 배경
사파리, Mac Chrome, Microsoft Edge 등 다양한 브라우저 환경에서도 사용자 경험을 잃지 않아야 한다.
IOS, Android OS 등 다양한 디바이스 환경에서도 사용자 경험을 잃지 않아야 한다.
## 작업
- input date 이슈
- input zoom
- 줌이 과하게 되는 이슈
- 100vh 이슈
- 화면을 넘어가던 이슈
## 진행사항
- Safari, Chrome, Microsoft Edge 대응
- IOS, Android OS 대응
<!--
## 공유사항
해당 작업을 수행함에 있어 주의해야 할 사항에 대한 설명
-->
<!--
아래 작업들을 완료 후 주석은 전부 제거
1. Assignees에 해당 작업과 관련된 팀원들만 배정되도록 수정
2. labels 목록 수정
3. Projects에 현재 진행 중인 스프린트에 해당되는 칸반보드 등록
4. 개별 작업을 진행하면서 진행사항의 체크리스트들을 칸반보드에서 하나씩 체크
-->
| infrastructure | 사파리 브라우저 이슈 해결 배경 사파리 mac chrome microsoft edge 등 다양한 브라우저 환경에서도 사용자 경험을 잃지 않아야 한다 ios android os 등 다양한 디바이스 환경에서도 사용자 경험을 잃지 않아야 한다 작업 input date 이슈 input zoom 줌이 과하게 되는 이슈 이슈 화면을 넘어가던 이슈 진행사항 safari chrome microsoft edge 대응 ios android os 대응 공유사항 해당 작업을 수행함에 있어 주의해야 할 사항에 대한 설명 아래 작업들을 완료 후 주석은 전부 제거 assignees에 해당 작업과 관련된 팀원들만 배정되도록 수정 labels 목록 수정 projects에 현재 진행 중인 스프린트에 해당되는 칸반보드 등록 개별 작업을 진행하면서 진행사항의 체크리스트들을 칸반보드에서 하나씩 체크 | 1 |
26,110 | 19,673,189,691 | IssuesEvent | 2022-01-11 09:37:48 | reapit/foundations | https://api.github.com/repos/reapit/foundations | closed | Housekeeping: Infra | front-end infrastructure task | **Background context or User story:**
_There are a number of bits of infra that can be tidied up - old POCs and redundant services that can be deprecated. Task is to review both dev and prod AWS and ensure any old infra is removed_
| 1.0 | Housekeeping: Infra - **Background context or User story:**
_There are a number of bits of infra that can be tidied up - old POCs and redundant services that can be deprecated. Task is to review both dev and prod AWS and ensure any old infra is removed_
| infrastructure | housekeeping infra background context or user story there are a number of bits of infra that can be tidied up old pocs and redundant services that can be deprecated task is to review both dev and prod aws and ensure any old infra is removed | 1 |
5,271 | 5,544,889,054 | IssuesEvent | 2017-03-22 20:15:41 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Add a test for devenv /build | Area-Infrastructure | We should have a test to test devenv /build, since ~~I~~ we have a knack for breaking it. | 1.0 | Add a test for devenv /build - We should have a test to test devenv /build, since ~~I~~ we have a knack for breaking it. | infrastructure | add a test for devenv build we should have a test to test devenv build since i we have a knack for breaking it | 1 |
5,027 | 5,395,444,054 | IssuesEvent | 2017-02-27 08:38:27 | camptocamp/c2cgeoportal | https://api.github.com/repos/camptocamp/c2cgeoportal | closed | OGC Proxy is gone? | Infrastructure Ongoing | Is OGC Proxy gone? Because after upgrade to `2.1.2.dev1` I get this error:
```
MakoRenderingException:
Traceback (most recent call last):
File ".build/venv/lib/python2.7/site-packages/pyramid_mako/__init__.py", line 148, in __call__
result = template.render_unicode(**system)
File ".build/venv/lib/python2.7/site-packages/mako/template.py", line 471, in render_unicode
as_unicode=True)
File ".build/venv/lib/python2.7/site-packages/mako/runtime.py", line 838, in _render
**_kwargs_for_callable(callable_, data))
File ".build/venv/lib/python2.7/site-packages/mako/runtime.py", line 873, in _render_context
_exec_template(inherit, lclcontext, args=args, kwargs=kwargs)
File ".build/venv/lib/python2.7/site-packages/mako/runtime.py", line 899, in _exec_template
callable_(context, *args, **kwargs)
File "nyon/templates/desktop.js", line 16, in render_body
OpenLayers.ProxyHost = "${request.route_url('ogcproxy') | n}?url=";
File ".build/venv/lib/python2.7/site-packages/pyramid/url.py", line 266, in route_url
raise KeyError('No such route named %s' % route_name)
KeyError: 'No such route named ogcproxy'
```
Full stack:
```
File '.build/venv/lib/python2.7/site-packages/weberror/evalexception.py', line 429 in respond
app_iter = self.application(environ, detect_start_response)
File '.build/venv/lib/python2.7/site-packages/fanstatic/publisher.py', line 219 in __call__
return self.app(environ, start_response)
File '.build/venv/lib/python2.7/site-packages/fanstatic/injector.py', line 54 in __call__
response = request.get_response(self.app)
File '.build/venv/lib/python2.7/site-packages/webob/request.py', line 1299 in send
application, catch_exc_info=False)
File '.build/venv/lib/python2.7/site-packages/webob/request.py', line 1263 in call_application
app_iter = application(self.environ, start_response)
File '.build/venv/lib/python2.7/site-packages/pyramid/router.py', line 223 in __call__
response = self.invoke_subrequest(request, use_tweens=True)
File '.build/venv/lib/python2.7/site-packages/pyramid/router.py', line 198 in invoke_subrequest
response = handle_request(request)
File '.build/venv/lib/python2.7/site-packages/c2cgeoportal/lib/cacheversion.py', line 62 in __call__
response = self.handler(request)
File '.build/venv/lib/python2.7/site-packages/pyramid/tweens.py', line 20 in excview_tween
response = handler(request)
File '.build/venv/lib/python2.7/site-packages/pyramid_tm/__init__.py', line 119 in tm_tween
reraise(*exc_info)
File '.build/venv/lib/python2.7/site-packages/pyramid_tm/__init__.py', line 98 in tm_tween
response = handler(request)
File '.build/venv/lib/python2.7/site-packages/pyramid/router.py', line 145 in handle_request
view_name
File '.build/venv/lib/python2.7/site-packages/pyramid/view.py', line 541 in _call_view
response = view_callable(context, request)
File '.build/venv/lib/python2.7/site-packages/c2cgeoportal/pyramid_.py', line 120 in new_f
return f(root, request)
File '.build/venv/lib/python2.7/site-packages/pyramid/config/views.py', line 377 in rendered_view
context)
File '.build/venv/lib/python2.7/site-packages/pyramid/renderers.py', line 450 in render_view
return self.render_to_response(response, system, request=request)
File '.build/venv/lib/python2.7/site-packages/pyramid/renderers.py', line 473 in render_to_response
result = self.render(value, system_values, request=request)
File '.build/venv/lib/python2.7/site-packages/pyramid/renderers.py', line 469 in render
result = renderer(value, system_values)
File '.build/venv/lib/python2.7/site-packages/pyramid_mako/__init__.py', line 156 in __call__
reraise(MakoRenderingException(errtext), None, exc_info[2])
File '.build/venv/lib/python2.7/site-packages/pyramid_mako/__init__.py', line 148 in __call__
result = template.render_unicode(**system)
File '.build/venv/lib/python2.7/site-packages/mako/template.py', line 471 in render_unicode
as_unicode=True)
File '.build/venv/lib/python2.7/site-packages/mako/runtime.py', line 838 in _render
**_kwargs_for_callable(callable_, data))
File '.build/venv/lib/python2.7/site-packages/mako/runtime.py', line 873 in _render_context
_exec_template(inherit, lclcontext, args=args, kwargs=kwargs)
File '.build/venv/lib/python2.7/site-packages/mako/runtime.py', line 899 in _exec_template
callable_(context, *args, **kwargs)
File '_desktop_js', line 43 in render_body
File '.build/venv/lib/python2.7/site-packages/pyramid/url.py', line 266 in route_url
raise KeyError('No such route named %s' % route_name)
``` | 1.0 | OGC Proxy is gone? - Is OGC Proxy gone? Because after upgrade to `2.1.2.dev1` I get this error:
```
MakoRenderingException:
Traceback (most recent call last):
File ".build/venv/lib/python2.7/site-packages/pyramid_mako/__init__.py", line 148, in __call__
result = template.render_unicode(**system)
File ".build/venv/lib/python2.7/site-packages/mako/template.py", line 471, in render_unicode
as_unicode=True)
File ".build/venv/lib/python2.7/site-packages/mako/runtime.py", line 838, in _render
**_kwargs_for_callable(callable_, data))
File ".build/venv/lib/python2.7/site-packages/mako/runtime.py", line 873, in _render_context
_exec_template(inherit, lclcontext, args=args, kwargs=kwargs)
File ".build/venv/lib/python2.7/site-packages/mako/runtime.py", line 899, in _exec_template
callable_(context, *args, **kwargs)
File "nyon/templates/desktop.js", line 16, in render_body
OpenLayers.ProxyHost = "${request.route_url('ogcproxy') | n}?url=";
File ".build/venv/lib/python2.7/site-packages/pyramid/url.py", line 266, in route_url
raise KeyError('No such route named %s' % route_name)
KeyError: 'No such route named ogcproxy'
```
Full stack:
```
File '.build/venv/lib/python2.7/site-packages/weberror/evalexception.py', line 429 in respond
app_iter = self.application(environ, detect_start_response)
File '.build/venv/lib/python2.7/site-packages/fanstatic/publisher.py', line 219 in __call__
return self.app(environ, start_response)
File '.build/venv/lib/python2.7/site-packages/fanstatic/injector.py', line 54 in __call__
response = request.get_response(self.app)
File '.build/venv/lib/python2.7/site-packages/webob/request.py', line 1299 in send
application, catch_exc_info=False)
File '.build/venv/lib/python2.7/site-packages/webob/request.py', line 1263 in call_application
app_iter = application(self.environ, start_response)
File '.build/venv/lib/python2.7/site-packages/pyramid/router.py', line 223 in __call__
response = self.invoke_subrequest(request, use_tweens=True)
File '.build/venv/lib/python2.7/site-packages/pyramid/router.py', line 198 in invoke_subrequest
response = handle_request(request)
File '.build/venv/lib/python2.7/site-packages/c2cgeoportal/lib/cacheversion.py', line 62 in __call__
response = self.handler(request)
File '.build/venv/lib/python2.7/site-packages/pyramid/tweens.py', line 20 in excview_tween
response = handler(request)
File '.build/venv/lib/python2.7/site-packages/pyramid_tm/__init__.py', line 119 in tm_tween
reraise(*exc_info)
File '.build/venv/lib/python2.7/site-packages/pyramid_tm/__init__.py', line 98 in tm_tween
response = handler(request)
File '.build/venv/lib/python2.7/site-packages/pyramid/router.py', line 145 in handle_request
view_name
File '.build/venv/lib/python2.7/site-packages/pyramid/view.py', line 541 in _call_view
response = view_callable(context, request)
File '.build/venv/lib/python2.7/site-packages/c2cgeoportal/pyramid_.py', line 120 in new_f
return f(root, request)
File '.build/venv/lib/python2.7/site-packages/pyramid/config/views.py', line 377 in rendered_view
context)
File '.build/venv/lib/python2.7/site-packages/pyramid/renderers.py', line 450 in render_view
return self.render_to_response(response, system, request=request)
File '.build/venv/lib/python2.7/site-packages/pyramid/renderers.py', line 473 in render_to_response
result = self.render(value, system_values, request=request)
File '.build/venv/lib/python2.7/site-packages/pyramid/renderers.py', line 469 in render
result = renderer(value, system_values)
File '.build/venv/lib/python2.7/site-packages/pyramid_mako/__init__.py', line 156 in __call__
reraise(MakoRenderingException(errtext), None, exc_info[2])
File '.build/venv/lib/python2.7/site-packages/pyramid_mako/__init__.py', line 148 in __call__
result = template.render_unicode(**system)
File '.build/venv/lib/python2.7/site-packages/mako/template.py', line 471 in render_unicode
as_unicode=True)
File '.build/venv/lib/python2.7/site-packages/mako/runtime.py', line 838 in _render
**_kwargs_for_callable(callable_, data))
File '.build/venv/lib/python2.7/site-packages/mako/runtime.py', line 873 in _render_context
_exec_template(inherit, lclcontext, args=args, kwargs=kwargs)
File '.build/venv/lib/python2.7/site-packages/mako/runtime.py', line 899 in _exec_template
callable_(context, *args, **kwargs)
File '_desktop_js', line 43 in render_body
File '.build/venv/lib/python2.7/site-packages/pyramid/url.py', line 266 in route_url
raise KeyError('No such route named %s' % route_name)
``` | infrastructure | ogc proxy is gone is ogc proxy gone because after upgrade to i get this error makorenderingexception traceback most recent call last file build venv lib site packages pyramid mako init py line in call result template render unicode system file build venv lib site packages mako template py line in render unicode as unicode true file build venv lib site packages mako runtime py line in render kwargs for callable callable data file build venv lib site packages mako runtime py line in render context exec template inherit lclcontext args args kwargs kwargs file build venv lib site packages mako runtime py line in exec template callable context args kwargs file nyon templates desktop js line in render body openlayers proxyhost request route url ogcproxy n url file build venv lib site packages pyramid url py line in route url raise keyerror no such route named s route name keyerror no such route named ogcproxy full stack file build venv lib site packages weberror evalexception py line in respond app iter self application environ detect start response file build venv lib site packages fanstatic publisher py line in call return self app environ start response file build venv lib site packages fanstatic injector py line in call response request get response self app file build venv lib site packages webob request py line in send application catch exc info false file build venv lib site packages webob request py line in call application app iter application self environ start response file build venv lib site packages pyramid router py line in call response self invoke subrequest request use tweens true file build venv lib site packages pyramid router py line in invoke subrequest response handle request request file build venv lib site packages lib cacheversion py line in call response self handler request file build venv lib site packages pyramid tweens py line in excview tween response handler request file build venv lib site packages pyramid tm init py line in tm tween reraise exc info file build venv lib site packages pyramid tm init py line in tm tween response handler request file build venv lib site packages pyramid router py line in handle request view name file build venv lib site packages pyramid view py line in call view response view callable context request file build venv lib site packages pyramid py line in new f return f root request file build venv lib site packages pyramid config views py line in rendered view context file build venv lib site packages pyramid renderers py line in render view return self render to response response system request request file build venv lib site packages pyramid renderers py line in render to response result self render value system values request request file build venv lib site packages pyramid renderers py line in render result renderer value system values file build venv lib site packages pyramid mako init py line in call reraise makorenderingexception errtext none exc info file build venv lib site packages pyramid mako init py line in call result template render unicode system file build venv lib site packages mako template py line in render unicode as unicode true file build venv lib site packages mako runtime py line in render kwargs for callable callable data file build venv lib site packages mako runtime py line in render context exec template inherit lclcontext args args kwargs kwargs file build venv lib site packages mako runtime py line in exec template callable context args kwargs file desktop js line in render body file build venv lib site packages pyramid url py line in route url raise keyerror no such route named s route name | 1 |
16,123 | 11,844,310,102 | IssuesEvent | 2020-03-24 05:22:17 | zachsanchez113/tableau | https://api.github.com/repos/zachsanchez113/tableau | closed | Create dev environment | infrastructure wontfix | This will be critical once we try and roll out a webapp, plus it would be helpful when making changes to the API | 1.0 | Create dev environment - This will be critical once we try and roll out a webapp, plus it would be helpful when making changes to the API | infrastructure | create dev environment this will be critical once we try and roll out a webapp plus it would be helpful when making changes to the api | 1 |
288,636 | 8,849,759,068 | IssuesEvent | 2019-01-08 11:10:17 | InFact-coop/create-refresh | https://api.github.com/repos/InFact-coop/create-refresh | closed | Connect sign up form to Mailchimp account | in-progress priority-2 question | - use Mailchimp API instead of typeform
- need details from Create.Refresh | 1.0 | Connect sign up form to Mailchimp account - - use Mailchimp API instead of typeform
- need details from Create.Refresh | non_infrastructure | connect sign up form to mailchimp account use mailchimp api instead of typeform need details from create refresh | 0 |
28,231 | 23,097,967,946 | IssuesEvent | 2022-07-26 21:43:52 | ampproject/amp-wp | https://api.github.com/repos/ampproject/amp-wp | closed | Close off service container after initialization | Infrastructure WS:Core | > * The service container currently allows anyone with access to it to `put()` new services into it, even overwriting existing ones. Is this something we want to restrict partially or fully?
**@westonruter:** Will we need that functionality? If not, I'd say remove it.
_Originally posted in https://github.com/ampproject/amp-wp/pull/4877#issuecomment-645553721_
We cannot remove the `put()`, as it is used for initially populating it. But we can add logic to close it off and prohibit further changes after the initial setup. | 1.0 | Close off service container after initialization - > * The service container currently allows anyone with access to it to `put()` new services into it, even overwriting existing ones. Is this something we want to restrict partially or fully?
**@westonruter:** Will we need that functionality? If not, I'd say remove it.
_Originally posted in https://github.com/ampproject/amp-wp/pull/4877#issuecomment-645553721_
We cannot remove the `put()`, as it is used for initially populating it. But we can add logic to close it off and prohibit further changes after the initial setup. | infrastructure | close off service container after initialization the service container currently allows anyone with access to it to put new services into it even overwriting existing ones is this something we want to restrict partially or fully westonruter will we need that functionality if not i d say remove it originally posted in we cannot remove the put as it is used for initially populating it but we can add logic to close it off and prohibit further changes after the initial setup | 1 |
33,238 | 27,323,264,203 | IssuesEvent | 2023-02-24 22:11:05 | UBCSailbot/sailbot_workspace | https://api.github.com/repos/UBCSailbot/sailbot_workspace | closed | Explore WSL v1.0.0 | devcontainer infrastructure | ### Purpose
<!-- What is the purpose of this issue? What does this issue hope to achieve? -->
WSL v1.0.0 is available for Windows 10 and 11 on the Microsoft Store. It should natively support GUI applications, which previously wasn't possible on windows 10. If we can confirm this to be the case, we should update the repository and its setup instructions accordingly to simplify the process.
### Changes
<!-- What changes will be made to the infrastructure of our GitHub organization? How will it be better? -->
- Remove the `DISPLAY` windows 10 hack in devcontainer.json
- In setup instructions in the docs repo
- Remove `DISPLAY` windows 10 hack from setup instructions
- Add instructions to use new WSL (I think it's as simple as `wsl --upgrade`)
- After these changes, the Windows installation instructions should look something like this
- Enable virtualization and WSL and hypervisor Windows features
- Install WSL and Ubuntu apps from Microsoft store
- Install docker desktop
- Troubleshooting
- Convert from WSL 1 to 2
- Use new WSL app on Microsoft store
- Set default distro to Ubuntu
### Resources
<!-- Link to any extra resources that might help with the enhancements and describe the relevance if not obvious. -->
- https://devblogs.microsoft.com/commandline/the-windows-subsystem-for-linux-in-the-microsoft-store-is-now-generally-available-on-windows-10-and-11/
| 1.0 | Explore WSL v1.0.0 - ### Purpose
<!-- What is the purpose of this issue? What does this issue hope to achieve? -->
WSL v1.0.0 is available for Windows 10 and 11 on the Microsoft Store. It should natively support GUI applications, which previously wasn't possible on windows 10. If we can confirm this to be the case, we should update the repository and its setup instructions accordingly to simplify the process.
### Changes
<!-- What changes will be made to the infrastructure of our GitHub organization? How will it be better? -->
- Remove the `DISPLAY` windows 10 hack in devcontainer.json
- In setup instructions in the docs repo
- Remove `DISPLAY` windows 10 hack from setup instructions
- Add instructions to use new WSL (I think it's as simple as `wsl --upgrade`)
- After these changes, the Windows installation instructions should look something like this
- Enable virtualization and WSL and hypervisor Windows features
- Install WSL and Ubuntu apps from Microsoft store
- Install docker desktop
- Troubleshooting
- Convert from WSL 1 to 2
- Use new WSL app on Microsoft store
- Set default distro to Ubuntu
### Resources
<!-- Link to any extra resources that might help with the enhancements and describe the relevance if not obvious. -->
- https://devblogs.microsoft.com/commandline/the-windows-subsystem-for-linux-in-the-microsoft-store-is-now-generally-available-on-windows-10-and-11/
| infrastructure | explore wsl purpose wsl is available for windows and on the microsoft store it should natively support gui applications which previously wasn t possible on windows if we can confirm this to be the case we should update the repository and its setup instructions accordingly to simplify the process changes remove the display windows hack in devcontainer json in setup instructions in the docs repo remove display windows hack from setup instructions add instructions to use new wsl i think it s as simple as wsl upgrade after these changes the windows installation instructions should look something like this enable virtualization and wsl and hypervisor windows features install wsl and ubuntu apps from microsoft store install docker desktop troubleshooting convert from wsl to use new wsl app on microsoft store set default distro to ubuntu resources | 1 |
265,634 | 20,105,630,661 | IssuesEvent | 2022-02-07 10:12:33 | Requisitos-de-Software/2021.2-Prime-video | https://api.github.com/repos/Requisitos-de-Software/2021.2-Prime-video | closed | Criar GitHub Pages | documentation | ## Criterios de aprovação:
- [x] Criar gh-pages.
- [x] Documentação atualizada no gh-pages. | 1.0 | Criar GitHub Pages - ## Criterios de aprovação:
- [x] Criar gh-pages.
- [x] Documentação atualizada no gh-pages. | non_infrastructure | criar github pages criterios de aprovação criar gh pages documentação atualizada no gh pages | 0 |
215,073 | 7,286,465,900 | IssuesEvent | 2018-02-23 09:50:55 | kcigeospatial/balt_co_ETL | https://api.github.com/repos/kcigeospatial/balt_co_ETL | closed | Stormwater - RestBMP - target RestBMP.BMP_DRAIN_ID does not populate | high priority item | For the outputted RestBMP feature class, the BMP_DRAIN_ID field is consistently empty. It should be populated with the MDE ID for the associated drainage area feature.
I've confirmed that for the tested data, there is a drainage area associated to the restoration BMP, and the DA has valid, populated data. | 1.0 | Stormwater - RestBMP - target RestBMP.BMP_DRAIN_ID does not populate - For the outputted RestBMP feature class, the BMP_DRAIN_ID field is consistently empty. It should be populated with the MDE ID for the associated drainage area feature.
I've confirmed that for the tested data, there is a drainage area associated to the restoration BMP, and the DA has valid, populated data. | non_infrastructure | stormwater restbmp target restbmp bmp drain id does not populate for the outputted restbmp feature class the bmp drain id field is consistently empty it should be populated with the mde id for the associated drainage area feature i ve confirmed that for the tested data there is a drainage area associated to the restoration bmp and the da has valid populated data | 0 |
89,195 | 17,795,185,111 | IssuesEvent | 2021-08-31 21:07:29 | skycoin/cx-game | https://api.github.com/repos/skycoin/cx-game | closed | Pipes | code | # Data
Planet needs an addition layer. This layer is dedicated to pipe tiles. Each pipe needs to store which neighbours it is connected to. A pipe may be joined with any combination of its neighboring pipes.
# Tools
Pipe placement tool is used by the player to place pipes. Pipe modification tool is a separate item which the player uses to adjust which neighbours a pipe is connected to.
| 1.0 | Pipes - # Data
Planet needs an addition layer. This layer is dedicated to pipe tiles. Each pipe needs to store which neighbours it is connected to. A pipe may be joined with any combination of its neighboring pipes.
# Tools
Pipe placement tool is used by the player to place pipes. Pipe modification tool is a separate item which the player uses to adjust which neighbours a pipe is connected to.
| non_infrastructure | pipes data planet needs an addition layer this layer is dedicated to pipe tiles each pipe needs to store which neighbours it is connected to a pipe may be joined with any combination of its neighboring pipes tools pipe placement tool is used by the player to place pipes pipe modification tool is a separate item which the player uses to adjust which neighbours a pipe is connected to | 0 |
727,076 | 25,022,437,533 | IssuesEvent | 2022-11-04 03:01:29 | milvus-io/milvus | https://api.github.com/repos/milvus-io/milvus | closed | [Bug]: Meta migrate may fail when upgrading | kind/bug priority/critical-urgent severity/critical triage/accepted | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Environment
```markdown
- Milvus version: 2.1.4 --> master-20221031-48f5c600
- Deployment mode(standalone or cluster):
- SDK version(e.g. pymilvus v2.0.0rc2):
- OS(Ubuntu or CentOS):
- CPU/Memory:
- GPU:
- Others:
```
### Current Behavior
```
+ ./migrate.sh -i cluster-upgrade-400 -n chaos-testing -s 2.1.4 -t 2.2.0 -m harbor.milvus.io/milvus/meta-migration:20221028-8cb5122d4 -w harbor.milvus.io/dockerhub/milvusdb/milvus:master-20221031-48f5c600
Migration milvus meta will take four steps:
1. Stop the milvus components
2. Backup the milvus meta
3. Migrate the milvus meta
4. Startup milvus components in new image version
Stop milvus deployments: cluster-upgrade-400-milvus-datacoord cluster-upgrade-400-milvus-datanode cluster-upgrade-400-milvus-indexcoord cluster-upgrade-400-milvus-indexnode cluster-upgrade-400-milvus-proxy cluster-upgrade-400-milvus-querycoord cluster-upgrade-400-milvus-querynode cluster-upgrade-400-milvus-rootcoord
Stopped...
Starting to migrate milvus meta...
persistentvolumeclaim/milvus-meta-migration-backup-cluster-upgrade-400 created
configmap/milvus-meta-migration-config-cluster-upgrade-400 created
Backuping meta...
Checking whether backup exists...
pod/milvus-meta-migration-backup-cluster-upgrade-400 created
Meta backup is done...
configmap/milvus-meta-migration-config-cluster-upgrade-400 annotated
Migrating meta...
pod/milvus-meta-migration-cluster-upgrade-400 created
Migration is failed...
Here is the log:
[0;33mCmd: run, SourceVersion: 2.1.4, TargetVersion: 2.2.0, BackupFilePath: /milvus/data/migration.bak, RunWithBackup: false[0m
[2022/10/31 08:35:09.493 +00:00] [DEBUG] [sessionutil/session_util.go:185] ["Session try to connect to etcd"]
[2022/10/31 08:35:09.497 +00:00] [DEBUG] [sessionutil/session_util.go:200] ["Session connect to etcd success"]
[2022/10/31 08:35:09.500 +00:00] [DEBUG] [sessionutil/session_util.go:282] ["Session get serverID success"] [key=id] [ServerId=15]
[2022/10/31 08:35:09.502 +00:00] [DEBUG] [sessionutil/session_util.go:422] ["SessionUtil GetSessions "] [prefix=rootcoord] [key=rootcoord] [address=10.102.7.219:53100]
[2022/10/31 08:35:09.502 +00:00] [DEBUG] [sessionutil/session_util.go:422] ["SessionUtil GetSessions "] [prefix=rootcoord] [key=rootcoord] [address=10.102.7.219:53100]
[0;31mthere are still sessions alive, prefix: rootcoord, num of alive sessions: 1[0m
```
```
+ kubectl get pods -o wide
+ grep cluster-upgrade-400
cluster-upgrade-400-etcd-0 1/1 Running 0 34m 10.102.7.227 devops-node11 <none> <none>
cluster-upgrade-400-etcd-1 1/1 Running 0 34m 10.102.10.73 devops-node20 <none> <none>
cluster-upgrade-400-etcd-2 1/1 Running 0 34m 10.102.6.169 devops-node10 <none> <none>
cluster-upgrade-400-kafka-0 2/2 Running 2 (33m ago) 34m 10.102.10.70 devops-node20 <none> <none>
cluster-upgrade-400-kafka-1 2/2 Running 1 (33m ago) 34m 10.102.7.229 devops-node11 <none> <none>
cluster-upgrade-400-kafka-2 2/2 Running 3 (33m ago) 34m 10.102.6.161 devops-node10 <none> <none>
cluster-upgrade-400-kafka-exporter-7fb6f5bd5b-b5tph 1/1 Running 4 (33m ago) 34m 10.102.6.155 devops-node10 <none> <none>
cluster-upgrade-400-milvus-rootcoord-74cdc7b8f4-vpxm5 1/1 Terminating 0 34m 10.102.7.219 devops-node11 <none> <none>
cluster-upgrade-400-minio-0 1/1 Running 0 34m 10.102.10.71 devops-node20 <none> <none>
cluster-upgrade-400-minio-1 1/1 Running 0 34m 10.102.7.231 devops-node11 <none> <none>
cluster-upgrade-400-minio-2 1/1 Running 0 34m 10.102.6.165 devops-node10 <none> <none>
cluster-upgrade-400-minio-3 1/1 Running 0 34m 10.102.8.198 devops-node12 <none> <none>
cluster-upgrade-400-zookeeper-0 1/1 Running 0 34m 10.102.5.11 devops-node21 <none> <none>
cluster-upgrade-400-zookeeper-1 1/1 Running 0 34m 10.102.6.164 devops-node10 <none> <none>
cluster-upgrade-400-zookeeper-2 1/1 Running 0 34m 10.102.7.230 devops-node11 <none> <none>
milvus-meta-migration-backup-cluster-upgrade-400 0/1 Completed 0 23s 10.102.6.181 devops-node10 <none> <none>
milvus-meta-migration-cluster-upgrade-400 0/1 Error 0 12s 10.102.6.182 devops-node10 <none> <none>
```
### Expected Behavior
Success to upgrade
### Steps To Reproduce
_No response_
### Milvus Log
_No response_
### Anything else?
_No response_ | 1.0 | [Bug]: Meta migrate may fail when upgrading - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Environment
```markdown
- Milvus version: 2.1.4 --> master-20221031-48f5c600
- Deployment mode(standalone or cluster):
- SDK version(e.g. pymilvus v2.0.0rc2):
- OS(Ubuntu or CentOS):
- CPU/Memory:
- GPU:
- Others:
```
### Current Behavior
```
+ ./migrate.sh -i cluster-upgrade-400 -n chaos-testing -s 2.1.4 -t 2.2.0 -m harbor.milvus.io/milvus/meta-migration:20221028-8cb5122d4 -w harbor.milvus.io/dockerhub/milvusdb/milvus:master-20221031-48f5c600
Migration milvus meta will take four steps:
1. Stop the milvus components
2. Backup the milvus meta
3. Migrate the milvus meta
4. Startup milvus components in new image version
Stop milvus deployments: cluster-upgrade-400-milvus-datacoord cluster-upgrade-400-milvus-datanode cluster-upgrade-400-milvus-indexcoord cluster-upgrade-400-milvus-indexnode cluster-upgrade-400-milvus-proxy cluster-upgrade-400-milvus-querycoord cluster-upgrade-400-milvus-querynode cluster-upgrade-400-milvus-rootcoord
Stopped...
Starting to migrate milvus meta...
persistentvolumeclaim/milvus-meta-migration-backup-cluster-upgrade-400 created
configmap/milvus-meta-migration-config-cluster-upgrade-400 created
Backuping meta...
Checking whether backup exists...
pod/milvus-meta-migration-backup-cluster-upgrade-400 created
Meta backup is done...
configmap/milvus-meta-migration-config-cluster-upgrade-400 annotated
Migrating meta...
pod/milvus-meta-migration-cluster-upgrade-400 created
Migration is failed...
Here is the log:
[0;33mCmd: run, SourceVersion: 2.1.4, TargetVersion: 2.2.0, BackupFilePath: /milvus/data/migration.bak, RunWithBackup: false[0m
[2022/10/31 08:35:09.493 +00:00] [DEBUG] [sessionutil/session_util.go:185] ["Session try to connect to etcd"]
[2022/10/31 08:35:09.497 +00:00] [DEBUG] [sessionutil/session_util.go:200] ["Session connect to etcd success"]
[2022/10/31 08:35:09.500 +00:00] [DEBUG] [sessionutil/session_util.go:282] ["Session get serverID success"] [key=id] [ServerId=15]
[2022/10/31 08:35:09.502 +00:00] [DEBUG] [sessionutil/session_util.go:422] ["SessionUtil GetSessions "] [prefix=rootcoord] [key=rootcoord] [address=10.102.7.219:53100]
[2022/10/31 08:35:09.502 +00:00] [DEBUG] [sessionutil/session_util.go:422] ["SessionUtil GetSessions "] [prefix=rootcoord] [key=rootcoord] [address=10.102.7.219:53100]
[0;31mthere are still sessions alive, prefix: rootcoord, num of alive sessions: 1[0m
```
```
+ kubectl get pods -o wide
+ grep cluster-upgrade-400
cluster-upgrade-400-etcd-0 1/1 Running 0 34m 10.102.7.227 devops-node11 <none> <none>
cluster-upgrade-400-etcd-1 1/1 Running 0 34m 10.102.10.73 devops-node20 <none> <none>
cluster-upgrade-400-etcd-2 1/1 Running 0 34m 10.102.6.169 devops-node10 <none> <none>
cluster-upgrade-400-kafka-0 2/2 Running 2 (33m ago) 34m 10.102.10.70 devops-node20 <none> <none>
cluster-upgrade-400-kafka-1 2/2 Running 1 (33m ago) 34m 10.102.7.229 devops-node11 <none> <none>
cluster-upgrade-400-kafka-2 2/2 Running 3 (33m ago) 34m 10.102.6.161 devops-node10 <none> <none>
cluster-upgrade-400-kafka-exporter-7fb6f5bd5b-b5tph 1/1 Running 4 (33m ago) 34m 10.102.6.155 devops-node10 <none> <none>
cluster-upgrade-400-milvus-rootcoord-74cdc7b8f4-vpxm5 1/1 Terminating 0 34m 10.102.7.219 devops-node11 <none> <none>
cluster-upgrade-400-minio-0 1/1 Running 0 34m 10.102.10.71 devops-node20 <none> <none>
cluster-upgrade-400-minio-1 1/1 Running 0 34m 10.102.7.231 devops-node11 <none> <none>
cluster-upgrade-400-minio-2 1/1 Running 0 34m 10.102.6.165 devops-node10 <none> <none>
cluster-upgrade-400-minio-3 1/1 Running 0 34m 10.102.8.198 devops-node12 <none> <none>
cluster-upgrade-400-zookeeper-0 1/1 Running 0 34m 10.102.5.11 devops-node21 <none> <none>
cluster-upgrade-400-zookeeper-1 1/1 Running 0 34m 10.102.6.164 devops-node10 <none> <none>
cluster-upgrade-400-zookeeper-2 1/1 Running 0 34m 10.102.7.230 devops-node11 <none> <none>
milvus-meta-migration-backup-cluster-upgrade-400 0/1 Completed 0 23s 10.102.6.181 devops-node10 <none> <none>
milvus-meta-migration-cluster-upgrade-400 0/1 Error 0 12s 10.102.6.182 devops-node10 <none> <none>
```
### Expected Behavior
Success to upgrade
### Steps To Reproduce
_No response_
### Milvus Log
_No response_
### Anything else?
_No response_ | non_infrastructure | meta migrate may fail when upgrading is there an existing issue for this i have searched the existing issues environment markdown milvus version master deployment mode standalone or cluster sdk version e g pymilvus os ubuntu or centos cpu memory gpu others current behavior migrate sh i cluster upgrade n chaos testing s t m harbor milvus io milvus meta migration w harbor milvus io dockerhub milvusdb milvus master migration milvus meta will take four steps stop the milvus components backup the milvus meta migrate the milvus meta startup milvus components in new image version stop milvus deployments cluster upgrade milvus datacoord cluster upgrade milvus datanode cluster upgrade milvus indexcoord cluster upgrade milvus indexnode cluster upgrade milvus proxy cluster upgrade milvus querycoord cluster upgrade milvus querynode cluster upgrade milvus rootcoord stopped starting to migrate milvus meta persistentvolumeclaim milvus meta migration backup cluster upgrade created configmap milvus meta migration config cluster upgrade created backuping meta checking whether backup exists pod milvus meta migration backup cluster upgrade created meta backup is done configmap milvus meta migration config cluster upgrade annotated migrating meta pod milvus meta migration cluster upgrade created migration is failed here is the log run sourceversion targetversion backupfilepath milvus data migration bak runwithbackup false are still sessions alive prefix rootcoord num of alive sessions kubectl get pods o wide grep cluster upgrade cluster upgrade etcd running devops cluster upgrade etcd running devops cluster upgrade etcd running devops cluster upgrade kafka running ago devops cluster upgrade kafka running ago devops cluster upgrade kafka running ago devops cluster upgrade kafka exporter running ago devops cluster upgrade milvus rootcoord terminating devops cluster upgrade minio running devops cluster upgrade minio running devops cluster upgrade minio running devops cluster upgrade minio running devops cluster upgrade zookeeper running devops cluster upgrade zookeeper running devops cluster upgrade zookeeper running devops milvus meta migration backup cluster upgrade completed devops milvus meta migration cluster upgrade error devops expected behavior success to upgrade steps to reproduce no response milvus log no response anything else no response | 0 |
6,452 | 6,420,406,728 | IssuesEvent | 2017-08-09 00:02:40 | dart-lang/site-webdev | https://api.github.com/repos/dart-lang/site-webdev | closed | Update Angular API documentation for 4.0.0-alpha+2 | Infrastructure | * https://webdev.dartlang.org/angular/version says `4.0.0-alpha`
* [Staging site](https://webdev-dartlang-org-dev.firebaseapp.com/angular/guide) is `4.0.0-alpha+1`
| 1.0 | Update Angular API documentation for 4.0.0-alpha+2 - * https://webdev.dartlang.org/angular/version says `4.0.0-alpha`
* [Staging site](https://webdev-dartlang-org-dev.firebaseapp.com/angular/guide) is `4.0.0-alpha+1`
| infrastructure | update angular api documentation for alpha says alpha is alpha | 1 |
8,426 | 7,422,795,938 | IssuesEvent | 2018-03-23 01:24:42 | connormlewis/idb | https://api.github.com/repos/connormlewis/idb | closed | Write acceptance tests | frontend infrastructure | We need acceptance tests for the frontend GUI. Tests will need to be done using Selenium.
Estimated time: 5 hours | 1.0 | Write acceptance tests - We need acceptance tests for the frontend GUI. Tests will need to be done using Selenium.
Estimated time: 5 hours | infrastructure | write acceptance tests we need acceptance tests for the frontend gui tests will need to be done using selenium estimated time hours | 1 |
801,187 | 28,456,251,001 | IssuesEvent | 2023-04-17 07:18:36 | plaidml/tpp-mlir | https://api.github.com/repos/plaidml/tpp-mlir | closed | Convert `linagl.generic { mul + add }` to `linalg.matmul` before blocking & tiling | high-priority | Our block and tile pass only looks at named ops, and upstream models emit generics, so we need to convert them to named ops before we try to block and fuse, otherwise we won't get decent performance from most upstream code.
This also helps our unpacked benchmarks, which is one of our main priorities right now.
@chelini | 1.0 | Convert `linagl.generic { mul + add }` to `linalg.matmul` before blocking & tiling - Our block and tile pass only looks at named ops, and upstream models emit generics, so we need to convert them to named ops before we try to block and fuse, otherwise we won't get decent performance from most upstream code.
This also helps our unpacked benchmarks, which is one of our main priorities right now.
@chelini | non_infrastructure | convert linagl generic mul add to linalg matmul before blocking tiling our block and tile pass only looks at named ops and upstream models emit generics so we need to convert them to named ops before we try to block and fuse otherwise we won t get decent performance from most upstream code this also helps our unpacked benchmarks which is one of our main priorities right now chelini | 0 |
368,263 | 10,869,003,091 | IssuesEvent | 2019-11-15 06:07:25 | horizontalsystems/HS-Design | https://api.github.com/repos/horizontalsystems/HS-Design | closed | Balance Tab : minor design improvements | minor priority | As the number of tokens supported by the app is growing we need to be able to differentiate between the same token on multiple blockchains i.e. Binance Token on Binance DEX from The Binance Token on Ethereum.
I propose following:
- [ ] both in collapsed/expanded coin cell append coin type next to coin symbol
-- if token is on Ethereum >> [coin symbol](ERC20)
-- if token is on EOS >> [coin symbol](EOSIO)
-- if token is on Binance DEX >> [coin symbol](BEP2) | 1.0 | Balance Tab : minor design improvements - As the number of tokens supported by the app is growing we need to be able to differentiate between the same token on multiple blockchains i.e. Binance Token on Binance DEX from The Binance Token on Ethereum.
I propose following:
- [ ] both in collapsed/expanded coin cell append coin type next to coin symbol
-- if token is on Ethereum >> [coin symbol](ERC20)
-- if token is on EOS >> [coin symbol](EOSIO)
-- if token is on Binance DEX >> [coin symbol](BEP2) | non_infrastructure | balance tab minor design improvements as the number of tokens supported by the app is growing we need to be able to differentiate between the same token on multiple blockchains i e binance token on binance dex from the binance token on ethereum i propose following both in collapsed expanded coin cell append coin type next to coin symbol if token is on ethereum if token is on eos eosio if token is on binance dex | 0 |
14,976 | 11,275,339,778 | IssuesEvent | 2020-01-14 20:33:01 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Update cross build-rootfs for modern world | area-Infrastructure untriaged | As part of building docker image for arm64 cross compiling I spent a lot of time looking at build--rootfs.sh. The following seem like changes which should be done...
+ Add latest Ubuntu version: Artful and Bionic
+ Remove unsupported Ubuntu versions: Vivid and Wily
+ Add latest lldbdev versions: 4.0, 5.0, 6.0
+ Update lldb default to 3.9 to be consistent with non-cross build
+ Drop old lldb versions 3.6 and 3.8
Similar changes should be made in corefx and coreclr
@janvorli FYI | 1.0 | Update cross build-rootfs for modern world - As part of building docker image for arm64 cross compiling I spent a lot of time looking at build--rootfs.sh. The following seem like changes which should be done...
+ Add latest Ubuntu version: Artful and Bionic
+ Remove unsupported Ubuntu versions: Vivid and Wily
+ Add latest lldbdev versions: 4.0, 5.0, 6.0
+ Update lldb default to 3.9 to be consistent with non-cross build
+ Drop old lldb versions 3.6 and 3.8
Similar changes should be made in corefx and coreclr
@janvorli FYI | infrastructure | update cross build rootfs for modern world as part of building docker image for cross compiling i spent a lot of time looking at build rootfs sh the following seem like changes which should be done add latest ubuntu version artful and bionic remove unsupported ubuntu versions vivid and wily add latest lldbdev versions update lldb default to to be consistent with non cross build drop old lldb versions and similar changes should be made in corefx and coreclr janvorli fyi | 1 |
13,235 | 10,165,217,547 | IssuesEvent | 2019-08-07 13:28:48 | bird-team/brisbane-bird-atlas | https://api.github.com/repos/bird-team/brisbane-bird-atlas | closed | Backup Old Assets at start of new rebuild | enhancement infrastructure | A random thought I was having today - I think it would be good if, at the start of each new `make assets` rebuild the system backed up the current set of assets to a .zip folder or similar, so that one can easily compare the two versions and ensure that the desired outcome has been reached.
This is especially useful when rebuilding assets after a data update, to ensure that it has actually gone through, or after a functional change to the code (e.g. changing the surveyor sheet formatting).
I think you could probably also put it in a folder with .gitignore (?) such that it is only kept locally on the user's computer and doesn't take up unnecessary space on the repo?
Something like:
`make assets`
*compress current \assets folder into a new `assets-backup.zip` folder which gets placed in old\assets-backup*
*delete current assets folder*
*continue on as per current code*
Does that make sense / sound possible? | 1.0 | Backup Old Assets at start of new rebuild - A random thought I was having today - I think it would be good if, at the start of each new `make assets` rebuild the system backed up the current set of assets to a .zip folder or similar, so that one can easily compare the two versions and ensure that the desired outcome has been reached.
This is especially useful when rebuilding assets after a data update, to ensure that it has actually gone through, or after a functional change to the code (e.g. changing the surveyor sheet formatting).
I think you could probably also put it in a folder with .gitignore (?) such that it is only kept locally on the user's computer and doesn't take up unnecessary space on the repo?
Something like:
`make assets`
*compress current \assets folder into a new `assets-backup.zip` folder which gets placed in old\assets-backup*
*delete current assets folder*
*continue on as per current code*
Does that make sense / sound possible? | infrastructure | backup old assets at start of new rebuild a random thought i was having today i think it would be good if at the start of each new make assets rebuild the system backed up the current set of assets to a zip folder or similar so that one can easily compare the two versions and ensure that the desired outcome has been reached this is especially useful when rebuilding assets after a data update to ensure that it has actually gone through or after a functional change to the code e g changing the surveyor sheet formatting i think you could probably also put it in a folder with gitignore such that it is only kept locally on the user s computer and doesn t take up unnecessary space on the repo something like make assets compress current assets folder into a new assets backup zip folder which gets placed in old assets backup delete current assets folder continue on as per current code does that make sense sound possible | 1 |
54,836 | 13,959,987,210 | IssuesEvent | 2020-10-24 18:54:30 | AeroScripts/QuestieDev | https://api.github.com/repos/AeroScripts/QuestieDev | closed | Wrong class colors | Type - Defect | 
How do I change it so the class colours are the same? It has orange for both of us but my other party member is a hunter. happens with no other addons on and using the latest version. | 1.0 | Wrong class colors - 
How do I change it so the class colours are the same? It has orange for both of us but my other party member is a hunter. happens with no other addons on and using the latest version. | non_infrastructure | wrong class colors how do i change it so the class colours are the same it has orange for both of us but my other party member is a hunter happens with no other addons on and using the latest version | 0 |
12,995 | 10,059,974,254 | IssuesEvent | 2019-07-22 17:42:20 | CodeSignal/dockerfiles | https://api.github.com/repos/CodeSignal/dockerfiles | opened | Create a Docker image for automated user tests | Infrastructure | https://app.asana.com/0/417895759505117/1132358108136274/f
Preferably not Galen. Needs Java, Selenium, some kind of driver -- WebdriverIO is fine, especially since we have experience with it already and it's a JS wrapper (no Java is great, and the sync API is nice too).
Consider the kinds of setup we do to get WebdriverIO running on CircleCI -- the need here is similar. | 1.0 | Create a Docker image for automated user tests - https://app.asana.com/0/417895759505117/1132358108136274/f
Preferably not Galen. Needs Java, Selenium, some kind of driver -- WebdriverIO is fine, especially since we have experience with it already and it's a JS wrapper (no Java is great, and the sync API is nice too).
Consider the kinds of setup we do to get WebdriverIO running on CircleCI -- the need here is similar. | infrastructure | create a docker image for automated user tests preferably not galen needs java selenium some kind of driver webdriverio is fine especially since we have experience with it already and it s a js wrapper no java is great and the sync api is nice too consider the kinds of setup we do to get webdriverio running on circleci the need here is similar | 1 |
72,639 | 9,603,119,916 | IssuesEvent | 2019-05-10 16:10:25 | cityofaustin/atd-knack-data-tracker | https://api.github.com/repos/cityofaustin/atd-knack-data-tracker | closed | Create User Stories for Signs Migration to Knack | Project: Signs Migration Team: Apps Type: Documentation Workgroup: SMB | - [x] Create Signs Migration to Knack user stories
- [x] Create Issues in Data Tracker Repo when completed | 1.0 | Create User Stories for Signs Migration to Knack - - [x] Create Signs Migration to Knack user stories
- [x] Create Issues in Data Tracker Repo when completed | non_infrastructure | create user stories for signs migration to knack create signs migration to knack user stories create issues in data tracker repo when completed | 0 |
283,766 | 30,913,542,999 | IssuesEvent | 2023-08-05 02:11:22 | hshivhare67/kernel_v4.19.72 | https://api.github.com/repos/hshivhare67/kernel_v4.19.72 | reopened | CVE-2020-25705 (High) detected in linuxlinux-4.19.282 | Mend: dependency security vulnerability | ## CVE-2020-25705 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72/commit/139c4e073703974ca0b05255c4cff6dcd52a8e31">139c4e073703974ca0b05255c4cff6dcd52a8e31</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw in ICMP packets in the Linux kernel may allow an attacker to quickly scan open UDP ports. This flaw allows an off-path remote attacker to effectively bypass source port UDP randomization. Software that relies on UDP source port randomization are indirectly affected as well on the Linux Based Products (RUGGEDCOM RM1224: All versions between v5.0 and v6.4, SCALANCE M-800: All versions between v5.0 and v6.4, SCALANCE S615: All versions between v5.0 and v6.4, SCALANCE SC-600: All versions prior to v2.1.3, SCALANCE W1750D: v8.3.0.1, v8.6.0, and v8.7.0, SIMATIC Cloud Connect 7: All versions, SIMATIC MV500 Family: All versions, SIMATIC NET CP 1243-1 (incl. SIPLUS variants): Versions 3.1.39 and later, SIMATIC NET CP 1243-7 LTE EU: Version
<p>Publish Date: 2020-11-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25705>CVE-2020-25705</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25705">https://www.linuxkernelcves.com/cves/CVE-2020-25705</a></p>
<p>Release Date: 2020-11-17</p>
<p>Fix Resolution: v4.4.241,v4.9.241,v4.14.203,v4.19.153,v5.4.73,v5.8.17,v5.9.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-25705 (High) detected in linuxlinux-4.19.282 - ## CVE-2020-25705 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.282</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/hshivhare67/kernel_v4.19.72/commit/139c4e073703974ca0b05255c4cff6dcd52a8e31">139c4e073703974ca0b05255c4cff6dcd52a8e31</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw in ICMP packets in the Linux kernel may allow an attacker to quickly scan open UDP ports. This flaw allows an off-path remote attacker to effectively bypass source port UDP randomization. Software that relies on UDP source port randomization are indirectly affected as well on the Linux Based Products (RUGGEDCOM RM1224: All versions between v5.0 and v6.4, SCALANCE M-800: All versions between v5.0 and v6.4, SCALANCE S615: All versions between v5.0 and v6.4, SCALANCE SC-600: All versions prior to v2.1.3, SCALANCE W1750D: v8.3.0.1, v8.6.0, and v8.7.0, SIMATIC Cloud Connect 7: All versions, SIMATIC MV500 Family: All versions, SIMATIC NET CP 1243-1 (incl. SIPLUS variants): Versions 3.1.39 and later, SIMATIC NET CP 1243-7 LTE EU: Version
<p>Publish Date: 2020-11-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25705>CVE-2020-25705</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25705">https://www.linuxkernelcves.com/cves/CVE-2020-25705</a></p>
<p>Release Date: 2020-11-17</p>
<p>Fix Resolution: v4.4.241,v4.9.241,v4.14.203,v4.19.153,v5.4.73,v5.8.17,v5.9.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details a flaw in icmp packets in the linux kernel may allow an attacker to quickly scan open udp ports this flaw allows an off path remote attacker to effectively bypass source port udp randomization software that relies on udp source port randomization are indirectly affected as well on the linux based products ruggedcom all versions between and scalance m all versions between and scalance all versions between and scalance sc all versions prior to scalance and simatic cloud connect all versions simatic family all versions simatic net cp incl siplus variants versions and later simatic net cp lte eu version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
29,984 | 14,360,520,641 | IssuesEvent | 2020-11-30 16:57:54 | xamarin/Xamarin.Forms | https://api.github.com/repos/xamarin/Xamarin.Forms | closed | [Bug] XF ListView and CollectionView leak on Android | a/collectionview a/listview a/performance s/needs-info ❓ s/unverified t/bug :bug: | ### Description
ListView and CollectionView remain in Android memory even if page that holds them is wiped by garbage collection.
Works fine with UWP. Not tried with iOS.
### Steps to Reproduce
1. Push page with listview to navigation stack
2. Pop page with listview from navigation stack
3. Clear page content so that nothing is referencing it anymore.
4. Page is sent for garbage collectionig, its children ListView/CollectionView are not.
### Expected Behavior
All children of page should be released from memory when page is released. Even ListView and CollectionView
### Actual Behavior
Page is released from memory but its children ListView and CollectionView are not.
### Basic Information
- Version with issue:
- Last known good version:
- IDE: VS 2019
- Platform Target Frameworks: 4.6, 4.8, 5.0-pre
- iOS: dunno
- Android: 9.0
### Screenshots
https://prnt.sc/utw0rn
https://prnt.sc/utw12j
https://prnt.sc/utw1db
https://prnt.sc/utw1oa
https://prnt.sc/utw1wu
https://prnt.sc/utw252
https://prnt.sc/utw2ez
https://prnt.sc/utw2rl
### Reproduction Link
https://github.com/tiitk/LV
### Workaround
No workaround found. | True | [Bug] XF ListView and CollectionView leak on Android - ### Description
ListView and CollectionView remain in Android memory even if page that holds them is wiped by garbage collection.
Works fine with UWP. Not tried with iOS.
### Steps to Reproduce
1. Push page with listview to navigation stack
2. Pop page with listview from navigation stack
3. Clear page content so that nothing is referencing it anymore.
4. Page is sent for garbage collectionig, its children ListView/CollectionView are not.
### Expected Behavior
All children of page should be released from memory when page is released. Even ListView and CollectionView
### Actual Behavior
Page is released from memory but its children ListView and CollectionView are not.
### Basic Information
- Version with issue:
- Last known good version:
- IDE: VS 2019
- Platform Target Frameworks: 4.6, 4.8, 5.0-pre
- iOS: dunno
- Android: 9.0
### Screenshots
https://prnt.sc/utw0rn
https://prnt.sc/utw12j
https://prnt.sc/utw1db
https://prnt.sc/utw1oa
https://prnt.sc/utw1wu
https://prnt.sc/utw252
https://prnt.sc/utw2ez
https://prnt.sc/utw2rl
### Reproduction Link
https://github.com/tiitk/LV
### Workaround
No workaround found. | non_infrastructure | xf listview and collectionview leak on android description listview and collectionview remain in android memory even if page that holds them is wiped by garbage collection works fine with uwp not tried with ios steps to reproduce push page with listview to navigation stack pop page with listview from navigation stack clear page content so that nothing is referencing it anymore page is sent for garbage collectionig its children listview collectionview are not expected behavior all children of page should be released from memory when page is released even listview and collectionview actual behavior page is released from memory but its children listview and collectionview are not basic information version with issue last known good version ide vs platform target frameworks pre ios dunno android screenshots reproduction link workaround no workaround found | 0 |
21,298 | 14,501,922,392 | IssuesEvent | 2020-12-11 20:14:02 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Skip building "3.0.0"-branded targeting pack in 3.0.2+ source-build | area-Infrastructure-installer | Source-build will be building the `Microsoft.NETCore.App.Ref` `3.0.0` targeting pack in dotnet/source-build-reference-packages for 3.0.1+, however in 3.0.2+ I also merged some Core-Setup infrastructure that will have this repo build a "3.0.0"-branded targeting pack when building any 3.0.2+ commit: https://github.com/dotnet/arcade/pull/4318.
The source-build version is preferred because it's more static: we can trust it not to change due to accidental (or intentional but unwanted) changes to future versions of Core-Setup. There is danger of Core-Setup 3.0.2 building a "3.0.0"-branded targeting pack that isn't actually the same as the actual `v3.0.0` targeting pack, causing behavior differences in the source-built SDK. See https://github.com/dotnet/source-build/pull/1389
This infra needs to be disabled so the Core-Setup "3.0.0"-branded targeting pack doesn't conflict with the source-build 3.0.0 package.
This can be done without an Arcade update or source-build customization by setting `SkipSetLastReleasedVersionForSourceBuild` to true here.
/cc @dseefeld @adaggarwal | 1.0 | Skip building "3.0.0"-branded targeting pack in 3.0.2+ source-build - Source-build will be building the `Microsoft.NETCore.App.Ref` `3.0.0` targeting pack in dotnet/source-build-reference-packages for 3.0.1+, however in 3.0.2+ I also merged some Core-Setup infrastructure that will have this repo build a "3.0.0"-branded targeting pack when building any 3.0.2+ commit: https://github.com/dotnet/arcade/pull/4318.
The source-build version is preferred because it's more static: we can trust it not to change due to accidental (or intentional but unwanted) changes to future versions of Core-Setup. There is danger of Core-Setup 3.0.2 building a "3.0.0"-branded targeting pack that isn't actually the same as the actual `v3.0.0` targeting pack, causing behavior differences in the source-built SDK. See https://github.com/dotnet/source-build/pull/1389
This infra needs to be disabled so the Core-Setup "3.0.0"-branded targeting pack doesn't conflict with the source-build 3.0.0 package.
This can be done without an Arcade update or source-build customization by setting `SkipSetLastReleasedVersionForSourceBuild` to true here.
/cc @dseefeld @adaggarwal | infrastructure | skip building branded targeting pack in source build source build will be building the microsoft netcore app ref targeting pack in dotnet source build reference packages for however in i also merged some core setup infrastructure that will have this repo build a branded targeting pack when building any commit the source build version is preferred because it s more static we can trust it not to change due to accidental or intentional but unwanted changes to future versions of core setup there is danger of core setup building a branded targeting pack that isn t actually the same as the actual targeting pack causing behavior differences in the source built sdk see this infra needs to be disabled so the core setup branded targeting pack doesn t conflict with the source build package this can be done without an arcade update or source build customization by setting skipsetlastreleasedversionforsourcebuild to true here cc dseefeld adaggarwal | 1 |
22,151 | 15,021,374,308 | IssuesEvent | 2021-02-01 15:44:36 | PostHog/posthog | https://api.github.com/repos/PostHog/posthog | closed | Joining Linode's One-Click marketplace | enhancement infrastructure | - [x] Write Linode StackScript (auto-deploy script) for PostHog. (https://github.com/PostHog/deployment/pull/11)
- [x] Request to become a Linode partner at https://www.linode.com/marketplace/app-partners/
- [ ] Prepare assets according to https://github.com/linode/Marketplace-Apps
- [ ] Get partnership approved by Linode.
- [ ] Submit assets to Linode for approval
- [ ] Write documentation for one-click Linode deployment
- [ ] Test it! | 1.0 | Joining Linode's One-Click marketplace - - [x] Write Linode StackScript (auto-deploy script) for PostHog. (https://github.com/PostHog/deployment/pull/11)
- [x] Request to become a Linode partner at https://www.linode.com/marketplace/app-partners/
- [ ] Prepare assets according to https://github.com/linode/Marketplace-Apps
- [ ] Get partnership approved by Linode.
- [ ] Submit assets to Linode for approval
- [ ] Write documentation for one-click Linode deployment
- [ ] Test it! | infrastructure | joining linode s one click marketplace write linode stackscript auto deploy script for posthog request to become a linode partner at prepare assets according to get partnership approved by linode submit assets to linode for approval write documentation for one click linode deployment test it | 1 |
33,561 | 27,583,461,635 | IssuesEvent | 2023-03-08 17:50:04 | nidak21/MLtoolsOverview | https://api.github.com/repos/nidak21/MLtoolsOverview | opened | MLtextTools: rename modules | infrastructure | 1. Rename:
- textTuningLib --> MLtuning
- tuningReportsLib --> MLtuningReports
- utilsLib --> MLtextUtils
- baseSampleDataLib --> MLbaseSample
1. Have to update all bin/scripts/templates to use new module names, update various comments, etc.
1. Rename automated test files, rename test directory to tests | 1.0 | MLtextTools: rename modules - 1. Rename:
- textTuningLib --> MLtuning
- tuningReportsLib --> MLtuningReports
- utilsLib --> MLtextUtils
- baseSampleDataLib --> MLbaseSample
1. Have to update all bin/scripts/templates to use new module names, update various comments, etc.
1. Rename automated test files, rename test directory to tests | infrastructure | mltexttools rename modules rename texttuninglib mltuning tuningreportslib mltuningreports utilslib mltextutils basesampledatalib mlbasesample have to update all bin scripts templates to use new module names update various comments etc rename automated test files rename test directory to tests | 1 |
131,329 | 27,876,512,903 | IssuesEvent | 2023-03-21 16:20:54 | Qiskit/qiskit.org | https://api.github.com/repos/Qiskit/qiskit.org | closed | [Code snippet] Create reusable component | feature code quality | ### Feature description
With some of the new designs coming up, we'll be reusing the Homepage code block styling (in the Providers page), along w/ the functionality to copy the code.
This presents a good opportunity to abstract the code into its own component, for consistency, reusability, and to handle the following requirements:
- code block for a single line
- code block for multiple lines
- component includes styling for code comments
- optionally includes a `Copy` button
### Additional context
_No response_ | 1.0 | [Code snippet] Create reusable component - ### Feature description
With some of the new designs coming up, we'll be reusing the Homepage code block styling (in the Providers page), along w/ the functionality to copy the code.
This presents a good opportunity to abstract the code into its own component, for consistency, reusability, and to handle the following requirements:
- code block for a single line
- code block for multiple lines
- component includes styling for code comments
- optionally includes a `Copy` button
### Additional context
_No response_ | non_infrastructure | create reusable component feature description with some of the new designs coming up we ll be reusing the homepage code block styling in the providers page along w the functionality to copy the code this presents a good opportunity to abstract the code into its own component for consistency reusability and to handle the following requirements code block for a single line code block for multiple lines component includes styling for code comments optionally includes a copy button additional context no response | 0 |
20,545 | 14,004,289,622 | IssuesEvent | 2020-10-28 16:54:51 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | [wasm-browser] unknown error: Chrome failed to start: exited abnormally. (chrome not reachable) | arch-wasm area-Infrastructure-mono blocking-clean-ci test bug | https://dev.azure.com/dnceng/public/_build/results?buildId=868621&view=logs&j=108d2c4a-8a62-5a58-8dad-8e1042acc93c&t=09a92b4c-3493-52a9-1172-0680b62c9e30
https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-43803-merge-48e510544eab4b8981/Microsoft.Extensions.DependencyModel.Tests/console.b31c959b.log?sv=2019-07-07&se=2020-11-17T12%3A40%3A20Z&sr=c&sp=rl&sig=hsagbbZGP1WmiA61t%2FZKutJ9xJt7ffIS82QRvSzuJbU%3D
```
Starting ChromeDriver 84.0.4147.0 (ff714bcb10c3356b1b1b2819148507a34fa4b1f8-refs/heads/master@{#768968}) on port 45939
Only local connections are allowed.
Please see https://chromedriver.chromium.org/security-considerations for suggestions on keeping ChromeDriver safe.
ChromeDriver was started successfully.
[12:41:07] crit: OpenQA.Selenium.WebDriverException: unknown error: Chrome failed to start: exited abnormally.
(chrome not reachable)
(The process started from chrome location /home/helixbot/work/9FCD08EB/p/chrome-linux/chrome is no longer running, so ChromeDriver is assuming that Chrome has crashed.)
at OpenQA.Selenium.Remote.RemoteWebDriver.UnpackAndThrowOnError(Response errorResponse)
at OpenQA.Selenium.Remote.RemoteWebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters)
at OpenQA.Selenium.Remote.RemoteWebDriver.StartSession(ICapabilities desiredCapabilities)
at OpenQA.Selenium.Remote.RemoteWebDriver..ctor(ICommandExecutor commandExecutor, ICapabilities desiredCapabilities)
at OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout)
at OpenQA.Selenium.Chrome.ChromeDriver..ctor(ChromeDriverService service, ChromeOptions options, TimeSpan commandTimeout)
at Microsoft.DotNet.XHarness.CLI.Commands.Wasm.WasmTestBrowserCommand.GetChromeDriver() in /_/src/Microsoft.DotNet.XHarness.CLI/Commands/WASM/Browser/WasmTestBrowserCommand.cs:line 90
at Microsoft.DotNet.XHarness.CLI.Commands.Wasm.WasmTestBrowserCommand.InvokeInternal(ILogger logger) in /_/src/Microsoft.DotNet.XHarness.CLI/Commands/WASM/Browser/WasmTestBrowserCommand.cs:line 53
at Microsoft.DotNet.XHarness.Common.CLI.Commands.XHarnessCommand.Invoke(IEnumerable`1 arguments) in /_/src/Microsoft.DotNet.XHarness.Common/CLI/Commands/XHarnessCommand.cs:line 120
XHarness exit code: 1001
XHarness artifacts: ./xharness-output
```
| 1.0 | [wasm-browser] unknown error: Chrome failed to start: exited abnormally. (chrome not reachable) - https://dev.azure.com/dnceng/public/_build/results?buildId=868621&view=logs&j=108d2c4a-8a62-5a58-8dad-8e1042acc93c&t=09a92b4c-3493-52a9-1172-0680b62c9e30
https://helixre8s23ayyeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-43803-merge-48e510544eab4b8981/Microsoft.Extensions.DependencyModel.Tests/console.b31c959b.log?sv=2019-07-07&se=2020-11-17T12%3A40%3A20Z&sr=c&sp=rl&sig=hsagbbZGP1WmiA61t%2FZKutJ9xJt7ffIS82QRvSzuJbU%3D
```
Starting ChromeDriver 84.0.4147.0 (ff714bcb10c3356b1b1b2819148507a34fa4b1f8-refs/heads/master@{#768968}) on port 45939
Only local connections are allowed.
Please see https://chromedriver.chromium.org/security-considerations for suggestions on keeping ChromeDriver safe.
ChromeDriver was started successfully.
[12:41:07] crit: OpenQA.Selenium.WebDriverException: unknown error: Chrome failed to start: exited abnormally.
(chrome not reachable)
(The process started from chrome location /home/helixbot/work/9FCD08EB/p/chrome-linux/chrome is no longer running, so ChromeDriver is assuming that Chrome has crashed.)
at OpenQA.Selenium.Remote.RemoteWebDriver.UnpackAndThrowOnError(Response errorResponse)
at OpenQA.Selenium.Remote.RemoteWebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters)
at OpenQA.Selenium.Remote.RemoteWebDriver.StartSession(ICapabilities desiredCapabilities)
at OpenQA.Selenium.Remote.RemoteWebDriver..ctor(ICommandExecutor commandExecutor, ICapabilities desiredCapabilities)
at OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout)
at OpenQA.Selenium.Chrome.ChromeDriver..ctor(ChromeDriverService service, ChromeOptions options, TimeSpan commandTimeout)
at Microsoft.DotNet.XHarness.CLI.Commands.Wasm.WasmTestBrowserCommand.GetChromeDriver() in /_/src/Microsoft.DotNet.XHarness.CLI/Commands/WASM/Browser/WasmTestBrowserCommand.cs:line 90
at Microsoft.DotNet.XHarness.CLI.Commands.Wasm.WasmTestBrowserCommand.InvokeInternal(ILogger logger) in /_/src/Microsoft.DotNet.XHarness.CLI/Commands/WASM/Browser/WasmTestBrowserCommand.cs:line 53
at Microsoft.DotNet.XHarness.Common.CLI.Commands.XHarnessCommand.Invoke(IEnumerable`1 arguments) in /_/src/Microsoft.DotNet.XHarness.Common/CLI/Commands/XHarnessCommand.cs:line 120
XHarness exit code: 1001
XHarness artifacts: ./xharness-output
```
| infrastructure | unknown error chrome failed to start exited abnormally chrome not reachable starting chromedriver refs heads master on port only local connections are allowed please see for suggestions on keeping chromedriver safe chromedriver was started successfully crit openqa selenium webdriverexception unknown error chrome failed to start exited abnormally chrome not reachable the process started from chrome location home helixbot work p chrome linux chrome is no longer running so chromedriver is assuming that chrome has crashed at openqa selenium remote remotewebdriver unpackandthrowonerror response errorresponse at openqa selenium remote remotewebdriver execute string drivercommandtoexecute dictionary parameters at openqa selenium remote remotewebdriver startsession icapabilities desiredcapabilities at openqa selenium remote remotewebdriver ctor icommandexecutor commandexecutor icapabilities desiredcapabilities at openqa selenium chromium chromiumdriver ctor chromiumdriverservice service chromiumoptions options timespan commandtimeout at openqa selenium chrome chromedriver ctor chromedriverservice service chromeoptions options timespan commandtimeout at microsoft dotnet xharness cli commands wasm wasmtestbrowsercommand getchromedriver in src microsoft dotnet xharness cli commands wasm browser wasmtestbrowsercommand cs line at microsoft dotnet xharness cli commands wasm wasmtestbrowsercommand invokeinternal ilogger logger in src microsoft dotnet xharness cli commands wasm browser wasmtestbrowsercommand cs line at microsoft dotnet xharness common cli commands xharnesscommand invoke ienumerable arguments in src microsoft dotnet xharness common cli commands xharnesscommand cs line xharness exit code xharness artifacts xharness output | 1 |
26,038 | 19,585,452,559 | IssuesEvent | 2022-01-05 05:59:05 | pixiebrix/playground | https://api.github.com/repos/pixiebrix/playground | closed | Rethink deployment strategy | infrastructure | We're only 6 projects in and the build is already taking 20 minutes 😰 https://vercel.com/pixiebrix/playground/G3rW9PbY9Y3ihbBrcGBgxAt2VjYJ
We should probably fix this before Vercel just stops building (and because 20 minutes is already a long time to find out whether something worked).
I don't know how to fix this yet, but I think it should **only build a project if it was changed.**
Possible solutions:
- https://turborepo.org/docs/getting-started (Vercel just bought this)
- Build on GitHub Actions (easier to skip parts of a build, needs some work to get PR preview builds)
- Ensure that Vercel’s build cache is working
Thoughts? We could also just use a regular host/S3/EC2, but preview builds and lack of server maintenance are a big plus for Vercel.
I sent a message to Vercel hoping for suggestions.
Platform limits:
- https://vercel.com/docs/concepts/deployments/build-step#technical-details | 1.0 | Rethink deployment strategy - We're only 6 projects in and the build is already taking 20 minutes 😰 https://vercel.com/pixiebrix/playground/G3rW9PbY9Y3ihbBrcGBgxAt2VjYJ
We should probably fix this before Vercel just stops building (and because 20 minutes is already a long time to find out whether something worked).
I don't know how to fix this yet, but I think it should **only build a project if it was changed.**
Possible solutions:
- https://turborepo.org/docs/getting-started (Vercel just bought this)
- Build on GitHub Actions (easier to skip parts of a build, needs some work to get PR preview builds)
- Ensure that Vercel’s build cache is working
Thoughts? We could also just use a regular host/S3/EC2, but preview builds and lack of server maintenance are a big plus for Vercel.
I sent a message to Vercel hoping for suggestions.
Platform limits:
- https://vercel.com/docs/concepts/deployments/build-step#technical-details | infrastructure | rethink deployment strategy we re only projects in and the build is already taking minutes 😰 we should probably fix this before vercel just stops building and because minutes is already a long time to find out whether something worked i don t know how to fix this yet but i think it should only build a project if it was changed possible solutions vercel just bought this build on github actions easier to skip parts of a build needs some work to get pr preview builds ensure that vercel’s build cache is working thoughts we could also just use a regular host but preview builds and lack of server maintenance are a big plus for vercel i sent a message to vercel hoping for suggestions platform limits | 1 |
27,632 | 22,050,068,221 | IssuesEvent | 2022-05-30 07:51:13 | DostEducation/RP_IVR_analytics | https://api.github.com/repos/DostEducation/RP_IVR_analytics | closed | [Infrastructure ERROR] RATE_LIMIT_EXCEEDED Cloud SQL API | infrastructure Infrastructure scale | **Describe the bug**
There are lots of webhook failures due to the following error at GCP (RATE_LIMIT_EXCEEDED).
```go
failed to refresh the ephemeral certificate for project-name:asia-south1:database-name: googleapi:
Error 429: Quota exceeded for quota metric 'Queries' and limit 'Queries per minute per user' of service 'sqladmin.googleapis.com' for consumer 'project_number:893******'.
Details: [{
"@type": "type.googleapis.com/google.rpc.ErrorInfo",
"domain": "googleapis.com",
"metadata": {
"consumer": "projects/893******",
"quota_limit": "defaultPerMinutePerUser",
"quota_metric": "sqladmin.googleapis.com/default",
"service": "sqladmin.googleapis.com"
},
"reason": "RATE_LIMIT_EXCEEDED"
}] , rateLimitExceeded
```
**Expected behavior**
The application should be able to handle load and should be able to process every webhook request
------
**Screenshots**
1. RP webhook failure for the last one Months (6th April to 6th May)
<img width="1186" alt="image" src="https://user-images.githubusercontent.com/39792899/167114818-e6e4b6bf-302e-47dd-a42c-3e0eb14b54fc.png">
2. KooKoo webhook failure for the last one Months (6th April to 6th May)
<img width="1167" alt="image" src="https://user-images.githubusercontent.com/39792899/167115690-81b0e520-0c89-4ebc-ad5a-3a0f3a8965d8.png">
------
**Additional context**
None
**Root cause**
1. The database size is growing rapidly, so the create/update queries are taking a long time and in some cases, connections are not getting free for new webhook calls. The cloud function is trying to create a new database connection in such a case and trying to create a new connection, which is hitting the limits.
**Impact**
The issue is increasing with time. For the last 1 month. Following are the infrastructure failure webhook stats
|Cloud function|Total webhook calls|Successful webhook calls|Failure webhook calls| Impact %|
|-------------|-------|-------|-----|------|
|RP webhook|574156| 572937|1219| 0.212|
|KooKoo webhook|1569953|1569668|285| 0.018|
**Query used**
```sql
SELECT COUNT(id)
FROM webhook_transaction_log
WHERE created_on::DATE > '2022-04-06';
SELECT COUNT(id)
FROM ivr_callback_transaction_log
WHERE created_on::DATE > '2022-04-06';
```
**Action items**
We need to plan the scaling as we are about to hit 30GB soon and we be delaying these queries further and hence will use more connections to handle the webhook load.
- [ ] #266
- [ ] Database scaling - Need to scale configuration for faster execution https://github.com/DostEducation/RP_IVR_analytics/issues/271
- [ ] We have the plan to implement queues for cloud function, already captured here - https://github.com/DostEducation/RP_IVR_callback/issues/31. Captured in https://github.com/DostEducation/RP_IVR_callback/issues/32
**Action items specific to Database issues**
Using `PC serverless connector`, we can solve the quota problem as the quota is limited to SQL APIs [stackoverflow reference](https://stackoverflow.com/a/69099505)
- [ ] #265
- [ ] #269
**Criticality**
- Medium: This is increasing rapidly and can cause more data loss in the future.
| 2.0 | [Infrastructure ERROR] RATE_LIMIT_EXCEEDED Cloud SQL API - **Describe the bug**
There are lots of webhook failures due to the following error at GCP (RATE_LIMIT_EXCEEDED).
```go
failed to refresh the ephemeral certificate for project-name:asia-south1:database-name: googleapi:
Error 429: Quota exceeded for quota metric 'Queries' and limit 'Queries per minute per user' of service 'sqladmin.googleapis.com' for consumer 'project_number:893******'.
Details: [{
"@type": "type.googleapis.com/google.rpc.ErrorInfo",
"domain": "googleapis.com",
"metadata": {
"consumer": "projects/893******",
"quota_limit": "defaultPerMinutePerUser",
"quota_metric": "sqladmin.googleapis.com/default",
"service": "sqladmin.googleapis.com"
},
"reason": "RATE_LIMIT_EXCEEDED"
}] , rateLimitExceeded
```
**Expected behavior**
The application should be able to handle load and should be able to process every webhook request
------
**Screenshots**
1. RP webhook failure for the last one Months (6th April to 6th May)
<img width="1186" alt="image" src="https://user-images.githubusercontent.com/39792899/167114818-e6e4b6bf-302e-47dd-a42c-3e0eb14b54fc.png">
2. KooKoo webhook failure for the last one Months (6th April to 6th May)
<img width="1167" alt="image" src="https://user-images.githubusercontent.com/39792899/167115690-81b0e520-0c89-4ebc-ad5a-3a0f3a8965d8.png">
------
**Additional context**
None
**Root cause**
1. The database size is growing rapidly, so the create/update queries are taking a long time and in some cases, connections are not getting free for new webhook calls. The cloud function is trying to create a new database connection in such a case and trying to create a new connection, which is hitting the limits.
**Impact**
The issue is increasing with time. For the last 1 month. Following are the infrastructure failure webhook stats
|Cloud function|Total webhook calls|Successful webhook calls|Failure webhook calls| Impact %|
|-------------|-------|-------|-----|------|
|RP webhook|574156| 572937|1219| 0.212|
|KooKoo webhook|1569953|1569668|285| 0.018|
**Query used**
```sql
SELECT COUNT(id)
FROM webhook_transaction_log
WHERE created_on::DATE > '2022-04-06';
SELECT COUNT(id)
FROM ivr_callback_transaction_log
WHERE created_on::DATE > '2022-04-06';
```
**Action items**
We need to plan the scaling as we are about to hit 30GB soon and we be delaying these queries further and hence will use more connections to handle the webhook load.
- [ ] #266
- [ ] Database scaling - Need to scale configuration for faster execution https://github.com/DostEducation/RP_IVR_analytics/issues/271
- [ ] We have the plan to implement queues for cloud function, already captured here - https://github.com/DostEducation/RP_IVR_callback/issues/31. Captured in https://github.com/DostEducation/RP_IVR_callback/issues/32
**Action items specific to Database issues**
Using `PC serverless connector`, we can solve the quota problem as the quota is limited to SQL APIs [stackoverflow reference](https://stackoverflow.com/a/69099505)
- [ ] #265
- [ ] #269
**Criticality**
- Medium: This is increasing rapidly and can cause more data loss in the future.
| infrastructure | rate limit exceeded cloud sql api describe the bug there are lots of webhook failures due to the following error at gcp rate limit exceeded go failed to refresh the ephemeral certificate for project name asia database name googleapi error quota exceeded for quota metric queries and limit queries per minute per user of service sqladmin googleapis com for consumer project number details type type googleapis com google rpc errorinfo domain googleapis com metadata consumer projects quota limit defaultperminuteperuser quota metric sqladmin googleapis com default service sqladmin googleapis com reason rate limit exceeded ratelimitexceeded expected behavior the application should be able to handle load and should be able to process every webhook request screenshots rp webhook failure for the last one months april to may img width alt image src kookoo webhook failure for the last one months april to may img width alt image src additional context none root cause the database size is growing rapidly so the create update queries are taking a long time and in some cases connections are not getting free for new webhook calls the cloud function is trying to create a new database connection in such a case and trying to create a new connection which is hitting the limits impact the issue is increasing with time for the last month following are the infrastructure failure webhook stats cloud function total webhook calls successful webhook calls failure webhook calls impact rp webhook kookoo webhook query used sql select count id from webhook transaction log where created on date select count id from ivr callback transaction log where created on date action items we need to plan the scaling as we are about to hit soon and we be delaying these queries further and hence will use more connections to handle the webhook load database scaling need to scale configuration for faster execution we have the plan to implement queues for cloud function already captured here captured in action items specific to database issues using pc serverless connector we can solve the quota problem as the quota is limited to sql apis criticality medium this is increasing rapidly and can cause more data loss in the future | 1 |
35,085 | 30,749,143,263 | IssuesEvent | 2023-07-28 17:31:35 | LSSTDESC/SprintWeek2023 | https://api.github.com/repos/LSSTDESC/SprintWeek2023 | opened | Get your pipeline+analysis tested, cared for, and ready for publication | Infrastructure Sprint | # Get your pipeline+analysis tested, cared for, and ready for publication
[One sentence description]
Contacts: Michael Wood-Vasey, Rob Knop
Day/Time:
Main communication channel:
GitHub repo:
Zoom room (if applicable):
## Goals and deliverable
We would like to help >2 people/teams get their analyses cleanly packaged, versioned, tested, and supported. We want to help make reproducibility easy, development testing quick, running efficient, versioning of data products simple, and describing all of the software and products for your paper simple.
Come find us and we will help.
## Resources and skills needed
Have an analysis you are doing. We are happy to help complete newbies through to experienced developers. Enthusiasm and a commitment to making your science clearly documented, reproducibility, and usable by other people (including future you).
## Detailed description
Come find us at the beginning of the week (or earlier) and we will sit down and help talk through the design of the pipeline+analysis, the requirements, the way you run things, the way you test things, and we will work together on a plan and implementation to help make things better for you, better for future you, and better for DESC.
[add additional details about the project]
| 1.0 | Get your pipeline+analysis tested, cared for, and ready for publication - # Get your pipeline+analysis tested, cared for, and ready for publication
[One sentence description]
Contacts: Michael Wood-Vasey, Rob Knop
Day/Time:
Main communication channel:
GitHub repo:
Zoom room (if applicable):
## Goals and deliverable
We would like to help >2 people/teams get their analyses cleanly packaged, versioned, tested, and supported. We want to help make reproducibility easy, development testing quick, running efficient, versioning of data products simple, and describing all of the software and products for your paper simple.
Come find us and we will help.
## Resources and skills needed
Have an analysis you are doing. We are happy to help complete newbies through to experienced developers. Enthusiasm and a commitment to making your science clearly documented, reproducibility, and usable by other people (including future you).
## Detailed description
Come find us at the beginning of the week (or earlier) and we will sit down and help talk through the design of the pipeline+analysis, the requirements, the way you run things, the way you test things, and we will work together on a plan and implementation to help make things better for you, better for future you, and better for DESC.
[add additional details about the project]
| infrastructure | get your pipeline analysis tested cared for and ready for publication get your pipeline analysis tested cared for and ready for publication contacts michael wood vasey rob knop day time main communication channel github repo zoom room if applicable goals and deliverable we would like to help people teams get their analyses cleanly packaged versioned tested and supported we want to help make reproducibility easy development testing quick running efficient versioning of data products simple and describing all of the software and products for your paper simple come find us and we will help resources and skills needed have an analysis you are doing we are happy to help complete newbies through to experienced developers enthusiasm and a commitment to making your science clearly documented reproducibility and usable by other people including future you detailed description come find us at the beginning of the week or earlier and we will sit down and help talk through the design of the pipeline analysis the requirements the way you run things the way you test things and we will work together on a plan and implementation to help make things better for you better for future you and better for desc | 1 |
394,446 | 27,029,421,266 | IssuesEvent | 2023-02-12 01:41:04 | xKabbe/denigma | https://api.github.com/repos/xKabbe/denigma | closed | :dna: [FEATURE] - Add Functionality To The `ViewSettings` Component :dna: | documentation enhancement frontend JavaScript / TypeScript test | # Description
The `ViewSettings` component is missing it's functionality.
# Expected Actions
- [x] Add functionality to `Zoom`
- [x] Add functionality to `Size`
# Definition of Done
- [x] `Code` implemented
- [x] `Tests` implemented and passing
- [x] `Documentation` / `Stories` implemented
- [x] `GitHub Actions` are passing
| 1.0 | :dna: [FEATURE] - Add Functionality To The `ViewSettings` Component :dna: - # Description
The `ViewSettings` component is missing it's functionality.
# Expected Actions
- [x] Add functionality to `Zoom`
- [x] Add functionality to `Size`
# Definition of Done
- [x] `Code` implemented
- [x] `Tests` implemented and passing
- [x] `Documentation` / `Stories` implemented
- [x] `GitHub Actions` are passing
| non_infrastructure | dna add functionality to the viewsettings component dna description the viewsettings component is missing it s functionality expected actions add functionality to zoom add functionality to size definition of done code implemented tests implemented and passing documentation stories implemented github actions are passing | 0 |
131,177 | 10,683,899,178 | IssuesEvent | 2019-10-22 09:20:05 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | The mysql_db module needs to use the --one-database flag for imports | affects_2.6 bug database has_pr module mysql support:community test | - Bug Report
##### COMPONENT NAME
mysql_db
##### ANSIBLE VERSION
This issue was discovered with 2.4.
```
ansible 2.6.0 (devel 45b5a486c6) last updated 2018/04/04 15:11:32 (GMT -400)
config file = /cygdrive/c/git/projects/project-bitbucket/ansible.cfg
configured module search path = [u'/opt/ansible/library']
ansible python module location = /opt/ansible/lib/ansible
executable location = /usr/bin/ansible
python version = 2.7.14 (default, Oct 31 2017, 21:12:13) [GCC 6.4.0]
```
##### CONFIGURATION
```
ANSIBLE_PIPELINING(env: ANSIBLE_PIPELINING) = True
ANSIBLE_SSH_ARGS(env: ANSIBLE_SSH_ARGS) = -o ControlMaster=no -o ForwardAgent=yes -o IdentitiesOnly=yes
DEFAULT_LOG_PATH(env: ANSIBLE_LOG_PATH) = /cygdrive/c/git/projects/project-test/logs/2018-04-04_14-31-22_ansible.log
DEFAULT_MODULE_PATH(env: ANSIBLE_LIBRARY) = [u'/opt/ansible/library']
DEFAULT_ROLES_PATH(env: ANSIBLE_ROLES_PATH) = [u'/cygdrive/c/git/roles']
```
##### OS / ENVIRONMENT
Most likely not relevant, but full disclosure:
Ansible source:
Window 10 with Cygwin
The Linux target, a REHL AMI on AWS:
```
$ uname -a
Linux test 3.10.0-693.21.1.el7.x86_64 #1 SMP Fri Feb 23 18:54:16 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
```
Using the MySQL variant MariaDB
```
$ mysql --version
mysql Ver 15.1 Distrib 5.5.56-MariaDB, for Linux (x86_64) using readline 5.1
```
##### SUMMARY
Our MySQL backups dump all the databases on the server into a single backup file
```
mysqldump --single-transaction --flush-logs --master-data=2 -u $USER -p$PASSWORD --max_allowed_packet=256M --all-databases | bzip2 -c > $BACKUP_PATH/mysql_backup_${DATE}.bz2
```
This results in a backup file that contains no CREATE DATABASE statements, but there is a USE <database> for each of the databases on the MySQL instance along with all of the schema and data needed to recover the database. The Ansible module mysql_db allows you to specify the database for imports, but instead of only recovering the database specified, it recovers the entire MySQL backup, over writing any other matching database on the server, including:
-
##### STEPS TO REPRODUCE
Dump your MySQL server using the --all-databases flag
```
mysqldump --single-transaction --flush-logs --master-data=2 -u $USER -p$PASSWORD --max_allowed_packet=256M --all-databases | bzip2 -c > $BACKUP_PATH/mysql_backup_${DATE}.bz2
```
- Create a new MySQL server, and add a couple users that do not exist and use a different root password.
- Verify the password and that the users exist, for example:
```
MariaDB [stash]> SELECT User FROM mysql.user;
+-----------+
| User |
+-----------+
| root |
| root |
| root |
| test |
+-----------+
4 rows in set (0.00 sec)
```
- Import the database dump into that new MySQL database, making sure to specify the database
```yaml
- name: "[restore] Restore the DB"
mysql_db:
login_user: "{{ mysql_backup_user }}"
login_password: "{{ mysql_backup_user_password }}"
target: "{{ mysql_backup_archive_name }}"
state: import
name: "{{ restore_db_name }}"
```
##### EXPECTED RESULTS
The single database specified would be restored and the rest of the databases on the server would be untouched.
##### ACTUAL RESULTS
All of the databases in the backup have been restored. The MySQL users have been overwritten and including the root password.
```
MariaDB [stash]> SELECT User FROM mysql.user;
+-----------+
| User |
+-----------+
| root |
| root |
| root |
| produser1 |
| produser2 |
| produser3 |
+-----------+
6 rows in set (0.00 sec)
```
##### THE REASON
The mysql_db module is using the mysql ```-D``` flag (```--database```), which only specifies the database to target:
https://dev.mysql.com/doc/refman/5.7/en/mysql-command-options.html#option_mysql_database
>
> --database=db_name, -D db_name
>
> The database to use. This is useful primarily in an option file.
Instead of the ```--one-database``` option:
https://dev.mysql.com/doc/refman/5.7/en/mysql-command-options.html#option_mysql_one-database
> --one-database, -o
>
> Ignore statements except those that occur while the default database is the one named on the command line. This option is rudimentary and should be used with care. Statement filtering is based only on USE statements.
>
> Initially, mysql executes statements in the input because specifying a database db_name on the command line is equivalent to inserting USE db_name at the beginning of the input. Then, for each USE statement encountered, mysql accepts or rejects following statements depending on whether the database named is the one on the command line. The content of the statements is immaterial.
>
So if you modify the lib/ansible/databases/mysql_db.py file from:
```
def db_import(module, host, user, password, db_name, target, all_databases, port, config_file, socket=None, ssl_cert=None, ssl_key=None, ssl_ca=None):
if not os.path.exists(target):
return module.fail_json(msg="target %s does not exist on the host" % target)
cmd = [module.get_bin_path('mysql', True)]
# --defaults-file must go first, or errors out
...
if not all_databases:
cmd.append("-D")
cmd.append(pipes.quote(db_name))
```
To:
```
def db_import(module, host, user, password, db_name, target, all_databases, port, config_file, socket=None, ssl_cert=None, ssl_key=None, ssl_ca=None):
if not os.path.exists(target):
return module.fail_json(msg="target %s does not exist on the host" % target)
cmd = [module.get_bin_path('mysql', True)]
# --defaults-file must go first, or errors out
...
if not all_databases:
cmd.append("--one-database %s" % pipes.quote(db_name))
```
This causes the expected behavior where only the specified database is created and modified. The ```--one-database``` db_name format implies the ```--database``` option as well, so if you use a SQL file with no ```USE db_name;``` lines, it will still be run against the named database. | 1.0 | The mysql_db module needs to use the --one-database flag for imports - - Bug Report
##### COMPONENT NAME
mysql_db
##### ANSIBLE VERSION
This issue was discovered with 2.4.
```
ansible 2.6.0 (devel 45b5a486c6) last updated 2018/04/04 15:11:32 (GMT -400)
config file = /cygdrive/c/git/projects/project-bitbucket/ansible.cfg
configured module search path = [u'/opt/ansible/library']
ansible python module location = /opt/ansible/lib/ansible
executable location = /usr/bin/ansible
python version = 2.7.14 (default, Oct 31 2017, 21:12:13) [GCC 6.4.0]
```
##### CONFIGURATION
```
ANSIBLE_PIPELINING(env: ANSIBLE_PIPELINING) = True
ANSIBLE_SSH_ARGS(env: ANSIBLE_SSH_ARGS) = -o ControlMaster=no -o ForwardAgent=yes -o IdentitiesOnly=yes
DEFAULT_LOG_PATH(env: ANSIBLE_LOG_PATH) = /cygdrive/c/git/projects/project-test/logs/2018-04-04_14-31-22_ansible.log
DEFAULT_MODULE_PATH(env: ANSIBLE_LIBRARY) = [u'/opt/ansible/library']
DEFAULT_ROLES_PATH(env: ANSIBLE_ROLES_PATH) = [u'/cygdrive/c/git/roles']
```
##### OS / ENVIRONMENT
Most likely not relevant, but full disclosure:
Ansible source:
Window 10 with Cygwin
The Linux target, a REHL AMI on AWS:
```
$ uname -a
Linux test 3.10.0-693.21.1.el7.x86_64 #1 SMP Fri Feb 23 18:54:16 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
```
Using the MySQL variant MariaDB
```
$ mysql --version
mysql Ver 15.1 Distrib 5.5.56-MariaDB, for Linux (x86_64) using readline 5.1
```
##### SUMMARY
Our MySQL backups dump all the databases on the server into a single backup file
```
mysqldump --single-transaction --flush-logs --master-data=2 -u $USER -p$PASSWORD --max_allowed_packet=256M --all-databases | bzip2 -c > $BACKUP_PATH/mysql_backup_${DATE}.bz2
```
This results in a backup file that contains no CREATE DATABASE statements, but there is a USE <database> for each of the databases on the MySQL instance along with all of the schema and data needed to recover the database. The Ansible module mysql_db allows you to specify the database for imports, but instead of only recovering the database specified, it recovers the entire MySQL backup, over writing any other matching database on the server, including:
-
##### STEPS TO REPRODUCE
Dump your MySQL server using the --all-databases flag
```
mysqldump --single-transaction --flush-logs --master-data=2 -u $USER -p$PASSWORD --max_allowed_packet=256M --all-databases | bzip2 -c > $BACKUP_PATH/mysql_backup_${DATE}.bz2
```
- Create a new MySQL server, and add a couple users that do not exist and use a different root password.
- Verify the password and that the users exist, for example:
```
MariaDB [stash]> SELECT User FROM mysql.user;
+-----------+
| User |
+-----------+
| root |
| root |
| root |
| test |
+-----------+
4 rows in set (0.00 sec)
```
- Import the database dump into that new MySQL database, making sure to specify the database
```yaml
- name: "[restore] Restore the DB"
mysql_db:
login_user: "{{ mysql_backup_user }}"
login_password: "{{ mysql_backup_user_password }}"
target: "{{ mysql_backup_archive_name }}"
state: import
name: "{{ restore_db_name }}"
```
##### EXPECTED RESULTS
The single database specified would be restored and the rest of the databases on the server would be untouched.
##### ACTUAL RESULTS
All of the databases in the backup have been restored. The MySQL users have been overwritten and including the root password.
```
MariaDB [stash]> SELECT User FROM mysql.user;
+-----------+
| User |
+-----------+
| root |
| root |
| root |
| produser1 |
| produser2 |
| produser3 |
+-----------+
6 rows in set (0.00 sec)
```
##### THE REASON
The mysql_db module is using the mysql ```-D``` flag (```--database```), which only specifies the database to target:
https://dev.mysql.com/doc/refman/5.7/en/mysql-command-options.html#option_mysql_database
>
> --database=db_name, -D db_name
>
> The database to use. This is useful primarily in an option file.
Instead of the ```--one-database``` option:
https://dev.mysql.com/doc/refman/5.7/en/mysql-command-options.html#option_mysql_one-database
> --one-database, -o
>
> Ignore statements except those that occur while the default database is the one named on the command line. This option is rudimentary and should be used with care. Statement filtering is based only on USE statements.
>
> Initially, mysql executes statements in the input because specifying a database db_name on the command line is equivalent to inserting USE db_name at the beginning of the input. Then, for each USE statement encountered, mysql accepts or rejects following statements depending on whether the database named is the one on the command line. The content of the statements is immaterial.
>
So if you modify the lib/ansible/databases/mysql_db.py file from:
```
def db_import(module, host, user, password, db_name, target, all_databases, port, config_file, socket=None, ssl_cert=None, ssl_key=None, ssl_ca=None):
if not os.path.exists(target):
return module.fail_json(msg="target %s does not exist on the host" % target)
cmd = [module.get_bin_path('mysql', True)]
# --defaults-file must go first, or errors out
...
if not all_databases:
cmd.append("-D")
cmd.append(pipes.quote(db_name))
```
To:
```
def db_import(module, host, user, password, db_name, target, all_databases, port, config_file, socket=None, ssl_cert=None, ssl_key=None, ssl_ca=None):
if not os.path.exists(target):
return module.fail_json(msg="target %s does not exist on the host" % target)
cmd = [module.get_bin_path('mysql', True)]
# --defaults-file must go first, or errors out
...
if not all_databases:
cmd.append("--one-database %s" % pipes.quote(db_name))
```
This causes the expected behavior where only the specified database is created and modified. The ```--one-database``` db_name format implies the ```--database``` option as well, so if you use a SQL file with no ```USE db_name;``` lines, it will still be run against the named database. | non_infrastructure | the mysql db module needs to use the one database flag for imports bug report component name mysql db ansible version this issue was discovered with ansible devel last updated gmt config file cygdrive c git projects project bitbucket ansible cfg configured module search path ansible python module location opt ansible lib ansible executable location usr bin ansible python version default oct configuration ansible pipelining env ansible pipelining true ansible ssh args env ansible ssh args o controlmaster no o forwardagent yes o identitiesonly yes default log path env ansible log path cygdrive c git projects project test logs ansible log default module path env ansible library default roles path env ansible roles path os environment most likely not relevant but full disclosure ansible source window with cygwin the linux target a rehl ami on aws uname a linux test smp fri feb utc gnu linux using the mysql variant mariadb mysql version mysql ver distrib mariadb for linux using readline summary our mysql backups dump all the databases on the server into a single backup file mysqldump single transaction flush logs master data u user p password max allowed packet all databases c backup path mysql backup date this results in a backup file that contains no create database statements but there is a use for each of the databases on the mysql instance along with all of the schema and data needed to recover the database the ansible module mysql db allows you to specify the database for imports but instead of only recovering the database specified it recovers the entire mysql backup over writing any other matching database on the server including steps to reproduce dump your mysql server using the all databases flag mysqldump single transaction flush logs master data u user p password max allowed packet all databases c backup path mysql backup date create a new mysql server and add a couple users that do not exist and use a different root password verify the password and that the users exist for example mariadb select user from mysql user user root root root test rows in set sec import the database dump into that new mysql database making sure to specify the database yaml name restore the db mysql db login user mysql backup user login password mysql backup user password target mysql backup archive name state import name restore db name expected results the single database specified would be restored and the rest of the databases on the server would be untouched actual results all of the databases in the backup have been restored the mysql users have been overwritten and including the root password mariadb select user from mysql user user root root root rows in set sec the reason the mysql db module is using the mysql d flag database which only specifies the database to target database db name d db name the database to use this is useful primarily in an option file instead of the one database option one database o ignore statements except those that occur while the default database is the one named on the command line this option is rudimentary and should be used with care statement filtering is based only on use statements initially mysql executes statements in the input because specifying a database db name on the command line is equivalent to inserting use db name at the beginning of the input then for each use statement encountered mysql accepts or rejects following statements depending on whether the database named is the one on the command line the content of the statements is immaterial so if you modify the lib ansible databases mysql db py file from def db import module host user password db name target all databases port config file socket none ssl cert none ssl key none ssl ca none if not os path exists target return module fail json msg target s does not exist on the host target cmd defaults file must go first or errors out if not all databases cmd append d cmd append pipes quote db name to def db import module host user password db name target all databases port config file socket none ssl cert none ssl key none ssl ca none if not os path exists target return module fail json msg target s does not exist on the host target cmd defaults file must go first or errors out if not all databases cmd append one database s pipes quote db name this causes the expected behavior where only the specified database is created and modified the one database db name format implies the database option as well so if you use a sql file with no use db name lines it will still be run against the named database | 0 |
21,077 | 14,354,607,079 | IssuesEvent | 2020-11-30 08:55:27 | JetBrains/azure-tools-for-intellij | https://api.github.com/repos/JetBrains/azure-tools-for-intellij | closed | Add Dockerfile with build container definition into the plugin repo | infrastructure | Currently, we use a container from `https://hub.docker.com/r/ivanpashchenko/rider-plugins-docker-image/dockerfile` to build the plugin. We are going to use a backend in plugin in nearest time (#234) that means that we will need to extend this container to setup `dotnet`, etc. We should rather create a Dockerfile with container definition and store it inside a plugin repo.
Please see all related steps in https://youtrack.jetbrains.com/issue/RIDER-45471 | 1.0 | Add Dockerfile with build container definition into the plugin repo - Currently, we use a container from `https://hub.docker.com/r/ivanpashchenko/rider-plugins-docker-image/dockerfile` to build the plugin. We are going to use a backend in plugin in nearest time (#234) that means that we will need to extend this container to setup `dotnet`, etc. We should rather create a Dockerfile with container definition and store it inside a plugin repo.
Please see all related steps in https://youtrack.jetbrains.com/issue/RIDER-45471 | infrastructure | add dockerfile with build container definition into the plugin repo currently we use a container from to build the plugin we are going to use a backend in plugin in nearest time that means that we will need to extend this container to setup dotnet etc we should rather create a dockerfile with container definition and store it inside a plugin repo please see all related steps in | 1 |
117,003 | 9,905,617,591 | IssuesEvent | 2019-06-27 12:03:54 | mesg-foundation/engine | https://api.github.com/repos/mesg-foundation/engine | closed | refactor all integration tests | integration tests low priority question | Currently we have integration tests like [this](https://github.com/mesg-foundation/core/blob/2fe4faf5b73c1c7cd64cb2e5239ec139a7ce6449/container/network_integration_test.go#L35) where error values just compared with nil to make a Test pass. In this case, instead of looking for err to be nil, we should be querying network list from docker and see if our method did really deleted that network.
Currently, I do not have a full research about the best approaches in integration tests for our case but we can definitely and should improve the current one. | 1.0 | refactor all integration tests - Currently we have integration tests like [this](https://github.com/mesg-foundation/core/blob/2fe4faf5b73c1c7cd64cb2e5239ec139a7ce6449/container/network_integration_test.go#L35) where error values just compared with nil to make a Test pass. In this case, instead of looking for err to be nil, we should be querying network list from docker and see if our method did really deleted that network.
Currently, I do not have a full research about the best approaches in integration tests for our case but we can definitely and should improve the current one. | non_infrastructure | refactor all integration tests currently we have integration tests like where error values just compared with nil to make a test pass in this case instead of looking for err to be nil we should be querying network list from docker and see if our method did really deleted that network currently i do not have a full research about the best approaches in integration tests for our case but we can definitely and should improve the current one | 0 |
115,336 | 24,750,334,011 | IssuesEvent | 2022-10-21 13:15:41 | UnitTestBot/UTBotJava | https://api.github.com/repos/UnitTestBot/UTBotJava | opened | Enum class is addressed incorrectly | bug codegen | **Description**
When Enum class is created by className, it addresses not class but the Enum constant.
**To Reproduce**
1. Run the 'UTBotJava' project in IntelliJ Idea
2. Install [plugin build](https://github.com/UnitTestBot/UTBotJava/actions/runs/3296912146) or one of the latest builds from main
3. Generate tests for utbot-sample/src/main/java/org/utbot/examples/enums
with No mocks, No parameterization.
4. Run the generated tests
**Expected behavior**
Assertions in generated tests are supposed to pass.
**Actual behavior**
There are assertion errors in tests - for every class in the package.
Please see the Class creation: `State$3` is addressing not the State enum, but the third constant in it.
**Visual proofs (screenshots, logs, images)**
~~~java
///region Test suites for executable org.utbot.examples.enums.ComplexEnumExamples.findState
///region FUZZER: SUCCESSFUL EXECUTIONS for method findState(int)
@Test
@DisplayName("findState: code > 0")
public void testFindState() throws ClassNotFoundException, IllegalAccessException {
ComplexEnumExamples complexEnumExamples = new ComplexEnumExamples();
State actual = complexEnumExamples.findState(1);
Class anonymousStateClazz = Class.forName("org.utbot.examples.enums.State$3");
Object expected = getEnumConstantByName(anonymousStateClazz, "UNKNOWN");
assertEquals(expected, actual);
}
~~~
same for Symbolic execution:
```java
///region Test suites for executable org.utbot.examples.enums.State.findStateByCode
///region SYMBOLIC EXECUTION: SUCCESSFUL EXECUTIONS for method findStateByCode(int)
/**
* @utbot.classUnderTest {@link State}
* @utbot.methodUnderTest {@link State#findStateByCode(int)}
*/
@Test
@DisplayName("findStateByCode: ")
public void testFindStateByCode() throws ClassNotFoundException, IllegalAccessException {
State actual = State.findStateByCode(255);
Class anonymousStateClazz = Class.forName("org.utbot.examples.enums.State$1");
Object expected = getEnumConstantByName(anonymousStateClazz, "OPEN");
assertEquals(expected, actual);
}
```
**Additional context**
Same issue with Fuzzing and / or Symbolic execution.
These tests are also failing when running in debug IDEA with runIde.
| 1.0 | Enum class is addressed incorrectly - **Description**
When Enum class is created by className, it addresses not class but the Enum constant.
**To Reproduce**
1. Run the 'UTBotJava' project in IntelliJ Idea
2. Install [plugin build](https://github.com/UnitTestBot/UTBotJava/actions/runs/3296912146) or one of the latest builds from main
3. Generate tests for utbot-sample/src/main/java/org/utbot/examples/enums
with No mocks, No parameterization.
4. Run the generated tests
**Expected behavior**
Assertions in generated tests are supposed to pass.
**Actual behavior**
There are assertion errors in tests - for every class in the package.
Please see the Class creation: `State$3` is addressing not the State enum, but the third constant in it.
**Visual proofs (screenshots, logs, images)**
~~~java
///region Test suites for executable org.utbot.examples.enums.ComplexEnumExamples.findState
///region FUZZER: SUCCESSFUL EXECUTIONS for method findState(int)
@Test
@DisplayName("findState: code > 0")
public void testFindState() throws ClassNotFoundException, IllegalAccessException {
ComplexEnumExamples complexEnumExamples = new ComplexEnumExamples();
State actual = complexEnumExamples.findState(1);
Class anonymousStateClazz = Class.forName("org.utbot.examples.enums.State$3");
Object expected = getEnumConstantByName(anonymousStateClazz, "UNKNOWN");
assertEquals(expected, actual);
}
~~~
same for Symbolic execution:
```java
///region Test suites for executable org.utbot.examples.enums.State.findStateByCode
///region SYMBOLIC EXECUTION: SUCCESSFUL EXECUTIONS for method findStateByCode(int)
/**
* @utbot.classUnderTest {@link State}
* @utbot.methodUnderTest {@link State#findStateByCode(int)}
*/
@Test
@DisplayName("findStateByCode: ")
public void testFindStateByCode() throws ClassNotFoundException, IllegalAccessException {
State actual = State.findStateByCode(255);
Class anonymousStateClazz = Class.forName("org.utbot.examples.enums.State$1");
Object expected = getEnumConstantByName(anonymousStateClazz, "OPEN");
assertEquals(expected, actual);
}
```
**Additional context**
Same issue with Fuzzing and / or Symbolic execution.
These tests are also failing when running in debug IDEA with runIde.
| non_infrastructure | enum class is addressed incorrectly description when enum class is created by classname it addresses not class but the enum constant to reproduce run the utbotjava project in intellij idea install or one of the latest builds from main generate tests for utbot sample src main java org utbot examples enums with no mocks no parameterization run the generated tests expected behavior assertions in generated tests are supposed to pass actual behavior there are assertion errors in tests for every class in the package please see the class creation state is addressing not the state enum but the third constant in it visual proofs screenshots logs images java region test suites for executable org utbot examples enums complexenumexamples findstate region fuzzer successful executions for method findstate int test displayname findstate code public void testfindstate throws classnotfoundexception illegalaccessexception complexenumexamples complexenumexamples new complexenumexamples state actual complexenumexamples findstate class anonymousstateclazz class forname org utbot examples enums state object expected getenumconstantbyname anonymousstateclazz unknown assertequals expected actual same for symbolic execution java region test suites for executable org utbot examples enums state findstatebycode region symbolic execution successful executions for method findstatebycode int utbot classundertest link state utbot methodundertest link state findstatebycode int test displayname findstatebycode public void testfindstatebycode throws classnotfoundexception illegalaccessexception state actual state findstatebycode class anonymousstateclazz class forname org utbot examples enums state object expected getenumconstantbyname anonymousstateclazz open assertequals expected actual additional context same issue with fuzzing and or symbolic execution these tests are also failing when running in debug idea with runide | 0 |
292,189 | 21,955,861,883 | IssuesEvent | 2022-05-24 12:03:46 | tabling-cloning/tabling-cloning-2022 | https://api.github.com/repos/tabling-cloning/tabling-cloning-2022 | closed | UX Flow Chart 공유하기 | documentation | 인당 4개의 UX 플로우를 제작하여 빠진 기능들을 추가하려고 합니다.
프로그램 기능들에 대한 부분은 해당 링크에서 참고하시면 될것 같습니다
=> https://github.com/tabling-cloning/tabling-cloning-2022/wiki/%5B%233%5D-2022-05-03
UX 플로우에 대한 부분을 자유롭게 적어주시고 추가될 기능들도 함께 토론해 주세요. | 1.0 | UX Flow Chart 공유하기 - 인당 4개의 UX 플로우를 제작하여 빠진 기능들을 추가하려고 합니다.
프로그램 기능들에 대한 부분은 해당 링크에서 참고하시면 될것 같습니다
=> https://github.com/tabling-cloning/tabling-cloning-2022/wiki/%5B%233%5D-2022-05-03
UX 플로우에 대한 부분을 자유롭게 적어주시고 추가될 기능들도 함께 토론해 주세요. | non_infrastructure | ux flow chart 공유하기 인당 ux 플로우를 제작하여 빠진 기능들을 추가하려고 합니다 프로그램 기능들에 대한 부분은 해당 링크에서 참고하시면 될것 같습니다 ux 플로우에 대한 부분을 자유롭게 적어주시고 추가될 기능들도 함께 토론해 주세요 | 0 |
34,661 | 30,257,308,527 | IssuesEvent | 2023-07-07 04:38:38 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | The event mechanism in RemoveBiomass in Plant.cs needs refactoring | interface/infrastructure refactor | /// <summary>Harvest the crop.</summary>
public void RemoveBiomass(string biomassRemoveType, RemovalFractions removalData = null)
{
Summary.WriteMessage(this, string.Format("Biomass removed from crop " + Name + " by " + biomassRemoveType.TrimEnd('e') + "ing"));
// Invoke specific defoliation events.
if (biomassRemoveType == "Harvest" && Harvesting != null)
Harvesting.Invoke(this, new EventArgs());
if (biomassRemoveType == "Prune" && Pruning != null)
Pruning.Invoke(this, new EventArgs());
if (biomassRemoveType == "LeafPluck" && LeafPlucking != null)
LeafPlucking.Invoke(this, new EventArgs());
if (biomassRemoveType == "Cut" && Cutting != null)
Cutting.Invoke(this, new EventArgs());
if (biomassRemoveType == "Graze" && Grazing != null)
Grazing.Invoke(this, new EventArgs());
// Set up the default BiomassRemovalData values
foreach (IOrgan organ in Organs)
{
// Get the default removal fractions
OrganBiomassRemovalType biomassRemoval = null;
if (removalData != null)
biomassRemoval = removalData.GetFractionsForOrgan(organ.Name);
organ.RemoveBiomass(biomassRemoveType, biomassRemoval);
}
// Reset the phenology if SetPhenologyStage specified.
if (removalData != null && removalData.SetPhenologyStage != 0)
Phenology.SetToStage(removalData.SetPhenologyStage);
// Reduce plant and stem population if thinning proportion specified
if (removalData != null && removalData.SetThinningProportion != 0 && Structure != null)
Structure.doThin(removalData.SetThinningProportion);
// Remove nodes from the main-stem
if (removalData != null && removalData.NodesToRemove > 0)
Structure.doNodeRemoval(removalData.NodesToRemove);
}
| 1.0 | The event mechanism in RemoveBiomass in Plant.cs needs refactoring - /// <summary>Harvest the crop.</summary>
public void RemoveBiomass(string biomassRemoveType, RemovalFractions removalData = null)
{
Summary.WriteMessage(this, string.Format("Biomass removed from crop " + Name + " by " + biomassRemoveType.TrimEnd('e') + "ing"));
// Invoke specific defoliation events.
if (biomassRemoveType == "Harvest" && Harvesting != null)
Harvesting.Invoke(this, new EventArgs());
if (biomassRemoveType == "Prune" && Pruning != null)
Pruning.Invoke(this, new EventArgs());
if (biomassRemoveType == "LeafPluck" && LeafPlucking != null)
LeafPlucking.Invoke(this, new EventArgs());
if (biomassRemoveType == "Cut" && Cutting != null)
Cutting.Invoke(this, new EventArgs());
if (biomassRemoveType == "Graze" && Grazing != null)
Grazing.Invoke(this, new EventArgs());
// Set up the default BiomassRemovalData values
foreach (IOrgan organ in Organs)
{
// Get the default removal fractions
OrganBiomassRemovalType biomassRemoval = null;
if (removalData != null)
biomassRemoval = removalData.GetFractionsForOrgan(organ.Name);
organ.RemoveBiomass(biomassRemoveType, biomassRemoval);
}
// Reset the phenology if SetPhenologyStage specified.
if (removalData != null && removalData.SetPhenologyStage != 0)
Phenology.SetToStage(removalData.SetPhenologyStage);
// Reduce plant and stem population if thinning proportion specified
if (removalData != null && removalData.SetThinningProportion != 0 && Structure != null)
Structure.doThin(removalData.SetThinningProportion);
// Remove nodes from the main-stem
if (removalData != null && removalData.NodesToRemove > 0)
Structure.doNodeRemoval(removalData.NodesToRemove);
}
| infrastructure | the event mechanism in removebiomass in plant cs needs refactoring harvest the crop public void removebiomass string biomassremovetype removalfractions removaldata null summary writemessage this string format biomass removed from crop name by biomassremovetype trimend e ing invoke specific defoliation events if biomassremovetype harvest harvesting null harvesting invoke this new eventargs if biomassremovetype prune pruning null pruning invoke this new eventargs if biomassremovetype leafpluck leafplucking null leafplucking invoke this new eventargs if biomassremovetype cut cutting null cutting invoke this new eventargs if biomassremovetype graze grazing null grazing invoke this new eventargs set up the default biomassremovaldata values foreach iorgan organ in organs get the default removal fractions organbiomassremovaltype biomassremoval null if removaldata null biomassremoval removaldata getfractionsfororgan organ name organ removebiomass biomassremovetype biomassremoval reset the phenology if setphenologystage specified if removaldata null removaldata setphenologystage phenology settostage removaldata setphenologystage reduce plant and stem population if thinning proportion specified if removaldata null removaldata setthinningproportion structure null structure dothin removaldata setthinningproportion remove nodes from the main stem if removaldata null removaldata nodestoremove structure donoderemoval removaldata nodestoremove | 1 |
1,883 | 3,413,290,924 | IssuesEvent | 2015-12-06 15:41:48 | servo/servo | https://api.github.com/repos/servo/servo | opened | Make travis run the unit tests | A-infrastructure A-testing E-easy | They're often forgotten when updating code. We should see if we can run them on Travis as well. It should only require adding a `./mach test-unit` entry to our `.travis.yml` file in the root directory. | 1.0 | Make travis run the unit tests - They're often forgotten when updating code. We should see if we can run them on Travis as well. It should only require adding a `./mach test-unit` entry to our `.travis.yml` file in the root directory. | infrastructure | make travis run the unit tests they re often forgotten when updating code we should see if we can run them on travis as well it should only require adding a mach test unit entry to our travis yml file in the root directory | 1 |
208,817 | 16,163,957,892 | IssuesEvent | 2021-05-01 06:01:14 | GreaterGoodCorp/SuperHelper | https://api.github.com/repos/GreaterGoodCorp/SuperHelper | closed | Missing full stop in command help messages | bug misc-documentation priority-low | Location:
- `--debug` option of `helper`
- `--list` option of `helper` | 1.0 | Missing full stop in command help messages - Location:
- `--debug` option of `helper`
- `--list` option of `helper` | non_infrastructure | missing full stop in command help messages location debug option of helper list option of helper | 0 |
7,379 | 6,928,234,596 | IssuesEvent | 2017-12-01 03:24:29 | twitter/reasonable-scala | https://api.github.com/repos/twitter/reasonable-scala | closed | Figure out the methodology for native benchmarks | Feature Infrastructure Performance | At the moment, the benchmarking infrastructure for RscNativeTypecheck is very minimal. We should bring it on par with the robustness that JMH provides (at the very minimum, add support for score errors) or replace it with a more advanced tool. | 1.0 | Figure out the methodology for native benchmarks - At the moment, the benchmarking infrastructure for RscNativeTypecheck is very minimal. We should bring it on par with the robustness that JMH provides (at the very minimum, add support for score errors) or replace it with a more advanced tool. | infrastructure | figure out the methodology for native benchmarks at the moment the benchmarking infrastructure for rscnativetypecheck is very minimal we should bring it on par with the robustness that jmh provides at the very minimum add support for score errors or replace it with a more advanced tool | 1 |
30,856 | 13,348,110,400 | IssuesEvent | 2020-08-29 16:52:47 | Azure/azure-sdk-for-java | https://api.github.com/repos/Azure/azure-sdk-for-java | closed | java.lang.IllegalStateException: RntbdServiceEndpoint is closed | Client Cosmos Service Attention bug cosmos:v4-item customer-reported needs-author-feedback needs-team-attention no-recent-activity | **Describe the bug**
The bug is with the below azure cosmos SDK
```
<dependency>
<groupId>com.azure</groupId>
<artifactId>azure-cosmos</artifactId>
<version>4.0.1-beta.2</version>
</dependency>
```
Getting below exception with **DIRECT** Connection mode and **pool size of 1000**:
**java.lang.IllegalStateException: RntbdServiceEndpoint({"id":1,"isClosed":true,"concurrentRequests":0,"remoteAddress":"cdb-ms-prod-westus2-fd15.documents.azure.com/<unresolved>:14096","channelPool":{"isClosed":true,"configuration":{"maxChannels":10,"maxRequestsPerChannel":30,"idleConnectionTimeout":0,"readDelayLimit":65000000000,"writeDelayLimit":10000000000},"state":{"channelsAcquired":0,"channelsAvailable":0,"requestQueueLength":0,"usedDirectMemory":16777216,"usedHeapMemory":16777216}}}) is closed**
This intermittent exception starts coming after some time of starting the spring boot application.
***Exception or Stack Trace***
java.lang.IllegalStateException: RntbdServiceEndpoint({"id":1,"isClosed":true,"concurrentRequests":0,"remoteAddress":"cdb-ms-prod-westus2-fd15.documents.azure.com/<unresolved>:14096","channelPool":{"isClosed":true,"configuration":{"maxChannels":10,"maxRequestsPerChannel":30,"idleConnectionTimeout":0,"readDelayLimit":65000000000,"writeDelayLimit":10000000000},"state":{"channelsAcquired":0,"channelsAvailable":0,"requestQueueLength":0,"usedDirectMemory":16777216,"usedHeapMemory":16777216}}}) is closed
at com.azure.cosmos.implementation.guava25.base.Preconditions.checkState(Preconditions.java:586) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.rntbd.RntbdServiceEndpoint.throwIfClosed(RntbdServiceEndpoint.java:227) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.rntbd.RntbdServiceEndpoint.request(RntbdServiceEndpoint.java:173) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.RntbdTransportClient.invokeStoreAsync(RntbdTransportClient.java:125) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.TransportClient.invokeResourceOperationAsync(TransportClient.java:21) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.StoreReader.readFromStoreAsync(StoreReader.java:639) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.StoreReader.readFromReplicas(StoreReader.java:212) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.StoreReader.lambda$readMultipleReplicasInternalAsync$10(StoreReader.java:394) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:46) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeatPredicate$RepeatPredicateSubscriber.resubscribe(FluxRepeatPredicate.java:113) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeatPredicate.subscribeOrReturn(FluxRepeatPredicate.java:49) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Mono.subscribe(Mono.java:4203) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxConcatArray$ConcatArraySubscriber.onComplete(FluxConcatArray.java:207) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxConcatArray.subscribe(FluxConcatArray.java:80) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxSwitchIfEmpty$SwitchIfEmptySubscriber.onComplete(FluxSwitchIfEmpty.java:75) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators.complete(Operators.java:135) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxEmpty.subscribe(FluxEmpty.java:41) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onNext(FluxFlatMap.java:418) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:73) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.checkTerminated(FluxFlatMap.java:838) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drainLoop(FluxFlatMap.java:600) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drain(FluxFlatMap.java:580) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onComplete(FluxFlatMap.java:457) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postCompleteDrain(DrainUtils.java:131) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postComplete(DrainUtils.java:186) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onComplete(FluxMapSignal.java:213) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoCacheTime.subscribeOrReturn(MonoCacheTime.java:140) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:57) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:150) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:73) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.checkTerminated(FluxFlatMap.java:838) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drainLoop(FluxFlatMap.java:600) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drain(FluxFlatMap.java:580) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onComplete(FluxFlatMap.java:457) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postCompleteDrain(DrainUtils.java:131) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postComplete(DrainUtils.java:186) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onComplete(FluxMapSignal.java:213) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoCacheTime.subscribeOrReturn(MonoCacheTime.java:140) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:57) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:150) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:73) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.checkTerminated(FluxFlatMap.java:838) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drainLoop(FluxFlatMap.java:600) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drain(FluxFlatMap.java:580) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onComplete(FluxFlatMap.java:457) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postCompleteDrain(DrainUtils.java:131) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postComplete(DrainUtils.java:186) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onComplete(FluxMapSignal.java:213) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxJust$WeakScalarSubscription.request(FluxJust.java:101) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.request(FluxMapSignal.java:225) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onSubscribe(FluxFlatMap.java:363) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onSubscribe(FluxMapSignal.java:115) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxJust.subscribe(FluxJust.java:70) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoDefer.subscribe(MonoDefer.java:52) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoDefer.subscribe(MonoDefer.java:52) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRetryWhen.subscribe(FluxRetryWhen.java:79) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoRetryWhen.subscribeOrReturn(MonoRetryWhen.java:46) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8311) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMergeSequential$MergeSequentialMain.onNext(FluxMergeSequential.java:230) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:169) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at com.azure.cosmos.implementation.query.Paginator.lambda$getPaginatedQueryResultAsObservable$1(Paginator.java:64) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:262) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:204) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMergeSequential$MergeSequentialMain.onSubscribe(FluxMergeSequential.java:191) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxGenerate.subscribe(FluxGenerate.java:83) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalFluxOperator.subscribe(InternalFluxOperator.java:62) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMergeSequential$MergeSequentialMain.onNext(FluxMergeSequential.java:230) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxIterable$IterableSubscription.slowPath(FluxIterable.java:267) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxIterable$IterableSubscription.request(FluxIterable.java:225) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMergeSequential$MergeSequentialMain.onSubscribe(FluxMergeSequential.java:191) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxIterable.subscribe(FluxIterable.java:161) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxIterable.subscribe(FluxIterable.java:86) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxConcatArray$ConcatArraySubscriber.onComplete(FluxConcatArray.java:207) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxConcatArray.subscribe(FluxConcatArray.java:80) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalFluxOperator.subscribe(InternalFluxOperator.java:62) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onNext(FluxFlatMap.java:418) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onComplete(FluxMapFuseable.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxJust$WeakScalarSubscription.request(FluxJust.java:101) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.request(FluxMapFuseable.java:162) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.request(MonoSingle.java:94) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onSubscribe(MonoFlatMap.java:230) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onSubscribe(MonoSingle.java:114) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onSubscribe(FluxMapFuseable.java:90) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxJust.subscribe(FluxJust.java:70) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFromFluxOperator.subscribe(MonoFromFluxOperator.java:81) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:150) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoTakeLastOne$TakeLastOneSubscriber.onComplete(MonoTakeLastOne.java:112) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxTakeUntil$TakeUntilPredicateSubscriber.onComplete(FluxTakeUntil.java:114) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxTakeUntil$TakeUntilPredicateSubscriber.onNext(FluxTakeUntil.java:92) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeat$RepeatSubscriber.onNext(FluxRepeat.java:83) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:73) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.checkTerminated(FluxFlatMap.java:838) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drainLoop(FluxFlatMap.java:600) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drain(FluxFlatMap.java:580) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onComplete(FluxFlatMap.java:457) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postCompleteDrain(DrainUtils.java:131) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postComplete(DrainUtils.java:186) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onComplete(FluxMapSignal.java:213) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoCacheTime.subscribeOrReturn(MonoCacheTime.java:140) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFromMonoOperator.subscribe(FluxFromMonoOperator.java:76) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeat$RepeatSubscriber.resubscribe(FluxRepeat.java:113) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeat$RepeatSubscriber.onComplete(FluxRepeat.java:97) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeat.subscribeOrReturn(FluxRepeat.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:57) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:150) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRetryWhen$RetryWhenMainSubscriber.onNext(FluxRetryWhen.java:162) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onComplete(Operators.java:1989) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.SerializedSubscriber.onComplete(SerializedSubscriber.java:146) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRetryWhen$RetryWhenMainSubscriber.onComplete(FluxRetryWhen.java:188) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onComplete(Operators.java:1989) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onComplete(FluxMapFuseable.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxSwitchIfEmpty$SwitchIfEmptySubscriber.onComplete(FluxSwitchIfEmpty.java:78) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxHandle$HandleSubscriber.onComplete(FluxHandle.java:206) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapConditionalSubscriber.onComplete(FluxMap.java:262) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onComplete(FluxDoFinally.java:138) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxHandleFuseable$HandleFuseableSubscriber.onComplete(FluxHandleFuseable.java:223) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxContextStart$ContextStartSubscriber.onComplete(FluxContextStart.java:115) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoCollectList$MonoCollectListSubscriber.onComplete(MonoCollectList.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxPeek$PeekSubscriber.onComplete(FluxPeek.java:252) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.netty.channel.FluxReceive.onInboundComplete(FluxReceive.java:366) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at reactor.netty.channel.ChannelOperations.onInboundComplete(ChannelOperations.java:367) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at reactor.netty.channel.ChannelOperations.terminate(ChannelOperations.java:423) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at reactor.netty.http.client.HttpClientOperations.onInboundNext(HttpClientOperations.java:607) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at reactor.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:96) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelRead(CombinedChannelDuplexHandler.java:436) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:324) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:296) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.CombinedChannelDuplexHandler.channelRead(CombinedChannelDuplexHandler.java:251) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1518) ~[netty-handler-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1267) ~[netty-handler-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1314) ~[netty-handler-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:501) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:440) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:276) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:650) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:576) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) ~[netty-common-4.1.49.Final.jar:4.1.49.Final]
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[netty-common-4.1.49.Final.jar:4.1.49.Final]
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[netty-common-4.1.49.Final.jar:4.1.49.Final]
at java.base/java.lang.Thread.run(Thread.java:832) ~[na:na]
**To Reproduce**
Use the above azure cosmos sdk and try to query a collection.
***Code Snippet***
Add the code snippet that causes the issue.
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
Output of **uname- a**
-t: cpu time (seconds) unlimited
-f: file size (blocks) unlimited
-d: data seg size (kbytes) unlimited
-s: stack size (kbytes) 8192
-c: core file size (blocks) 0
-v: address space (kbytes) unlimited
-l: locked-in-memory size (kbytes) unlimited
-u: processes 1392
-n: file descriptors 2560
**Setup (please complete the following information):**
- OS: iOS
- IDE : eclipse
- Version of the Library used
4.0.1-beta.2
**Additional context**
Add any other context about the problem here.
**Information Checklist**
Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report
- [x] Bug Description Added
- [x] Repro Steps Added
- [x] Setup information Added
| 1.0 | java.lang.IllegalStateException: RntbdServiceEndpoint is closed - **Describe the bug**
The bug is with the below azure cosmos SDK
```
<dependency>
<groupId>com.azure</groupId>
<artifactId>azure-cosmos</artifactId>
<version>4.0.1-beta.2</version>
</dependency>
```
Getting below exception with **DIRECT** Connection mode and **pool size of 1000**:
**java.lang.IllegalStateException: RntbdServiceEndpoint({"id":1,"isClosed":true,"concurrentRequests":0,"remoteAddress":"cdb-ms-prod-westus2-fd15.documents.azure.com/<unresolved>:14096","channelPool":{"isClosed":true,"configuration":{"maxChannels":10,"maxRequestsPerChannel":30,"idleConnectionTimeout":0,"readDelayLimit":65000000000,"writeDelayLimit":10000000000},"state":{"channelsAcquired":0,"channelsAvailable":0,"requestQueueLength":0,"usedDirectMemory":16777216,"usedHeapMemory":16777216}}}) is closed**
This intermittent exception starts coming after some time of starting the spring boot application.
***Exception or Stack Trace***
java.lang.IllegalStateException: RntbdServiceEndpoint({"id":1,"isClosed":true,"concurrentRequests":0,"remoteAddress":"cdb-ms-prod-westus2-fd15.documents.azure.com/<unresolved>:14096","channelPool":{"isClosed":true,"configuration":{"maxChannels":10,"maxRequestsPerChannel":30,"idleConnectionTimeout":0,"readDelayLimit":65000000000,"writeDelayLimit":10000000000},"state":{"channelsAcquired":0,"channelsAvailable":0,"requestQueueLength":0,"usedDirectMemory":16777216,"usedHeapMemory":16777216}}}) is closed
at com.azure.cosmos.implementation.guava25.base.Preconditions.checkState(Preconditions.java:586) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.rntbd.RntbdServiceEndpoint.throwIfClosed(RntbdServiceEndpoint.java:227) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.rntbd.RntbdServiceEndpoint.request(RntbdServiceEndpoint.java:173) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.RntbdTransportClient.invokeStoreAsync(RntbdTransportClient.java:125) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.TransportClient.invokeResourceOperationAsync(TransportClient.java:21) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.StoreReader.readFromStoreAsync(StoreReader.java:639) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.StoreReader.readFromReplicas(StoreReader.java:212) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at com.azure.cosmos.implementation.directconnectivity.StoreReader.lambda$readMultipleReplicasInternalAsync$10(StoreReader.java:394) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:46) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeatPredicate$RepeatPredicateSubscriber.resubscribe(FluxRepeatPredicate.java:113) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeatPredicate.subscribeOrReturn(FluxRepeatPredicate.java:49) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Mono.subscribe(Mono.java:4203) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxConcatArray$ConcatArraySubscriber.onComplete(FluxConcatArray.java:207) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxConcatArray.subscribe(FluxConcatArray.java:80) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxSwitchIfEmpty$SwitchIfEmptySubscriber.onComplete(FluxSwitchIfEmpty.java:75) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators.complete(Operators.java:135) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxEmpty.subscribe(FluxEmpty.java:41) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onNext(FluxFlatMap.java:418) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:73) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.checkTerminated(FluxFlatMap.java:838) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drainLoop(FluxFlatMap.java:600) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drain(FluxFlatMap.java:580) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onComplete(FluxFlatMap.java:457) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postCompleteDrain(DrainUtils.java:131) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postComplete(DrainUtils.java:186) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onComplete(FluxMapSignal.java:213) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoCacheTime.subscribeOrReturn(MonoCacheTime.java:140) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:57) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:150) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:73) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.checkTerminated(FluxFlatMap.java:838) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drainLoop(FluxFlatMap.java:600) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drain(FluxFlatMap.java:580) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onComplete(FluxFlatMap.java:457) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postCompleteDrain(DrainUtils.java:131) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postComplete(DrainUtils.java:186) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onComplete(FluxMapSignal.java:213) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoCacheTime.subscribeOrReturn(MonoCacheTime.java:140) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:57) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:150) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:73) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.checkTerminated(FluxFlatMap.java:838) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drainLoop(FluxFlatMap.java:600) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drain(FluxFlatMap.java:580) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onComplete(FluxFlatMap.java:457) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postCompleteDrain(DrainUtils.java:131) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postComplete(DrainUtils.java:186) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onComplete(FluxMapSignal.java:213) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxJust$WeakScalarSubscription.request(FluxJust.java:101) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.request(FluxMapSignal.java:225) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onSubscribe(FluxFlatMap.java:363) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onSubscribe(FluxMapSignal.java:115) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxJust.subscribe(FluxJust.java:70) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoDefer.subscribe(MonoDefer.java:52) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:64) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoDefer.subscribe(MonoDefer.java:52) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRetryWhen.subscribe(FluxRetryWhen.java:79) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoRetryWhen.subscribeOrReturn(MonoRetryWhen.java:46) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8311) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMergeSequential$MergeSequentialMain.onNext(FluxMergeSequential.java:230) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.next(FluxGenerate.java:169) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at com.azure.cosmos.implementation.query.Paginator.lambda$getPaginatedQueryResultAsObservable$1(Paginator.java:64) ~[azure-cosmos-4.0.1-beta.2.jar:na]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.slowPath(FluxGenerate.java:262) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxGenerate$GenerateSubscription.request(FluxGenerate.java:204) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMergeSequential$MergeSequentialMain.onSubscribe(FluxMergeSequential.java:191) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxGenerate.subscribe(FluxGenerate.java:83) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalFluxOperator.subscribe(InternalFluxOperator.java:62) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMergeSequential$MergeSequentialMain.onNext(FluxMergeSequential.java:230) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxIterable$IterableSubscription.slowPath(FluxIterable.java:267) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxIterable$IterableSubscription.request(FluxIterable.java:225) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMergeSequential$MergeSequentialMain.onSubscribe(FluxMergeSequential.java:191) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxIterable.subscribe(FluxIterable.java:161) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxIterable.subscribe(FluxIterable.java:86) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxConcatArray$ConcatArraySubscriber.onComplete(FluxConcatArray.java:207) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxConcatArray.subscribe(FluxConcatArray.java:80) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalFluxOperator.subscribe(InternalFluxOperator.java:62) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Flux.subscribe(Flux.java:8325) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onNext(FluxFlatMap.java:418) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onComplete(FluxMapFuseable.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxJust$WeakScalarSubscription.request(FluxJust.java:101) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.request(FluxMapFuseable.java:162) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.request(MonoSingle.java:94) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onSubscribe(MonoFlatMap.java:230) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onSubscribe(MonoSingle.java:114) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onSubscribe(FluxMapFuseable.java:90) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxJust.subscribe(FluxJust.java:70) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFromFluxOperator.subscribe(MonoFromFluxOperator.java:81) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:150) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoTakeLastOne$TakeLastOneSubscriber.onComplete(MonoTakeLastOne.java:112) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxTakeUntil$TakeUntilPredicateSubscriber.onComplete(FluxTakeUntil.java:114) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxTakeUntil$TakeUntilPredicateSubscriber.onNext(FluxTakeUntil.java:92) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeat$RepeatSubscriber.onNext(FluxRepeat.java:83) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxOnErrorResume$ResumeSubscriber.onNext(FluxOnErrorResume.java:73) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onNext(FluxMapFuseable.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.checkTerminated(FluxFlatMap.java:838) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drainLoop(FluxFlatMap.java:600) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.drain(FluxFlatMap.java:580) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFlatMap$FlatMapMain.onComplete(FluxFlatMap.java:457) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postCompleteDrain(DrainUtils.java:131) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.DrainUtils.postComplete(DrainUtils.java:186) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapSignal$FluxMapSignalSubscriber.onComplete(FluxMapSignal.java:213) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoCacheTime.subscribeOrReturn(MonoCacheTime.java:140) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxFromMonoOperator.subscribe(FluxFromMonoOperator.java:76) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDefer.subscribe(FluxDefer.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeat$RepeatSubscriber.resubscribe(FluxRepeat.java:113) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeat$RepeatSubscriber.onComplete(FluxRepeat.java:97) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRepeat.subscribeOrReturn(FluxRepeat.java:54) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.InternalMonoOperator.subscribe(InternalMonoOperator.java:57) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:150) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.SerializedSubscriber.onNext(SerializedSubscriber.java:99) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRetryWhen$RetryWhenMainSubscriber.onNext(FluxRetryWhen.java:162) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapMain.onNext(MonoFlatMap.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onComplete(Operators.java:1989) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.SerializedSubscriber.onComplete(SerializedSubscriber.java:146) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxRetryWhen$RetryWhenMainSubscriber.onComplete(FluxRetryWhen.java:188) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MultiSubscriptionSubscriber.onComplete(Operators.java:1989) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMapFuseable$MapFuseableSubscriber.onComplete(FluxMapFuseable.java:144) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoFlatMap$FlatMapInner.onNext(MonoFlatMap.java:241) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1755) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoSingle$SingleSubscriber.onComplete(MonoSingle.java:171) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxSwitchIfEmpty$SwitchIfEmptySubscriber.onComplete(FluxSwitchIfEmpty.java:78) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxHandle$HandleSubscriber.onComplete(FluxHandle.java:206) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapConditionalSubscriber.onComplete(FluxMap.java:262) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxDoFinally$DoFinallySubscriber.onComplete(FluxDoFinally.java:138) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxHandleFuseable$HandleFuseableSubscriber.onComplete(FluxHandleFuseable.java:223) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxContextStart$ContextStartSubscriber.onComplete(FluxContextStart.java:115) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.Operators$MonoSubscriber.complete(Operators.java:1756) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.MonoCollectList$MonoCollectListSubscriber.onComplete(MonoCollectList.java:121) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxPeek$PeekSubscriber.onComplete(FluxPeek.java:252) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.core.publisher.FluxMap$MapSubscriber.onComplete(FluxMap.java:136) ~[reactor-core-3.3.5.RELEASE.jar:3.3.5.RELEASE]
at reactor.netty.channel.FluxReceive.onInboundComplete(FluxReceive.java:366) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at reactor.netty.channel.ChannelOperations.onInboundComplete(ChannelOperations.java:367) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at reactor.netty.channel.ChannelOperations.terminate(ChannelOperations.java:423) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at reactor.netty.http.client.HttpClientOperations.onInboundNext(HttpClientOperations.java:607) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at reactor.netty.channel.ChannelOperationsHandler.channelRead(ChannelOperationsHandler.java:96) ~[reactor-netty-0.9.7.RELEASE.jar:0.9.7.RELEASE]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.CombinedChannelDuplexHandler$DelegatingChannelHandlerContext.fireChannelRead(CombinedChannelDuplexHandler.java:436) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:324) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:296) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.CombinedChannelDuplexHandler.channelRead(CombinedChannelDuplexHandler.java:251) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1518) ~[netty-handler-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1267) ~[netty-handler-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1314) ~[netty-handler-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:501) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:440) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:276) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:714) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:650) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:576) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final]
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) ~[netty-common-4.1.49.Final.jar:4.1.49.Final]
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) ~[netty-common-4.1.49.Final.jar:4.1.49.Final]
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) ~[netty-common-4.1.49.Final.jar:4.1.49.Final]
at java.base/java.lang.Thread.run(Thread.java:832) ~[na:na]
**To Reproduce**
Use the above azure cosmos sdk and try to query a collection.
***Code Snippet***
Add the code snippet that causes the issue.
**Expected behavior**
A clear and concise description of what you expected to happen.
**Screenshots**
Output of **uname- a**
-t: cpu time (seconds) unlimited
-f: file size (blocks) unlimited
-d: data seg size (kbytes) unlimited
-s: stack size (kbytes) 8192
-c: core file size (blocks) 0
-v: address space (kbytes) unlimited
-l: locked-in-memory size (kbytes) unlimited
-u: processes 1392
-n: file descriptors 2560
**Setup (please complete the following information):**
- OS: iOS
- IDE : eclipse
- Version of the Library used
4.0.1-beta.2
**Additional context**
Add any other context about the problem here.
**Information Checklist**
Kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report
- [x] Bug Description Added
- [x] Repro Steps Added
- [x] Setup information Added
| non_infrastructure | java lang illegalstateexception rntbdserviceendpoint is closed describe the bug the bug is with the below azure cosmos sdk com azure azure cosmos beta getting below exception with direct connection mode and pool size of java lang illegalstateexception rntbdserviceendpoint id isclosed true concurrentrequests remoteaddress cdb ms prod documents azure com channelpool isclosed true configuration maxchannels maxrequestsperchannel idleconnectiontimeout readdelaylimit writedelaylimit state channelsacquired channelsavailable requestqueuelength useddirectmemory usedheapmemory is closed this intermittent exception starts coming after some time of starting the spring boot application exception or stack trace java lang illegalstateexception rntbdserviceendpoint id isclosed true concurrentrequests remoteaddress cdb ms prod documents azure com channelpool isclosed true configuration maxchannels maxrequestsperchannel idleconnectiontimeout readdelaylimit writedelaylimit state channelsacquired channelsavailable requestqueuelength useddirectmemory usedheapmemory is closed at com azure cosmos implementation base preconditions checkstate preconditions java at com azure cosmos implementation directconnectivity rntbd rntbdserviceendpoint throwifclosed rntbdserviceendpoint java at com azure cosmos implementation directconnectivity rntbd rntbdserviceendpoint request rntbdserviceendpoint java at com azure cosmos implementation directconnectivity rntbdtransportclient invokestoreasync rntbdtransportclient java at com azure cosmos implementation directconnectivity transportclient invokeresourceoperationasync transportclient java at com azure cosmos implementation directconnectivity storereader readfromstoreasync storereader java at com azure cosmos implementation directconnectivity storereader readfromreplicas storereader java at com azure cosmos implementation directconnectivity storereader lambda readmultiplereplicasinternalasync storereader java at reactor core publisher fluxdefer subscribe fluxdefer java at reactor core publisher fluxrepeatpredicate repeatpredicatesubscriber resubscribe fluxrepeatpredicate java at reactor core publisher fluxrepeatpredicate subscribeorreturn fluxrepeatpredicate java at reactor core publisher mono subscribe mono java at reactor core publisher fluxconcatarray concatarraysubscriber oncomplete fluxconcatarray java at reactor core publisher fluxconcatarray subscribe fluxconcatarray java at reactor core publisher fluxdefer subscribe fluxdefer java at reactor core publisher flux subscribe flux java at reactor core publisher fluxswitchifempty switchifemptysubscriber oncomplete fluxswitchifempty java at reactor core publisher operators complete operators java at reactor core publisher fluxempty subscribe fluxempty java at reactor core publisher flux subscribe flux java at reactor core publisher fluxflatmap flatmapmain onnext fluxflatmap java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapinner onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher fluxflatmap flatmapmain checkterminated fluxflatmap java at reactor core publisher fluxflatmap flatmapmain drainloop fluxflatmap java at reactor core publisher fluxflatmap flatmapmain drain fluxflatmap java at reactor core publisher fluxflatmap flatmapmain oncomplete fluxflatmap java at reactor core publisher drainutils postcompletedrain drainutils java at reactor core publisher drainutils postcomplete drainutils java at reactor core publisher fluxmapsignal fluxmapsignalsubscriber oncomplete fluxmapsignal java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monocachetime subscribeorreturn monocachetime java at reactor core publisher internalmonooperator subscribe internalmonooperator java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapinner onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher fluxflatmap flatmapmain checkterminated fluxflatmap java at reactor core publisher fluxflatmap flatmapmain drainloop fluxflatmap java at reactor core publisher fluxflatmap flatmapmain drain fluxflatmap java at reactor core publisher fluxflatmap flatmapmain oncomplete fluxflatmap java at reactor core publisher drainutils postcompletedrain drainutils java at reactor core publisher drainutils postcomplete drainutils java at reactor core publisher fluxmapsignal fluxmapsignalsubscriber oncomplete fluxmapsignal java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monocachetime subscribeorreturn monocachetime java at reactor core publisher internalmonooperator subscribe internalmonooperator java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher fluxflatmap flatmapmain checkterminated fluxflatmap java at reactor core publisher fluxflatmap flatmapmain drainloop fluxflatmap java at reactor core publisher fluxflatmap flatmapmain drain fluxflatmap java at reactor core publisher fluxflatmap flatmapmain oncomplete fluxflatmap java at reactor core publisher drainutils postcompletedrain drainutils java at reactor core publisher drainutils postcomplete drainutils java at reactor core publisher fluxmapsignal fluxmapsignalsubscriber oncomplete fluxmapsignal java at reactor core publisher fluxjust weakscalarsubscription request fluxjust java at reactor core publisher fluxmapsignal fluxmapsignalsubscriber request fluxmapsignal java at reactor core publisher fluxflatmap flatmapmain onsubscribe fluxflatmap java at reactor core publisher fluxmapsignal fluxmapsignalsubscriber onsubscribe fluxmapsignal java at reactor core publisher fluxjust subscribe fluxjust java at reactor core publisher internalmonooperator subscribe internalmonooperator java at reactor core publisher monodefer subscribe monodefer java at reactor core publisher internalmonooperator subscribe internalmonooperator java at reactor core publisher monodefer subscribe monodefer java at reactor core publisher fluxretrywhen subscribe fluxretrywhen java at reactor core publisher monoretrywhen subscribeorreturn monoretrywhen java at reactor core publisher flux subscribe flux java at reactor core publisher fluxmergesequential mergesequentialmain onnext fluxmergesequential java at reactor core publisher fluxgenerate generatesubscription next fluxgenerate java at com azure cosmos implementation query paginator lambda getpaginatedqueryresultasobservable paginator java at reactor core publisher fluxgenerate generatesubscription slowpath fluxgenerate java at reactor core publisher fluxgenerate generatesubscription request fluxgenerate java at reactor core publisher fluxmergesequential mergesequentialmain onsubscribe fluxmergesequential java at reactor core publisher fluxgenerate subscribe fluxgenerate java at reactor core publisher internalfluxoperator subscribe internalfluxoperator java at reactor core publisher fluxdefer subscribe fluxdefer java at reactor core publisher flux subscribe flux java at reactor core publisher fluxmergesequential mergesequentialmain onnext fluxmergesequential java at reactor core publisher fluxiterable iterablesubscription slowpath fluxiterable java at reactor core publisher fluxiterable iterablesubscription request fluxiterable java at reactor core publisher fluxmergesequential mergesequentialmain onsubscribe fluxmergesequential java at reactor core publisher fluxiterable subscribe fluxiterable java at reactor core publisher fluxiterable subscribe fluxiterable java at reactor core publisher flux subscribe flux java at reactor core publisher fluxconcatarray concatarraysubscriber oncomplete fluxconcatarray java at reactor core publisher fluxconcatarray subscribe fluxconcatarray java at reactor core publisher internalfluxoperator subscribe internalfluxoperator java at reactor core publisher fluxdefer subscribe fluxdefer java at reactor core publisher flux subscribe flux java at reactor core publisher fluxflatmap flatmapmain onnext fluxflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapinner onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapinner onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapinner onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher fluxmapfuseable mapfuseablesubscriber oncomplete fluxmapfuseable java at reactor core publisher fluxjust weakscalarsubscription request fluxjust java at reactor core publisher fluxmapfuseable mapfuseablesubscriber request fluxmapfuseable java at reactor core publisher monosingle singlesubscriber request monosingle java at reactor core publisher monoflatmap flatmapinner onsubscribe monoflatmap java at reactor core publisher monosingle singlesubscriber onsubscribe monosingle java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onsubscribe fluxmapfuseable java at reactor core publisher fluxjust subscribe fluxjust java at reactor core publisher monofromfluxoperator subscribe monofromfluxoperator java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monotakelastone takelastonesubscriber oncomplete monotakelastone java at reactor core publisher fluxtakeuntil takeuntilpredicatesubscriber oncomplete fluxtakeuntil java at reactor core publisher fluxtakeuntil takeuntilpredicatesubscriber onnext fluxtakeuntil java at reactor core publisher fluxrepeat repeatsubscriber onnext fluxrepeat java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher fluxonerrorresume resumesubscriber onnext fluxonerrorresume java at reactor core publisher fluxmapfuseable mapfuseablesubscriber onnext fluxmapfuseable java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher fluxflatmap flatmapmain checkterminated fluxflatmap java at reactor core publisher fluxflatmap flatmapmain drainloop fluxflatmap java at reactor core publisher fluxflatmap flatmapmain drain fluxflatmap java at reactor core publisher fluxflatmap flatmapmain oncomplete fluxflatmap java at reactor core publisher drainutils postcompletedrain drainutils java at reactor core publisher drainutils postcomplete drainutils java at reactor core publisher fluxmapsignal fluxmapsignalsubscriber oncomplete fluxmapsignal java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monocachetime subscribeorreturn monocachetime java at reactor core publisher fluxfrommonooperator subscribe fluxfrommonooperator java at reactor core publisher fluxdefer subscribe fluxdefer java at reactor core publisher fluxrepeat repeatsubscriber resubscribe fluxrepeat java at reactor core publisher fluxrepeat repeatsubscriber oncomplete fluxrepeat java at reactor core publisher fluxrepeat subscribeorreturn fluxrepeat java at reactor core publisher internalmonooperator subscribe internalmonooperator java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher serializedsubscriber onnext serializedsubscriber java at reactor core publisher fluxretrywhen retrywhenmainsubscriber onnext fluxretrywhen java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapmain onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher fluxmap mapsubscriber oncomplete fluxmap java at reactor core publisher fluxmap mapsubscriber oncomplete fluxmap java at reactor core publisher operators multisubscriptionsubscriber oncomplete operators java at reactor core publisher serializedsubscriber oncomplete serializedsubscriber java at reactor core publisher fluxretrywhen retrywhenmainsubscriber oncomplete fluxretrywhen java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher operators multisubscriptionsubscriber oncomplete operators java at reactor core publisher fluxmapfuseable mapfuseablesubscriber oncomplete fluxmapfuseable java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monoflatmap flatmapinner onnext monoflatmap java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monosingle singlesubscriber oncomplete monosingle java at reactor core publisher fluxmap mapsubscriber oncomplete fluxmap java at reactor core publisher fluxswitchifempty switchifemptysubscriber oncomplete fluxswitchifempty java at reactor core publisher fluxhandle handlesubscriber oncomplete fluxhandle java at reactor core publisher fluxmap mapconditionalsubscriber oncomplete fluxmap java at reactor core publisher fluxdofinally dofinallysubscriber oncomplete fluxdofinally java at reactor core publisher fluxhandlefuseable handlefuseablesubscriber oncomplete fluxhandlefuseable java at reactor core publisher fluxcontextstart contextstartsubscriber oncomplete fluxcontextstart java at reactor core publisher operators monosubscriber complete operators java at reactor core publisher monocollectlist monocollectlistsubscriber oncomplete monocollectlist java at reactor core publisher fluxpeek peeksubscriber oncomplete fluxpeek java at reactor core publisher fluxmap mapsubscriber oncomplete fluxmap java at reactor netty channel fluxreceive oninboundcomplete fluxreceive java at reactor netty channel channeloperations oninboundcomplete channeloperations java at reactor netty channel channeloperations terminate channeloperations java at reactor netty http client httpclientoperations oninboundnext httpclientoperations java at reactor netty channel channeloperationshandler channelread channeloperationshandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel combinedchannelduplexhandler delegatingchannelhandlercontext firechannelread combinedchannelduplexhandler java at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel combinedchannelduplexhandler channelread combinedchannelduplexhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler ssl sslhandler unwrap sslhandler java at io netty handler ssl sslhandler decodejdkcompatible sslhandler java at io netty handler ssl sslhandler decode sslhandler java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java to reproduce use the above azure cosmos sdk and try to query a collection code snippet add the code snippet that causes the issue expected behavior a clear and concise description of what you expected to happen screenshots output of uname a t cpu time seconds unlimited f file size blocks unlimited d data seg size kbytes unlimited s stack size kbytes c core file size blocks v address space kbytes unlimited l locked in memory size kbytes unlimited u processes n file descriptors setup please complete the following information os ios ide eclipse version of the library used beta additional context add any other context about the problem here information checklist kindly make sure that you have added all the following information above and checkoff the required fields otherwise we will treat the issuer as an incomplete report bug description added repro steps added setup information added | 0 |
2,281 | 5,108,101,563 | IssuesEvent | 2017-01-05 16:44:29 | jlm2017/jlm-video-subtitles | https://api.github.com/repos/jlm2017/jlm-video-subtitles | closed | [Subtitles] [FR] La revue de la semaine n°1 : pauvreté, Hayange, démocratie, Alstom, Juppé et retraites. | Language: French Process: [6] Approved | # Video title
La revue de la semaine n°1 : pauvreté, Hayange, démocratie, Alstom, Juppé et retraites.
# URL
https://www.youtube.com/watch?v=ynfJBfJKzFw
# Youtube subtitles language
Français
# Duration
21:16
# Subtitles URL
https://www.youtube.com/timedtext_editor?tab=captions&v=ynfJBfJKzFw&ui=hd&action_mde_edit_form=1&ref=player&lang=fr&bl=vmp | 1.0 | [Subtitles] [FR] La revue de la semaine n°1 : pauvreté, Hayange, démocratie, Alstom, Juppé et retraites. - # Video title
La revue de la semaine n°1 : pauvreté, Hayange, démocratie, Alstom, Juppé et retraites.
# URL
https://www.youtube.com/watch?v=ynfJBfJKzFw
# Youtube subtitles language
Français
# Duration
21:16
# Subtitles URL
https://www.youtube.com/timedtext_editor?tab=captions&v=ynfJBfJKzFw&ui=hd&action_mde_edit_form=1&ref=player&lang=fr&bl=vmp | non_infrastructure | la revue de la semaine n° pauvreté hayange démocratie alstom juppé et retraites video title la revue de la semaine n° pauvreté hayange démocratie alstom juppé et retraites url youtube subtitles language français duration subtitles url | 0 |
12,094 | 9,594,279,447 | IssuesEvent | 2019-05-09 13:39:18 | facebook/react-native | https://api.github.com/repos/facebook/react-native | closed | [META] add test case for template | Contributor Partner 🔩Test Infrastructure | Test template is working in ci system. With this the when release new version should be more confident.
I may not have time for this recently, any pr is welcome. | 1.0 | [META] add test case for template - Test template is working in ci system. With this the when release new version should be more confident.
I may not have time for this recently, any pr is welcome. | infrastructure | add test case for template test template is working in ci system with this the when release new version should be more confident i may not have time for this recently any pr is welcome | 1 |
18,904 | 3,733,933,808 | IssuesEvent | 2016-03-08 03:06:57 | duttydev/sample-rest-api | https://api.github.com/repos/duttydev/sample-rest-api | closed | Test Coverage for PaginationUtil | 4 - Done unit-test |
<!---
@huboard:{"order":1.0,"milestone_order":1.0,"custom_state":""}
-->
| 1.0 | Test Coverage for PaginationUtil -
<!---
@huboard:{"order":1.0,"milestone_order":1.0,"custom_state":""}
-->
| non_infrastructure | test coverage for paginationutil huboard order milestone order custom state | 0 |
150,255 | 5,748,431,445 | IssuesEvent | 2017-04-25 00:41:00 | coreos/bugs | https://api.github.com/repos/coreos/bugs | closed | Random containerd crashes | area/usability component/docker kind/bug priority/P0 team/os | # Issue Report #
## Bug ##
We are observing containerd crashes across multiple nodes randomly. Each node have around 8 to 10 containers. containerd is spiking to ~100% CPU usage and after sometime it is crashing by killing all container-shim processes. Our node is able to work as expected when containerd is down. But as a result, docker is becoming unresponsive after couple of days.
We have no idea about how to reproduce this since this is happening on some random nodes.
All the nodes in our network have same configuration.
### Container Linux Version ###
```
$ cat /etc/os-release
NAME="Container Linux by CoreOS"
ID=coreos
VERSION=1298.6.0
VERSION_ID=1298.6.0
BUILD_ID=2017-03-14-2119
PRETTY_NAME="Container Linux by CoreOS 1298.6.0 (Ladybug)"
ANSI_COLOR="38;5;75"
HOME_URL="https://coreos.com/"
BUG_REPORT_URL="https://github.com/coreos/bugs/issues"
```
docker version:
```
Client:
Version: 1.12.6
API version: 1.24
Go version: go1.6.3
Git commit: d5236f0
Built: Tue Mar 14 21:10:49 2017
OS/Arch: linux/amd64
Server:
Version: 1.12.6
API version: 1.24
Go version: go1.6.3
Git commit: d5236f0
Built: Tue Mar 14 21:10:49 2017
OS/Arch: linux/amd64
```
docker info:
```
Containers: 9
Running: 9
Paused: 0
Stopped: 0
Images: 14
Server Version: 1.12.6
Storage Driver: overlay
Backing Filesystem: extfs
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: null host bridge overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Security Options: seccomp selinux
Kernel Version: 4.9.9-coreos-r1
Operating System: Container Linux by CoreOS 1298.6.0 (Ladybug)
OSType: linux
Architecture: x86_64
CPUs: 70
Total Memory: 58.98 GiB
Name: Server01
ID: PYQ3:JC2K:TNDY:QT6J:5RCU:6V64:TRPC:E44I:5WUB:B3RW:V3VF:GLGI
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Insecure Registries:
127.0.0.0/8
```
Docker and containerd journalctl logs:
sudo journalctl --no-pager -b -u docker -u containerd
[journal_b_docker.txt](https://github.com/coreos/bugs/files/914504/journal_b_docker.txt)
Docker daemon journalctl logs using sigusr1 signal at the time of containerd crash:
sudo kill -s SIGUSR1 <dockerd pid>
[journal_dockerd_sigusr1.txt](https://github.com/coreos/bugs/files/914508/journal_dockerd_sigusr1.txt)
If you need any specific info, I'll be happy to provide. | 1.0 | Random containerd crashes - # Issue Report #
## Bug ##
We are observing containerd crashes across multiple nodes randomly. Each node have around 8 to 10 containers. containerd is spiking to ~100% CPU usage and after sometime it is crashing by killing all container-shim processes. Our node is able to work as expected when containerd is down. But as a result, docker is becoming unresponsive after couple of days.
We have no idea about how to reproduce this since this is happening on some random nodes.
All the nodes in our network have same configuration.
### Container Linux Version ###
```
$ cat /etc/os-release
NAME="Container Linux by CoreOS"
ID=coreos
VERSION=1298.6.0
VERSION_ID=1298.6.0
BUILD_ID=2017-03-14-2119
PRETTY_NAME="Container Linux by CoreOS 1298.6.0 (Ladybug)"
ANSI_COLOR="38;5;75"
HOME_URL="https://coreos.com/"
BUG_REPORT_URL="https://github.com/coreos/bugs/issues"
```
docker version:
```
Client:
Version: 1.12.6
API version: 1.24
Go version: go1.6.3
Git commit: d5236f0
Built: Tue Mar 14 21:10:49 2017
OS/Arch: linux/amd64
Server:
Version: 1.12.6
API version: 1.24
Go version: go1.6.3
Git commit: d5236f0
Built: Tue Mar 14 21:10:49 2017
OS/Arch: linux/amd64
```
docker info:
```
Containers: 9
Running: 9
Paused: 0
Stopped: 0
Images: 14
Server Version: 1.12.6
Storage Driver: overlay
Backing Filesystem: extfs
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: null host bridge overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Security Options: seccomp selinux
Kernel Version: 4.9.9-coreos-r1
Operating System: Container Linux by CoreOS 1298.6.0 (Ladybug)
OSType: linux
Architecture: x86_64
CPUs: 70
Total Memory: 58.98 GiB
Name: Server01
ID: PYQ3:JC2K:TNDY:QT6J:5RCU:6V64:TRPC:E44I:5WUB:B3RW:V3VF:GLGI
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Registry: https://index.docker.io/v1/
Insecure Registries:
127.0.0.0/8
```
Docker and containerd journalctl logs:
sudo journalctl --no-pager -b -u docker -u containerd
[journal_b_docker.txt](https://github.com/coreos/bugs/files/914504/journal_b_docker.txt)
Docker daemon journalctl logs using sigusr1 signal at the time of containerd crash:
sudo kill -s SIGUSR1 <dockerd pid>
[journal_dockerd_sigusr1.txt](https://github.com/coreos/bugs/files/914508/journal_dockerd_sigusr1.txt)
If you need any specific info, I'll be happy to provide. | non_infrastructure | random containerd crashes issue report bug we are observing containerd crashes across multiple nodes randomly each node have around to containers containerd is spiking to cpu usage and after sometime it is crashing by killing all container shim processes our node is able to work as expected when containerd is down but as a result docker is becoming unresponsive after couple of days we have no idea about how to reproduce this since this is happening on some random nodes all the nodes in our network have same configuration container linux version cat etc os release name container linux by coreos id coreos version version id build id pretty name container linux by coreos ladybug ansi color home url bug report url docker version client version api version go version git commit built tue mar os arch linux server version api version go version git commit built tue mar os arch linux docker info containers running paused stopped images server version storage driver overlay backing filesystem extfs logging driver json file cgroup driver cgroupfs plugins volume local network null host bridge overlay swarm inactive runtimes runc default runtime runc security options seccomp selinux kernel version coreos operating system container linux by coreos ladybug ostype linux architecture cpus total memory gib name id tndy trpc glgi docker root dir var lib docker debug mode client false debug mode server false registry insecure registries docker and containerd journalctl logs sudo journalctl no pager b u docker u containerd docker daemon journalctl logs using signal at the time of containerd crash sudo kill s if you need any specific info i ll be happy to provide | 0 |
26,182 | 19,716,474,221 | IssuesEvent | 2022-01-13 11:29:29 | google/web-stories-wp | https://api.github.com/repos/google/web-stories-wp | closed | Failing AMP validation unit tests | Type: Bug P0 Type: Infrastructure Pod: WP & Infra | <!--
NOTE: For help requests, support questions, or general feedback,
please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/
-->
## Bug Description
<!-- A clear and concise description of what the bug is. -->
There were some changes on the AMP infrastructure side, causing the AMP validation tests to fail.
## Expected Behaviour
<!-- Please describe clearly and concisely what the expected behaviour should be. -->
## Steps to Reproduce
<!-- Please provide detailed steps on how to reproduce the bug. -->
## Screenshots
<!--
If applicable, please add screenshots to help explain your problem.
Bonus points for videos!
-->
## Additional Context
<!-- Please complete the following information. -->
- Plugin Version:
- WordPress Version:
- Operating System:
- Browser:
| 1.0 | Failing AMP validation unit tests - <!--
NOTE: For help requests, support questions, or general feedback,
please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/
-->
## Bug Description
<!-- A clear and concise description of what the bug is. -->
There were some changes on the AMP infrastructure side, causing the AMP validation tests to fail.
## Expected Behaviour
<!-- Please describe clearly and concisely what the expected behaviour should be. -->
## Steps to Reproduce
<!-- Please provide detailed steps on how to reproduce the bug. -->
## Screenshots
<!--
If applicable, please add screenshots to help explain your problem.
Bonus points for videos!
-->
## Additional Context
<!-- Please complete the following information. -->
- Plugin Version:
- WordPress Version:
- Operating System:
- Browser:
| infrastructure | failing amp validation unit tests note for help requests support questions or general feedback please use the wordpress org forums instead bug description there were some changes on the amp infrastructure side causing the amp validation tests to fail expected behaviour steps to reproduce screenshots if applicable please add screenshots to help explain your problem bonus points for videos additional context plugin version wordpress version operating system browser | 1 |
10,678 | 2,963,012,844 | IssuesEvent | 2015-07-10 07:11:51 | dhowe/AdNauseam | https://api.github.com/repos/dhowe/AdNauseam | closed | New log file needs design love | 2 - Awaiting Design Enhancement PRIORITY: High |

which may also require additional JSON formatting in the addon | 1.0 | New log file needs design love -

which may also require additional JSON formatting in the addon | non_infrastructure | new log file needs design love which may also require additional json formatting in the addon | 0 |
147,675 | 23,250,946,455 | IssuesEvent | 2022-08-04 03:37:09 | MozillaFoundation/Design | https://api.github.com/repos/MozillaFoundation/Design | closed | [RegretsReporter] Design Discovery Phase | design YouTube Regrets | **Preparation**
- [x] Watch video recording
- [x] Review last years work
- [x] Review roadmap
- [x] Start setting up docs for moodboard and concepts | 1.0 | [RegretsReporter] Design Discovery Phase - **Preparation**
- [x] Watch video recording
- [x] Review last years work
- [x] Review roadmap
- [x] Start setting up docs for moodboard and concepts | non_infrastructure | design discovery phase preparation watch video recording review last years work review roadmap start setting up docs for moodboard and concepts | 0 |
23,450 | 16,191,902,990 | IssuesEvent | 2021-05-04 09:40:01 | google/web-stories-wp | https://api.github.com/repos/google/web-stories-wp | closed | Update `@percy/puppeteer` to 2.0 | FIXIT P3 Pod: WP & Infra Type: Infrastructure | <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Task Description
Need to follow the upgrading guide at https://github.com/percy/percy-puppeteer#upgrading
| 1.0 | Update `@percy/puppeteer` to 2.0 - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Task Description
Need to follow the upgrading guide at https://github.com/percy/percy-puppeteer#upgrading
| infrastructure | update percy puppeteer to task description need to follow the upgrading guide at | 1 |
210,338 | 23,751,455,314 | IssuesEvent | 2022-08-31 21:04:05 | aayant-mend/onboardtraining | https://api.github.com/repos/aayant-mend/onboardtraining | reopened | camel-zookeeper-3.4.3.jar: 9 vulnerabilities (highest severity is: 7.5) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>camel-zookeeper-3.4.3.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/28.2-jre/guava-28.2-jre.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-25649](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25649) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | jackson-databind-2.10.0.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2020-36518](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | jackson-databind-2.10.0.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2020-11612](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11612) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | netty-codec-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2021-37136](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37136) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | netty-codec-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2021-37137](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37137) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | netty-codec-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [WS-2020-0408](https://github.com/netty/netty/issues/10362) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | netty-handler-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2022-24823](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | netty-common-4.1.45.Final.jar | Transitive | N/A | ❌ |
| [CVE-2021-21290](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21290) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | netty-handler-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2020-8908](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.3 | guava-28.2-jre.jar | Transitive | 3.11.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-25649</summary>
### Vulnerable Library - <b>jackson-databind-2.10.0.jar</b></p>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.0/jackson-databind-2.10.0.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- curator-x-discovery-4.3.0.jar
- :x: **jackson-databind-2.10.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A flaw was found in FasterXML Jackson Databind, where it did not have entity expansion secured properly. This flaw allows vulnerability to XML external entity (XXE) attacks. The highest threat from this vulnerability is data integrity.
<p>Publish Date: 2020-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25649>CVE-2020-25649</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-12-03</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.10.5.1</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-36518</summary>
### Vulnerable Library - <b>jackson-databind-2.10.0.jar</b></p>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.0/jackson-databind-2.10.0.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- curator-x-discovery-4.3.0.jar
- :x: **jackson-databind-2.10.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.
Mend Note: After conducting further research, Mend has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.
<p>Publish Date: 2022-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-11</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.12.6.1</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-11612</summary>
### Vulnerable Library - <b>netty-codec-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec/4.1.45.Final/netty-codec-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- netty-handler-4.1.45.Final.jar
- :x: **netty-codec-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The ZlibDecoders in Netty 4.1.x before 4.1.46 allow for unbounded memory allocation while decoding a ZlibEncoded byte stream. An attacker could send a large ZlibEncoded byte stream to the Netty server, forcing the server to allocate all of its free memory to a single decoder.
<p>Publish Date: 2020-04-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11612>CVE-2020-11612</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://netty.io/news/2020/02/28/4-1-46-Final.html">https://netty.io/news/2020/02/28/4-1-46-Final.html</a></p>
<p>Release Date: 2020-04-07</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.46.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-37136</summary>
### Vulnerable Library - <b>netty-codec-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec/4.1.45.Final/netty-codec-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- netty-handler-4.1.45.Final.jar
- :x: **netty-codec-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Bzip2 decompression decoder function doesn't allow setting size restrictions on the decompressed output data (which affects the allocation size used during decompression). All users of Bzip2Decoder are affected. The malicious input can trigger an OOME and so a DoS attack
<p>Publish Date: 2021-10-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37136>CVE-2021-37136</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-grg4-wf29-r9vv">https://github.com/netty/netty/security/advisories/GHSA-grg4-wf29-r9vv</a></p>
<p>Release Date: 2021-10-19</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.68.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-37137</summary>
### Vulnerable Library - <b>netty-codec-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec/4.1.45.Final/netty-codec-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- netty-handler-4.1.45.Final.jar
- :x: **netty-codec-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Snappy frame decoder function doesn't restrict the chunk length which may lead to excessive memory usage. Beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well. This vulnerability can be triggered by supplying malicious input that decompresses to a very big size (via a network stream or a file) or by sending a huge skippable chunk.
<p>Publish Date: 2021-10-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37137>CVE-2021-37137</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9vjp-v76f-g363">https://github.com/advisories/GHSA-9vjp-v76f-g363</a></p>
<p>Release Date: 2021-10-19</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.68.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0408</summary>
### Vulnerable Library - <b>netty-handler-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-handler/4.1.45.Final/netty-handler-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- :x: **netty-handler-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue was found in all versions of io.netty:netty-all. Host verification in Netty is disabled by default. This can lead to MITM attack in which an attacker can forge valid SSL/TLS certificates for a different hostname in order to intercept traffic that doesn’t intend for him. This is an issue because the certificate is not matched with the host.
<p>Publish Date: 2020-06-22
<p>URL: <a href=https://github.com/netty/netty/issues/10362>WS-2020-0408</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.4</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2020-0408">https://nvd.nist.gov/vuln/detail/WS-2020-0408</a></p>
<p>Release Date: 2020-06-22</p>
<p>Fix Resolution (io.netty:netty-handler): 4.1.69.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-24823</summary>
### Vulnerable Library - <b>netty-common-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-common/4.1.45.Final/netty-common-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- netty-handler-4.1.45.Final.jar
- :x: **netty-common-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user.
<p>Publish Date: 2022-05-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823>CVE-2022-24823</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p>
<p>Release Date: 2022-05-06</p>
<p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-21290</summary>
### Vulnerable Library - <b>netty-handler-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-handler/4.1.45.Final/netty-handler-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- :x: **netty-handler-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty before version 4.1.59.Final there is a vulnerability on Unix-like systems involving an insecure temp file. When netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. On unix-like systems, the temporary directory is shared between all user. As such, writing to this directory using APIs that do not explicitly set the file/directory permissions can lead to information disclosure. Of note, this does not impact modern MacOS Operating Systems. The method "File.createTempFile" on unix-like systems creates a random file, but, by default will create this file with the permissions "-rw-r--r--". Thus, if sensitive information is written to this file, other local users can read this information. This is the case in netty's "AbstractDiskHttpData" is vulnerable. This has been fixed in version 4.1.59.Final. As a workaround, one may specify your own "java.io.tmpdir" when you start the JVM or use "DefaultHttpDataFactory.setBaseDir(...)" to set the directory to something that is only readable by the current user.
<p>Publish Date: 2021-02-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21290>CVE-2021-21290</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-5mcr-gq6c-3hq2">https://github.com/netty/netty/security/advisories/GHSA-5mcr-gq6c-3hq2</a></p>
<p>Release Date: 2021-02-08</p>
<p>Fix Resolution (io.netty:netty-handler): 4.1.59.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2020-8908</summary>
### Vulnerable Library - <b>guava-28.2-jre.jar</b></p>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.</p>
<p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/28.2-jre/guava-28.2-jre.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- :x: **guava-28.2-jre.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A temp directory creation vulnerability exists in all versions of Guava, allowing an attacker with access to the machine to potentially access data in a temporary directory created by the Guava API com.google.common.io.Files.createTempDir(). By default, on unix-like systems, the created directory is world-readable (readable by an attacker with access to the system). The method in question has been marked @Deprecated in versions 30.0 and later and should not be used. For Android developers, we recommend choosing a temporary directory API provided by Android, such as context.getCacheDir(). For other Java developers, we recommend migrating to the Java 7 API java.nio.file.Files.createTempDirectory() which explicitly configures permissions of 700, or configuring the Java runtime's java.io.tmpdir system property to point to a location whose permissions are appropriately configured.
<p>Publish Date: 2020-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908>CVE-2020-8908</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908</a></p>
<p>Release Date: 2020-12-10</p>
<p>Fix Resolution (com.google.guava:guava): 30.0-android</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | camel-zookeeper-3.4.3.jar: 9 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>camel-zookeeper-3.4.3.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/28.2-jre/guava-28.2-jre.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-25649](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25649) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | jackson-databind-2.10.0.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2020-36518](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | jackson-databind-2.10.0.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2020-11612](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11612) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | netty-codec-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2021-37136](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37136) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | netty-codec-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2021-37137](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37137) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | netty-codec-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [WS-2020-0408](https://github.com/netty/netty/issues/10362) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.4 | netty-handler-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2022-24823](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | netty-common-4.1.45.Final.jar | Transitive | N/A | ❌ |
| [CVE-2021-21290](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21290) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | netty-handler-4.1.45.Final.jar | Transitive | 3.11.0 | ❌ |
| [CVE-2020-8908](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908) | <img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Low | 3.3 | guava-28.2-jre.jar | Transitive | 3.11.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-25649</summary>
### Vulnerable Library - <b>jackson-databind-2.10.0.jar</b></p>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.0/jackson-databind-2.10.0.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- curator-x-discovery-4.3.0.jar
- :x: **jackson-databind-2.10.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A flaw was found in FasterXML Jackson Databind, where it did not have entity expansion secured properly. This flaw allows vulnerability to XML external entity (XXE) attacks. The highest threat from this vulnerability is data integrity.
<p>Publish Date: 2020-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25649>CVE-2020-25649</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-12-03</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.10.5.1</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-36518</summary>
### Vulnerable Library - <b>jackson-databind-2.10.0.jar</b></p>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.0/jackson-databind-2.10.0.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- curator-x-discovery-4.3.0.jar
- :x: **jackson-databind-2.10.0.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.
Mend Note: After conducting further research, Mend has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.
<p>Publish Date: 2022-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-11</p>
<p>Fix Resolution (com.fasterxml.jackson.core:jackson-databind): 2.12.6.1</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-11612</summary>
### Vulnerable Library - <b>netty-codec-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec/4.1.45.Final/netty-codec-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- netty-handler-4.1.45.Final.jar
- :x: **netty-codec-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The ZlibDecoders in Netty 4.1.x before 4.1.46 allow for unbounded memory allocation while decoding a ZlibEncoded byte stream. An attacker could send a large ZlibEncoded byte stream to the Netty server, forcing the server to allocate all of its free memory to a single decoder.
<p>Publish Date: 2020-04-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11612>CVE-2020-11612</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://netty.io/news/2020/02/28/4-1-46-Final.html">https://netty.io/news/2020/02/28/4-1-46-Final.html</a></p>
<p>Release Date: 2020-04-07</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.46.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-37136</summary>
### Vulnerable Library - <b>netty-codec-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec/4.1.45.Final/netty-codec-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- netty-handler-4.1.45.Final.jar
- :x: **netty-codec-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Bzip2 decompression decoder function doesn't allow setting size restrictions on the decompressed output data (which affects the allocation size used during decompression). All users of Bzip2Decoder are affected. The malicious input can trigger an OOME and so a DoS attack
<p>Publish Date: 2021-10-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37136>CVE-2021-37136</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-grg4-wf29-r9vv">https://github.com/netty/netty/security/advisories/GHSA-grg4-wf29-r9vv</a></p>
<p>Release Date: 2021-10-19</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.68.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-37137</summary>
### Vulnerable Library - <b>netty-codec-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-codec/4.1.45.Final/netty-codec-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- netty-handler-4.1.45.Final.jar
- :x: **netty-codec-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The Snappy frame decoder function doesn't restrict the chunk length which may lead to excessive memory usage. Beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well. This vulnerability can be triggered by supplying malicious input that decompresses to a very big size (via a network stream or a file) or by sending a huge skippable chunk.
<p>Publish Date: 2021-10-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37137>CVE-2021-37137</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-9vjp-v76f-g363">https://github.com/advisories/GHSA-9vjp-v76f-g363</a></p>
<p>Release Date: 2021-10-19</p>
<p>Fix Resolution (io.netty:netty-codec): 4.1.68.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0408</summary>
### Vulnerable Library - <b>netty-handler-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-handler/4.1.45.Final/netty-handler-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- :x: **netty-handler-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue was found in all versions of io.netty:netty-all. Host verification in Netty is disabled by default. This can lead to MITM attack in which an attacker can forge valid SSL/TLS certificates for a different hostname in order to intercept traffic that doesn’t intend for him. This is an issue because the certificate is not matched with the host.
<p>Publish Date: 2020-06-22
<p>URL: <a href=https://github.com/netty/netty/issues/10362>WS-2020-0408</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.4</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/WS-2020-0408">https://nvd.nist.gov/vuln/detail/WS-2020-0408</a></p>
<p>Release Date: 2020-06-22</p>
<p>Fix Resolution (io.netty:netty-handler): 4.1.69.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-24823</summary>
### Vulnerable Library - <b>netty-common-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-common/4.1.45.Final/netty-common-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- netty-handler-4.1.45.Final.jar
- :x: **netty-common-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user.
<p>Publish Date: 2022-05-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823>CVE-2022-24823</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p>
<p>Release Date: 2022-05-06</p>
<p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-21290</summary>
### Vulnerable Library - <b>netty-handler-4.1.45.Final.jar</b></p>
<p>Netty is an asynchronous event-driven network application framework for
rapid development of maintainable high performance protocol servers and
clients.</p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/netty/netty-handler/4.1.45.Final/netty-handler-4.1.45.Final.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- zookeeper-3.5.7.jar
- :x: **netty-handler-4.1.45.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Netty is an open-source, asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers & clients. In Netty before version 4.1.59.Final there is a vulnerability on Unix-like systems involving an insecure temp file. When netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. On unix-like systems, the temporary directory is shared between all user. As such, writing to this directory using APIs that do not explicitly set the file/directory permissions can lead to information disclosure. Of note, this does not impact modern MacOS Operating Systems. The method "File.createTempFile" on unix-like systems creates a random file, but, by default will create this file with the permissions "-rw-r--r--". Thus, if sensitive information is written to this file, other local users can read this information. This is the case in netty's "AbstractDiskHttpData" is vulnerable. This has been fixed in version 4.1.59.Final. As a workaround, one may specify your own "java.io.tmpdir" when you start the JVM or use "DefaultHttpDataFactory.setBaseDir(...)" to set the directory to something that is only readable by the current user.
<p>Publish Date: 2021-02-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21290>CVE-2021-21290</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/netty/netty/security/advisories/GHSA-5mcr-gq6c-3hq2">https://github.com/netty/netty/security/advisories/GHSA-5mcr-gq6c-3hq2</a></p>
<p>Release Date: 2021-02-08</p>
<p>Fix Resolution (io.netty:netty-handler): 4.1.59.Final</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> CVE-2020-8908</summary>
### Vulnerable Library - <b>guava-28.2-jre.jar</b></p>
<p>Guava is a suite of core and expanded libraries that include
utility classes, google's collections, io classes, and much
much more.</p>
<p>Library home page: <a href="https://github.com/google/guava">https://github.com/google/guava</a></p>
<p>Path to dependency file: /Java/Maven/simple-pom-2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/guava/guava/28.2-jre/guava-28.2-jre.jar</p>
<p>
Dependency Hierarchy:
- camel-zookeeper-3.4.3.jar (Root Library)
- :x: **guava-28.2-jre.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/aayant-mend/onboardtraining/commit/4839f6588961e746880b27503fdce27cafb1e42e">4839f6588961e746880b27503fdce27cafb1e42e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A temp directory creation vulnerability exists in all versions of Guava, allowing an attacker with access to the machine to potentially access data in a temporary directory created by the Guava API com.google.common.io.Files.createTempDir(). By default, on unix-like systems, the created directory is world-readable (readable by an attacker with access to the system). The method in question has been marked @Deprecated in versions 30.0 and later and should not be used. For Android developers, we recommend choosing a temporary directory API provided by Android, such as context.getCacheDir(). For other Java developers, we recommend migrating to the Java 7 API java.nio.file.Files.createTempDirectory() which explicitly configures permissions of 700, or configuring the Java runtime's java.io.tmpdir system property to point to a location whose permissions are appropriately configured.
<p>Publish Date: 2020-12-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8908>CVE-2020-8908</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>3.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8908</a></p>
<p>Release Date: 2020-12-10</p>
<p>Fix Resolution (com.google.guava:guava): 30.0-android</p>
<p>Direct dependency fix Resolution (org.apache.camel:camel-zookeeper): 3.11.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_infrastructure | camel zookeeper jar vulnerabilities highest severity is vulnerable library camel zookeeper jar path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository com google guava guava jre guava jre jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high jackson databind jar transitive high jackson databind jar transitive high netty codec final jar transitive high netty codec final jar transitive high netty codec final jar transitive high netty handler final jar transitive medium netty common final jar transitive n a medium netty handler final jar transitive low guava jre jar transitive details cve vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy camel zookeeper jar root library curator x discovery jar x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details a flaw was found in fasterxml jackson databind where it did not have entity expansion secured properly this flaw allows vulnerability to xml external entity xxe attacks the highest threat from this vulnerability is data integrity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution org apache camel camel zookeeper step up your open source security game with mend cve vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy camel zookeeper jar root library curator x discovery jar x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects mend note after conducting further research mend has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution com fasterxml jackson core jackson databind direct dependency fix resolution org apache camel camel zookeeper step up your open source security game with mend cve vulnerable library netty codec final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository io netty netty codec final netty codec final jar dependency hierarchy camel zookeeper jar root library zookeeper jar netty handler final jar x netty codec final jar vulnerable library found in head commit a href found in base branch main vulnerability details the zlibdecoders in netty x before allow for unbounded memory allocation while decoding a zlibencoded byte stream an attacker could send a large zlibencoded byte stream to the netty server forcing the server to allocate all of its free memory to a single decoder publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec final direct dependency fix resolution org apache camel camel zookeeper step up your open source security game with mend cve vulnerable library netty codec final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository io netty netty codec final netty codec final jar dependency hierarchy camel zookeeper jar root library zookeeper jar netty handler final jar x netty codec final jar vulnerable library found in head commit a href found in base branch main vulnerability details the decompression decoder function doesn t allow setting size restrictions on the decompressed output data which affects the allocation size used during decompression all users of are affected the malicious input can trigger an oome and so a dos attack publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec final direct dependency fix resolution org apache camel camel zookeeper step up your open source security game with mend cve vulnerable library netty codec final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository io netty netty codec final netty codec final jar dependency hierarchy camel zookeeper jar root library zookeeper jar netty handler final jar x netty codec final jar vulnerable library found in head commit a href found in base branch main vulnerability details the snappy frame decoder function doesn t restrict the chunk length which may lead to excessive memory usage beside this it also may buffer reserved skippable chunks until the whole chunk was received which may lead to excessive memory usage as well this vulnerability can be triggered by supplying malicious input that decompresses to a very big size via a network stream or a file or by sending a huge skippable chunk publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty codec final direct dependency fix resolution org apache camel camel zookeeper step up your open source security game with mend ws vulnerable library netty handler final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository io netty netty handler final netty handler final jar dependency hierarchy camel zookeeper jar root library zookeeper jar x netty handler final jar vulnerable library found in head commit a href found in base branch main vulnerability details an issue was found in all versions of io netty netty all host verification in netty is disabled by default this can lead to mitm attack in which an attacker can forge valid ssl tls certificates for a different hostname in order to intercept traffic that doesn’t intend for him this is an issue because the certificate is not matched with the host publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty handler final direct dependency fix resolution org apache camel camel zookeeper step up your open source security game with mend cve vulnerable library netty common final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository io netty netty common final netty common final jar dependency hierarchy camel zookeeper jar root library zookeeper jar netty handler final jar x netty common final jar vulnerable library found in head commit a href found in base branch main vulnerability details netty is an open source asynchronous event driven network application framework the package io netty netty codec http prior to version final contains an insufficient fix for cve when netty s multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled this only impacts applications running on java version and lower additionally this vulnerability impacts code running on unix like systems and very old versions of mac osx and windows as they all share the system temporary directory between all users version final contains a patch for this vulnerability as a workaround specify one s own java io tmpdir when starting the jvm or use defaulthttpdatafactory setbasedir to set the directory to something that is only readable by the current user publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all io netty netty common final step up your open source security game with mend cve vulnerable library netty handler final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository io netty netty handler final netty handler final jar dependency hierarchy camel zookeeper jar root library zookeeper jar x netty handler final jar vulnerable library found in head commit a href found in base branch main vulnerability details netty is an open source asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers clients in netty before version final there is a vulnerability on unix like systems involving an insecure temp file when netty s multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled on unix like systems the temporary directory is shared between all user as such writing to this directory using apis that do not explicitly set the file directory permissions can lead to information disclosure of note this does not impact modern macos operating systems the method file createtempfile on unix like systems creates a random file but by default will create this file with the permissions rw r r thus if sensitive information is written to this file other local users can read this information this is the case in netty s abstractdiskhttpdata is vulnerable this has been fixed in version final as a workaround one may specify your own java io tmpdir when you start the jvm or use defaulthttpdatafactory setbasedir to set the directory to something that is only readable by the current user publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty handler final direct dependency fix resolution org apache camel camel zookeeper step up your open source security game with mend cve vulnerable library guava jre jar guava is a suite of core and expanded libraries that include utility classes google s collections io classes and much much more library home page a href path to dependency file java maven simple pom pom xml path to vulnerable library home wss scanner repository com google guava guava jre guava jre jar dependency hierarchy camel zookeeper jar root library x guava jre jar vulnerable library found in head commit a href found in base branch main vulnerability details a temp directory creation vulnerability exists in all versions of guava allowing an attacker with access to the machine to potentially access data in a temporary directory created by the guava api com google common io files createtempdir by default on unix like systems the created directory is world readable readable by an attacker with access to the system the method in question has been marked deprecated in versions and later and should not be used for android developers we recommend choosing a temporary directory api provided by android such as context getcachedir for other java developers we recommend migrating to the java api java nio file files createtempdirectory which explicitly configures permissions of or configuring the java runtime s java io tmpdir system property to point to a location whose permissions are appropriately configured publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com google guava guava android direct dependency fix resolution org apache camel camel zookeeper step up your open source security game with mend | 0 |
698,846 | 23,993,697,395 | IssuesEvent | 2022-09-14 05:11:39 | pombase/canto | https://api.github.com/repos/pombase/canto | closed | warning for exact duplicate annotations | low priority quick workflow improvements FlyBase |
AS a sanity check I would like a warning when I try to create duplicate annotations

| 1.0 | warning for exact duplicate annotations -
AS a sanity check I would like a warning when I try to create duplicate annotations

| non_infrastructure | warning for exact duplicate annotations as a sanity check i would like a warning when i try to create duplicate annotations | 0 |
1,811 | 3,385,346,079 | IssuesEvent | 2015-11-27 10:57:47 | OpenSCAP/scap-security-guide | https://api.github.com/repos/OpenSCAP/scap-security-guide | closed | [RFE] Add ability to specify list of concrete profiles that should be included in the produced benchmark | enhancement Infrastructure | Will describe current situation for ```RHEL/6``` product below (the situation for other products is similar / analogous). As of right now the determination which profiles, available for the specific product, will be included in the produced benchmark is performed by specifying the list of profiles to be included in the:
```<product>/input/guide.xslt``` XSLT transformation.
The current form of this file for ```RHEL/6``` product (listing only issue interesting part of this transformation):
```
<!-- adding profiles here -->
<xsl:if test=" number($withtest) = number(0) ">
<xsl:apply-templates select="document('profiles/test.xml')" />
</xsl:if>
<xsl:apply-templates select="document('profiles/CS2.xml')" />
<xsl:apply-templates select="document('profiles/common.xml')" />
<!-- <xsl:apply-templates select="document('profiles/desktop.xml')" /> -->
<xsl:apply-templates select="document('profiles/server.xml')" />
<!-- <xsl:apply-templates select="document('profiles/ftp.xml')" /> -->
<xsl:apply-templates select="document('profiles/stig-rhel6-server-upstream.xml')" />
<xsl:apply-templates select="document('profiles/usgcb-rhel6-server.xml')" />
<xsl:apply-templates select="document('profiles/rht-ccp.xml')" />
<xsl:apply-templates select="document('profiles/CSCF-RHEL6-MLS.xml')" />
<xsl:apply-templates select="document('profiles/C2S.xml')" />
<xsl:apply-templates select="document('profiles/pci-dss.xml')" />
<xsl:apply-templates select="document('profiles/nist-CL-IL-AL.xml')" />
```
Since during the build process it might be desired the users building the content to be able to specify the final list of profiles to be included in the benchmark, the build process should be modified / enhanced to allow this functionality.
The expectation being issuing a build command like the following:
```make content PROFILES=C2S, desktop,usgcb-rhel6-server```
(IOW calling ```make content``` target with ```PROFILES=```` variable set to list of profile ids, split by comma)
would produce a benchmark containing only these profiles included.
Side note:
* In the default mode (PROFILES = '' or not set) all available profiles for particular product would be included in the benchmark,
* In the case ```PROFILES``` is explicitly set, ```make content``` would include just those profiles into the benchmark.
| 1.0 | [RFE] Add ability to specify list of concrete profiles that should be included in the produced benchmark - Will describe current situation for ```RHEL/6``` product below (the situation for other products is similar / analogous). As of right now the determination which profiles, available for the specific product, will be included in the produced benchmark is performed by specifying the list of profiles to be included in the:
```<product>/input/guide.xslt``` XSLT transformation.
The current form of this file for ```RHEL/6``` product (listing only issue interesting part of this transformation):
```
<!-- adding profiles here -->
<xsl:if test=" number($withtest) = number(0) ">
<xsl:apply-templates select="document('profiles/test.xml')" />
</xsl:if>
<xsl:apply-templates select="document('profiles/CS2.xml')" />
<xsl:apply-templates select="document('profiles/common.xml')" />
<!-- <xsl:apply-templates select="document('profiles/desktop.xml')" /> -->
<xsl:apply-templates select="document('profiles/server.xml')" />
<!-- <xsl:apply-templates select="document('profiles/ftp.xml')" /> -->
<xsl:apply-templates select="document('profiles/stig-rhel6-server-upstream.xml')" />
<xsl:apply-templates select="document('profiles/usgcb-rhel6-server.xml')" />
<xsl:apply-templates select="document('profiles/rht-ccp.xml')" />
<xsl:apply-templates select="document('profiles/CSCF-RHEL6-MLS.xml')" />
<xsl:apply-templates select="document('profiles/C2S.xml')" />
<xsl:apply-templates select="document('profiles/pci-dss.xml')" />
<xsl:apply-templates select="document('profiles/nist-CL-IL-AL.xml')" />
```
Since during the build process it might be desired the users building the content to be able to specify the final list of profiles to be included in the benchmark, the build process should be modified / enhanced to allow this functionality.
The expectation being issuing a build command like the following:
```make content PROFILES=C2S, desktop,usgcb-rhel6-server```
(IOW calling ```make content``` target with ```PROFILES=```` variable set to list of profile ids, split by comma)
would produce a benchmark containing only these profiles included.
Side note:
* In the default mode (PROFILES = '' or not set) all available profiles for particular product would be included in the benchmark,
* In the case ```PROFILES``` is explicitly set, ```make content``` would include just those profiles into the benchmark.
| infrastructure | add ability to specify list of concrete profiles that should be included in the produced benchmark will describe current situation for rhel product below the situation for other products is similar analogous as of right now the determination which profiles available for the specific product will be included in the produced benchmark is performed by specifying the list of profiles to be included in the input guide xslt xslt transformation the current form of this file for rhel product listing only issue interesting part of this transformation since during the build process it might be desired the users building the content to be able to specify the final list of profiles to be included in the benchmark the build process should be modified enhanced to allow this functionality the expectation being issuing a build command like the following make content profiles desktop usgcb server iow calling make content target with profiles variable set to list of profile ids split by comma would produce a benchmark containing only these profiles included side note in the default mode profiles or not set all available profiles for particular product would be included in the benchmark in the case profiles is explicitly set make content would include just those profiles into the benchmark | 1 |
7,896 | 7,125,867,895 | IssuesEvent | 2018-01-20 02:25:32 | glen3b/CyPatScoreboardBot | https://api.github.com/repos/glen3b/CyPatScoreboardBot | opened | Respond to server ratelimits | feature infrastructure | The CCS server sends HTTP 429 responses when you've been sending too many requests - we should be more aggressive with our requests, but honor these when sent by the server. We'll also need to retry the request in that case after our cooldown period. | 1.0 | Respond to server ratelimits - The CCS server sends HTTP 429 responses when you've been sending too many requests - we should be more aggressive with our requests, but honor these when sent by the server. We'll also need to retry the request in that case after our cooldown period. | infrastructure | respond to server ratelimits the ccs server sends http responses when you ve been sending too many requests we should be more aggressive with our requests but honor these when sent by the server we ll also need to retry the request in that case after our cooldown period | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.