Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,185
| 18,955,159,595
|
IssuesEvent
|
2021-11-18 19:20:28
|
googleapis/repo-automation-bots
|
https://api.github.com/repos/googleapis/repo-automation-bots
|
opened
|
unpin probot for owl-bot
|
type: process bot: owl-bot
|
#2912 added a pin on probot dependency to `12.1.0` because `12.1.1` gives me a compilation error.
|
1.0
|
unpin probot for owl-bot - #2912 added a pin on probot dependency to `12.1.0` because `12.1.1` gives me a compilation error.
|
process
|
unpin probot for owl bot added a pin on probot dependency to because gives me a compilation error
| 1
|
4,532
| 2,559,371,772
|
IssuesEvent
|
2015-02-05 00:11:17
|
tvkanters/Dopestreamer
|
https://api.github.com/repos/tvkanters/Dopestreamer
|
opened
|
Add viewer count
|
enhancement low-priority
|
There should be an viewer count in the main stream which sums the Hitbox, Twitch, Vacker live and Vacker restream viewcounts.
|
1.0
|
Add viewer count - There should be an viewer count in the main stream which sums the Hitbox, Twitch, Vacker live and Vacker restream viewcounts.
|
non_process
|
add viewer count there should be an viewer count in the main stream which sums the hitbox twitch vacker live and vacker restream viewcounts
| 0
|
6,765
| 9,888,280,189
|
IssuesEvent
|
2019-06-25 11:09:59
|
linnovate/root
|
https://api.github.com/repos/linnovate/root
|
closed
|
search office :can not click folder and signature
|
2.0.7 Fixed Process bug Search
|
go to offices
open new office
go to search and search for this office
Click on the office, and then click Folder or Signature
result : clicking on a signature or folder does not work
go to offices
open new office:

go to search and search for this office

Click on the office, and then click Folder or Signature

|
1.0
|
search office :can not click folder and signature - go to offices
open new office
go to search and search for this office
Click on the office, and then click Folder or Signature
result : clicking on a signature or folder does not work
go to offices
open new office:

go to search and search for this office

Click on the office, and then click Folder or Signature

|
process
|
search office can not click folder and signature go to offices open new office go to search and search for this office click on the office and then click folder or signature result clicking on a signature or folder does not work go to offices open new office go to search and search for this office click on the office and then click folder or signature
| 1
|
81,509
| 23,479,804,250
|
IssuesEvent
|
2022-08-17 09:33:20
|
godotengine/godot-proposals
|
https://api.github.com/repos/godotengine/godot-proposals
|
reopened
|
Tree-shaking compiler to reduce size and load time of exported games
|
archived topic:buildsystem
|
### Describe the project you are working on
Many small HTML5 and iOS/Android games, where it's crucial to have fast download and load times.
### Describe the problem or limitation you are having in your project
Godot has an ever-growing bundle of amazing functionality, which is hugely helpful for making games. However, most of this functionality is not used, leading to bloated game bundles that take longer for the player to download. For instance, even the simplest HTML5 games made with Godot need 15 MB for the engine, whereas the game logic and assets might only be 2-3 MB.
These simple games can also take 10+ seconds to load on old devices because the device has to load a lot of unneeded functionality into memory. This will likely become an increasingly important issue as more functionality is added to Godot.
### Describe the feature / enhancement and how it helps to overcome the problem or limitation
Add [tree-shaking](https://en.wikipedia.org/wiki/Tree_shaking) functionality to the compiler so only the part of the Godot engine that are actually used by the game are included. This would dramatically reduce size and load times.
### Describe how your proposal will work, with code, pseudo-code, mock-ups, and/or diagrams
This would work similar to [webpack's tree-shaking](https://webpack.js.org/guides/tree-shaking/) in that Godot would identify which functions are being used and remove the rest from the exported bundle.
This is similar to [disabling 3D mentioned in the docs](https://docs.godotengine.org/en/stable/development/compiling/optimizing_for_size.html#disabling-3d), but more refined and with much bigger benefit because it would disable lots of other unused functionality besides 3D. That being said, it might be cool to just do a simple initial version that just automatically detects if 3D is used and disables it otherwise.
Implementing tree-shaking would likely take quite a bit of effort, but would improve the bundle size and load time of every single game exported with Godot (there’s no game that uses all the functionality).
### If this enhancement will not be used often, can it be worked around with a few lines of script?
No.
### Is there a reason why this should be core and not an add-on in the asset library?
AFAIK this would only be possible to do in core.
|
1.0
|
Tree-shaking compiler to reduce size and load time of exported games - ### Describe the project you are working on
Many small HTML5 and iOS/Android games, where it's crucial to have fast download and load times.
### Describe the problem or limitation you are having in your project
Godot has an ever-growing bundle of amazing functionality, which is hugely helpful for making games. However, most of this functionality is not used, leading to bloated game bundles that take longer for the player to download. For instance, even the simplest HTML5 games made with Godot need 15 MB for the engine, whereas the game logic and assets might only be 2-3 MB.
These simple games can also take 10+ seconds to load on old devices because the device has to load a lot of unneeded functionality into memory. This will likely become an increasingly important issue as more functionality is added to Godot.
### Describe the feature / enhancement and how it helps to overcome the problem or limitation
Add [tree-shaking](https://en.wikipedia.org/wiki/Tree_shaking) functionality to the compiler so only the part of the Godot engine that are actually used by the game are included. This would dramatically reduce size and load times.
### Describe how your proposal will work, with code, pseudo-code, mock-ups, and/or diagrams
This would work similar to [webpack's tree-shaking](https://webpack.js.org/guides/tree-shaking/) in that Godot would identify which functions are being used and remove the rest from the exported bundle.
This is similar to [disabling 3D mentioned in the docs](https://docs.godotengine.org/en/stable/development/compiling/optimizing_for_size.html#disabling-3d), but more refined and with much bigger benefit because it would disable lots of other unused functionality besides 3D. That being said, it might be cool to just do a simple initial version that just automatically detects if 3D is used and disables it otherwise.
Implementing tree-shaking would likely take quite a bit of effort, but would improve the bundle size and load time of every single game exported with Godot (there’s no game that uses all the functionality).
### If this enhancement will not be used often, can it be worked around with a few lines of script?
No.
### Is there a reason why this should be core and not an add-on in the asset library?
AFAIK this would only be possible to do in core.
|
non_process
|
tree shaking compiler to reduce size and load time of exported games describe the project you are working on many small and ios android games where it s crucial to have fast download and load times describe the problem or limitation you are having in your project godot has an ever growing bundle of amazing functionality which is hugely helpful for making games however most of this functionality is not used leading to bloated game bundles that take longer for the player to download for instance even the simplest games made with godot need mb for the engine whereas the game logic and assets might only be mb these simple games can also take seconds to load on old devices because the device has to load a lot of unneeded functionality into memory this will likely become an increasingly important issue as more functionality is added to godot describe the feature enhancement and how it helps to overcome the problem or limitation add functionality to the compiler so only the part of the godot engine that are actually used by the game are included this would dramatically reduce size and load times describe how your proposal will work with code pseudo code mock ups and or diagrams this would work similar to in that godot would identify which functions are being used and remove the rest from the exported bundle this is similar to but more refined and with much bigger benefit because it would disable lots of other unused functionality besides that being said it might be cool to just do a simple initial version that just automatically detects if is used and disables it otherwise implementing tree shaking would likely take quite a bit of effort but would improve the bundle size and load time of every single game exported with godot there’s no game that uses all the functionality if this enhancement will not be used often can it be worked around with a few lines of script no is there a reason why this should be core and not an add on in the asset library afaik this would only be possible to do in core
| 0
|
649,766
| 21,320,384,885
|
IssuesEvent
|
2022-04-17 01:25:57
|
WeaponMechanics/MechanicsMain
|
https://api.github.com/repos/WeaponMechanics/MechanicsMain
|
closed
|
Add Vanilla Command Arguments/Validation
|
priority: low will add working on it
|
### Link to code
https://github.com/WeaponMechanics/MechanicsMain/tree/master/MechanicsCore/src/main/java/me/deecaad/core/commands
### Related Issues
_No response_
### Improvements
Lets say you are making a CSGO or Valorant or similar server. You need to give weapons to each member of a team. You cannot do this using `/wm give` since it doesn't accept command arguments like `@a[type=PLAYER,team=blue]`. However, in vanilla MC, commands handle this automatically.
MC also shows validation through colors, making command usage more responsive on the fly. To handle this,
you need an NMS based command api.
Existing:
* https://github.com/JorelAli/CommandAPI
* https://github.com/lucko/commodore
|
1.0
|
Add Vanilla Command Arguments/Validation - ### Link to code
https://github.com/WeaponMechanics/MechanicsMain/tree/master/MechanicsCore/src/main/java/me/deecaad/core/commands
### Related Issues
_No response_
### Improvements
Lets say you are making a CSGO or Valorant or similar server. You need to give weapons to each member of a team. You cannot do this using `/wm give` since it doesn't accept command arguments like `@a[type=PLAYER,team=blue]`. However, in vanilla MC, commands handle this automatically.
MC also shows validation through colors, making command usage more responsive on the fly. To handle this,
you need an NMS based command api.
Existing:
* https://github.com/JorelAli/CommandAPI
* https://github.com/lucko/commodore
|
non_process
|
add vanilla command arguments validation link to code related issues no response improvements lets say you are making a csgo or valorant or similar server you need to give weapons to each member of a team you cannot do this using wm give since it doesn t accept command arguments like a however in vanilla mc commands handle this automatically mc also shows validation through colors making command usage more responsive on the fly to handle this you need an nms based command api existing
| 0
|
296,191
| 25,535,616,349
|
IssuesEvent
|
2022-11-29 11:44:54
|
ToolJet/ToolJet
|
https://api.github.com/repos/ToolJet/ToolJet
|
closed
|
Add data-cy for table column edit options.
|
test cypress
|
### Specify the kind of test
<!--
Provide the kind of test
-->
Cypress E2E
### Describe the test
<!--
Provide a clear description of the test
-->
Add data -cy to help test cases for column edit options.
|
1.0
|
Add data-cy for table column edit options. - ### Specify the kind of test
<!--
Provide the kind of test
-->
Cypress E2E
### Describe the test
<!--
Provide a clear description of the test
-->
Add data -cy to help test cases for column edit options.
|
non_process
|
add data cy for table column edit options specify the kind of test provide the kind of test cypress describe the test provide a clear description of the test add data cy to help test cases for column edit options
| 0
|
7,139
| 10,281,285,005
|
IssuesEvent
|
2019-08-26 08:07:07
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Add the "-overwrite" flag to the "clip raster by mask" GDAL tool
|
Feature Request Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
When atempting to overwrite a previous raster (in this case) using clip raster by mask, it is not over written even when QGIS prompts you to overwrite.
**How to Reproduce**
Raster > Extraction > Clip Raster by Mask layer
Save the file to a directory
Go through the same process, save as same file name, and agree to overwrite the file.
Run clipping process again.
See error message below.
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error -->
**QGIS and OS versions**
QGIS: 3.4.10-Madeira
OS: Ubuntu Xenial
GDAL/OGR: 2.2.2
<!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here -->
**Additional context**
<!-- Add any other context about the problem here. -->
Error Message:
```
Clip Raster by Mask Layer
ERROR 1: Output dataset /filename_clipped.tif exists,
but some command line options were provided indicating a new dataset
should be created. Please delete existing dataset and run again.
```
I think here when running the tool the ```-overwrite``` needs to be specified?
|
1.0
|
Add the "-overwrite" flag to the "clip raster by mask" GDAL tool - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
When atempting to overwrite a previous raster (in this case) using clip raster by mask, it is not over written even when QGIS prompts you to overwrite.
**How to Reproduce**
Raster > Extraction > Clip Raster by Mask layer
Save the file to a directory
Go through the same process, save as same file name, and agree to overwrite the file.
Run clipping process again.
See error message below.
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error -->
**QGIS and OS versions**
QGIS: 3.4.10-Madeira
OS: Ubuntu Xenial
GDAL/OGR: 2.2.2
<!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here -->
**Additional context**
<!-- Add any other context about the problem here. -->
Error Message:
```
Clip Raster by Mask Layer
ERROR 1: Output dataset /filename_clipped.tif exists,
but some command line options were provided indicating a new dataset
should be created. Please delete existing dataset and run again.
```
I think here when running the tool the ```-overwrite``` needs to be specified?
|
process
|
add the overwrite flag to the clip raster by mask gdal tool bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug when atempting to overwrite a previous raster in this case using clip raster by mask it is not over written even when qgis prompts you to overwrite how to reproduce raster extraction clip raster by mask layer save the file to a directory go through the same process save as same file name and agree to overwrite the file run clipping process again see error message below steps sample datasets and qgis project file to reproduce the behavior screencasts or screenshots welcome go to click on scroll down to see error qgis and os versions qgis madeira os ubuntu xenial gdal ogr additional context error message clip raster by mask layer error output dataset filename clipped tif exists but some command line options were provided indicating a new dataset should be created please delete existing dataset and run again i think here when running the tool the overwrite needs to be specified
| 1
|
10,810
| 13,609,288,898
|
IssuesEvent
|
2020-09-23 04:50:19
|
googleapis/java-accessapproval
|
https://api.github.com/repos/googleapis/java-accessapproval
|
closed
|
Dependency Dashboard
|
api: accessapproval type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-accessapproval-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-accessapproval to v1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-accessapproval-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-accessapproval to v1
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any chore deps update dependency com google cloud google cloud accessapproval to chore deps update dependency com google cloud libraries bom to check this box to trigger a request for renovate to run again on this repository
| 1
|
5,870
| 8,691,508,758
|
IssuesEvent
|
2018-12-04 01:40:39
|
knative/serving
|
https://api.github.com/repos/knative/serving
|
closed
|
We should only use knative-releases for public released artifacts
|
area/test-and-release kind/feature kind/process
|
<!--
Pro-tip: You can leave this block commented, and it still works!
Select the appropriate areas for your issue:
/area test-and-release
Classify what kind of issue this is:
/kind feature
/kind process
/assign @adrcunha
-->
## Expected Behavior
We should knative-releases bucket exclusively for public releases (vX.Y.Z releases) to avoid accident cleanups of nightly and other ephemeral releases.
Nightly should be published to knative-nightly bucket.
## Actual Behavior
Nightly is currently published to knative-releases GCS/GCR buckets.
|
1.0
|
We should only use knative-releases for public released artifacts - <!--
Pro-tip: You can leave this block commented, and it still works!
Select the appropriate areas for your issue:
/area test-and-release
Classify what kind of issue this is:
/kind feature
/kind process
/assign @adrcunha
-->
## Expected Behavior
We should knative-releases bucket exclusively for public releases (vX.Y.Z releases) to avoid accident cleanups of nightly and other ephemeral releases.
Nightly should be published to knative-nightly bucket.
## Actual Behavior
Nightly is currently published to knative-releases GCS/GCR buckets.
|
process
|
we should only use knative releases for public released artifacts pro tip you can leave this block commented and it still works select the appropriate areas for your issue area test and release classify what kind of issue this is kind feature kind process assign adrcunha expected behavior we should knative releases bucket exclusively for public releases vx y z releases to avoid accident cleanups of nightly and other ephemeral releases nightly should be published to knative nightly bucket actual behavior nightly is currently published to knative releases gcs gcr buckets
| 1
|
432,333
| 30,278,937,578
|
IssuesEvent
|
2023-07-07 23:12:07
|
houghj16/ShareX
|
https://api.github.com/repos/houghj16/ShareX
|
opened
|
Update the routing for the OneDrive
|
documentation
|
#3
#6
@houghj16
Description is very important
```[tasklist]
### Tasks
- [ ] #7
```
|
1.0
|
Update the routing for the OneDrive - #3
#6
@houghj16
Description is very important
```[tasklist]
### Tasks
- [ ] #7
```
|
non_process
|
update the routing for the onedrive description is very important tasks
| 0
|
173,030
| 14,399,550,475
|
IssuesEvent
|
2020-12-03 11:05:01
|
kubernetes-sigs/external-dns
|
https://api.github.com/repos/kubernetes-sigs/external-dns
|
closed
|
documentation for fast dns changed
|
kind/bug kind/documentation
|
<!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!
-->
**What happened**:
Documentation for dast dns changes as it become lagacy so old links in https://github.com/kubernetes-sigs/external-dns/blob/master/docs/tutorials/akamai-fastdns.md broken
**What you expected to happen**:
Fastdns documentation shoud redirect to correct documentation page
|
1.0
|
documentation for fast dns changed - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks!
-->
**What happened**:
Documentation for dast dns changes as it become lagacy so old links in https://github.com/kubernetes-sigs/external-dns/blob/master/docs/tutorials/akamai-fastdns.md broken
**What you expected to happen**:
Fastdns documentation shoud redirect to correct documentation page
|
non_process
|
documentation for fast dns changed please use this template while reporting a bug and provide as much info as possible not doing so may result in your bug not being addressed in a timely manner thanks what happened documentation for dast dns changes as it become lagacy so old links in broken what you expected to happen fastdns documentation shoud redirect to correct documentation page
| 0
|
302,930
| 26,174,182,533
|
IssuesEvent
|
2023-01-02 07:15:55
|
BoBAdministration/QA-Bug-Reports
|
https://api.github.com/repos/BoBAdministration/QA-Bug-Reports
|
closed
|
Tapping E makes you auto drink
|
Fixed-PendingTesting
|
**Describe the Bug**
When you tap E (or use, whatever keybind you have) at a water source you can infinitely drink.
**To Reproduce**
1. Logged onto a test server
2. Go on any creture
3. Go to any water source
4. Tap the button to drink, not hold
**Expected behavior**
Tapping the use button while trying to drink should make it that you only drink for a second
**Actual behavior**
Tapping the use button at water makes you drink forever
**Screenshots & Video**
Showed Pred on stream
**Branch Version**
Tester and Live
**Additional Information**
I first thought you could drain ponds with the bug but turns out it only takes water until you're full water/sat but you stay in the drinking animation, so not as huge of an issue in that case.
|
1.0
|
Tapping E makes you auto drink - **Describe the Bug**
When you tap E (or use, whatever keybind you have) at a water source you can infinitely drink.
**To Reproduce**
1. Logged onto a test server
2. Go on any creture
3. Go to any water source
4. Tap the button to drink, not hold
**Expected behavior**
Tapping the use button while trying to drink should make it that you only drink for a second
**Actual behavior**
Tapping the use button at water makes you drink forever
**Screenshots & Video**
Showed Pred on stream
**Branch Version**
Tester and Live
**Additional Information**
I first thought you could drain ponds with the bug but turns out it only takes water until you're full water/sat but you stay in the drinking animation, so not as huge of an issue in that case.
|
non_process
|
tapping e makes you auto drink describe the bug when you tap e or use whatever keybind you have at a water source you can infinitely drink to reproduce logged onto a test server go on any creture go to any water source tap the button to drink not hold expected behavior tapping the use button while trying to drink should make it that you only drink for a second actual behavior tapping the use button at water makes you drink forever screenshots video showed pred on stream branch version tester and live additional information i first thought you could drain ponds with the bug but turns out it only takes water until you re full water sat but you stay in the drinking animation so not as huge of an issue in that case
| 0
|
8,378
| 11,525,777,586
|
IssuesEvent
|
2020-02-15 10:55:57
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Make it possible to start a process suspened, and later resume it.
|
api-suggestion area-System.Diagnostics.Process
|
This would make it a lot easier to (for example) attach it to a job.
|
1.0
|
Make it possible to start a process suspened, and later resume it. - This would make it a lot easier to (for example) attach it to a job.
|
process
|
make it possible to start a process suspened and later resume it this would make it a lot easier to for example attach it to a job
| 1
|
80,989
| 15,613,862,545
|
IssuesEvent
|
2021-03-19 17:00:02
|
CliMA/RRTMGP.jl
|
https://api.github.com/repos/CliMA/RRTMGP.jl
|
closed
|
Add code coverage back in
|
code quality
|
It looks like code-coverage is no longer being reported, and we should add this back in.
|
1.0
|
Add code coverage back in - It looks like code-coverage is no longer being reported, and we should add this back in.
|
non_process
|
add code coverage back in it looks like code coverage is no longer being reported and we should add this back in
| 0
|
172,153
| 13,263,670,777
|
IssuesEvent
|
2020-08-21 01:16:58
|
omegaup/omegaup
|
https://api.github.com/repos/omegaup/omegaup
|
closed
|
[FEATURE] Hacer obligatorio el campo de lenguaje cuando creas un concurso
|
UX Task feature-request omegaUp for Contests
|
En https://omegaup.com/contest/new/:
* Por default todos los lenguajes deben de estar seleccionados.
* Si el usuario desmarca todos los lenguajes se le muestra un error.
|
1.0
|
[FEATURE] Hacer obligatorio el campo de lenguaje cuando creas un concurso - En https://omegaup.com/contest/new/:
* Por default todos los lenguajes deben de estar seleccionados.
* Si el usuario desmarca todos los lenguajes se le muestra un error.
|
non_process
|
hacer obligatorio el campo de lenguaje cuando creas un concurso en por default todos los lenguajes deben de estar seleccionados si el usuario desmarca todos los lenguajes se le muestra un error
| 0
|
506,413
| 14,664,483,700
|
IssuesEvent
|
2020-12-29 12:06:56
|
eXpandFramework/eXpand
|
https://api.github.com/repos/eXpandFramework/eXpand
|
closed
|
How can I debug this?
|
Priority Question ❤ Backer
|
From time to time, I found this exception in production environment (web app), any help to debug it?
System.ArgumentException: Nombre de tipo duplicado en un ensamblado.
en System.Reflection.Emit.ModuleBuilder.CheckTypeNameConflict(String strTypeName, Type enclosingType)
en System.Reflection.Emit.AssemblyBuilderData.CheckTypeNameConflict(String strTypeName, TypeBuilder enclosingType)
en System.Reflection.Emit.TypeBuilder.Init(String fullname, TypeAttributes attr, Type parent, Type[] interfaces, ModuleBuilder module, PackingSize iPackingSize, Int32 iTypeSize, TypeBuilder enclosingType)
en System.Reflection.Emit.ModuleBuilder.DefineType(String name, TypeAttributes attr, Type parent)
en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.NewControllerType[T](String id) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 170
en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.RegisterAction[TController,TAction](ApplicationModulesManager applicationModulesManager, String id, Func`2 actionBase) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 147
en Xpand.XAF.Modules.PositionInListview.SwapPositionInListViewService.<>c__DisplayClass4_0.b__0(String actionnId) en D:\a\1\s\src\Modules\PositionInListview\SwapPositionInListViewService.cs:línea 33
en System.Reactive.Linq.ObservableImpl.SelectMany`2.ObservableSelector._.OnNext(TSource value)
WEB
Void CheckTypeNameConflict(System.String, System.Type)
en System.Reflection.Emit.ModuleBuilder.CheckTypeNameConflict(String strTypeName, Type enclosingType)
en System.Reflection.Emit.AssemblyBuilderData.CheckTypeNameConflict(String strTypeName, TypeBuilder enclosingType)
en System.Reflection.Emit.TypeBuilder.Init(String fullname, TypeAttributes attr, Type parent, Type[] interfaces, ModuleBuilder module, PackingSize iPackingSize, Int32 iTypeSize, TypeBuilder enclosingType)
en System.Reflection.Emit.ModuleBuilder.DefineType(String name, TypeAttributes attr, Type parent)
en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.NewControllerType[T](String id) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 170
en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.RegisterAction[TController,TAction](ApplicationModulesManager applicationModulesManager, String id, Func`2 actionBase) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 147
en Xpand.XAF.Modules.PositionInListview.SwapPositionInListViewService.<>c__DisplayClass4_0.b__0(String actionnId) en D:\a\1\s\src\Modules\PositionInListview\SwapPositionInListViewService.cs:línea 33
en System.Reactive.Linq.ObservableImpl.SelectMany`2.ObservableSelector._.OnNext(TSource value)
|
1.0
|
How can I debug this? - From time to time, I found this exception in production environment (web app), any help to debug it?
System.ArgumentException: Nombre de tipo duplicado en un ensamblado.
en System.Reflection.Emit.ModuleBuilder.CheckTypeNameConflict(String strTypeName, Type enclosingType)
en System.Reflection.Emit.AssemblyBuilderData.CheckTypeNameConflict(String strTypeName, TypeBuilder enclosingType)
en System.Reflection.Emit.TypeBuilder.Init(String fullname, TypeAttributes attr, Type parent, Type[] interfaces, ModuleBuilder module, PackingSize iPackingSize, Int32 iTypeSize, TypeBuilder enclosingType)
en System.Reflection.Emit.ModuleBuilder.DefineType(String name, TypeAttributes attr, Type parent)
en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.NewControllerType[T](String id) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 170
en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.RegisterAction[TController,TAction](ApplicationModulesManager applicationModulesManager, String id, Func`2 actionBase) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 147
en Xpand.XAF.Modules.PositionInListview.SwapPositionInListViewService.<>c__DisplayClass4_0.b__0(String actionnId) en D:\a\1\s\src\Modules\PositionInListview\SwapPositionInListViewService.cs:línea 33
en System.Reactive.Linq.ObservableImpl.SelectMany`2.ObservableSelector._.OnNext(TSource value)
WEB
Void CheckTypeNameConflict(System.String, System.Type)
en System.Reflection.Emit.ModuleBuilder.CheckTypeNameConflict(String strTypeName, Type enclosingType)
en System.Reflection.Emit.AssemblyBuilderData.CheckTypeNameConflict(String strTypeName, TypeBuilder enclosingType)
en System.Reflection.Emit.TypeBuilder.Init(String fullname, TypeAttributes attr, Type parent, Type[] interfaces, ModuleBuilder module, PackingSize iPackingSize, Int32 iTypeSize, TypeBuilder enclosingType)
en System.Reflection.Emit.ModuleBuilder.DefineType(String name, TypeAttributes attr, Type parent)
en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.NewControllerType[T](String id) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 170
en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.RegisterAction[TController,TAction](ApplicationModulesManager applicationModulesManager, String id, Func`2 actionBase) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 147
en Xpand.XAF.Modules.PositionInListview.SwapPositionInListViewService.<>c__DisplayClass4_0.b__0(String actionnId) en D:\a\1\s\src\Modules\PositionInListview\SwapPositionInListViewService.cs:línea 33
en System.Reactive.Linq.ObservableImpl.SelectMany`2.ObservableSelector._.OnNext(TSource value)
|
non_process
|
how can i debug this from time to time i found this exception in production environment web app any help to debug it system argumentexception nombre de tipo duplicado en un ensamblado en system reflection emit modulebuilder checktypenameconflict string strtypename type enclosingtype en system reflection emit assemblybuilderdata checktypenameconflict string strtypename typebuilder enclosingtype en system reflection emit typebuilder init string fullname typeattributes attr type parent type interfaces modulebuilder module packingsize ipackingsize itypesize typebuilder enclosingtype en system reflection emit modulebuilder definetype string name typeattributes attr type parent en xpand xaf modules reactive services actions actionsservice newcontrollertype string id en d a s src modules reactive services actions registeraction cs línea en xpand xaf modules reactive services actions actionsservice registeraction applicationmodulesmanager applicationmodulesmanager string id func actionbase en d a s src modules reactive services actions registeraction cs línea en xpand xaf modules positioninlistview swappositioninlistviewservice c b string actionnid en d a s src modules positioninlistview swappositioninlistviewservice cs línea en system reactive linq observableimpl selectmany observableselector onnext tsource value web void checktypenameconflict system string system type en system reflection emit modulebuilder checktypenameconflict string strtypename type enclosingtype en system reflection emit assemblybuilderdata checktypenameconflict string strtypename typebuilder enclosingtype en system reflection emit typebuilder init string fullname typeattributes attr type parent type interfaces modulebuilder module packingsize ipackingsize itypesize typebuilder enclosingtype en system reflection emit modulebuilder definetype string name typeattributes attr type parent en xpand xaf modules reactive services actions actionsservice newcontrollertype string id en d a s src modules reactive services actions registeraction cs línea en xpand xaf modules reactive services actions actionsservice registeraction applicationmodulesmanager applicationmodulesmanager string id func actionbase en d a s src modules reactive services actions registeraction cs línea en xpand xaf modules positioninlistview swappositioninlistviewservice c b string actionnid en d a s src modules positioninlistview swappositioninlistviewservice cs línea en system reactive linq observableimpl selectmany observableselector onnext tsource value
| 0
|
12,550
| 14,976,333,471
|
IssuesEvent
|
2021-01-28 07:53:00
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Audit Logs] "studyVersion" is displayed null for the event in Response module
|
Bug P2 Process: Fixed Response datastore
|
Event:
STUDY_METADATA_RECEIVED

|
1.0
|
[Audit Logs] "studyVersion" is displayed null for the event in Response module - Event:
STUDY_METADATA_RECEIVED

|
process
|
studyversion is displayed null for the event in response module event study metadata received
| 1
|
11,640
| 14,496,618,192
|
IssuesEvent
|
2020-12-11 13:04:05
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
[Bug report]wx:model-value-path=""报错
|
processing
|



可以发现valuepath 这里为true 调用reduce就报错了,wx:model-value-path="[]" 没问题的
另外建议文档-api里 wx:model-value-path的链一下对应指南里的部分
|
1.0
|
[Bug report]wx:model-value-path=""报错 - 


可以发现valuepath 这里为true 调用reduce就报错了,wx:model-value-path="[]" 没问题的
另外建议文档-api里 wx:model-value-path的链一下对应指南里的部分
|
process
|
wx model value path 报错 可以发现valuepath 这里为true 调用reduce就报错了 wx model value path 没问题的 另外建议文档 api里 wx model value path的链一下对应指南里的部分
| 1
|
287,675
| 8,818,181,568
|
IssuesEvent
|
2018-12-31 09:36:41
|
Veil-Project/veil
|
https://api.github.com/repos/Veil-Project/veil
|
opened
|
Daemon Crashes and corrupts wallet while mining.
|
core high priority wallet
|
ok you can get the debug log here https://veil.suprnova.cc/debug.log.2
and the wallet here https://veil.suprnova.cc/wallet.dat.2
just try to start the daemon with this wallet and it won't work
it will go to 100% cpu usage and simply do nothing
the only obvious errors i see in the logs are
2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error
2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error
2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error
2018-12-31T07:18:12Z UpdateTip: new best=f74862d1fbbd29087c606b7d7ea0333045d4015b52ab30cc2c7999c1f3a597cd height=9702 version=0x20000000 log2_work=60.34344 tx=16463 date='2018-12-31T06:34:48Z' progress=0.999463 cache=0.4MiB(2315txo)
2018-12-31T07:18:12Z ERROR: FindTx: txid mismatch
2018-12-31T07:18:12Z ERROR: FindTx: txid mismatch
|
1.0
|
Daemon Crashes and corrupts wallet while mining. - ok you can get the debug log here https://veil.suprnova.cc/debug.log.2
and the wallet here https://veil.suprnova.cc/wallet.dat.2
just try to start the daemon with this wallet and it won't work
it will go to 100% cpu usage and simply do nothing
the only obvious errors i see in the logs are
2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error
2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error
2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error
2018-12-31T07:18:12Z UpdateTip: new best=f74862d1fbbd29087c606b7d7ea0333045d4015b52ab30cc2c7999c1f3a597cd height=9702 version=0x20000000 log2_work=60.34344 tx=16463 date='2018-12-31T06:34:48Z' progress=0.999463 cache=0.4MiB(2315txo)
2018-12-31T07:18:12Z ERROR: FindTx: txid mismatch
2018-12-31T07:18:12Z ERROR: FindTx: txid mismatch
|
non_process
|
daemon crashes and corrupts wallet while mining ok you can get the debug log here and the wallet here just try to start the daemon with this wallet and it won t work it will go to cpu usage and simply do nothing the only obvious errors i see in the logs are error findtx deserialize or i o error readcompactsize size too large iostream error error findtx deserialize or i o error readcompactsize size too large iostream error error findtx deserialize or i o error readcompactsize size too large iostream error updatetip new best height version work tx date progress cache error findtx txid mismatch error findtx txid mismatch
| 0
|
12,273
| 3,061,890,203
|
IssuesEvent
|
2015-08-16 01:12:01
|
oppia/oppia
|
https://api.github.com/repos/oppia/oppia
|
closed
|
Bring the "featured exploration" flow within the site
|
feature: important ref: frontend/editor TODO: design doc
|
```
What steps will reproduce the problem?
1. Create a new exploration and publish it.
2. Click on "Nominate for featured status".
What is the expected output? What do you see instead?
A modal pops up and says "please write to this forum". This seems like too much
hassle and is a bit of a weird flow.
Instead it would be nicer for the nomination to be recorded in the moderator
queue, and an email to be automatically sent to moderators, when someone clicks
the button. There should also be a mechanism for the person who clicked the
button to get a reply from the moderator.
```
Original issue reported on code.google.com by `s...@seanlip.org` on 1 Dec 2014 at 5:02
|
1.0
|
Bring the "featured exploration" flow within the site - ```
What steps will reproduce the problem?
1. Create a new exploration and publish it.
2. Click on "Nominate for featured status".
What is the expected output? What do you see instead?
A modal pops up and says "please write to this forum". This seems like too much
hassle and is a bit of a weird flow.
Instead it would be nicer for the nomination to be recorded in the moderator
queue, and an email to be automatically sent to moderators, when someone clicks
the button. There should also be a mechanism for the person who clicked the
button to get a reply from the moderator.
```
Original issue reported on code.google.com by `s...@seanlip.org` on 1 Dec 2014 at 5:02
|
non_process
|
bring the featured exploration flow within the site what steps will reproduce the problem create a new exploration and publish it click on nominate for featured status what is the expected output what do you see instead a modal pops up and says please write to this forum this seems like too much hassle and is a bit of a weird flow instead it would be nicer for the nomination to be recorded in the moderator queue and an email to be automatically sent to moderators when someone clicks the button there should also be a mechanism for the person who clicked the button to get a reply from the moderator original issue reported on code google com by s seanlip org on dec at
| 0
|
7,370
| 10,512,610,703
|
IssuesEvent
|
2019-09-27 18:21:40
|
openopps/openopps-platform
|
https://api.github.com/repos/openopps/openopps-platform
|
closed
|
Bug: Experience dates display incorrectly on application review page
|
Apply Process Bug State Dept.
|
Environment: Production
Issue: Work experience start and end dates are displaying incorrectly on the application review page
Steps to reproduce:
Jot down work experience start and end dates in USAJOBS
Apply for an internship in Open Opps
Go to the review page of your application and the start and end dates display one month off
Related ticket: 3947
|
1.0
|
Bug: Experience dates display incorrectly on application review page - Environment: Production
Issue: Work experience start and end dates are displaying incorrectly on the application review page
Steps to reproduce:
Jot down work experience start and end dates in USAJOBS
Apply for an internship in Open Opps
Go to the review page of your application and the start and end dates display one month off
Related ticket: 3947
|
process
|
bug experience dates display incorrectly on application review page environment production issue work experience start and end dates are displaying incorrectly on the application review page steps to reproduce jot down work experience start and end dates in usajobs apply for an internship in open opps go to the review page of your application and the start and end dates display one month off related ticket
| 1
|
5,341
| 8,167,601,147
|
IssuesEvent
|
2018-08-26 01:10:17
|
MobileOrg/mobileorg
|
https://api.github.com/repos/MobileOrg/mobileorg
|
opened
|
Fastlane tools
|
development process
|
Use fastlane.tools for build/release to automate more. Releasing to testflight should be fully automated.
* https://fastlane.tools/
|
1.0
|
Fastlane tools - Use fastlane.tools for build/release to automate more. Releasing to testflight should be fully automated.
* https://fastlane.tools/
|
process
|
fastlane tools use fastlane tools for build release to automate more releasing to testflight should be fully automated
| 1
|
5,444
| 8,306,330,490
|
IssuesEvent
|
2018-09-22 17:33:11
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
turn off undo for process buffer
|
process
|
I always get a warning when running loops that print something:
`Warning (undo): Buffer `*julia*' undo info was 12560552 bytes long.
The undo info was discarded because it exceeded `undo-outer-limit'.`
I cannot think of a case where undoing something in the process buffer would be useful. So why not disable undo there completely? This also speeds things up if printing a lot into the REPL.
This can easily be done by `(add-hook 'inferior-ess-mode-hook 'buffer-disable-undo)` in the `.emacs`, but I think, there should be an option/documentation in ESS to do this automatically for process buffers.
|
1.0
|
turn off undo for process buffer - I always get a warning when running loops that print something:
`Warning (undo): Buffer `*julia*' undo info was 12560552 bytes long.
The undo info was discarded because it exceeded `undo-outer-limit'.`
I cannot think of a case where undoing something in the process buffer would be useful. So why not disable undo there completely? This also speeds things up if printing a lot into the REPL.
This can easily be done by `(add-hook 'inferior-ess-mode-hook 'buffer-disable-undo)` in the `.emacs`, but I think, there should be an option/documentation in ESS to do this automatically for process buffers.
|
process
|
turn off undo for process buffer i always get a warning when running loops that print something warning undo buffer julia undo info was bytes long the undo info was discarded because it exceeded undo outer limit i cannot think of a case where undoing something in the process buffer would be useful so why not disable undo there completely this also speeds things up if printing a lot into the repl this can easily be done by add hook inferior ess mode hook buffer disable undo in the emacs but i think there should be an option documentation in ess to do this automatically for process buffers
| 1
|
1,184
| 3,687,011,074
|
IssuesEvent
|
2016-02-25 05:22:41
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Keyscope + branch filtering potential bug
|
bug needs-reproduction P2 preprocess/filtering
|
Related to this discussion I had no the DITA specs list:
https://lists.oasis-open.org/archives/dita-comment/201602/msg00001.html
So in one place of the DITA Map I have:
<topichead navtitle="Test" collection-type="choice">
<ditavalref href="topics/getting_started/install_windows.ditaval">
<ditavalmeta>
<dvrResourcePrefix>windows-</dvrResourcePrefix>
<dvrKeyscopePrefix>windows_scope-</dvrKeyscopePrefix>
</ditavalmeta>
</ditavalref>
<topicref href="topics/install.dita"
keys="topicref_install"
keyscope="scope_install"/>
</topichead>
and in some other place I have a keyref like:
<topicref
keyref="windows_scope-scope_install.topicref_install"/>
From what I understand the keyref is invalid because the "windows_scope-" prefix must only be applied on the @keyscope attribute specified on the topichead in which the ditavalref is placed.
But the DITA OT 2.x processes the keyref as valid and finds the target topic.
As the @keyscope is missing on the topichead, according to the specs the @keyscope on the topichead will become "windows-" so the proper way to have the keyref would be:
<topicref
keyref="windows_scope-.scope_install_select_parallel.topicref_install_select_parallel"/>
which currently does not work using DITA OT 2.x.
What's your opinion on this?
|
1.0
|
Keyscope + branch filtering potential bug - Related to this discussion I had no the DITA specs list:
https://lists.oasis-open.org/archives/dita-comment/201602/msg00001.html
So in one place of the DITA Map I have:
<topichead navtitle="Test" collection-type="choice">
<ditavalref href="topics/getting_started/install_windows.ditaval">
<ditavalmeta>
<dvrResourcePrefix>windows-</dvrResourcePrefix>
<dvrKeyscopePrefix>windows_scope-</dvrKeyscopePrefix>
</ditavalmeta>
</ditavalref>
<topicref href="topics/install.dita"
keys="topicref_install"
keyscope="scope_install"/>
</topichead>
and in some other place I have a keyref like:
<topicref
keyref="windows_scope-scope_install.topicref_install"/>
From what I understand the keyref is invalid because the "windows_scope-" prefix must only be applied on the @keyscope attribute specified on the topichead in which the ditavalref is placed.
But the DITA OT 2.x processes the keyref as valid and finds the target topic.
As the @keyscope is missing on the topichead, according to the specs the @keyscope on the topichead will become "windows-" so the proper way to have the keyref would be:
<topicref
keyref="windows_scope-.scope_install_select_parallel.topicref_install_select_parallel"/>
which currently does not work using DITA OT 2.x.
What's your opinion on this?
|
process
|
keyscope branch filtering potential bug related to this discussion i had no the dita specs list so in one place of the dita map i have windows windows scope topicref href topics install dita keys topicref install keyscope scope install and in some other place i have a keyref like topicref keyref windows scope scope install topicref install from what i understand the keyref is invalid because the windows scope prefix must only be applied on the keyscope attribute specified on the topichead in which the ditavalref is placed but the dita ot x processes the keyref as valid and finds the target topic as the keyscope is missing on the topichead according to the specs the keyscope on the topichead will become windows so the proper way to have the keyref would be topicref keyref windows scope scope install select parallel topicref install select parallel which currently does not work using dita ot x what s your opinion on this
| 1
|
180,557
| 13,937,210,133
|
IssuesEvent
|
2020-10-22 13:53:32
|
root-project/root
|
https://api.github.com/repos/root-project/root
|
closed
|
Unable to install pytest on MacOS with python2
|
bug in:Testing
|
We removed the pytest shipped with roottest because the source code was from 2014 and incompatible with py3.9 (see #6597). However, this poses now the issue on MacOs with python2 that we have to install pytest. Without a virtual environment, MacOS does not allow to pip packages. Since roottest fails on configuration level without pytest, roottest is currently broken in this configuration.
@axel @oshadura What should we do? Our CI always runs roottest against python3, so we currently don't see the issue in our infrastructure. I see three options:
1. Ditch testing of python2 on MacOS and rely on the test coverage of other platforms (python2 is anyway dead)
2. Use a venv overlay in Jenkins for the MacOS nodes (haven't tested but it should work and is binary compatible with the system python)
3. We change the hard failure of roottest to a soft failure.
|
1.0
|
Unable to install pytest on MacOS with python2 - We removed the pytest shipped with roottest because the source code was from 2014 and incompatible with py3.9 (see #6597). However, this poses now the issue on MacOs with python2 that we have to install pytest. Without a virtual environment, MacOS does not allow to pip packages. Since roottest fails on configuration level without pytest, roottest is currently broken in this configuration.
@axel @oshadura What should we do? Our CI always runs roottest against python3, so we currently don't see the issue in our infrastructure. I see three options:
1. Ditch testing of python2 on MacOS and rely on the test coverage of other platforms (python2 is anyway dead)
2. Use a venv overlay in Jenkins for the MacOS nodes (haven't tested but it should work and is binary compatible with the system python)
3. We change the hard failure of roottest to a soft failure.
|
non_process
|
unable to install pytest on macos with we removed the pytest shipped with roottest because the source code was from and incompatible with see however this poses now the issue on macos with that we have to install pytest without a virtual environment macos does not allow to pip packages since roottest fails on configuration level without pytest roottest is currently broken in this configuration axel oshadura what should we do our ci always runs roottest against so we currently don t see the issue in our infrastructure i see three options ditch testing of on macos and rely on the test coverage of other platforms is anyway dead use a venv overlay in jenkins for the macos nodes haven t tested but it should work and is binary compatible with the system python we change the hard failure of roottest to a soft failure
| 0
|
537,762
| 15,736,848,622
|
IssuesEvent
|
2021-03-30 01:34:35
|
musescore/MuseScore
|
https://api.github.com/repos/musescore/MuseScore
|
opened
|
[MU4 Issue] Scores should open on first page even if previous score was closed on a different page
|
Low Priority
|
**Describe the bug**
If user closes score on the second or upper page and then open another score, it will be opened from those page which was opened on a previous closed score
**To Reproduce**
Steps to reproduce the behavior:
1. Create a score with at least 2 pages
2. Close the score on the 2-nd page or higher
3. Open new score with at least 2 pages > score will be opened on the 2-nd page
**Expected behavior**
Scores should open on first page even if previous score was closed on a different page
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Desktop (please complete the following information):**
MacOS
**Additional context**
Add any other context about the problem here.
|
1.0
|
[MU4 Issue] Scores should open on first page even if previous score was closed on a different page - **Describe the bug**
If user closes score on the second or upper page and then open another score, it will be opened from those page which was opened on a previous closed score
**To Reproduce**
Steps to reproduce the behavior:
1. Create a score with at least 2 pages
2. Close the score on the 2-nd page or higher
3. Open new score with at least 2 pages > score will be opened on the 2-nd page
**Expected behavior**
Scores should open on first page even if previous score was closed on a different page
**Screenshots**
If applicable, add screenshots to help explain your problem.
**Desktop (please complete the following information):**
MacOS
**Additional context**
Add any other context about the problem here.
|
non_process
|
scores should open on first page even if previous score was closed on a different page describe the bug if user closes score on the second or upper page and then open another score it will be opened from those page which was opened on a previous closed score to reproduce steps to reproduce the behavior create a score with at least pages close the score on the nd page or higher open new score with at least pages score will be opened on the nd page expected behavior scores should open on first page even if previous score was closed on a different page screenshots if applicable add screenshots to help explain your problem desktop please complete the following information macos additional context add any other context about the problem here
| 0
|
14,681
| 17,797,906,223
|
IssuesEvent
|
2021-09-01 02:01:23
|
Leviatan-Analytics/LA-data-processing
|
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
|
closed
|
Improve ward detection model [5]
|
Data Processing Week 4 Sprint 3
|
Re label images and find ways to improve ward detection model metrics:
- Processing time
- Model accuracy
|
1.0
|
Improve ward detection model [5] - Re label images and find ways to improve ward detection model metrics:
- Processing time
- Model accuracy
|
process
|
improve ward detection model re label images and find ways to improve ward detection model metrics processing time model accuracy
| 1
|
69,026
| 7,122,171,911
|
IssuesEvent
|
2018-01-19 10:50:38
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
rancher-cli install creates empty stack without services
|
area/cli status/resolved status/to-test
|
I've a local rancher installation. I've created an evironment and a API Key for it. I also have a custom catalog. If I now try to create a new stack using the following command line:
``
rancher catalog install customcat/testitem:1.0 --name test
``
the stack is successfully created, but does not contain any services. If I do the same using the GUI I get the expected services in my stack. Using --debug doesn't give any useful additional informations and the logs are empty either.
Am I missing something?
Rancher-CLI Version: 0.6.2 (current version offered as download)
---
| Useful | Info |
| :-- | :-- |
|Versions|Rancher `v1.6.0` Cattle: `v0.179.7` UI: `v1.6.1` Rancher-CLI: `v0.6.2`|
|Access|`ldap` `admin`|
|Orchestration|`Cattle`|
|Route|`stacks.index`|
|
1.0
|
rancher-cli install creates empty stack without services - I've a local rancher installation. I've created an evironment and a API Key for it. I also have a custom catalog. If I now try to create a new stack using the following command line:
``
rancher catalog install customcat/testitem:1.0 --name test
``
the stack is successfully created, but does not contain any services. If I do the same using the GUI I get the expected services in my stack. Using --debug doesn't give any useful additional informations and the logs are empty either.
Am I missing something?
Rancher-CLI Version: 0.6.2 (current version offered as download)
---
| Useful | Info |
| :-- | :-- |
|Versions|Rancher `v1.6.0` Cattle: `v0.179.7` UI: `v1.6.1` Rancher-CLI: `v0.6.2`|
|Access|`ldap` `admin`|
|Orchestration|`Cattle`|
|Route|`stacks.index`|
|
non_process
|
rancher cli install creates empty stack without services i ve a local rancher installation i ve created an evironment and a api key for it i also have a custom catalog if i now try to create a new stack using the following command line rancher catalog install customcat testitem name test the stack is successfully created but does not contain any services if i do the same using the gui i get the expected services in my stack using debug doesn t give any useful additional informations and the logs are empty either am i missing something rancher cli version current version offered as download useful info versions rancher cattle ui rancher cli access ldap admin orchestration cattle route stacks index
| 0
|
9,594
| 12,543,045,137
|
IssuesEvent
|
2020-06-05 14:56:21
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Improve validation message for @@unique
|
kind/improvement process/candidate topic: errors
|
```prisma
// Specify a multi-field unique attribute that includes a relation field
model Post {
id Int @default(autoincrement())
author User @relation(fields: [authorId], references: [id])
authorId Int
title String
published Boolean @default(false)
@@unique([author, title])
}
model User {
id Int @id @default(autoincrement())
email String @unique
posts Post[]
}
```
Current message for this schema
`Error validating model "Post": The unique index definition refers to the relation fields author. Index definitions must reference only scalar fields.`
Here the fix is to replace `@@unique([author, title])` by `@@unique([authorId, title])` so the message could mention it maybe?
Discussion with @do4gr https://prisma-company.slack.com/archives/C5Z9TH6N9/p1591348814019400
|
1.0
|
Improve validation message for @@unique - ```prisma
// Specify a multi-field unique attribute that includes a relation field
model Post {
id Int @default(autoincrement())
author User @relation(fields: [authorId], references: [id])
authorId Int
title String
published Boolean @default(false)
@@unique([author, title])
}
model User {
id Int @id @default(autoincrement())
email String @unique
posts Post[]
}
```
Current message for this schema
`Error validating model "Post": The unique index definition refers to the relation fields author. Index definitions must reference only scalar fields.`
Here the fix is to replace `@@unique([author, title])` by `@@unique([authorId, title])` so the message could mention it maybe?
Discussion with @do4gr https://prisma-company.slack.com/archives/C5Z9TH6N9/p1591348814019400
|
process
|
improve validation message for unique prisma specify a multi field unique attribute that includes a relation field model post id int default autoincrement author user relation fields references authorid int title string published boolean default false unique model user id int id default autoincrement email string unique posts post current message for this schema error validating model post the unique index definition refers to the relation fields author index definitions must reference only scalar fields here the fix is to replace unique by unique so the message could mention it maybe discussion with
| 1
|
89,670
| 18,019,568,097
|
IssuesEvent
|
2021-09-16 17:36:22
|
WordPress/openverse-frontend
|
https://api.github.com/repos/WordPress/openverse-frontend
|
closed
|
[Bug] Managing playback of multiple media files
|
🟧 priority: high 🛠 goal: fix 💻 aspect: code
|
## Description
<!-- Concisely describe the bug. -->
The current setup allows for multiple audio files to be played concurrently, which is a bad user experience.
## Reproduction
<!-- Provide detailed steps to reproduce the bug. -->
1. View any page with multiple audio players
2. Press play on multiple audio players
3. Listen to the resulting 'chaos orchestra'
## Expectation
<!-- Concisely describe what you expected to happen. -->
When pressing 'play' on an audio file, if there is _already_ an active audio file it should be paused.
## Screenshots
<!-- Add screenshots to show the problem; or delete the section entirely. -->
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
I have proposed a solution in #183.
- [ ] 🙋 I would be interested in resolving this bug.
|
1.0
|
[Bug] Managing playback of multiple media files - ## Description
<!-- Concisely describe the bug. -->
The current setup allows for multiple audio files to be played concurrently, which is a bad user experience.
## Reproduction
<!-- Provide detailed steps to reproduce the bug. -->
1. View any page with multiple audio players
2. Press play on multiple audio players
3. Listen to the resulting 'chaos orchestra'
## Expectation
<!-- Concisely describe what you expected to happen. -->
When pressing 'play' on an audio file, if there is _already_ an active audio file it should be paused.
## Screenshots
<!-- Add screenshots to show the problem; or delete the section entirely. -->
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
I have proposed a solution in #183.
- [ ] 🙋 I would be interested in resolving this bug.
|
non_process
|
managing playback of multiple media files description the current setup allows for multiple audio files to be played concurrently which is a bad user experience reproduction view any page with multiple audio players press play on multiple audio players listen to the resulting chaos orchestra expectation when pressing play on an audio file if there is already an active audio file it should be paused screenshots resolution i have proposed a solution in 🙋 i would be interested in resolving this bug
| 0
|
7,806
| 10,960,891,835
|
IssuesEvent
|
2019-11-27 14:25:50
|
codeuniversity/smag-mvp
|
https://api.github.com/repos/codeuniversity/smag-mvp
|
opened
|
Figure out favorite bands of users and find associated photos
|
Image Processing
|
As part of the interests, it would be nice to find musicians the user is interested and get representative pictures of them.
|
1.0
|
Figure out favorite bands of users and find associated photos - As part of the interests, it would be nice to find musicians the user is interested and get representative pictures of them.
|
process
|
figure out favorite bands of users and find associated photos as part of the interests it would be nice to find musicians the user is interested and get representative pictures of them
| 1
|
57,596
| 14,163,858,424
|
IssuesEvent
|
2020-11-12 03:26:55
|
woocommerce/woocommerce-admin
|
https://api.github.com/repos/woocommerce/woocommerce-admin
|
closed
|
e2e Testing: Set up Puppeteer master issue
|
Build [Type] Task [estimate] 13
|
End to end (e2e) testing automates user flows of navigating apps by simulating clicks and selections.
Add Puppeteer infrastructure and tests so that we can check that reports, pages, and filtering are functioning as they should.
Why Puppeteer? Core will eventually migrate in that direction (p7bje6-1ne-p2).
### Tasks
- [x] Implement initial Puppeteer architecture #4343
- [x] Write a simple test to test Puppeteer config #4343
- [x] Integrate [Gutenberg’s WP util functions package](https://github.com/WordPress/gutenberg/tree/master/packages/e2e-test-utils) #4343
- [x] Write documentation on e2e test suite config #4343
- [ ] Create baseline tests for each report and page. Ensure elements are loading correctly.
- [ ] Set up Testing in different browsers
- [x] Integrate with Travis CI #4343
- [ ] Integrate with Slack/email/(something else?) to deliver a notice with screenshot of failed test
- [ ] Identify and create tests for complex flows, ie filtering or settings manipulation.
|
1.0
|
e2e Testing: Set up Puppeteer master issue - End to end (e2e) testing automates user flows of navigating apps by simulating clicks and selections.
Add Puppeteer infrastructure and tests so that we can check that reports, pages, and filtering are functioning as they should.
Why Puppeteer? Core will eventually migrate in that direction (p7bje6-1ne-p2).
### Tasks
- [x] Implement initial Puppeteer architecture #4343
- [x] Write a simple test to test Puppeteer config #4343
- [x] Integrate [Gutenberg’s WP util functions package](https://github.com/WordPress/gutenberg/tree/master/packages/e2e-test-utils) #4343
- [x] Write documentation on e2e test suite config #4343
- [ ] Create baseline tests for each report and page. Ensure elements are loading correctly.
- [ ] Set up Testing in different browsers
- [x] Integrate with Travis CI #4343
- [ ] Integrate with Slack/email/(something else?) to deliver a notice with screenshot of failed test
- [ ] Identify and create tests for complex flows, ie filtering or settings manipulation.
|
non_process
|
testing set up puppeteer master issue end to end testing automates user flows of navigating apps by simulating clicks and selections add puppeteer infrastructure and tests so that we can check that reports pages and filtering are functioning as they should why puppeteer core will eventually migrate in that direction tasks implement initial puppeteer architecture write a simple test to test puppeteer config integrate write documentation on test suite config create baseline tests for each report and page ensure elements are loading correctly set up testing in different browsers integrate with travis ci integrate with slack email something else to deliver a notice with screenshot of failed test identify and create tests for complex flows ie filtering or settings manipulation
| 0
|
182,691
| 21,673,922,027
|
IssuesEvent
|
2022-05-08 12:05:29
|
turkdevops/vscode
|
https://api.github.com/repos/turkdevops/vscode
|
closed
|
WS-2018-0069 (High) detected in is-my-json-valid-2.16.1.tgz - autoclosed
|
security vulnerability
|
## WS-2018-0069 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.16.1.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.1.tgz</a></p>
<p>Path to dependency file: /extensions/emmet/package.json</p>
<p>Path to vulnerable library: /extensions/emmet/node_modules/is-my-json-valid/package.json,/extensions/vscode-api-tests/node_modules/is-my-json-valid/package.json,/extensions/vscode-colorize-tests/node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- vscode-1.1.5.tgz (Root Library)
- gulp-remote-src-0.4.3.tgz
- request-2.79.0.tgz
- har-validator-2.0.6.tgz
- :x: **is-my-json-valid-2.16.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/vscode/commit/9fd6b056a06e14655f4f0b0f631d670b24878828">9fd6b056a06e14655f4f0b0f631d670b24878828</a></p>
<p>Found in base branch: <b>webview-views</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function.
<p>Publish Date: 2018-02-14
<p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976>WS-2018-0069</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/572">https://nodesecurity.io/advisories/572</a></p>
<p>Release Date: 2018-02-14</p>
<p>Fix Resolution (is-my-json-valid): 2.17.2</p>
<p>Direct dependency fix Resolution (vscode): 1.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2018-0069 (High) detected in is-my-json-valid-2.16.1.tgz - autoclosed - ## WS-2018-0069 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.16.1.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.1.tgz</a></p>
<p>Path to dependency file: /extensions/emmet/package.json</p>
<p>Path to vulnerable library: /extensions/emmet/node_modules/is-my-json-valid/package.json,/extensions/vscode-api-tests/node_modules/is-my-json-valid/package.json,/extensions/vscode-colorize-tests/node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- vscode-1.1.5.tgz (Root Library)
- gulp-remote-src-0.4.3.tgz
- request-2.79.0.tgz
- har-validator-2.0.6.tgz
- :x: **is-my-json-valid-2.16.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/vscode/commit/9fd6b056a06e14655f4f0b0f631d670b24878828">9fd6b056a06e14655f4f0b0f631d670b24878828</a></p>
<p>Found in base branch: <b>webview-views</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function.
<p>Publish Date: 2018-02-14
<p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976>WS-2018-0069</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/572">https://nodesecurity.io/advisories/572</a></p>
<p>Release Date: 2018-02-14</p>
<p>Fix Resolution (is-my-json-valid): 2.17.2</p>
<p>Direct dependency fix Resolution (vscode): 1.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws high detected in is my json valid tgz autoclosed ws high severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file extensions emmet package json path to vulnerable library extensions emmet node modules is my json valid package json extensions vscode api tests node modules is my json valid package json extensions vscode colorize tests node modules is my json valid package json dependency hierarchy vscode tgz root library gulp remote src tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch webview views vulnerability details version of is my json valid before or are vulnerable to regular expression denial of service redos via the email validation function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid direct dependency fix resolution vscode step up your open source security game with whitesource
| 0
|
11,100
| 13,129,646,588
|
IssuesEvent
|
2020-08-06 14:12:45
|
sekiguchi-nagisa/ydsh
|
https://api.github.com/repos/sekiguchi-nagisa/ydsh
|
closed
|
prune redundant paths in source globbing
|
Core incompatible change
|
glob in source statement matches redundant paths when having ``..`` patterns as the following
```sh
> echo /*r*/../r*
/proc/../root /proc/../run /run/../root /run/../run /srv/../root /srv/../run /usr/../root /usr/../run /var/../root /var/../run
```
this is correct behavior in posix shell, but in source statement, only require unique match results.
|
True
|
prune redundant paths in source globbing - glob in source statement matches redundant paths when having ``..`` patterns as the following
```sh
> echo /*r*/../r*
/proc/../root /proc/../run /run/../root /run/../run /srv/../root /srv/../run /usr/../root /usr/../run /var/../root /var/../run
```
this is correct behavior in posix shell, but in source statement, only require unique match results.
|
non_process
|
prune redundant paths in source globbing glob in source statement matches redundant paths when having patterns as the following sh echo r r proc root proc run run root run run srv root srv run usr root usr run var root var run this is correct behavior in posix shell but in source statement only require unique match results
| 0
|
1,367
| 3,925,094,232
|
IssuesEvent
|
2016-04-22 17:37:49
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
opened
|
NTR: ‘(positive/negative) regulation of right ventricular blood pressure'
|
BHF-UCL miRNA New term request RNA processes
|
Dear Editors,
I have recently been annotating a paper (PMID:22161164), which relied on the measurement of right systolic ventricular pressure (RSVP) as an indicator of the lung (pulmonary) blood pressure.
Basic scientists as well as clinicians use the measurement of RVBP (e.g. in PMID:22161164) to estimate lung blood pressure. For instance, normal human RVBP and lung blood pressure are usually both close to 20/15 mmHg.
However, although in normal circumstances RVBP and lung blood pressure will give the same numerical value, they are not the same pressures.
RVBP is a function of the volume of the blood in the RV and the resistance of the ventricular walls, therefore as ventricles contract (systole), the pressure inside them will increase. However, until RVBP is sufficiently high to open the valve and allow the blood to start flowing into the pulmonary artery, the pressure in the artery will remain lower than in the RV. Whereas once the stroke of blood has been ejected into the pulmonary artery, the pressure in this artery (and then the lungs) will be higher than in the RV, which will have just emptied.
In view of the above, we have had several group discussions about whether we should request a new term: ‘(positive/negative) regulation of right ventricular blood pressure (RVBP)’ as a child term of ‘regulation of blood pressure’.
It would additionally be helpful to add a comment in QiuckGO to this ‘regulation of RVBP’ term to clarify that scientists and clinicians typically use RVSP as a measurement of the lung/pulmonary blood pressure, as the measurements of these pressures are typically the same. Nonetheless, regulation of RVSP is not synonymous with the regulation of lung/pulmonary blood pressure. The same comment could also be added to the already existing ‘regulation of lung blood pressure’ terms.
Perhaps alternatively (or in addition), ‘regulation of RVBP’ could have a ‘part of’ relationship with ‘regulation of lung blood pressure’, as it is such a major factor influencing the latter?
The blood pressure (positive and negative) regulation terms, which currently exist in GO include:
Id: GO:0008217
Name: regulation of blood pressure
Id: GO:0003073
Name: regulation of systemic arterial blood pressure
Is_a: GO:0008217 regulation of blood pressure
Id: GO:0014916
Name: regulation of lung blood pressure
Is_a: GO:0008217 regulation of blood pressure
Synonym: regulation of pulmonary blood pressure (exact)
(Link to ancestor chart: https://www.ebi.ac.uk/QuickGO/GMultiTerm#a=64%2400m1020P03f4&tab=chart&c=&chartTab=settings)
Unlike the systemic circulation, the pulmonary circulation is a low-pressure system, therefore there are no distinct differences between the right ventricular blood pressure (RVBP) and the pulmonary (lung) pressure in terms of measurements (both are ~20/15 mmHg in human). However, they are not synonymous. And additional factors, which do not affect RVBP (e.g. artery elasticity), will affect the lung blood pressure.
Therefore, although the annotated paper (PMID:22161164) focused on pulmonary hypertension, we decided that ‘regulation of lung blood pressure’ was too far downstream to annotate to this term, based on the presented data and the measurements of RVBP.
Instead, we annotated the entity rno-miR-17-5p to positive regulation of blood pressure (GO:0045777) IMP occurs_in UBERON:0002080 heart right ventricle.
However, having RVBP regulation terms for this type of annotations would have been very helpful.
I will look forward to hearing from you.
Thank you,
Barbara
GOC:BHF_miRNA
GOC:BHF
GOC:bc
@rachhuntley
@RLovering
|
1.0
|
NTR: ‘(positive/negative) regulation of right ventricular blood pressure' - Dear Editors,
I have recently been annotating a paper (PMID:22161164), which relied on the measurement of right systolic ventricular pressure (RSVP) as an indicator of the lung (pulmonary) blood pressure.
Basic scientists as well as clinicians use the measurement of RVBP (e.g. in PMID:22161164) to estimate lung blood pressure. For instance, normal human RVBP and lung blood pressure are usually both close to 20/15 mmHg.
However, although in normal circumstances RVBP and lung blood pressure will give the same numerical value, they are not the same pressures.
RVBP is a function of the volume of the blood in the RV and the resistance of the ventricular walls, therefore as ventricles contract (systole), the pressure inside them will increase. However, until RVBP is sufficiently high to open the valve and allow the blood to start flowing into the pulmonary artery, the pressure in the artery will remain lower than in the RV. Whereas once the stroke of blood has been ejected into the pulmonary artery, the pressure in this artery (and then the lungs) will be higher than in the RV, which will have just emptied.
In view of the above, we have had several group discussions about whether we should request a new term: ‘(positive/negative) regulation of right ventricular blood pressure (RVBP)’ as a child term of ‘regulation of blood pressure’.
It would additionally be helpful to add a comment in QiuckGO to this ‘regulation of RVBP’ term to clarify that scientists and clinicians typically use RVSP as a measurement of the lung/pulmonary blood pressure, as the measurements of these pressures are typically the same. Nonetheless, regulation of RVSP is not synonymous with the regulation of lung/pulmonary blood pressure. The same comment could also be added to the already existing ‘regulation of lung blood pressure’ terms.
Perhaps alternatively (or in addition), ‘regulation of RVBP’ could have a ‘part of’ relationship with ‘regulation of lung blood pressure’, as it is such a major factor influencing the latter?
The blood pressure (positive and negative) regulation terms, which currently exist in GO include:
Id: GO:0008217
Name: regulation of blood pressure
Id: GO:0003073
Name: regulation of systemic arterial blood pressure
Is_a: GO:0008217 regulation of blood pressure
Id: GO:0014916
Name: regulation of lung blood pressure
Is_a: GO:0008217 regulation of blood pressure
Synonym: regulation of pulmonary blood pressure (exact)
(Link to ancestor chart: https://www.ebi.ac.uk/QuickGO/GMultiTerm#a=64%2400m1020P03f4&tab=chart&c=&chartTab=settings)
Unlike the systemic circulation, the pulmonary circulation is a low-pressure system, therefore there are no distinct differences between the right ventricular blood pressure (RVBP) and the pulmonary (lung) pressure in terms of measurements (both are ~20/15 mmHg in human). However, they are not synonymous. And additional factors, which do not affect RVBP (e.g. artery elasticity), will affect the lung blood pressure.
Therefore, although the annotated paper (PMID:22161164) focused on pulmonary hypertension, we decided that ‘regulation of lung blood pressure’ was too far downstream to annotate to this term, based on the presented data and the measurements of RVBP.
Instead, we annotated the entity rno-miR-17-5p to positive regulation of blood pressure (GO:0045777) IMP occurs_in UBERON:0002080 heart right ventricle.
However, having RVBP regulation terms for this type of annotations would have been very helpful.
I will look forward to hearing from you.
Thank you,
Barbara
GOC:BHF_miRNA
GOC:BHF
GOC:bc
@rachhuntley
@RLovering
|
process
|
ntr ‘ positive negative regulation of right ventricular blood pressure dear editors i have recently been annotating a paper pmid which relied on the measurement of right systolic ventricular pressure rsvp as an indicator of the lung pulmonary blood pressure basic scientists as well as clinicians use the measurement of rvbp e g in pmid to estimate lung blood pressure for instance normal human rvbp and lung blood pressure are usually both close to mmhg however although in normal circumstances rvbp and lung blood pressure will give the same numerical value they are not the same pressures rvbp is a function of the volume of the blood in the rv and the resistance of the ventricular walls therefore as ventricles contract systole the pressure inside them will increase however until rvbp is sufficiently high to open the valve and allow the blood to start flowing into the pulmonary artery the pressure in the artery will remain lower than in the rv whereas once the stroke of blood has been ejected into the pulmonary artery the pressure in this artery and then the lungs will be higher than in the rv which will have just emptied in view of the above we have had several group discussions about whether we should request a new term ‘ positive negative regulation of right ventricular blood pressure rvbp ’ as a child term of ‘regulation of blood pressure’ it would additionally be helpful to add a comment in qiuckgo to this ‘regulation of rvbp’ term to clarify that scientists and clinicians typically use rvsp as a measurement of the lung pulmonary blood pressure as the measurements of these pressures are typically the same nonetheless regulation of rvsp is not synonymous with the regulation of lung pulmonary blood pressure the same comment could also be added to the already existing ‘regulation of lung blood pressure’ terms perhaps alternatively or in addition ‘regulation of rvbp’ could have a ‘part of’ relationship with ‘regulation of lung blood pressure’ as it is such a major factor influencing the latter the blood pressure positive and negative regulation terms which currently exist in go include id go name regulation of blood pressure id go name regulation of systemic arterial blood pressure is a go regulation of blood pressure id go name regulation of lung blood pressure is a go regulation of blood pressure synonym regulation of pulmonary blood pressure exact link to ancestor chart unlike the systemic circulation the pulmonary circulation is a low pressure system therefore there are no distinct differences between the right ventricular blood pressure rvbp and the pulmonary lung pressure in terms of measurements both are mmhg in human however they are not synonymous and additional factors which do not affect rvbp e g artery elasticity will affect the lung blood pressure therefore although the annotated paper pmid focused on pulmonary hypertension we decided that ‘regulation of lung blood pressure’ was too far downstream to annotate to this term based on the presented data and the measurements of rvbp instead we annotated the entity rno mir to positive regulation of blood pressure go imp occurs in uberon heart right ventricle however having rvbp regulation terms for this type of annotations would have been very helpful i will look forward to hearing from you thank you barbara goc bhf mirna goc bhf goc bc rachhuntley rlovering
| 1
|
12,065
| 14,739,732,895
|
IssuesEvent
|
2021-01-07 07:49:09
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Cerida Investment Corp. Invoice for account 118-SR0634
|
anc-process anp-important ant-bug has attachment
|
In GitLab by @kdjstudios on Sep 17, 2018, 09:12
**Submitted by:** "Amanda Jennings " <amanda.jennings@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-09-17-21134/conversation
**Server:** internal
**Client/Site:** Sarasota
**Account:** 118-SR0634
**Issue:**
Amanda wrote:
> One of our customers got this message, I am unable to test this portion of SAB to see what I get.
I have tried to pay this invoice on-line (three times!) but your web site is broken:

Please let me know when you get it fixed so I can pay with our company credit card through your web site.
|
1.0
|
Cerida Investment Corp. Invoice for account 118-SR0634 - In GitLab by @kdjstudios on Sep 17, 2018, 09:12
**Submitted by:** "Amanda Jennings " <amanda.jennings@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-09-17-21134/conversation
**Server:** internal
**Client/Site:** Sarasota
**Account:** 118-SR0634
**Issue:**
Amanda wrote:
> One of our customers got this message, I am unable to test this portion of SAB to see what I get.
I have tried to pay this invoice on-line (three times!) but your web site is broken:

Please let me know when you get it fixed so I can pay with our company credit card through your web site.
|
process
|
cerida investment corp invoice for account in gitlab by kdjstudios on sep submitted by amanda jennings helpdesk server internal client site sarasota account issue amanda wrote one of our customers got this message i am unable to test this portion of sab to see what i get i have tried to pay this invoice on line three times but your web site is broken uploads image png please let me know when you get it fixed so i can pay with our company credit card through your web site
| 1
|
387,565
| 26,727,075,465
|
IssuesEvent
|
2023-01-29 20:58:02
|
AntiMicroX/antimicrox
|
https://api.github.com/repos/AntiMicroX/antimicrox
|
closed
|
Add CONTRIBUTING.md file to repository
|
documentation
|
It could be done based on https://github.com/AntiMicroX/antimicrox/wiki/Coding-guidelines
Iw may be very useful to do it as soon as possible because of Hacktoberfest
https://hacktoberfest.com/participation/#maintainers
|
1.0
|
Add CONTRIBUTING.md file to repository - It could be done based on https://github.com/AntiMicroX/antimicrox/wiki/Coding-guidelines
Iw may be very useful to do it as soon as possible because of Hacktoberfest
https://hacktoberfest.com/participation/#maintainers
|
non_process
|
add contributing md file to repository it could be done based on iw may be very useful to do it as soon as possible because of hacktoberfest
| 0
|
10
| 2,496,235,514
|
IssuesEvent
|
2015-01-06 18:01:39
|
vivo-isf/vivo-isf-ontology
|
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
|
closed
|
learning or memory
|
biological_process imported
|
_From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 15:45:16_
GO:0007611
parent: neurological system process
\<a href="http://purl.obolibrary.org/obo/GO_0050877" rel="nofollow">http://purl.obolibrary.org/obo/GO_0050877</a>
_Original issue: http://code.google.com/p/eagle-i/issues/detail?id=160_
|
1.0
|
learning or memory - _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 15:45:16_
GO:0007611
parent: neurological system process
\<a href="http://purl.obolibrary.org/obo/GO_0050877" rel="nofollow">http://purl.obolibrary.org/obo/GO_0050877</a>
_Original issue: http://code.google.com/p/eagle-i/issues/detail?id=160_
|
process
|
learning or memory from on november go parent neurological system process original issue
| 1
|
16,724
| 3,352,844,664
|
IssuesEvent
|
2015-11-18 00:58:22
|
eloipuertas/ES2015A
|
https://api.github.com/repos/eloipuertas/ES2015A
|
closed
|
Placing trees. Part 1
|
design Group D
|
### DESCRIPTION:
Map with trees. Six different models.
### OUTCOME EXPECTED / ACCEPTANCE CRITERIA:
Trees in harmony with the environment. Part (1)
### Estimated time effort: 2h
|
1.0
|
Placing trees. Part 1 - ### DESCRIPTION:
Map with trees. Six different models.
### OUTCOME EXPECTED / ACCEPTANCE CRITERIA:
Trees in harmony with the environment. Part (1)
### Estimated time effort: 2h
|
non_process
|
placing trees part description map with trees six different models outcome expected acceptance criteria trees in harmony with the environment part estimated time effort
| 0
|
16,449
| 21,327,655,217
|
IssuesEvent
|
2022-04-18 02:27:50
|
emily-writes-poems/emily-writes-poems-processing
|
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
|
closed
|
set current feature
|
processing
|
select from list of all features to set one as the current one.
also option to deselect so there is no current feature.
|
1.0
|
set current feature - select from list of all features to set one as the current one.
also option to deselect so there is no current feature.
|
process
|
set current feature select from list of all features to set one as the current one also option to deselect so there is no current feature
| 1
|
289,010
| 24,950,216,732
|
IssuesEvent
|
2022-11-01 06:17:49
|
redhat-developer/vscode-openshift-tools
|
https://api.github.com/repos/redhat-developer/vscode-openshift-tools
|
closed
|
Application Explorer options (buttons) are missing and only "loaded-context" item is present when no kubeconfig is set
|
priority/blocker kind/bug odov3-test
|
In a situation when I do not have any kubeconfig on ~/.kube/config or env. var set. the extension shows only "loaded-context" tree item in application explorer instead of Login, Create component, etc.
|
1.0
|
Application Explorer options (buttons) are missing and only "loaded-context" item is present when no kubeconfig is set - In a situation when I do not have any kubeconfig on ~/.kube/config or env. var set. the extension shows only "loaded-context" tree item in application explorer instead of Login, Create component, etc.
|
non_process
|
application explorer options buttons are missing and only loaded context item is present when no kubeconfig is set in a situation when i do not have any kubeconfig on kube config or env var set the extension shows only loaded context tree item in application explorer instead of login create component etc
| 0
|
292,681
| 22,034,308,335
|
IssuesEvent
|
2022-05-28 10:18:24
|
KorBasilion/OSS-Basic-Project
|
https://api.github.com/repos/KorBasilion/OSS-Basic-Project
|
closed
|
현재의 가계부 삭제 로직의 재구성 예정
|
documentation
|

현재 가계부 삭제방식은 사용자가 직접 추가할 때 처럼 내용을 모두 입력받아 해당내용과 일치한 행에 대한 정보들을 삭제하는 로직으로 구성되어있으나 앞선 이슈에서 언급했듯이 이 방법은 다소 불편할 것으로 예상되어 행 더블클릭시 삭제 창에 해당 행의 정보를 자동으로 입력시킨 뒤 사용자로부터 삭제 명령만 받을 수 있게끔 구현 예정입니다.
관련 내용을 찾아 본 결과
https://blog.naver.com/happyyhj/221849262839
https://stackoverflow.com/questions/61404261/tkinter-selecting-an-item-from-a-treeview-using-single-click-instead-of-double
위 링크에 내용이 있어 참고 차 첨부드립니다.
|
1.0
|
현재의 가계부 삭제 로직의 재구성 예정 -

현재 가계부 삭제방식은 사용자가 직접 추가할 때 처럼 내용을 모두 입력받아 해당내용과 일치한 행에 대한 정보들을 삭제하는 로직으로 구성되어있으나 앞선 이슈에서 언급했듯이 이 방법은 다소 불편할 것으로 예상되어 행 더블클릭시 삭제 창에 해당 행의 정보를 자동으로 입력시킨 뒤 사용자로부터 삭제 명령만 받을 수 있게끔 구현 예정입니다.
관련 내용을 찾아 본 결과
https://blog.naver.com/happyyhj/221849262839
https://stackoverflow.com/questions/61404261/tkinter-selecting-an-item-from-a-treeview-using-single-click-instead-of-double
위 링크에 내용이 있어 참고 차 첨부드립니다.
|
non_process
|
현재의 가계부 삭제 로직의 재구성 예정 현재 가계부 삭제방식은 사용자가 직접 추가할 때 처럼 내용을 모두 입력받아 해당내용과 일치한 행에 대한 정보들을 삭제하는 로직으로 구성되어있으나 앞선 이슈에서 언급했듯이 이 방법은 다소 불편할 것으로 예상되어 행 더블클릭시 삭제 창에 해당 행의 정보를 자동으로 입력시킨 뒤 사용자로부터 삭제 명령만 받을 수 있게끔 구현 예정입니다 관련 내용을 찾아 본 결과 위 링크에 내용이 있어 참고 차 첨부드립니다
| 0
|
243,355
| 20,381,284,490
|
IssuesEvent
|
2022-02-21 22:15:22
|
Qiskit/platypus
|
https://api.github.com/repos/Qiskit/platypus
|
opened
|
Validate Grader Integration Feature with Professors
|
User Testing
|
Success Metric:
- Number of problem sets or exams created
- Number of scores returned
1. Reach out to 3 - 5 Professors
2. Ask to integrate with ongoing course
3. Measure success metrics as described above
|
1.0
|
Validate Grader Integration Feature with Professors - Success Metric:
- Number of problem sets or exams created
- Number of scores returned
1. Reach out to 3 - 5 Professors
2. Ask to integrate with ongoing course
3. Measure success metrics as described above
|
non_process
|
validate grader integration feature with professors success metric number of problem sets or exams created number of scores returned reach out to professors ask to integrate with ongoing course measure success metrics as described above
| 0
|
10,485
| 13,252,927,498
|
IssuesEvent
|
2020-08-20 06:35:02
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
Introduce Arena / Bump Allocators for Expression Evaluation
|
PCP-S1 difficulty/hard sig/coprocessor status/help-wanted
|
## Description
In Coprocessor evaluation framework, each column occupies a `Vec`. Column is the source, the immediate container and the destination of all evaluation. For example, consider expression `a+b+c`, a vector `d` will be created to store results of `a+b` and then a vector `e` will be created to store results of `d+c`. It would be efficient to use a single memory pool to store these vectors so that there can be no allocation cost and deallocation cost, which can improve performance. The memory pool could be an arena allocator or a bump allocator.
## Difficulty
* Hard
## Score
* 3000
## Mentor(s)
- @breeswish
- @sticnarf
- @andylokandy
## Recommended Skills
- Unsafe Rust programming
|
1.0
|
Introduce Arena / Bump Allocators for Expression Evaluation - ## Description
In Coprocessor evaluation framework, each column occupies a `Vec`. Column is the source, the immediate container and the destination of all evaluation. For example, consider expression `a+b+c`, a vector `d` will be created to store results of `a+b` and then a vector `e` will be created to store results of `d+c`. It would be efficient to use a single memory pool to store these vectors so that there can be no allocation cost and deallocation cost, which can improve performance. The memory pool could be an arena allocator or a bump allocator.
## Difficulty
* Hard
## Score
* 3000
## Mentor(s)
- @breeswish
- @sticnarf
- @andylokandy
## Recommended Skills
- Unsafe Rust programming
|
process
|
introduce arena bump allocators for expression evaluation description in coprocessor evaluation framework each column occupies a vec column is the source the immediate container and the destination of all evaluation for example consider expression a b c a vector d will be created to store results of a b and then a vector e will be created to store results of d c it would be efficient to use a single memory pool to store these vectors so that there can be no allocation cost and deallocation cost which can improve performance the memory pool could be an arena allocator or a bump allocator difficulty hard score mentor s breeswish sticnarf andylokandy recommended skills unsafe rust programming
| 1
|
828,467
| 31,830,425,731
|
IssuesEvent
|
2023-09-14 10:12:50
|
MuntashirAkon/AppManager
|
https://api.github.com/repos/MuntashirAkon/AppManager
|
closed
|
WakeLock under-locked
|
Bug Severity: 2 Priority: 1 Status: Accepted
|
### Please check before submitting an issue
- [X] I know what my device, OS and App Manager versions are
- [X] I know how to take logs
- [ ] I know how to reproduce the issue which may not be specific to my device
### Describe the bug
AppManager crashed while performing 1-click ops immediately after restarting the device. (1 time crash)
### To Reproduce
1. Restart the device
2. Immediately open AM > 1-Click Ops > Clear data from uninstall apps & Trim caches in all apps
3. See error
### Expected behavior
_No response_
### Screenshots
_No response_
### Logs
```
java.lang.RuntimeException: WakeLock under-locked AppManager::1-click_ops
at android.os.PowerManager$WakeLock.release(PowerManager.java:2899)
at android.os.PowerManager$WakeLock.release(PowerManager.java:2861)
at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity.clearData(OneClickOpsActivity.java:404)
at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity.$r8$lambda$PAAA15TnXXKPI_Vc49jE5M68oeE(Unknown Source:0)
at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity$$ExternalSyntheticLambda15.onChanged(Unknown Source:4)
at androidx.lifecycle.LiveData.considerNotify(LiveData.java:133)
at androidx.lifecycle.LiveData.dispatchingValue(LiveData.java:151)
at androidx.lifecycle.LiveData.setValue(LiveData.java:309)
at androidx.lifecycle.MutableLiveData.setValue(MutableLiveData.java:50)
at androidx.lifecycle.LiveData$1.run(LiveData.java:93)
at android.os.Handler.handleCallback(Handler.java:978)
at android.os.Handler.dispatchMessage(Handler.java:104)
at android.os.Looper.loopOnce(Looper.java:238)
at android.os.Looper.loop(Looper.java:357)
at android.app.ActivityThread.main(ActivityThread.java:8090)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:548)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1026)
```
### Device info
Device Info:
App version: 4.0.0-alpha01-DEBUG
App version code: 434
Android build version: 8049b-fa5cf
Android release version: 12
Android SDK version: 31
Android build ID: S2RUBS32.51-15-9-7
Device brand: motorola
Device manufacturer: motorola
Device name: corfur
Device model: moto g71 5G
Device product name: corfur_g
Device hardware name: qcom
ABIs: [arm64-v8a, armeabi-v7a, armeabi]
ABIs (32bit): [armeabi-v7a, armeabi]
ABIs (64bit): [arm64-v8a]
System language: en-IN
In-App Language: auto
Mode: root
Inferred Mode: Root
### Additional context
_No response_
|
1.0
|
WakeLock under-locked - ### Please check before submitting an issue
- [X] I know what my device, OS and App Manager versions are
- [X] I know how to take logs
- [ ] I know how to reproduce the issue which may not be specific to my device
### Describe the bug
AppManager crashed while performing 1-click ops immediately after restarting the device. (1 time crash)
### To Reproduce
1. Restart the device
2. Immediately open AM > 1-Click Ops > Clear data from uninstall apps & Trim caches in all apps
3. See error
### Expected behavior
_No response_
### Screenshots
_No response_
### Logs
```
java.lang.RuntimeException: WakeLock under-locked AppManager::1-click_ops
at android.os.PowerManager$WakeLock.release(PowerManager.java:2899)
at android.os.PowerManager$WakeLock.release(PowerManager.java:2861)
at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity.clearData(OneClickOpsActivity.java:404)
at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity.$r8$lambda$PAAA15TnXXKPI_Vc49jE5M68oeE(Unknown Source:0)
at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity$$ExternalSyntheticLambda15.onChanged(Unknown Source:4)
at androidx.lifecycle.LiveData.considerNotify(LiveData.java:133)
at androidx.lifecycle.LiveData.dispatchingValue(LiveData.java:151)
at androidx.lifecycle.LiveData.setValue(LiveData.java:309)
at androidx.lifecycle.MutableLiveData.setValue(MutableLiveData.java:50)
at androidx.lifecycle.LiveData$1.run(LiveData.java:93)
at android.os.Handler.handleCallback(Handler.java:978)
at android.os.Handler.dispatchMessage(Handler.java:104)
at android.os.Looper.loopOnce(Looper.java:238)
at android.os.Looper.loop(Looper.java:357)
at android.app.ActivityThread.main(ActivityThread.java:8090)
at java.lang.reflect.Method.invoke(Native Method)
at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:548)
at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1026)
```
### Device info
Device Info:
App version: 4.0.0-alpha01-DEBUG
App version code: 434
Android build version: 8049b-fa5cf
Android release version: 12
Android SDK version: 31
Android build ID: S2RUBS32.51-15-9-7
Device brand: motorola
Device manufacturer: motorola
Device name: corfur
Device model: moto g71 5G
Device product name: corfur_g
Device hardware name: qcom
ABIs: [arm64-v8a, armeabi-v7a, armeabi]
ABIs (32bit): [armeabi-v7a, armeabi]
ABIs (64bit): [arm64-v8a]
System language: en-IN
In-App Language: auto
Mode: root
Inferred Mode: Root
### Additional context
_No response_
|
non_process
|
wakelock under locked please check before submitting an issue i know what my device os and app manager versions are i know how to take logs i know how to reproduce the issue which may not be specific to my device describe the bug appmanager crashed while performing click ops immediately after restarting the device time crash to reproduce restart the device immediately open am click ops clear data from uninstall apps trim caches in all apps see error expected behavior no response screenshots no response logs java lang runtimeexception wakelock under locked appmanager click ops at android os powermanager wakelock release powermanager java at android os powermanager wakelock release powermanager java at io github muntashirakon appmanager oneclickops oneclickopsactivity cleardata oneclickopsactivity java at io github muntashirakon appmanager oneclickops oneclickopsactivity lambda unknown source at io github muntashirakon appmanager oneclickops oneclickopsactivity onchanged unknown source at androidx lifecycle livedata considernotify livedata java at androidx lifecycle livedata dispatchingvalue livedata java at androidx lifecycle livedata setvalue livedata java at androidx lifecycle mutablelivedata setvalue mutablelivedata java at androidx lifecycle livedata run livedata java at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper looponce looper java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java device info device info app version debug app version code android build version android release version android sdk version android build id device brand motorola device manufacturer motorola device name corfur device model moto device product name corfur g device hardware name qcom abis abis abis system language en in in app language auto mode root inferred mode root additional context no response
| 0
|
23,943
| 2,664,916,681
|
IssuesEvent
|
2015-03-20 17:14:08
|
QuiteRSS/quiterss
|
https://api.github.com/repos/QuiteRSS/quiterss
|
opened
|
Не показывать уведомления в полноэкранном режиме
|
2–5 stars bug imported OpSys-Linux Priority-Low
|
_From [raistlin...@gmail.com](https://code.google.com/u/109057101288473400433/) on January 24, 2013 04:23:53_
Хочется такую опцию, ибо при просмотре фильма это окошко - не к месту.
_Original issue: http://code.google.com/p/quite-rss/issues/detail?id=106_
|
1.0
|
Не показывать уведомления в полноэкранном режиме - _From [raistlin...@gmail.com](https://code.google.com/u/109057101288473400433/) on January 24, 2013 04:23:53_
Хочется такую опцию, ибо при просмотре фильма это окошко - не к месту.
_Original issue: http://code.google.com/p/quite-rss/issues/detail?id=106_
|
non_process
|
не показывать уведомления в полноэкранном режиме from on january хочется такую опцию ибо при просмотре фильма это окошко не к месту original issue
| 0
|
18,891
| 24,833,454,031
|
IssuesEvent
|
2022-10-26 06:48:34
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
改进:在写小程序分包时,编译后页面使用到的组件引用路径优化
|
processing
|
**问题描述**
我司现有原生语法写的微信小程序,近期需要升级新增页面,由于大小限制,我便使用了分包机制,我正在使用mpx写小程序的几个新加页面,并且将其编译文件生成的目录由原来的dist改到了另外一个小程序的分包路径下了,按照mpx的单文件形式进行了编译处理,生成的页面所引用的组件路径都是以 / 开头的(例如:/components/chat5bda63f4/lib/iview/input/index),但实际呢该组件存在的路径是 /我的分包名/components/chat5bda63f4/lib/iview/input/index 这样,我就要手动将分包下生成的components目录复制到 根路径下 或者 将生成的页面的包引用路径 手动更改为合适的位置;这样问题就来了,现有的小程序空间已经无法装下那些生成组件了,必须放到子包路径下,所以,第一种方法不可行,那么第二种方法,我就要不断的修改生成页面的组件引用路径,源码每次修改自动编译后都会重置,这显然是不现实的,希望大佬们可以增加配置项,将生成页面的组件引用路径可以改为相对路径的形式,这样,我在子包里使用框架生成的页面就毫无问题了。
|
1.0
|
改进:在写小程序分包时,编译后页面使用到的组件引用路径优化 - **问题描述**
我司现有原生语法写的微信小程序,近期需要升级新增页面,由于大小限制,我便使用了分包机制,我正在使用mpx写小程序的几个新加页面,并且将其编译文件生成的目录由原来的dist改到了另外一个小程序的分包路径下了,按照mpx的单文件形式进行了编译处理,生成的页面所引用的组件路径都是以 / 开头的(例如:/components/chat5bda63f4/lib/iview/input/index),但实际呢该组件存在的路径是 /我的分包名/components/chat5bda63f4/lib/iview/input/index 这样,我就要手动将分包下生成的components目录复制到 根路径下 或者 将生成的页面的包引用路径 手动更改为合适的位置;这样问题就来了,现有的小程序空间已经无法装下那些生成组件了,必须放到子包路径下,所以,第一种方法不可行,那么第二种方法,我就要不断的修改生成页面的组件引用路径,源码每次修改自动编译后都会重置,这显然是不现实的,希望大佬们可以增加配置项,将生成页面的组件引用路径可以改为相对路径的形式,这样,我在子包里使用框架生成的页面就毫无问题了。
|
process
|
改进:在写小程序分包时,编译后页面使用到的组件引用路径优化 问题描述 我司现有原生语法写的微信小程序,近期需要升级新增页面,由于大小限制,我便使用了分包机制,我正在使用mpx写小程序的几个新加页面,并且将其编译文件生成的目录由原来的dist改到了另外一个小程序的分包路径下了,按照mpx的单文件形式进行了编译处理,生成的页面所引用的组件路径都是以 开头的(例如: components lib iview input index),但实际呢该组件存在的路径是 我的分包名 components lib iview input index 这样,我就要手动将分包下生成的components目录复制到 根路径下 或者 将生成的页面的包引用路径 手动更改为合适的位置;这样问题就来了,现有的小程序空间已经无法装下那些生成组件了,必须放到子包路径下,所以,第一种方法不可行,那么第二种方法,我就要不断的修改生成页面的组件引用路径,源码每次修改自动编译后都会重置,这显然是不现实的,希望大佬们可以增加配置项,将生成页面的组件引用路径可以改为相对路径的形式,这样,我在子包里使用框架生成的页面就毫无问题了。
| 1
|
531,080
| 15,440,189,323
|
IssuesEvent
|
2021-03-08 02:37:41
|
AY2021S2-CS2103-W16-3/tp
|
https://api.github.com/repos/AY2021S2-CS2103-W16-3/tp
|
closed
|
💡 Resize GUI textbox to fit messages
|
priority::3 type::enhancement
|
### Describe the issue
*What is the problem, if any? What is the desired behavior?*
Textbox in HippoCampus GUI currently too small to fit help messages.
It is currently scrollable, but preferably resized to show more.
Or at least allow it to be resized with window.
________________________________________________________________
### Additional context
*Describe possible and/or alternative solutions, to highlight required scope of work.*
|
1.0
|
💡 Resize GUI textbox to fit messages - ### Describe the issue
*What is the problem, if any? What is the desired behavior?*
Textbox in HippoCampus GUI currently too small to fit help messages.
It is currently scrollable, but preferably resized to show more.
Or at least allow it to be resized with window.
________________________________________________________________
### Additional context
*Describe possible and/or alternative solutions, to highlight required scope of work.*
|
non_process
|
💡 resize gui textbox to fit messages describe the issue what is the problem if any what is the desired behavior textbox in hippocampus gui currently too small to fit help messages it is currently scrollable but preferably resized to show more or at least allow it to be resized with window additional context describe possible and or alternative solutions to highlight required scope of work
| 0
|
51,336
| 13,635,095,445
|
IssuesEvent
|
2020-09-25 01:52:47
|
nasifimtiazohi/openmrs-core-2.3.0
|
https://api.github.com/repos/nasifimtiazohi/openmrs-core-2.3.0
|
opened
|
CVE-2019-17571 (High) detected in log4j-1.2.17.jar
|
security vulnerability
|
## CVE-2019-17571 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to dependency file: openmrs-core-2.3.0/web/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-core-2.3.0/commit/51205552c13cbb164119762e0de32dd074f8ba6d">51205552c13cbb164119762e0de32dd074f8ba6d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0-alpha1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-17571 (High) detected in log4j-1.2.17.jar - ## CVE-2019-17571 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to dependency file: openmrs-core-2.3.0/web/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-core-2.3.0/commit/51205552c13cbb164119762e0de32dd074f8ba6d">51205552c13cbb164119762e0de32dd074f8ba6d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0-alpha1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in jar cve high severity vulnerability vulnerable library jar apache path to dependency file openmrs core web pom xml path to vulnerable library home wss scanner repository jar canner repository jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core step up your open source security game with whitesource
| 0
|
15,884
| 20,072,970,399
|
IssuesEvent
|
2022-02-04 09:28:34
|
amor71/LiuAlgoTrader
|
https://api.github.com/repos/amor71/LiuAlgoTrader
|
closed
|
migrate the momentum-scanner to support Alpaca
|
enhancement in-process
|
**Is your feature request related to a problem? Please describe.**
The current `momentum-scanner` is using Polygon, and the older way the platform was loading data. Need to migrate the scanner to support both Polygon and Alpaca
|
1.0
|
migrate the momentum-scanner to support Alpaca - **Is your feature request related to a problem? Please describe.**
The current `momentum-scanner` is using Polygon, and the older way the platform was loading data. Need to migrate the scanner to support both Polygon and Alpaca
|
process
|
migrate the momentum scanner to support alpaca is your feature request related to a problem please describe the current momentum scanner is using polygon and the older way the platform was loading data need to migrate the scanner to support both polygon and alpaca
| 1
|
15,537
| 19,703,299,750
|
IssuesEvent
|
2022-01-12 18:54:31
|
googleapis/python-translate
|
https://api.github.com/repos/googleapis/python-translate
|
opened
|
Your .repo-metadata.json file has a problem 🤒
|
type: process repo-metadata: lint
|
You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'translation' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
1.0
|
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* api_shortname 'translation' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions.
|
process
|
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname translation invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
| 1
|
633,353
| 20,252,751,757
|
IssuesEvent
|
2022-02-14 19:37:19
|
wasmerio/wasmer
|
https://api.github.com/repos/wasmerio/wasmer
|
closed
|
OOB table access with singlepass backend
|
🐞 bug priority-high
|
<!-- Thanks for the bug report! -->
### Describe the bug
singlepass backend somehow emits instructions that can cause unexpected OOB in table access.
llvm and cranelift backend doesn't reproduce this issue. Also this issue has been introduced since 2.2.0-rc1
```console
$ echo "`wasmer -V` | `rustc -V` | `uname -m`"
wasmer 2.2.0-rc1 | rustc 1.56.1 (59eed8a2a 2021-11-01) | x86_64
```
### Steps to reproduce
[VarArgs.swift.wasm.zip](https://github.com/wasmerio/wasmer/files/8011496/VarArgs.swift.wasm.zip)
```console
$ wasmer run --singlepass VarArgs.swift.wasm
```
### Expected behavior
```console
$ wasmer run --singlepass VarArgs.swift.wasm
The answer to life and everything is 42, 42, -42, 3.141593
dig it: 0 0 -1 1 -2 2 -3 3 -4 4 -5 5 -6 6 -7 7 -8 8 -9 9 -10 10 -11 11
pointers: '0x12345670' '0x12345671' '0x12345672' '0x12345673' '0x12345674'
a 1.1 1 2.2 1 3.3 1 a
b 1 1.1 1 2.2 1 3.3 1 b
c 1 2 1 3 1 4444444444444444 1 c
d 1 1 2 1 3 1 4444444444444444 1 d
e 10 20 10 30 10 4040404040404040 10 e
f 10 10 20 10 30 10 4040404040404040 10 f
rdar-32547102: 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 10.0 11.0
a 1.1 1 2.2 1 4.5 1 1.1 a
b 1 1.1 1 2.2 1 4.5 1 1.1 b
done.
```
### Actual behavior
```console
$ wasmer run --singlepass VarArgs.swift.wasm
error: failed to run `VarArgs.swift.wasm`
│ 1: RuntimeError: undefined element: out of bounds table access
╰─▶ 2: table_get_oob
```
### Additional context
This issue was found in SwiftWasm's CI: https://github.com/swiftwasm/swift/pull/4203
|
1.0
|
OOB table access with singlepass backend - <!-- Thanks for the bug report! -->
### Describe the bug
singlepass backend somehow emits instructions that can cause unexpected OOB in table access.
llvm and cranelift backend doesn't reproduce this issue. Also this issue has been introduced since 2.2.0-rc1
```console
$ echo "`wasmer -V` | `rustc -V` | `uname -m`"
wasmer 2.2.0-rc1 | rustc 1.56.1 (59eed8a2a 2021-11-01) | x86_64
```
### Steps to reproduce
[VarArgs.swift.wasm.zip](https://github.com/wasmerio/wasmer/files/8011496/VarArgs.swift.wasm.zip)
```console
$ wasmer run --singlepass VarArgs.swift.wasm
```
### Expected behavior
```console
$ wasmer run --singlepass VarArgs.swift.wasm
The answer to life and everything is 42, 42, -42, 3.141593
dig it: 0 0 -1 1 -2 2 -3 3 -4 4 -5 5 -6 6 -7 7 -8 8 -9 9 -10 10 -11 11
pointers: '0x12345670' '0x12345671' '0x12345672' '0x12345673' '0x12345674'
a 1.1 1 2.2 1 3.3 1 a
b 1 1.1 1 2.2 1 3.3 1 b
c 1 2 1 3 1 4444444444444444 1 c
d 1 1 2 1 3 1 4444444444444444 1 d
e 10 20 10 30 10 4040404040404040 10 e
f 10 10 20 10 30 10 4040404040404040 10 f
rdar-32547102: 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 10.0 11.0
a 1.1 1 2.2 1 4.5 1 1.1 a
b 1 1.1 1 2.2 1 4.5 1 1.1 b
done.
```
### Actual behavior
```console
$ wasmer run --singlepass VarArgs.swift.wasm
error: failed to run `VarArgs.swift.wasm`
│ 1: RuntimeError: undefined element: out of bounds table access
╰─▶ 2: table_get_oob
```
### Additional context
This issue was found in SwiftWasm's CI: https://github.com/swiftwasm/swift/pull/4203
|
non_process
|
oob table access with singlepass backend describe the bug singlepass backend somehow emits instructions that can cause unexpected oob in table access llvm and cranelift backend doesn t reproduce this issue also this issue has been introduced since console echo wasmer v rustc v uname m wasmer rustc steps to reproduce console wasmer run singlepass varargs swift wasm expected behavior console wasmer run singlepass varargs swift wasm the answer to life and everything is dig it pointers a a b b c c d d e e f f rdar a a b b done actual behavior console wasmer run singlepass varargs swift wasm error failed to run varargs swift wasm │ runtimeerror undefined element out of bounds table access ╰─▶ table get oob additional context this issue was found in swiftwasm s ci
| 0
|
19,896
| 26,347,038,309
|
IssuesEvent
|
2023-01-10 23:18:07
|
googleapis/nodejs-compute
|
https://api.github.com/repos/googleapis/nodejs-compute
|
closed
|
test(flakiness): reenable "instances with custom machine type tests" tests
|
priority: p2 type: process api: compute samples
|
The `instances with custom machine type tests` have been turned off due to flakiness.
Rather than fixing integration tests, it would be worth investigating using mocking library (sinon or nock) to test library behavior (_reducing the number of integration tests_).
Refs: #791
|
1.0
|
test(flakiness): reenable "instances with custom machine type tests" tests - The `instances with custom machine type tests` have been turned off due to flakiness.
Rather than fixing integration tests, it would be worth investigating using mocking library (sinon or nock) to test library behavior (_reducing the number of integration tests_).
Refs: #791
|
process
|
test flakiness reenable instances with custom machine type tests tests the instances with custom machine type tests have been turned off due to flakiness rather than fixing integration tests it would be worth investigating using mocking library sinon or nock to test library behavior reducing the number of integration tests refs
| 1
|
14,240
| 10,721,167,915
|
IssuesEvent
|
2019-10-26 23:34:45
|
anikau31/systemc-clang
|
https://api.github.com/repos/anikau31/systemc-clang
|
closed
|
catch2 on mac
|
testing infrastructure
|
Now, there are additional errors being generated from the catch header when compiling on a mac with clang 11.0.0. The prior one was "raise" on line 10630 being undeclared. The new errors after the last push are line 10639 signalstack being undeclared, 10645, 10657, 10660 sigaction having the wrong number of arguments (3, when 1 is required).
|
1.0
|
catch2 on mac - Now, there are additional errors being generated from the catch header when compiling on a mac with clang 11.0.0. The prior one was "raise" on line 10630 being undeclared. The new errors after the last push are line 10639 signalstack being undeclared, 10645, 10657, 10660 sigaction having the wrong number of arguments (3, when 1 is required).
|
non_process
|
on mac now there are additional errors being generated from the catch header when compiling on a mac with clang the prior one was raise on line being undeclared the new errors after the last push are line signalstack being undeclared sigaction having the wrong number of arguments when is required
| 0
|
2,213
| 5,051,502,272
|
IssuesEvent
|
2016-12-20 22:05:12
|
cfpb/design-manual
|
https://api.github.com/repos/cfpb/design-manual
|
closed
|
include links to incomplete issues on the manual site
|
1 - Working process and planning
|
cfpb.github.io/design-manual/ is how most non-employees are viewing our design manual work, but there's so much more work that hasn't been published yet, living in our issues.
I propose two updates:
1. Include links to the related github issues on each page of the design manual site
2. Create pages on the design manual site for incomplete issues. This will make it easier for the public to know what we're working on and join our conversations
|
1.0
|
include links to incomplete issues on the manual site - cfpb.github.io/design-manual/ is how most non-employees are viewing our design manual work, but there's so much more work that hasn't been published yet, living in our issues.
I propose two updates:
1. Include links to the related github issues on each page of the design manual site
2. Create pages on the design manual site for incomplete issues. This will make it easier for the public to know what we're working on and join our conversations
|
process
|
include links to incomplete issues on the manual site cfpb github io design manual is how most non employees are viewing our design manual work but there s so much more work that hasn t been published yet living in our issues i propose two updates include links to the related github issues on each page of the design manual site create pages on the design manual site for incomplete issues this will make it easier for the public to know what we re working on and join our conversations
| 1
|
9,214
| 12,247,259,933
|
IssuesEvent
|
2020-05-05 15:37:42
|
labnote-ant/labnote
|
https://api.github.com/repos/labnote-ant/labnote
|
closed
|
Set ingredient's volume, weight, pressure to null for Mix
|
process-view
|
The ingredients' volume, weight, and pressure in the output of Mix process are set to null.
It will be changed in the major updates.
|
1.0
|
Set ingredient's volume, weight, pressure to null for Mix - The ingredients' volume, weight, and pressure in the output of Mix process are set to null.
It will be changed in the major updates.
|
process
|
set ingredient s volume weight pressure to null for mix the ingredients volume weight and pressure in the output of mix process are set to null it will be changed in the major updates
| 1
|
11,183
| 13,957,696,324
|
IssuesEvent
|
2020-10-24 08:11:49
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
PT: Harvesting
|
Geoportal Harvesting process PT - Portugal
|
Geoportal team,
The harvesting to the Portuguese catalogue is still not running daily... It´s taking at least 1 week for us to check the results.
Can you please start a harvesting do the Portuguese catalogue?
Thank you!
Best regards,
Vanda Marcos
|
1.0
|
PT: Harvesting - Geoportal team,
The harvesting to the Portuguese catalogue is still not running daily... It´s taking at least 1 week for us to check the results.
Can you please start a harvesting do the Portuguese catalogue?
Thank you!
Best regards,
Vanda Marcos
|
process
|
pt harvesting geoportal team the harvesting to the portuguese catalogue is still not running daily it acute s taking at least week for us to check the results can you please start a harvesting do the portuguese catalogue thank you best regards vanda marcos
| 1
|
530,561
| 15,434,031,600
|
IssuesEvent
|
2021-03-07 00:55:42
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
opened
|
[Coverity CID :219573] Untrusted value as argument in samples/net/sockets/coap_client/src/coap-client.c
|
Coverity bug priority: low
|
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/samples/net/sockets/coap_client/src/coap-client.c
Category: Insecure data handling
Function: `process_simple_coap_reply`
Component: Samples
CID: [219573](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219573)
Details:
https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/samples/net/sockets/coap_client/src/coap-client.c#L124
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
|
1.0
|
[Coverity CID :219573] Untrusted value as argument in samples/net/sockets/coap_client/src/coap-client.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/samples/net/sockets/coap_client/src/coap-client.c
Category: Insecure data handling
Function: `process_simple_coap_reply`
Component: Samples
CID: [219573](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219573)
Details:
https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/samples/net/sockets/coap_client/src/coap-client.c#L124
Please fix or provide comments in coverity using the link:
https://scan9.coverity.com/reports.htm#v32951/p12996.
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
|
non_process
|
untrusted value as argument in samples net sockets coap client src coap client c static code scan issues found in file category insecure data handling function process simple coap reply component samples cid details please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
| 0
|
78,211
| 9,670,301,658
|
IssuesEvent
|
2019-05-21 19:33:53
|
apollographql/apollo-client
|
https://api.github.com/repos/apollographql/apollo-client
|
closed
|
Apollo-client 2.5 state management resolvers not firing correctly
|
:wilted_flower: needs-reproduction ✍️ working-as-designed 🏓 waiting-response 🛬 fixed-in-prerelease
|
**Intended outcome:**
I am working towards replacing our app's state management with apollo-client 2.5's local state management (because graphql is awesome).
I have the following query
```
query GetClient {
client(id: "0y9FVRmIiDIIhjr1czVi", live: true) @client {
id
name
queueCount
updated
tickets @client {
id
deleted
comments @client {
id
author_email
public
}
}
}
}
```
Here is what my resolver looks like:
```
Query: {
client: async (parent, { id, live }, context, info) => {
// DO SOME STUFF
var data = {} // DATA is actually returned after some stuff is done
return data;
}
},
Client: {
tickets: async (client, { live }, context, info) => {
// DO SOME STUFF
var data = {} // DATA is actually returned after some stuff is done
return data;
}
},
Ticket: {
comments: async (parent, { live }, context, info) => {
// DO SOME STUFF
var data = {} // DATA is actually returned after some stuff is done
return data;
}
}
```
What I expect for the query to do is to start in the Query --> client section, and then after I return the data from there, to go to the Client --> tickets section of the resolver to run the logic there, and once each ticket is retrieved, it goes down to the Ticket --> comments section.
**Actual outcome:**
The Query --> client section fires, but neither the Client --> tickets nor Ticket --> comments section fire :(. I have tried removing the @clients decorators on the tickets and comments parts of the query, still no luck.
**How to reproduce the issue:**
See above code
**Versions**
System:
OS: macOS 10.14.4
Binaries:
Node: 10.15.3 - ~/.nvm/versions/node/v10.15.3/bin/node
npm: 6.4.1 - ~/.nvm/versions/node/v10.15.3/bin/npm
Browsers:
Chrome: 74.0.3729.157
Firefox: 62.0
Safari: 12.1
npmPackages:
apollo-boost: ^0.3.1 => 0.3.1
apollo-client: ^2.5.1 => 2.5.1
react-apollo: ^2.5.5 => 2.5.5
|
1.0
|
Apollo-client 2.5 state management resolvers not firing correctly - **Intended outcome:**
I am working towards replacing our app's state management with apollo-client 2.5's local state management (because graphql is awesome).
I have the following query
```
query GetClient {
client(id: "0y9FVRmIiDIIhjr1czVi", live: true) @client {
id
name
queueCount
updated
tickets @client {
id
deleted
comments @client {
id
author_email
public
}
}
}
}
```
Here is what my resolver looks like:
```
Query: {
client: async (parent, { id, live }, context, info) => {
// DO SOME STUFF
var data = {} // DATA is actually returned after some stuff is done
return data;
}
},
Client: {
tickets: async (client, { live }, context, info) => {
// DO SOME STUFF
var data = {} // DATA is actually returned after some stuff is done
return data;
}
},
Ticket: {
comments: async (parent, { live }, context, info) => {
// DO SOME STUFF
var data = {} // DATA is actually returned after some stuff is done
return data;
}
}
```
What I expect for the query to do is to start in the Query --> client section, and then after I return the data from there, to go to the Client --> tickets section of the resolver to run the logic there, and once each ticket is retrieved, it goes down to the Ticket --> comments section.
**Actual outcome:**
The Query --> client section fires, but neither the Client --> tickets nor Ticket --> comments section fire :(. I have tried removing the @clients decorators on the tickets and comments parts of the query, still no luck.
**How to reproduce the issue:**
See above code
**Versions**
System:
OS: macOS 10.14.4
Binaries:
Node: 10.15.3 - ~/.nvm/versions/node/v10.15.3/bin/node
npm: 6.4.1 - ~/.nvm/versions/node/v10.15.3/bin/npm
Browsers:
Chrome: 74.0.3729.157
Firefox: 62.0
Safari: 12.1
npmPackages:
apollo-boost: ^0.3.1 => 0.3.1
apollo-client: ^2.5.1 => 2.5.1
react-apollo: ^2.5.5 => 2.5.5
|
non_process
|
apollo client state management resolvers not firing correctly intended outcome i am working towards replacing our app s state management with apollo client s local state management because graphql is awesome i have the following query query getclient client id live true client id name queuecount updated tickets client id deleted comments client id author email public here is what my resolver looks like query client async parent id live context info do some stuff var data data is actually returned after some stuff is done return data client tickets async client live context info do some stuff var data data is actually returned after some stuff is done return data ticket comments async parent live context info do some stuff var data data is actually returned after some stuff is done return data what i expect for the query to do is to start in the query client section and then after i return the data from there to go to the client tickets section of the resolver to run the logic there and once each ticket is retrieved it goes down to the ticket comments section actual outcome the query client section fires but neither the client tickets nor ticket comments section fire i have tried removing the clients decorators on the tickets and comments parts of the query still no luck how to reproduce the issue see above code versions system os macos binaries node nvm versions node bin node npm nvm versions node bin npm browsers chrome firefox safari npmpackages apollo boost apollo client react apollo
| 0
|
60,127
| 6,671,984,739
|
IssuesEvent
|
2017-10-04 09:41:24
|
zalando/zalenium
|
https://api.github.com/repos/zalando/zalenium
|
closed
|
Ability to group videos
|
waiting-retest
|
Is there a way at the moment to group videos by a constant? like for example, the date, or the build server build number?
|
1.0
|
Ability to group videos - Is there a way at the moment to group videos by a constant? like for example, the date, or the build server build number?
|
non_process
|
ability to group videos is there a way at the moment to group videos by a constant like for example the date or the build server build number
| 0
|
12,447
| 14,934,725,065
|
IssuesEvent
|
2021-01-25 10:55:17
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Open study > Enrollment status is incorrectly displayed as 'Not eligible' in participant details page
|
Bug P2 Participant manager datastore Process: Fixed Process: Tested dev
|
**Steps:**
1. Enroll into any open study successfully
2. WIthdraw from the study
3. Fail the eligibility test for the same study
4. Observe the Enrollment status value in participant details page
**Actual:** Enrollment status is incorrect displayed as 'Not eligible' in participant details page
**Expected:** Enrollment status should displayt 'Withdrawn' in participant details page
Note: Issue not observed in Enrollment registry page

|
2.0
|
[PM] Open study > Enrollment status is incorrectly displayed as 'Not eligible' in participant details page - **Steps:**
1. Enroll into any open study successfully
2. WIthdraw from the study
3. Fail the eligibility test for the same study
4. Observe the Enrollment status value in participant details page
**Actual:** Enrollment status is incorrect displayed as 'Not eligible' in participant details page
**Expected:** Enrollment status should displayt 'Withdrawn' in participant details page
Note: Issue not observed in Enrollment registry page

|
process
|
open study enrollment status is incorrectly displayed as not eligible in participant details page steps enroll into any open study successfully withdraw from the study fail the eligibility test for the same study observe the enrollment status value in participant details page actual enrollment status is incorrect displayed as not eligible in participant details page expected enrollment status should displayt withdrawn in participant details page note issue not observed in enrollment registry page
| 1
|
6,184
| 9,101,753,499
|
IssuesEvent
|
2019-02-20 11:58:45
|
FACK1/ReservationSystem
|
https://api.github.com/repos/FACK1/ReservationSystem
|
opened
|
Get the events of the requested service ( front-end )
|
inProcess technical
|
* When an `service card` in the services page is clicked:
- [ ] Make a request from the front side that should return the `service details` in the bookevent page, and details view page
|
1.0
|
Get the events of the requested service ( front-end ) - * When an `service card` in the services page is clicked:
- [ ] Make a request from the front side that should return the `service details` in the bookevent page, and details view page
|
process
|
get the events of the requested service front end when an service card in the services page is clicked make a request from the front side that should return the service details in the bookevent page and details view page
| 1
|
16,111
| 11,840,366,991
|
IssuesEvent
|
2020-03-23 18:44:27
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
opened
|
Set minimum CMake version to 3.14.5
|
area-Infrastructure
|
See discussion in https://github.com/dotnet/runtime/pull/33959/commits/16e1d309221e70d4f3d484e8c8a4fb2a4347f83c#r396389297
The documented minimum CMake version is 3.15.5 already but a few places still check for 3.14:
```bash
$ git grep "_minimum_required\|3\.14\.2" # search for _minimum_required OR 3.14.2
docs/workflow/requirements/linux-requirements.md:Building the repo requires CMake 3.14.2 or newer on Linux. Add Kitware's APT feed to your configuration for a newer version of CMake. See their instructions at <https://apt.kitware.com/>. Also, add LLVM/s APT feed to your configuration for a newer version of CMake. See their instructions as <http://apt.llvm.org/>.
eng/native/build-commons.sh: if [[ "$(version "$cmake_version")" -lt "$(version 3.14.2)" ]]; then
eng/native/build-commons.sh: echo "Please install CMake 3.14.2 or newer from http://www.cmake.org/download/ or https://apt.kitware.com and ensure it is on your path."; exit 1;
global.json: "cmake": "3.14.2",
src/coreclr/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/coreclr/src/pal/tests/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/coreclr/tests/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/installer/corehost/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/installer/corehost/cli/test/mockhostfxr/CMakeLists.txt:cmake_minimum_required (VERSION 2.6)
src/libraries/Native/Unix/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/libraries/Native/Windows/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/libraries/Native/Windows/clrcompression/zlib/CMakeLists.txt:cmake_minimum_required(VERSION 2.4.4)
src/mono/CMakeLists.txt:cmake_minimum_required (VERSION 3.10)
src/mono/mono/btls/CMakeLists.txt:cmake_minimum_required (VERSION 2.8.10)
src/mono/netcore/sample/iOS/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.5)
```
|
1.0
|
Set minimum CMake version to 3.14.5 - See discussion in https://github.com/dotnet/runtime/pull/33959/commits/16e1d309221e70d4f3d484e8c8a4fb2a4347f83c#r396389297
The documented minimum CMake version is 3.15.5 already but a few places still check for 3.14:
```bash
$ git grep "_minimum_required\|3\.14\.2" # search for _minimum_required OR 3.14.2
docs/workflow/requirements/linux-requirements.md:Building the repo requires CMake 3.14.2 or newer on Linux. Add Kitware's APT feed to your configuration for a newer version of CMake. See their instructions at <https://apt.kitware.com/>. Also, add LLVM/s APT feed to your configuration for a newer version of CMake. See their instructions as <http://apt.llvm.org/>.
eng/native/build-commons.sh: if [[ "$(version "$cmake_version")" -lt "$(version 3.14.2)" ]]; then
eng/native/build-commons.sh: echo "Please install CMake 3.14.2 or newer from http://www.cmake.org/download/ or https://apt.kitware.com and ensure it is on your path."; exit 1;
global.json: "cmake": "3.14.2",
src/coreclr/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/coreclr/src/pal/tests/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/coreclr/tests/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/installer/corehost/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/installer/corehost/cli/test/mockhostfxr/CMakeLists.txt:cmake_minimum_required (VERSION 2.6)
src/libraries/Native/Unix/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/libraries/Native/Windows/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2)
src/libraries/Native/Windows/clrcompression/zlib/CMakeLists.txt:cmake_minimum_required(VERSION 2.4.4)
src/mono/CMakeLists.txt:cmake_minimum_required (VERSION 3.10)
src/mono/mono/btls/CMakeLists.txt:cmake_minimum_required (VERSION 2.8.10)
src/mono/netcore/sample/iOS/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.5)
```
|
non_process
|
set minimum cmake version to see discussion in the documented minimum cmake version is already but a few places still check for bash git grep minimum required search for minimum required or docs workflow requirements linux requirements md building the repo requires cmake or newer on linux add kitware s apt feed to your configuration for a newer version of cmake see their instructions at also add llvm s apt feed to your configuration for a newer version of cmake see their instructions as eng native build commons sh if then eng native build commons sh echo please install cmake or newer from or and ensure it is on your path exit global json cmake src coreclr cmakelists txt cmake minimum required version src coreclr src pal tests cmakelists txt cmake minimum required version src coreclr tests cmakelists txt cmake minimum required version src installer corehost cmakelists txt cmake minimum required version src installer corehost cli test mockhostfxr cmakelists txt cmake minimum required version src libraries native unix cmakelists txt cmake minimum required version src libraries native windows cmakelists txt cmake minimum required version src libraries native windows clrcompression zlib cmakelists txt cmake minimum required version src mono cmakelists txt cmake minimum required version src mono mono btls cmakelists txt cmake minimum required version src mono netcore sample ios cmakelists txt cmake minimum required version
| 0
|
799
| 3,276,146,500
|
IssuesEvent
|
2015-10-26 18:13:36
|
hammerlab/pileup.js
|
https://api.github.com/repos/hammerlab/pileup.js
|
closed
|
Code coverage doesn't match source
|
bug process
|
See, for example, coverage for [`RemoteFile.js`][2]. The lines covered don't make logical sense. This was broken by my recent [switch to babelify][1].
The issue is that the coverage is being generated for the babel-ified source, which doesn't necessarily match up 1-1 line for line with the original source.
[1]: https://github.com/hammerlab/pileup.js/pull/316
[2]: https://coveralls.io/builds/3932994/source?filename=src%2Fmain%2FRemoteFile.js
|
1.0
|
Code coverage doesn't match source - See, for example, coverage for [`RemoteFile.js`][2]. The lines covered don't make logical sense. This was broken by my recent [switch to babelify][1].
The issue is that the coverage is being generated for the babel-ified source, which doesn't necessarily match up 1-1 line for line with the original source.
[1]: https://github.com/hammerlab/pileup.js/pull/316
[2]: https://coveralls.io/builds/3932994/source?filename=src%2Fmain%2FRemoteFile.js
|
process
|
code coverage doesn t match source see for example coverage for the lines covered don t make logical sense this was broken by my recent the issue is that the coverage is being generated for the babel ified source which doesn t necessarily match up line for line with the original source
| 1
|
19,682
| 26,033,039,027
|
IssuesEvent
|
2022-12-22 00:04:28
|
nkdAgility/azure-devops-migration-tools
|
https://api.github.com/repos/nkdAgility/azure-devops-migration-tools
|
closed
|
Edited discussion comments can be lost when migrating work items
|
question no-issue-activity Refactor WorkItem Processor
|
## Describe your issue:
When migrating work items between Azure DevOps Services instances using "ReplayRevisions": true there is a situation where a work item discussion comment can be lost.
To replicate:
- Add a discussion comment
- Add another discussion comment
- Edit the first discussion comment
When the work item is migrated the edited comment will not exist in the target.
## Source Details
- **Source Version:** Azure DevOps Services
- **Source Language:** Eng
## Target Details
- **Target Version:** Azure DevOps Services
- **Target Language:** Eng
|
1.0
|
Edited discussion comments can be lost when migrating work items - ## Describe your issue:
When migrating work items between Azure DevOps Services instances using "ReplayRevisions": true there is a situation where a work item discussion comment can be lost.
To replicate:
- Add a discussion comment
- Add another discussion comment
- Edit the first discussion comment
When the work item is migrated the edited comment will not exist in the target.
## Source Details
- **Source Version:** Azure DevOps Services
- **Source Language:** Eng
## Target Details
- **Target Version:** Azure DevOps Services
- **Target Language:** Eng
|
process
|
edited discussion comments can be lost when migrating work items describe your issue when migrating work items between azure devops services instances using replayrevisions true there is a situation where a work item discussion comment can be lost to replicate add a discussion comment add another discussion comment edit the first discussion comment when the work item is migrated the edited comment will not exist in the target source details source version azure devops services source language eng target details target version azure devops services target language eng
| 1
|
10,050
| 13,044,161,659
|
IssuesEvent
|
2020-07-29 03:47:25
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `SubDateDatetimeString` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `SubDateDatetimeString` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `SubDateDatetimeString` from TiDB -
## Description
Port the scalar function `SubDateDatetimeString` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function subdatedatetimestring from tidb description port the scalar function subdatedatetimestring from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
193,326
| 6,884,032,685
|
IssuesEvent
|
2017-11-21 11:30:15
|
EyeSeeTea/dhis2-android-dashboard
|
https://api.github.com/repos/EyeSeeTea/dhis2-android-dashboard
|
closed
|
On login, toast are crazy
|
priority - critical
|
- [ ] When no connexion, it doesn't say a thing
- [ ] When URL is introduced wrong, it says you have no connexion
|
1.0
|
On login, toast are crazy - - [ ] When no connexion, it doesn't say a thing
- [ ] When URL is introduced wrong, it says you have no connexion
|
non_process
|
on login toast are crazy when no connexion it doesn t say a thing when url is introduced wrong it says you have no connexion
| 0
|
38,086
| 8,408,655,828
|
IssuesEvent
|
2018-10-12 02:55:47
|
fga-eps-mds/2018.2-NaturalSearch
|
https://api.github.com/repos/fga-eps-mds/2018.2-NaturalSearch
|
closed
|
US10-Eu, como desenvolvedor, desejo remover informações desnecessárias de proponentes que serão utilizadas.
|
Code User Story eps mds
|
Critérios de aceitação:
- [ ] tratar os dados de proponentes para que sejam obtidos apenas os valores desejados.
- [ ] visualizar os dados tratados.
|
1.0
|
US10-Eu, como desenvolvedor, desejo remover informações desnecessárias de proponentes que serão utilizadas. - Critérios de aceitação:
- [ ] tratar os dados de proponentes para que sejam obtidos apenas os valores desejados.
- [ ] visualizar os dados tratados.
|
non_process
|
eu como desenvolvedor desejo remover informações desnecessárias de proponentes que serão utilizadas critérios de aceitação tratar os dados de proponentes para que sejam obtidos apenas os valores desejados visualizar os dados tratados
| 0
|
216,815
| 16,820,001,766
|
IssuesEvent
|
2021-06-17 12:02:12
|
ARMmbed/mbed-os
|
https://api.github.com/repos/ARMmbed/mbed-os
|
opened
|
Findout a place to keep mbed-os/UNITTESTS/target_h
|
UNITTESTS component: core component: test
|
<!--
************************************** WARNING **************************************
The ciarcom bot parses this header automatically. Any deviation from the
template may cause the bot to automatically correct this header or may result in a
warning message, requesting updates.
Please ensure all sections of the template below are filled in and no changes
are made to the template format. Only bugs should be raised here as issues.
Questions or enhancements should instead be raised on our forums:
https://forums.mbed.com/ .
*************************************************************************************
-->
### Description of defect
Findout a place to keep mbed-os/UNITTESTS/target_h
<!--
Add detailed description of what you are reporting.
Good example: https://os.mbed.com/docs/mbed-os/latest/contributing/workflow.html
-->
#### Target(s) affected by this defect ?
N/A
#### Toolchain(s) (name and version) displaying this defect ?
N/A
#### What version of Mbed-os are you using (tag or sha) ?
<!--
For a released version please provide the release tag (this can be found as per the instructions below)
mbed-os version can be found in /platform/mbed_version.h. The tag can be reconstructed as follows:
mbed-os-MBED_MAJOR_VERSION.MBED_MINOR_VERSION.MBED_PATCH_VERSION
Master branch is indicated by 'mbed-os-99.99.99
For an issue found on Master please provide the sha being used.
-->
Latest
#### What version(s) of tools are you using. List all that apply (E.g. mbed-cli)
N/A
#### How is this defect reproduced ?
N/A
<!--
For connectivity issues it is recommended to enable mbed trace and attach the traces
generated by the logger to the issue.
-->
|
2.0
|
Findout a place to keep mbed-os/UNITTESTS/target_h - <!--
************************************** WARNING **************************************
The ciarcom bot parses this header automatically. Any deviation from the
template may cause the bot to automatically correct this header or may result in a
warning message, requesting updates.
Please ensure all sections of the template below are filled in and no changes
are made to the template format. Only bugs should be raised here as issues.
Questions or enhancements should instead be raised on our forums:
https://forums.mbed.com/ .
*************************************************************************************
-->
### Description of defect
Findout a place to keep mbed-os/UNITTESTS/target_h
<!--
Add detailed description of what you are reporting.
Good example: https://os.mbed.com/docs/mbed-os/latest/contributing/workflow.html
-->
#### Target(s) affected by this defect ?
N/A
#### Toolchain(s) (name and version) displaying this defect ?
N/A
#### What version of Mbed-os are you using (tag or sha) ?
<!--
For a released version please provide the release tag (this can be found as per the instructions below)
mbed-os version can be found in /platform/mbed_version.h. The tag can be reconstructed as follows:
mbed-os-MBED_MAJOR_VERSION.MBED_MINOR_VERSION.MBED_PATCH_VERSION
Master branch is indicated by 'mbed-os-99.99.99
For an issue found on Master please provide the sha being used.
-->
Latest
#### What version(s) of tools are you using. List all that apply (E.g. mbed-cli)
N/A
#### How is this defect reproduced ?
N/A
<!--
For connectivity issues it is recommended to enable mbed trace and attach the traces
generated by the logger to the issue.
-->
|
non_process
|
findout a place to keep mbed os unittests target h warning the ciarcom bot parses this header automatically any deviation from the template may cause the bot to automatically correct this header or may result in a warning message requesting updates please ensure all sections of the template below are filled in and no changes are made to the template format only bugs should be raised here as issues questions or enhancements should instead be raised on our forums description of defect findout a place to keep mbed os unittests target h add detailed description of what you are reporting good example target s affected by this defect n a toolchain s name and version displaying this defect n a what version of mbed os are you using tag or sha for a released version please provide the release tag this can be found as per the instructions below mbed os version can be found in platform mbed version h the tag can be reconstructed as follows mbed os mbed major version mbed minor version mbed patch version master branch is indicated by mbed os for an issue found on master please provide the sha being used latest what version s of tools are you using list all that apply e g mbed cli n a how is this defect reproduced n a for connectivity issues it is recommended to enable mbed trace and attach the traces generated by the logger to the issue
| 0
|
892
| 3,354,875,271
|
IssuesEvent
|
2015-11-18 14:22:03
|
kerubistan/kerub
|
https://api.github.com/repos/kerubistan/kerub
|
closed
|
generalize 'host assignment'
|
cleanup component:data processing priority: normal
|
*Host* assignment should be just *assignment*
* modify data structure
* all references should be called *assignment*
|
1.0
|
generalize 'host assignment' - *Host* assignment should be just *assignment*
* modify data structure
* all references should be called *assignment*
|
process
|
generalize host assignment host assignment should be just assignment modify data structure all references should be called assignment
| 1
|
7,000
| 10,145,505,138
|
IssuesEvent
|
2019-08-05 04:43:11
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
joins across multiple tables
|
.Proposal Query Language (MBQL) Query Processor
|
one of the biggest feature requests here is the ability to join more than one hop from the starting table. usually in filters, but sometimes in group-bys too. i know this is a nontrivial feature, but i can imagine it would be at least somewhat doable, UI wise, by just including connections in the filter column list, and collapsing when necessary if you keep following foreign keys. (and when rendering the filter "chip.")
background: i've been doing "metabase office hours" here for the last couple days, and i've collected a handful of bug reports/feature requests. i'll try to search first to make sure i don't dupe existing issues, but apologies in advance if i do.
people are loving metabase here, btw! here are a few messages i've gotten on slack, verbatim:
> hey! everyone is raving about metabase, but my invite is not working. i want in on the new hotness! any chance i can get another invite?
> oh wow nice! this thing is impressive. i _love_ that you can do raw sql!
> holy crap metabase is amazing :rainbow: :sparkles:
⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
|
1.0
|
joins across multiple tables - one of the biggest feature requests here is the ability to join more than one hop from the starting table. usually in filters, but sometimes in group-bys too. i know this is a nontrivial feature, but i can imagine it would be at least somewhat doable, UI wise, by just including connections in the filter column list, and collapsing when necessary if you keep following foreign keys. (and when rendering the filter "chip.")
background: i've been doing "metabase office hours" here for the last couple days, and i've collected a handful of bug reports/feature requests. i'll try to search first to make sure i don't dupe existing issues, but apologies in advance if i do.
people are loving metabase here, btw! here are a few messages i've gotten on slack, verbatim:
> hey! everyone is raving about metabase, but my invite is not working. i want in on the new hotness! any chance i can get another invite?
> oh wow nice! this thing is impressive. i _love_ that you can do raw sql!
> holy crap metabase is amazing :rainbow: :sparkles:
⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
|
process
|
joins across multiple tables one of the biggest feature requests here is the ability to join more than one hop from the starting table usually in filters but sometimes in group bys too i know this is a nontrivial feature but i can imagine it would be at least somewhat doable ui wise by just including connections in the filter column list and collapsing when necessary if you keep following foreign keys and when rendering the filter chip background i ve been doing metabase office hours here for the last couple days and i ve collected a handful of bug reports feature requests i ll try to search first to make sure i don t dupe existing issues but apologies in advance if i do people are loving metabase here btw here are a few messages i ve gotten on slack verbatim hey everyone is raving about metabase but my invite is not working i want in on the new hotness any chance i can get another invite oh wow nice this thing is impressive i love that you can do raw sql holy crap metabase is amazing rainbow sparkles ⬇️ please click the 👍 reaction instead of leaving a or 👍 comment
| 1
|
161,269
| 13,821,975,254
|
IssuesEvent
|
2020-10-13 03:51:40
|
Hbdkfk/Automated-Prostate-Cancer-Detection-using-MRI
|
https://api.github.com/repos/Hbdkfk/Automated-Prostate-Cancer-Detection-using-MRI
|
opened
|
EDA and further researches
|
documentation
|
-[ ] check number of patient
-[ ] distribution of series (not all the patient has same amount of series, select patient ids that have all series), zones, benign/cancerous findings
-[ ] search for the ROI sizes used
-[ ] see if other teams have dealt with the edge detection issues for CNN
|
1.0
|
EDA and further researches - -[ ] check number of patient
-[ ] distribution of series (not all the patient has same amount of series, select patient ids that have all series), zones, benign/cancerous findings
-[ ] search for the ROI sizes used
-[ ] see if other teams have dealt with the edge detection issues for CNN
|
non_process
|
eda and further researches check number of patient distribution of series not all the patient has same amount of series select patient ids that have all series zones benign cancerous findings search for the roi sizes used see if other teams have dealt with the edge detection issues for cnn
| 0
|
8,669
| 11,802,675,811
|
IssuesEvent
|
2020-03-18 22:06:07
|
phokz/mod-auth-external
|
https://api.github.com/repos/phokz/mod-auth-external
|
opened
|
Now that we have official Windows builds, we should update INSTALL
|
process-item
|
Now that we have official Windows builds, we should update INSTALL with example steps and requirements. The current one claims it may not work. :)
The Wiki was [previously updated](https://github.com/phokz/mod-auth-external/wiki/Installation), but could use more details (example, requirements) and maybe reorganization would be good.
|
1.0
|
Now that we have official Windows builds, we should update INSTALL - Now that we have official Windows builds, we should update INSTALL with example steps and requirements. The current one claims it may not work. :)
The Wiki was [previously updated](https://github.com/phokz/mod-auth-external/wiki/Installation), but could use more details (example, requirements) and maybe reorganization would be good.
|
process
|
now that we have official windows builds we should update install now that we have official windows builds we should update install with example steps and requirements the current one claims it may not work the wiki was but could use more details example requirements and maybe reorganization would be good
| 1
|
11,904
| 14,698,310,116
|
IssuesEvent
|
2021-01-04 06:08:23
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
"Extract by location" atributes bug when two predicates are selected
|
Bug Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
When using "Extract by location" from _Processings_, choosing two predicates as seen below causes duplicates rows ... when object intersects and is within.
Geometries are unique but I get duplicates rows in attributes.
**How to Reproduce**
1. Open Processings toolbox
2. Click on "Extract by location"
3. Choose any input layer : here "polygons-2"
4. Check predicates : 'intersects' and 'are within'
5. Choose second layer `By comparing features from` "polygon"

6. Open attributes table from created layer
7. Observe duplicated rows in attributes when object match both 'intersects' and 'within' conditions !

**QGIS and OS versions**
Version de QGIS | 3.10.0-A Coruña | Révision du code | 6c816b4204
-- | -- | -- | --
Compilé avec Qt | 5.9.5 | Utilisant Qt | 5.9.5
Compilé avec GDAL/OGR | 2.2.3 | Utilisé avec GDAL/OGR | 2.2.3
Compilé avec GEOS | 3.6.2-CAPI-1.10.2 | Utilisé avec GEOS | 3.6.2-CAPI-1.10.2 4d2925d6
Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0
Version du client PostgreSQL | 10.10 (Ubuntu 10.10-0ubuntu0.18.04.1) | Version de SpatiaLite | 4.3.0a
Version de QWT | 6.1.3 | Version de QScintilla2 | 2.10.2
Version de PROJ.4 : | 493
OS Version | Ubuntu 18.04.3 LTS
Extensions Python actives | VectorBender; AnotherDXF2Shape; mask; Qgis2threejs; GeoCoding; vector_tiles_reader; qconsolidate3; shapetools; Hqgis; DataPlotly; french_locator_filter; SRTM-Downloader; Beeline; OSMtools; QNEAT3; loadthemall; db_manager; MetaSearch; processing
**Additional context**
<!-- Add any other context about the problem here. -->
|
1.0
|
"Extract by location" atributes bug when two predicates are selected - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
When using "Extract by location" from _Processings_, choosing two predicates as seen below causes duplicates rows ... when object intersects and is within.
Geometries are unique but I get duplicates rows in attributes.
**How to Reproduce**
1. Open Processings toolbox
2. Click on "Extract by location"
3. Choose any input layer : here "polygons-2"
4. Check predicates : 'intersects' and 'are within'
5. Choose second layer `By comparing features from` "polygon"

6. Open attributes table from created layer
7. Observe duplicated rows in attributes when object match both 'intersects' and 'within' conditions !

**QGIS and OS versions**
Version de QGIS | 3.10.0-A Coruña | Révision du code | 6c816b4204
-- | -- | -- | --
Compilé avec Qt | 5.9.5 | Utilisant Qt | 5.9.5
Compilé avec GDAL/OGR | 2.2.3 | Utilisé avec GDAL/OGR | 2.2.3
Compilé avec GEOS | 3.6.2-CAPI-1.10.2 | Utilisé avec GEOS | 3.6.2-CAPI-1.10.2 4d2925d6
Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0
Version du client PostgreSQL | 10.10 (Ubuntu 10.10-0ubuntu0.18.04.1) | Version de SpatiaLite | 4.3.0a
Version de QWT | 6.1.3 | Version de QScintilla2 | 2.10.2
Version de PROJ.4 : | 493
OS Version | Ubuntu 18.04.3 LTS
Extensions Python actives | VectorBender; AnotherDXF2Shape; mask; Qgis2threejs; GeoCoding; vector_tiles_reader; qconsolidate3; shapetools; Hqgis; DataPlotly; french_locator_filter; SRTM-Downloader; Beeline; OSMtools; QNEAT3; loadthemall; db_manager; MetaSearch; processing
**Additional context**
<!-- Add any other context about the problem here. -->
|
process
|
extract by location atributes bug when two predicates are selected bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug when using extract by location from processings choosing two predicates as seen below causes duplicates rows when object intersects and is within geometries are unique but i get duplicates rows in attributes how to reproduce open processings toolbox click on extract by location choose any input layer here polygons check predicates intersects and are within choose second layer by comparing features from polygon open attributes table from created layer observe duplicated rows in attributes when object match both intersects and within conditions qgis and os versions version de qgis a coruña révision du code compilé avec qt utilisant qt compilé avec gdal ogr utilisé avec gdal ogr compilé avec geos capi utilisé avec geos capi compiled against sqlite running against sqlite version du client postgresql ubuntu version de spatialite version de qwt version de version de proj os version ubuntu lts extensions python actives vectorbender mask geocoding vector tiles reader shapetools hqgis dataplotly french locator filter srtm downloader beeline osmtools loadthemall db manager metasearch processing additional context
| 1
|
525,381
| 15,252,099,894
|
IssuesEvent
|
2021-02-20 01:31:30
|
rubyforgood/casa
|
https://api.github.com/repos/rubyforgood/casa
|
closed
|
Do not show inactive assignment or inactive case to volunteers. Do not include in supervisor weekly digest.
|
:clipboard: Supervisor :email: Email :no_good_man: User Permissions :raised_hands: Volunteer Priority: Medium
|
**What is the problem, and what should happen instead?**
function `case_assignments_with_cases`
contains `case_assignments.includes(:casa_case)`
called by `views/volunteers/edit.html.erb` on line 80
and `views/supervisor_mailer/weekly_digest.html.erb `on line 15
**Acceptance criteria:**
- [ ] Volunteer Edit should NOT include casa cases from which the volunteer has been unassigned (inactive assignment)
- [ ] Write a test
- [ ] Volunteer Edit should NOT include **inactive** casa cases (inactive case)
- [ ] Write a test
- [ ] Supervisor email should NOT include cases which are inactive at the time that the mailer is generated (inactive case)
- [ ] Write a test
- [ ] Supervisor email should NOT include cases which have been unassigned from volunteers (inactive assignment)
- [ ] Write a test
- [ ] If a volunteer performs actions
- - and is unassigned from a supervisor in the same week
- - and **not** assigned to another supervisor
then the supervisor should see the activity in the email
and the email should indicate that the volunteer is unassigned
- [ ] If a volunteer performs actions
- - and is unassigned from a supervisor in the same week
- - and **is** assigned to another supervisor, then the volunteer's old supervisor should not see the volunteer's activity in the email
- - and the volunteer's new supervisor should see the volunteer's activity in their email
Thanks @FireLemons for finding this! https://rubyforgood.slack.com/archives/CVB0QJGVD/p1607574310223400
|
1.0
|
Do not show inactive assignment or inactive case to volunteers. Do not include in supervisor weekly digest. - **What is the problem, and what should happen instead?**
function `case_assignments_with_cases`
contains `case_assignments.includes(:casa_case)`
called by `views/volunteers/edit.html.erb` on line 80
and `views/supervisor_mailer/weekly_digest.html.erb `on line 15
**Acceptance criteria:**
- [ ] Volunteer Edit should NOT include casa cases from which the volunteer has been unassigned (inactive assignment)
- [ ] Write a test
- [ ] Volunteer Edit should NOT include **inactive** casa cases (inactive case)
- [ ] Write a test
- [ ] Supervisor email should NOT include cases which are inactive at the time that the mailer is generated (inactive case)
- [ ] Write a test
- [ ] Supervisor email should NOT include cases which have been unassigned from volunteers (inactive assignment)
- [ ] Write a test
- [ ] If a volunteer performs actions
- - and is unassigned from a supervisor in the same week
- - and **not** assigned to another supervisor
then the supervisor should see the activity in the email
and the email should indicate that the volunteer is unassigned
- [ ] If a volunteer performs actions
- - and is unassigned from a supervisor in the same week
- - and **is** assigned to another supervisor, then the volunteer's old supervisor should not see the volunteer's activity in the email
- - and the volunteer's new supervisor should see the volunteer's activity in their email
Thanks @FireLemons for finding this! https://rubyforgood.slack.com/archives/CVB0QJGVD/p1607574310223400
|
non_process
|
do not show inactive assignment or inactive case to volunteers do not include in supervisor weekly digest what is the problem and what should happen instead function case assignments with cases contains case assignments includes casa case called by views volunteers edit html erb on line and views supervisor mailer weekly digest html erb on line acceptance criteria volunteer edit should not include casa cases from which the volunteer has been unassigned inactive assignment write a test volunteer edit should not include inactive casa cases inactive case write a test supervisor email should not include cases which are inactive at the time that the mailer is generated inactive case write a test supervisor email should not include cases which have been unassigned from volunteers inactive assignment write a test if a volunteer performs actions and is unassigned from a supervisor in the same week and not assigned to another supervisor then the supervisor should see the activity in the email and the email should indicate that the volunteer is unassigned if a volunteer performs actions and is unassigned from a supervisor in the same week and is assigned to another supervisor then the volunteer s old supervisor should not see the volunteer s activity in the email and the volunteer s new supervisor should see the volunteer s activity in their email thanks firelemons for finding this
| 0
|
256,496
| 27,561,680,730
|
IssuesEvent
|
2023-03-07 22:39:42
|
samqws-marketing/electronicarts_ava-capture
|
https://api.github.com/repos/samqws-marketing/electronicarts_ava-capture
|
closed
|
CVE-2021-44906 (High) detected in minimist-1.2.5.tgz - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2021-44906 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p>
<p>Path to dependency file: /website-frontend/package.json</p>
<p>Path to vulnerable library: /website-frontend/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- awesome-typescript-loader-5.2.1.tgz (Root Library)
- mkdirp-0.5.5.tgz
- :x: **minimist-1.2.5.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: minimist - 1.2.6</p>
</p>
</details>
<p></p>
|
True
|
CVE-2021-44906 (High) detected in minimist-1.2.5.tgz - autoclosed - ## CVE-2021-44906 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary>
<p>parse argument options</p>
<p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p>
<p>Path to dependency file: /website-frontend/package.json</p>
<p>Path to vulnerable library: /website-frontend/node_modules/minimist/package.json</p>
<p>
Dependency Hierarchy:
- awesome-typescript-loader-5.2.1.tgz (Root Library)
- mkdirp-0.5.5.tgz
- :x: **minimist-1.2.5.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95).
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-44906>CVE-2021-44906</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: minimist - 1.2.6</p>
</p>
</details>
<p></p>
|
non_process
|
cve high detected in minimist tgz autoclosed cve high severity vulnerability vulnerable library minimist tgz parse argument options library home page a href path to dependency file website frontend package json path to vulnerable library website frontend node modules minimist package json dependency hierarchy awesome typescript loader tgz root library mkdirp tgz x minimist tgz vulnerable library found in base branch master vulnerability details minimist is vulnerable to prototype pollution via file index js function setkey lines publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimist
| 0
|
26,900
| 4,826,052,555
|
IssuesEvent
|
2016-11-07 08:49:09
|
bridgedotnet/Bridge
|
https://api.github.com/repos/bridgedotnet/Bridge
|
opened
|
Build: Module 'XXX' has no exported member 'YYY'.
|
defect
|
http://forums.bridge.net/forum/bridge-net-pro/bugs/2976-build-module-xxx-has-no-exported-member-yyy
### Expected
```js
module DictionaryMap$2
```
### Actual
```js
export interface DictionaryMap$2<T1,T2> {
getForward(): Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T1,T2>;
setForward(value: Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T1,T2>): void;
getReverse(): Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T2,T1>;
setReverse(value: Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T2,T1>): void;
add(t1: T1, t2: T2): void;
}
export function DictionaryMap$2<T1,T2>(T1: {prototype: T1}, T2: {prototype: T2}): {
prototype: DictionaryMap$2<T1,T2>;
ctor: {
new (): DictionaryMap$2<T1,T2>
};
$ctor1: {
new (initialValues: System.Collections.Generic.KeyValuePair$2<T1,T2>[]): DictionaryMap$2<T1,T2>
};
}
module DictionaryMap {
export interface Indexer$2<T1,T2,T3,T4> {
getItem(index: T3): T4;
setItem(index: T3, value: T4): void;
containsKey(index: T3): boolean;
}
export function Indexer$2<T1,T2,T3,T4>(T1: {prototype: T1}, T2: {prototype: T2}, T3: {prototype: T3}, T4: {prototype: T4}): {
prototype: Indexer$2<T1,T2,T3,T4>;
new (dictionary: System.Collections.Generic.Dictionary$2<T3,T4>): Indexer$2<T1,T2,T3,T4>;
}
}
```
### Steps To Reproduce
[Deck](http://deck.net/REPLACE-WITH-YOUR-DECK-LINK)
```cs
public class DictionaryMap<T1, T2>
{
private Dictionary<T1, T2> _forward = new Dictionary<T1, T2>();
private Dictionary<T2, T1> _reverse = new Dictionary<T2, T1>();
public DictionaryMap()
{
this.Forward = new Indexer<T1, T2>(_forward);
this.Reverse = new Indexer<T2, T1>(_reverse);
}
public DictionaryMap(params KeyValuePair<T1, T2>[] initialValues): this()
{
foreach (KeyValuePair<T1, T2> value in initialValues)
Add(value.Key, value.Value);
}
public class Indexer<T3, T4>
{
private Dictionary<T3, T4> _dictionary;
public Indexer(Dictionary<T3, T4> dictionary)
{
_dictionary = dictionary;
}
public T4 this[T3 index]
{
get { return _dictionary[index]; }
set { _dictionary[index] = value; }
}
public bool ContainsKey(T3 index)
{
return _dictionary.ContainsKey(index);
}
}
public void Add(T1 t1, T2 t2)
{
_forward.Add(t1, t2);
_reverse.Add(t2, t1);
}
public Indexer<T1, T2> Forward { get; private set; }
public Indexer<T2, T1> Reverse { get; private set; }
}
```
|
1.0
|
Build: Module 'XXX' has no exported member 'YYY'. - http://forums.bridge.net/forum/bridge-net-pro/bugs/2976-build-module-xxx-has-no-exported-member-yyy
### Expected
```js
module DictionaryMap$2
```
### Actual
```js
export interface DictionaryMap$2<T1,T2> {
getForward(): Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T1,T2>;
setForward(value: Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T1,T2>): void;
getReverse(): Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T2,T1>;
setReverse(value: Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T2,T1>): void;
add(t1: T1, t2: T2): void;
}
export function DictionaryMap$2<T1,T2>(T1: {prototype: T1}, T2: {prototype: T2}): {
prototype: DictionaryMap$2<T1,T2>;
ctor: {
new (): DictionaryMap$2<T1,T2>
};
$ctor1: {
new (initialValues: System.Collections.Generic.KeyValuePair$2<T1,T2>[]): DictionaryMap$2<T1,T2>
};
}
module DictionaryMap {
export interface Indexer$2<T1,T2,T3,T4> {
getItem(index: T3): T4;
setItem(index: T3, value: T4): void;
containsKey(index: T3): boolean;
}
export function Indexer$2<T1,T2,T3,T4>(T1: {prototype: T1}, T2: {prototype: T2}, T3: {prototype: T3}, T4: {prototype: T4}): {
prototype: Indexer$2<T1,T2,T3,T4>;
new (dictionary: System.Collections.Generic.Dictionary$2<T3,T4>): Indexer$2<T1,T2,T3,T4>;
}
}
```
### Steps To Reproduce
[Deck](http://deck.net/REPLACE-WITH-YOUR-DECK-LINK)
```cs
public class DictionaryMap<T1, T2>
{
private Dictionary<T1, T2> _forward = new Dictionary<T1, T2>();
private Dictionary<T2, T1> _reverse = new Dictionary<T2, T1>();
public DictionaryMap()
{
this.Forward = new Indexer<T1, T2>(_forward);
this.Reverse = new Indexer<T2, T1>(_reverse);
}
public DictionaryMap(params KeyValuePair<T1, T2>[] initialValues): this()
{
foreach (KeyValuePair<T1, T2> value in initialValues)
Add(value.Key, value.Value);
}
public class Indexer<T3, T4>
{
private Dictionary<T3, T4> _dictionary;
public Indexer(Dictionary<T3, T4> dictionary)
{
_dictionary = dictionary;
}
public T4 this[T3 index]
{
get { return _dictionary[index]; }
set { _dictionary[index] = value; }
}
public bool ContainsKey(T3 index)
{
return _dictionary.ContainsKey(index);
}
}
public void Add(T1 t1, T2 t2)
{
_forward.Add(t1, t2);
_reverse.Add(t2, t1);
}
public Indexer<T1, T2> Forward { get; private set; }
public Indexer<T2, T1> Reverse { get; private set; }
}
```
|
non_process
|
build module xxx has no exported member yyy expected js module dictionarymap actual js export interface dictionarymap getforward mosaic infrastructure general dictionarymap indexer setforward value mosaic infrastructure general dictionarymap indexer void getreverse mosaic infrastructure general dictionarymap indexer setreverse value mosaic infrastructure general dictionarymap indexer void add void export function dictionarymap prototype prototype prototype dictionarymap ctor new dictionarymap new initialvalues system collections generic keyvaluepair dictionarymap module dictionarymap export interface indexer getitem index setitem index value void containskey index boolean export function indexer prototype prototype prototype prototype prototype indexer new dictionary system collections generic dictionary indexer steps to reproduce cs public class dictionarymap private dictionary forward new dictionary private dictionary reverse new dictionary public dictionarymap this forward new indexer forward this reverse new indexer reverse public dictionarymap params keyvaluepair initialvalues this foreach keyvaluepair value in initialvalues add value key value value public class indexer private dictionary dictionary public indexer dictionary dictionary dictionary dictionary public this get return dictionary set dictionary value public bool containskey index return dictionary containskey index public void add forward add reverse add public indexer forward get private set public indexer reverse get private set
| 0
|
19,960
| 26,441,240,973
|
IssuesEvent
|
2023-01-16 00:28:48
|
hsmusic/hsmusic-wiki
|
https://api.github.com/repos/hsmusic/hsmusic-wiki
|
closed
|
Totally empty documents (end of file only?) break YAML processing
|
type: bug scope: data processing
|
I.e, `---` at the end of an album yaml apparently breaks stuff. Wow yikes!!
Thanks to koba and Niklink for spotting this.
|
1.0
|
Totally empty documents (end of file only?) break YAML processing - I.e, `---` at the end of an album yaml apparently breaks stuff. Wow yikes!!
Thanks to koba and Niklink for spotting this.
|
process
|
totally empty documents end of file only break yaml processing i e at the end of an album yaml apparently breaks stuff wow yikes thanks to koba and niklink for spotting this
| 1
|
19,152
| 11,151,025,416
|
IssuesEvent
|
2019-12-24 02:01:42
|
microsoft/vscode-cpptools
|
https://api.github.com/repos/microsoft/vscode-cpptools
|
opened
|
C/C++ IntelliSense "expression must have a pointer type" is incorrectly shown on a pointer type (wrong location)
|
Language Service Visual Studio bug
|
Use
```
class c2
{
public:
int i;
};
class c1
{
public:
c2 func() { return c2(); }
};
void main(void)
{
std::shared_ptr<c1> ptr = std::make_shared<c1>();
ptr->func()->i;
}
```
Bug: "expression must have a pointer type" is shown on the "ptr", which is the wrong location and confusing. It should be shown after the func() call, which is the expression that doesn't have the pointer type, which is where the cl.exe compiler errors at.
Reported at https://developercommunity.visualstudio.com/content/problem/866095/cc-intellisense-expression-must-have-a-pointer-typ.html .
|
1.0
|
C/C++ IntelliSense "expression must have a pointer type" is incorrectly shown on a pointer type (wrong location) - Use
```
class c2
{
public:
int i;
};
class c1
{
public:
c2 func() { return c2(); }
};
void main(void)
{
std::shared_ptr<c1> ptr = std::make_shared<c1>();
ptr->func()->i;
}
```
Bug: "expression must have a pointer type" is shown on the "ptr", which is the wrong location and confusing. It should be shown after the func() call, which is the expression that doesn't have the pointer type, which is where the cl.exe compiler errors at.
Reported at https://developercommunity.visualstudio.com/content/problem/866095/cc-intellisense-expression-must-have-a-pointer-typ.html .
|
non_process
|
c c intellisense expression must have a pointer type is incorrectly shown on a pointer type wrong location use class public int i class public func return void main void std shared ptr ptr std make shared ptr func i bug expression must have a pointer type is shown on the ptr which is the wrong location and confusing it should be shown after the func call which is the expression that doesn t have the pointer type which is where the cl exe compiler errors at reported at
| 0
|
290,313
| 25,047,544,576
|
IssuesEvent
|
2022-11-05 13:01:54
|
dotnet/maui
|
https://api.github.com/repos/dotnet/maui
|
closed
|
[Critical] Command + V and more crashes the application (macOS 13)
|
t/bug platform/macOS 🍏 area/controls 🎮 control-entry s/needs-attention s/try-latest-version
|
### Description
When developing my own application, I noticed a very serious bug: when I tried to copy and paste certain text, the application stalled and then crashed. I can note that when testing on Windows (version 11) everything worked fine.
### Steps to Reproduce
1. Create a File > New .NET MAUI App
2. Add a `Entry` like so: `<Entry Text="this is a bug" />`
3. Click the added entry, press in keyboard Command + V, Command + C, Keyboard arrows left or right (macOS system) and observe the bug 🐞
Expected result: The text insertion will be successful.
Actual result: the application is slowing down/crashing.
There were no errors when trying to run this bug on Windows.
### Link to public reproduction project repository
none
### Version with bug
6.0.400
### Last version that worked well
Unknown/Other
### Affected platforms
macOS
### Affected platform versions
macOS 13.0 Ventura Beta (22A5373b)
### Did you find any workaround?
Sadly, no.
### Relevant log output
```shell
2022-10-22 22:39:46.959 BugMauiApp[29575:379247] Warning: observer object was not disposed manually with Dispose()
2022-10-22 22:39:46.959 BugMauiApp[29575:379247] Warning: observer object was not disposed manually with Dispose()
2022-10-22 22:39:51.040 BugMauiApp[29575:379239] *** Terminating app due to uncaught exception 'NSInvalidArgumentException', reason: '-[<_UIObscurableTextContentStorage: 0x600001445950> offsetFromLocation:toLocation:] received invalid locations 0 (null)'
*** First throw call stack:
(
0 CoreFoundation 0x00007ff80bb31cd3 __exceptionPreprocess + 242
1 libobjc.A.dylib 0x00007ff80b68210a objc_exception_throw + 48
2 CoreFoundation 0x00007ff80bb31b38 +[NSException raise:format:] + 206
3 UIFoundation 0x00007ff80fb8fbc0 -[NSTextContentStorage offsetFromLocation:toLocation:] + 290
4 UIKitCore 0x00007ff919906647 -[_UITextKit2LayoutController offsetFromPosition:toPosition:] + 234
5 UIKitCore 0x00007ff91a0f6430 -[UITextInputController offsetFromPosition:toPosition:] + 85
6 UIKitCore 0x00007ff91a104fea __91-[_UITextInputControllerTokenizer _closestTokenSubrangeForPosition:granularity:downstream:]_block_invoke + 175
7 UIFoundation 0x00007ff80fb67b5c -[NSTextStorage coordinateReading:] + 42
8 UIKitCore 0x00007ff91a104cf0 -[_UITextInputControllerTokenizer _closestTokenSubrangeForPosition:granularity:downstream:] + 412
9 UIKitCore 0x00007ff919d2864b -[UITextInputStringTokenizer positionFromPosition:toBoundary:inDirection:] + 100
10 UIKitCore 0x00007ff919c984c6 -[UIKBRTIPartner _queryUIKitDocumentRequest:completion:] + 1329
11 UIKitCore 0x00007ff919c96aee -[UIKBRTIPartner _queryDocumentRequest:completion:] + 159
12 UIKitCore 0x00007ff919c90f13 -[UIKBRTIPartner _updateRTIStateWithCompletion:] + 559
13 UIKitCore 0x00007ff919c8f912 -[UIKBRTIPartner updateStateWithCompletion:] + 65
14 UIKitCore 0x00007ff919c8ba37 -[UIKBRTIPartner documentStateChanged:] + 126
15 UIKitCore 0x00007ff919be8200 -[UIKeyboardImpl setDocumentState:] + 263
16 UIKitCore 0x00007ff919c03cbf -[UIKeyboardImpl updateKeyboardStateForDeletion] + 337
17 UIKitCore 0x00007ff9198c31b3 -[UIKBInputDelegateManager _deleteBackwardAndNotify:reinsertText:] + 651
18 UIKitCore 0x00007ff919c03b5d -[UIKeyboardImpl deleteBackwardAndNotify:reinsertText:] + 64
19 UIKitCore 0x00007ff919be7865 -[UIKeyboardImpl performTextOperationActionSelector:] + 1272
20 UIKitCore 0x00007ff919c95477 -[UIKBRTIPartner _queued_performTextOperations:resultHandler:] + 1408
21 libdispatch.dylib 0x00007ff80b85e7fb _dispatch_call_block_and_release + 12
22 libdispatch.dylib 0x00007ff80b85fa44 _dispatch_client_callout + 8
23 libdispatch.dylib 0x00007ff80b86c7b9 _dispatch_main_queue_drain + 952
24 libdispatch.dylib 0x00007ff80b86c3f3 _dispatch_main_queue_callback_4CF + 31
25 CoreFoundation 0x00007ff80baf844d __CFRUNLOOP_IS_SERVICING_THE_MAIN_DISPATCH_QUEUE__ + 9
26 CoreFoundation 0x00007ff80bab8edb __CFRunLoopRun + 2498
27 CoreFoundation 0x00007ff80bab7e9f CFRunLoopRunSpecific + 560
28 HIToolbox 0x00007ff8158eabd6 RunCurrentEventLoopInMode + 292
29 HIToolbox 0x00007ff8158ea806 ReceiveNextEventCommon + 199
30 HIToolbox 0x00007ff8158ea723 _BlockUntilNextEventMatchingListInModeWithFilter + 70
31 AppKit 0x00007ff80eb19b87 _DPSNextEvent + 909
32 AppKit 0x00007ff80eb18a08 -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:] + 1219
33 AppKit 0x00007ff80eb0b043 -[NSApplication run] + 586
34 AppKit 0x00007ff80eadf0a8 NSApplicationMain + 817
35 AppKit 0x00007ff80ed9f6d7 _NSApplicationMainWithInfoDictionary + 16
36 UIKitMacHelper 0x00007ff82300ce97 UINSApplicationMain + 1413
37 UIKitCore 0x00007ff918f615b9 UIApplicationMain + 144
38 libmonosgen-2.0.dylib 0x000000010e6d43b0 do_icall + 288
39 libmonosgen-2.0.dylib 0x000000010e6d3388 do_icall_wrapper + 248
40 libmonosgen-2.0.dylib 0x000000010e6c54d1 interp_exec_method + 3393
41 libmonosgen-2.0.dylib 0x000000010e6c35ff interp_runtime_invoke + 239
42 libmonosgen-2.0.dylib 0x000000010e5c20c4 mono_jit_runtime_invoke + 1220
43 libmonosgen-2.0.dylib 0x000000010e4de2b8 mono_runtime_invoke_checked + 136
44 libmonosgen-2.0.dylib 0x000000010e4e620c do_exec_main_checked + 92
45 libmonosgen-2.0.dylib 0x000000010e61ea92 mono_jit_exec + 354
46 libxamarin-dotnet-debug.dylib 0x000000010dea4b99 xamarin_main + 809
47 BugMauiApp 0x0000000104d99ba4 main + 52
48 dyld 0x00000002054fa310 start + 2432
)
libc++abi: terminating with uncaught exception of type NSException
=================================================================
Native Crash Reporting
=================================================================
Got a SIGABRT while executing native code. This usually indicates
a fatal error in the mono runtime or one of the native libraries
used by your application.
=================================================================
=================================================================
Native stacktrace:
=================================================================
0x10e6c1e85 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_dump_native_crash_info
0x10e663d6e - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_handle_native_crash
0x10e6c1782 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : sigabrt_signal_handler
0x7ff80ba09c1d - /usr/lib/system/libsystem_platform.dylib : _sigtramp
0x0 - Unknown
0x7ff80b928ca5 - /usr/lib/system/libsystem_c.dylib : abort
0x7ff80b999082 - /usr/lib/libc++abi.dylib : abort_message
0x7ff80b98a23d - /usr/lib/libc++abi.dylib : _ZL29demangling_unexpected_handlerv
0x7ff80b689250 - /usr/lib/libobjc.A.dylib : _ZL15_objc_terminatev
0x7ff80b9984a5 - /usr/lib/libc++abi.dylib : _ZSt11__terminatePFvvE
0x7ff80b998456 - /usr/lib/libc++abi.dylib : _ZSt9terminatev
0x7ff80b85fa58 - /usr/lib/system/libdispatch.dylib : _dispatch_client_callout
0x7ff80b86c7b9 - /usr/lib/system/libdispatch.dylib : _dispatch_main_queue_drain
0x7ff80b86c3f3 - /usr/lib/system/libdispatch.dylib : _dispatch_main_queue_callback_4CF
0x7ff80baf844d - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : __CFRUNLOOP_IS_SERVICING_THE_MAIN_DISPATCH_QUEUE__
0x7ff80bab8edb - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : __CFRunLoopRun
0x7ff80bab7e9f - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : CFRunLoopRunSpecific
0x7ff8158eabd6 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : RunCurrentEventLoopInMode
0x7ff8158ea806 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : ReceiveNextEventCommon
0x7ff8158ea723 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : _BlockUntilNextEventMatchingListInModeWithFilter
0x7ff80eb19b87 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : _DPSNextEvent
0x7ff80eb18a08 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:]
0x7ff80eb0b043 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : -[NSApplication run]
0x7ff80eadf0a8 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : NSApplicationMain
0x7ff80ed9f6d7 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : _NSApplicationMainWithInfoDictionary
0x7ff82300ce97 - /System/Library/PrivateFrameworks/UIKitMacHelper.framework/Versions/A/UIKitMacHelper : UINSApplicationMain
0x7ff918f615b9 - /System/iOSSupport/System/Library/PrivateFrameworks/UIKitCore.framework/Versions/A/UIKitCore : UIApplicationMain
0x10e6d43b0 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_icall
0x10e6d3388 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_icall_wrapper
0x10e6c54d1 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : interp_exec_method
0x10e6c35ff - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : interp_runtime_invoke
0x10e5c20c4 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_jit_runtime_invoke
0x10e4de2b8 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_runtime_invoke_checked
0x10e4e620c - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_exec_main_checked
0x10e61ea92 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_jit_exec
0x10dea4b99 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libxamarin-dotnet-debug.dylib : xamarin_main
0x104d99ba4 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MacOS/BugMauiApp : main
0x2054fa310 - Unknown
=================================================================
Telemetry Dumper:
=================================================================
Pkilling 0x13128519680x from 0x8679686848x
Pkilling 0x13126410240x from 0x8679686848x
Could not exec mono-hang-watchdog, expected on path '/Users/runner/work/1/s/artifacts/obj/mono/MacCatalyst.x64.Release/out/etc/../bin/mono-hang-watchdog' (errno 2)
Entering thread summarizer pause from 0x8679686848x
Finished thread summarizer pause from 0x8679686848x.
Failed to create breadcrumb file (null)/crash_hash_0x29e26bbc7a
Waiting for dumping threads to resume
=================================================================
Basic Fault Address Reporting
=================================================================
Memory around native instruction pointer (0x7ff80b9a730e):0x7ff80b9a72fe ff ff c3 90 90 90 b8 48 01 00 02 49 89 ca 0f 05 .......H...I....
0x7ff80b9a730e 73 08 48 89 c7 e9 aa 9a ff ff c3 90 90 90 b8 53 s.H............S
0x7ff80b9a731e 00 00 02 49 89 ca 0f 05 73 08 48 89 c7 e9 92 9a ...I....s.H.....
0x7ff80b9a732e ff ff c3 90 90 90 b8 83 01 00 02 49 89 ca 0f 05 ...........I....
=================================================================
Managed Stacktrace:
=================================================================
at <unknown> <0xffffffff>
at UIKit.UIApplication:UIApplicationMain <0x0019e>
at UIKit.UIApplication:Main <0x00106>
at BugMauiApp.Program:Main <0x00016>
at <Module>:runtime_invoke_direct_void_string[] <0x00074>
at <unknown> <0x00000>
=================================================================
```
|
1.0
|
[Critical] Command + V and more crashes the application (macOS 13) - ### Description
When developing my own application, I noticed a very serious bug: when I tried to copy and paste certain text, the application stalled and then crashed. I can note that when testing on Windows (version 11) everything worked fine.
### Steps to Reproduce
1. Create a File > New .NET MAUI App
2. Add a `Entry` like so: `<Entry Text="this is a bug" />`
3. Click the added entry, press in keyboard Command + V, Command + C, Keyboard arrows left or right (macOS system) and observe the bug 🐞
Expected result: The text insertion will be successful.
Actual result: the application is slowing down/crashing.
There were no errors when trying to run this bug on Windows.
### Link to public reproduction project repository
none
### Version with bug
6.0.400
### Last version that worked well
Unknown/Other
### Affected platforms
macOS
### Affected platform versions
macOS 13.0 Ventura Beta (22A5373b)
### Did you find any workaround?
Sadly, no.
### Relevant log output
```shell
2022-10-22 22:39:46.959 BugMauiApp[29575:379247] Warning: observer object was not disposed manually with Dispose()
2022-10-22 22:39:46.959 BugMauiApp[29575:379247] Warning: observer object was not disposed manually with Dispose()
2022-10-22 22:39:51.040 BugMauiApp[29575:379239] *** Terminating app due to uncaught exception 'NSInvalidArgumentException', reason: '-[<_UIObscurableTextContentStorage: 0x600001445950> offsetFromLocation:toLocation:] received invalid locations 0 (null)'
*** First throw call stack:
(
0 CoreFoundation 0x00007ff80bb31cd3 __exceptionPreprocess + 242
1 libobjc.A.dylib 0x00007ff80b68210a objc_exception_throw + 48
2 CoreFoundation 0x00007ff80bb31b38 +[NSException raise:format:] + 206
3 UIFoundation 0x00007ff80fb8fbc0 -[NSTextContentStorage offsetFromLocation:toLocation:] + 290
4 UIKitCore 0x00007ff919906647 -[_UITextKit2LayoutController offsetFromPosition:toPosition:] + 234
5 UIKitCore 0x00007ff91a0f6430 -[UITextInputController offsetFromPosition:toPosition:] + 85
6 UIKitCore 0x00007ff91a104fea __91-[_UITextInputControllerTokenizer _closestTokenSubrangeForPosition:granularity:downstream:]_block_invoke + 175
7 UIFoundation 0x00007ff80fb67b5c -[NSTextStorage coordinateReading:] + 42
8 UIKitCore 0x00007ff91a104cf0 -[_UITextInputControllerTokenizer _closestTokenSubrangeForPosition:granularity:downstream:] + 412
9 UIKitCore 0x00007ff919d2864b -[UITextInputStringTokenizer positionFromPosition:toBoundary:inDirection:] + 100
10 UIKitCore 0x00007ff919c984c6 -[UIKBRTIPartner _queryUIKitDocumentRequest:completion:] + 1329
11 UIKitCore 0x00007ff919c96aee -[UIKBRTIPartner _queryDocumentRequest:completion:] + 159
12 UIKitCore 0x00007ff919c90f13 -[UIKBRTIPartner _updateRTIStateWithCompletion:] + 559
13 UIKitCore 0x00007ff919c8f912 -[UIKBRTIPartner updateStateWithCompletion:] + 65
14 UIKitCore 0x00007ff919c8ba37 -[UIKBRTIPartner documentStateChanged:] + 126
15 UIKitCore 0x00007ff919be8200 -[UIKeyboardImpl setDocumentState:] + 263
16 UIKitCore 0x00007ff919c03cbf -[UIKeyboardImpl updateKeyboardStateForDeletion] + 337
17 UIKitCore 0x00007ff9198c31b3 -[UIKBInputDelegateManager _deleteBackwardAndNotify:reinsertText:] + 651
18 UIKitCore 0x00007ff919c03b5d -[UIKeyboardImpl deleteBackwardAndNotify:reinsertText:] + 64
19 UIKitCore 0x00007ff919be7865 -[UIKeyboardImpl performTextOperationActionSelector:] + 1272
20 UIKitCore 0x00007ff919c95477 -[UIKBRTIPartner _queued_performTextOperations:resultHandler:] + 1408
21 libdispatch.dylib 0x00007ff80b85e7fb _dispatch_call_block_and_release + 12
22 libdispatch.dylib 0x00007ff80b85fa44 _dispatch_client_callout + 8
23 libdispatch.dylib 0x00007ff80b86c7b9 _dispatch_main_queue_drain + 952
24 libdispatch.dylib 0x00007ff80b86c3f3 _dispatch_main_queue_callback_4CF + 31
25 CoreFoundation 0x00007ff80baf844d __CFRUNLOOP_IS_SERVICING_THE_MAIN_DISPATCH_QUEUE__ + 9
26 CoreFoundation 0x00007ff80bab8edb __CFRunLoopRun + 2498
27 CoreFoundation 0x00007ff80bab7e9f CFRunLoopRunSpecific + 560
28 HIToolbox 0x00007ff8158eabd6 RunCurrentEventLoopInMode + 292
29 HIToolbox 0x00007ff8158ea806 ReceiveNextEventCommon + 199
30 HIToolbox 0x00007ff8158ea723 _BlockUntilNextEventMatchingListInModeWithFilter + 70
31 AppKit 0x00007ff80eb19b87 _DPSNextEvent + 909
32 AppKit 0x00007ff80eb18a08 -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:] + 1219
33 AppKit 0x00007ff80eb0b043 -[NSApplication run] + 586
34 AppKit 0x00007ff80eadf0a8 NSApplicationMain + 817
35 AppKit 0x00007ff80ed9f6d7 _NSApplicationMainWithInfoDictionary + 16
36 UIKitMacHelper 0x00007ff82300ce97 UINSApplicationMain + 1413
37 UIKitCore 0x00007ff918f615b9 UIApplicationMain + 144
38 libmonosgen-2.0.dylib 0x000000010e6d43b0 do_icall + 288
39 libmonosgen-2.0.dylib 0x000000010e6d3388 do_icall_wrapper + 248
40 libmonosgen-2.0.dylib 0x000000010e6c54d1 interp_exec_method + 3393
41 libmonosgen-2.0.dylib 0x000000010e6c35ff interp_runtime_invoke + 239
42 libmonosgen-2.0.dylib 0x000000010e5c20c4 mono_jit_runtime_invoke + 1220
43 libmonosgen-2.0.dylib 0x000000010e4de2b8 mono_runtime_invoke_checked + 136
44 libmonosgen-2.0.dylib 0x000000010e4e620c do_exec_main_checked + 92
45 libmonosgen-2.0.dylib 0x000000010e61ea92 mono_jit_exec + 354
46 libxamarin-dotnet-debug.dylib 0x000000010dea4b99 xamarin_main + 809
47 BugMauiApp 0x0000000104d99ba4 main + 52
48 dyld 0x00000002054fa310 start + 2432
)
libc++abi: terminating with uncaught exception of type NSException
=================================================================
Native Crash Reporting
=================================================================
Got a SIGABRT while executing native code. This usually indicates
a fatal error in the mono runtime or one of the native libraries
used by your application.
=================================================================
=================================================================
Native stacktrace:
=================================================================
0x10e6c1e85 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_dump_native_crash_info
0x10e663d6e - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_handle_native_crash
0x10e6c1782 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : sigabrt_signal_handler
0x7ff80ba09c1d - /usr/lib/system/libsystem_platform.dylib : _sigtramp
0x0 - Unknown
0x7ff80b928ca5 - /usr/lib/system/libsystem_c.dylib : abort
0x7ff80b999082 - /usr/lib/libc++abi.dylib : abort_message
0x7ff80b98a23d - /usr/lib/libc++abi.dylib : _ZL29demangling_unexpected_handlerv
0x7ff80b689250 - /usr/lib/libobjc.A.dylib : _ZL15_objc_terminatev
0x7ff80b9984a5 - /usr/lib/libc++abi.dylib : _ZSt11__terminatePFvvE
0x7ff80b998456 - /usr/lib/libc++abi.dylib : _ZSt9terminatev
0x7ff80b85fa58 - /usr/lib/system/libdispatch.dylib : _dispatch_client_callout
0x7ff80b86c7b9 - /usr/lib/system/libdispatch.dylib : _dispatch_main_queue_drain
0x7ff80b86c3f3 - /usr/lib/system/libdispatch.dylib : _dispatch_main_queue_callback_4CF
0x7ff80baf844d - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : __CFRUNLOOP_IS_SERVICING_THE_MAIN_DISPATCH_QUEUE__
0x7ff80bab8edb - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : __CFRunLoopRun
0x7ff80bab7e9f - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : CFRunLoopRunSpecific
0x7ff8158eabd6 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : RunCurrentEventLoopInMode
0x7ff8158ea806 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : ReceiveNextEventCommon
0x7ff8158ea723 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : _BlockUntilNextEventMatchingListInModeWithFilter
0x7ff80eb19b87 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : _DPSNextEvent
0x7ff80eb18a08 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:]
0x7ff80eb0b043 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : -[NSApplication run]
0x7ff80eadf0a8 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : NSApplicationMain
0x7ff80ed9f6d7 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : _NSApplicationMainWithInfoDictionary
0x7ff82300ce97 - /System/Library/PrivateFrameworks/UIKitMacHelper.framework/Versions/A/UIKitMacHelper : UINSApplicationMain
0x7ff918f615b9 - /System/iOSSupport/System/Library/PrivateFrameworks/UIKitCore.framework/Versions/A/UIKitCore : UIApplicationMain
0x10e6d43b0 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_icall
0x10e6d3388 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_icall_wrapper
0x10e6c54d1 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : interp_exec_method
0x10e6c35ff - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : interp_runtime_invoke
0x10e5c20c4 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_jit_runtime_invoke
0x10e4de2b8 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_runtime_invoke_checked
0x10e4e620c - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_exec_main_checked
0x10e61ea92 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_jit_exec
0x10dea4b99 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libxamarin-dotnet-debug.dylib : xamarin_main
0x104d99ba4 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MacOS/BugMauiApp : main
0x2054fa310 - Unknown
=================================================================
Telemetry Dumper:
=================================================================
Pkilling 0x13128519680x from 0x8679686848x
Pkilling 0x13126410240x from 0x8679686848x
Could not exec mono-hang-watchdog, expected on path '/Users/runner/work/1/s/artifacts/obj/mono/MacCatalyst.x64.Release/out/etc/../bin/mono-hang-watchdog' (errno 2)
Entering thread summarizer pause from 0x8679686848x
Finished thread summarizer pause from 0x8679686848x.
Failed to create breadcrumb file (null)/crash_hash_0x29e26bbc7a
Waiting for dumping threads to resume
=================================================================
Basic Fault Address Reporting
=================================================================
Memory around native instruction pointer (0x7ff80b9a730e):0x7ff80b9a72fe ff ff c3 90 90 90 b8 48 01 00 02 49 89 ca 0f 05 .......H...I....
0x7ff80b9a730e 73 08 48 89 c7 e9 aa 9a ff ff c3 90 90 90 b8 53 s.H............S
0x7ff80b9a731e 00 00 02 49 89 ca 0f 05 73 08 48 89 c7 e9 92 9a ...I....s.H.....
0x7ff80b9a732e ff ff c3 90 90 90 b8 83 01 00 02 49 89 ca 0f 05 ...........I....
=================================================================
Managed Stacktrace:
=================================================================
at <unknown> <0xffffffff>
at UIKit.UIApplication:UIApplicationMain <0x0019e>
at UIKit.UIApplication:Main <0x00106>
at BugMauiApp.Program:Main <0x00016>
at <Module>:runtime_invoke_direct_void_string[] <0x00074>
at <unknown> <0x00000>
=================================================================
```
|
non_process
|
command v and more crashes the application macos description when developing my own application i noticed a very serious bug when i tried to copy and paste certain text the application stalled and then crashed i can note that when testing on windows version everything worked fine steps to reproduce create a file new net maui app add a entry like so click the added entry press in keyboard command v command c keyboard arrows left or right macos system and observe the bug 🐞 expected result the text insertion will be successful actual result the application is slowing down crashing there were no errors when trying to run this bug on windows link to public reproduction project repository none version with bug last version that worked well unknown other affected platforms macos affected platform versions macos ventura beta did you find any workaround sadly no relevant log output shell bugmauiapp warning observer object was not disposed manually with dispose bugmauiapp warning observer object was not disposed manually with dispose bugmauiapp terminating app due to uncaught exception nsinvalidargumentexception reason received invalid locations null first throw call stack corefoundation exceptionpreprocess libobjc a dylib objc exception throw corefoundation uifoundation uikitcore uikitcore uikitcore block invoke uifoundation uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore libdispatch dylib dispatch call block and release libdispatch dylib dispatch client callout libdispatch dylib dispatch main queue drain libdispatch dylib dispatch main queue callback corefoundation cfrunloop is servicing the main dispatch queue corefoundation cfrunlooprun corefoundation cfrunlooprunspecific hitoolbox runcurrenteventloopinmode hitoolbox receivenexteventcommon hitoolbox blockuntilnexteventmatchinglistinmodewithfilter appkit dpsnextevent appkit appkit appkit nsapplicationmain appkit nsapplicationmainwithinfodictionary uikitmachelper uinsapplicationmain uikitcore uiapplicationmain libmonosgen dylib do icall libmonosgen dylib do icall wrapper libmonosgen dylib interp exec method libmonosgen dylib interp runtime invoke libmonosgen dylib mono jit runtime invoke libmonosgen dylib mono runtime invoke checked libmonosgen dylib do exec main checked libmonosgen dylib mono jit exec libxamarin dotnet debug dylib xamarin main bugmauiapp main dyld start libc abi terminating with uncaught exception of type nsexception native crash reporting got a sigabrt while executing native code this usually indicates a fatal error in the mono runtime or one of the native libraries used by your application native stacktrace users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono dump native crash info users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono handle native crash users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib sigabrt signal handler usr lib system libsystem platform dylib sigtramp unknown usr lib system libsystem c dylib abort usr lib libc abi dylib abort message usr lib libc abi dylib unexpected handlerv usr lib libobjc a dylib objc terminatev usr lib libc abi dylib terminatepfvve usr lib libc abi dylib usr lib system libdispatch dylib dispatch client callout usr lib system libdispatch dylib dispatch main queue drain usr lib system libdispatch dylib dispatch main queue callback system library frameworks corefoundation framework versions a corefoundation cfrunloop is servicing the main dispatch queue system library frameworks corefoundation framework versions a corefoundation cfrunlooprun system library frameworks corefoundation framework versions a corefoundation cfrunlooprunspecific system library frameworks carbon framework versions a frameworks hitoolbox framework versions a hitoolbox runcurrenteventloopinmode system library frameworks carbon framework versions a frameworks hitoolbox framework versions a hitoolbox receivenexteventcommon system library frameworks carbon framework versions a frameworks hitoolbox framework versions a hitoolbox blockuntilnexteventmatchinglistinmodewithfilter system library frameworks appkit framework versions c appkit dpsnextevent system library frameworks appkit framework versions c appkit system library frameworks appkit framework versions c appkit system library frameworks appkit framework versions c appkit nsapplicationmain system library frameworks appkit framework versions c appkit nsapplicationmainwithinfodictionary system library privateframeworks uikitmachelper framework versions a uikitmachelper uinsapplicationmain system iossupport system library privateframeworks uikitcore framework versions a uikitcore uiapplicationmain users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib do icall users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib do icall wrapper users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib interp exec method users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib interp runtime invoke users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono jit runtime invoke users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono runtime invoke checked users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib do exec main checked users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono jit exec users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libxamarin dotnet debug dylib xamarin main users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents macos bugmauiapp main unknown telemetry dumper pkilling from pkilling from could not exec mono hang watchdog expected on path users runner work s artifacts obj mono maccatalyst release out etc bin mono hang watchdog errno entering thread summarizer pause from finished thread summarizer pause from failed to create breadcrumb file null crash hash waiting for dumping threads to resume basic fault address reporting memory around native instruction pointer ff ff ca h i aa ff ff s h s ca i s h ff ff ca i managed stacktrace at at uikit uiapplication uiapplicationmain at uikit uiapplication main at bugmauiapp program main at runtime invoke direct void string at
| 0
|
20,992
| 27,855,441,283
|
IssuesEvent
|
2023-03-20 22:22:19
|
microsoft/vscode
|
https://api.github.com/repos/microsoft/vscode
|
closed
|
Enable remote terminals to be run in systemd, where available
|
feature-request terminal-process
|
<!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
On remote Linux machines, we'd like to be able to run the VS Code terminals within a systemd slice. This enables us to be able to set resource usage limits and direct OOM kills to a particular slice so we can manage processes in our infra. Everything created under a slice parents up to the main slice so the sub-processes (such as large build jobs) can be managed separately while having vscode-server run within a different slice to enable protection for vscode-server and avoid possible connectivity loss.
|
1.0
|
Enable remote terminals to be run in systemd, where available - <!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ -->
<!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ -->
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
On remote Linux machines, we'd like to be able to run the VS Code terminals within a systemd slice. This enables us to be able to set resource usage limits and direct OOM kills to a particular slice so we can manage processes in our infra. Everything created under a slice parents up to the main slice so the sub-processes (such as large build jobs) can be managed separately while having vscode-server run within a different slice to enable protection for vscode-server and avoid possible connectivity loss.
|
process
|
enable remote terminals to be run in systemd where available on remote linux machines we d like to be able to run the vs code terminals within a systemd slice this enables us to be able to set resource usage limits and direct oom kills to a particular slice so we can manage processes in our infra everything created under a slice parents up to the main slice so the sub processes such as large build jobs can be managed separately while having vscode server run within a different slice to enable protection for vscode server and avoid possible connectivity loss
| 1
|
761,012
| 26,663,328,869
|
IssuesEvent
|
2023-01-25 23:39:25
|
cryostatio/cryostat
|
https://api.github.com/repos/cryostatio/cryostat
|
closed
|
[Bug] "View Report" failure/hang/timeout on archived recordings if not previously generated
|
bug high-priority
|
## Current Behavior:
Report generation fails, appears to hang and then time out on generation in some situations.
## Expected Behavior:
Report generation succeeds and the document is visible in a new tab.
## Steps To Reproduce:
1. `sh smoketest.sh`
2. Select a sample app and start a new recording (say 10 seconds, or any) with any template (I used Profiling). Leave "archive on stop" enabled.
3. Wait for recording to complete and be archived
4. Go to Archives > All Archives > sample target > recording > click three dots menu, "view report". Appears to hang processing and then time out and fail.
5. Go to Archives > All Targets > sample target > recording > three dots "view report" and observe similar behaviour. Logs show:
```
...
Jan 25, 2023 4:19:53 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData
WARNING: Thread Thread[vert.x-eventloop-thread-2,5,main] has been blocked for 27467 ms, time limit is 2000 ms
io.vertx.core.VertxException: Thread blocked
at java.base@17.0.5/jdk.internal.misc.Unsafe.park(Native Method)
at java.base@17.0.5/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252)
at java.base@17.0.5/java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1866)
at java.base@17.0.5/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3463)
at java.base@17.0.5/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3434)
at java.base@17.0.5/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1939)
at java.base@17.0.5/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095)
at app//io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99)
at app//io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58)
at app//io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148)
at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99)
at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72)
at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189)
at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48)
at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123)
at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65)
at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68)
at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37)
at app//io.cryostat.net.web.WebServer$$Lambda$842/0x00000008013b0238.handle(Unknown Source)
at app//io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168)
at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57)
at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30)
at app//io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55)
at app//io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158)
at app//io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145)
at app//io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157)
at app//io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93)
at app//io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at app//io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373)
at app//io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236)
at app//io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285)
at app//io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519)
at app//io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458)
at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at app//io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base@17.0.5/java.lang.Thread.run(Thread.java:833)
Jan 25, 2023 4:19:54 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData
WARNING: Thread Thread[vert.x-eventloop-thread-2,5,main] has been blocked for 28467 ms, time limit is 2000 ms
io.vertx.core.VertxException: Thread blocked
at java.base@17.0.5/jdk.internal.misc.Unsafe.park(Native Method)
at java.base@17.0.5/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252)
at java.base@17.0.5/java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1866)
at java.base@17.0.5/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3463)
at java.base@17.0.5/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3434)
at java.base@17.0.5/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1939)
at java.base@17.0.5/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095)
at app//io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99)
at app//io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58)
at app//io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148)
at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99)
at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72)
at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189)
at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48)
at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123)
at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65)
at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68)
at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37)
at app//io.cryostat.net.web.WebServer$$Lambda$842/0x00000008013b0238.handle(Unknown Source)
at app//io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168)
at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57)
at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30)
at app//io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55)
at app//io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158)
at app//io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145)
at app//io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157)
at app//io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93)
at app//io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at app//io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373)
at app//io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236)
at app//io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285)
at app//io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519)
at app//io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458)
at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at app//io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base@17.0.5/java.lang.Thread.run(Thread.java:833)
Jan 25, 2023 4:19:55 PM io.cryostat.core.log.Logger error
SEVERE: Exception thrown
java.util.concurrent.TimeoutException
at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1960)
at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095)
at io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99)
at io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58)
at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148)
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99)
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72)
at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189)
at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48)
at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123)
at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65)
at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68)
at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37)
at io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168)
at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57)
at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30)
at io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55)
at io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158)
at io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145)
at io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157)
at io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93)
at io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373)
at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236)
at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285)
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519)
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:833)
Jan 25, 2023 4:19:55 PM io.cryostat.core.log.Logger error
SEVERE: HTTP 500: java.util.concurrent.TimeoutException
io.vertx.ext.web.handler.HttpException: Internal Server Error
Caused by: io.cryostat.net.web.http.api.v2.ApiException: java.util.concurrent.TimeoutException
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:119)
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72)
at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189)
at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48)
at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123)
at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65)
at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68)
at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37)
at io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168)
at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57)
at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30)
at io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55)
at io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158)
at io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145)
at io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157)
at io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93)
at io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373)
at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236)
at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285)
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519)
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.util.concurrent.ExecutionException: java.util.concurrent.TimeoutException
at java.base/java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:396)
at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2096)
at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:149)
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99)
... 65 more
Caused by: java.util.concurrent.TimeoutException
at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1960)
at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095)
at io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99)
at io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58)
at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148)
... 66 more
Jan 25, 2023 4:19:55 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData
SEVERE: 10.0.2.100 - - [Wed, 25 Jan 2023 16:19:55 GMT] 29020ms "GET /api/beta/fs/reports/GAYHANDKNZVEEQKRIEZVO3SMMZTS24KEPJJDOS3BI5QU233TORRXOWKWPFFDE2KXKAZW6PI=/es-andrewazor-demo-Main_foo_20230125T161921Z.jfr/jwt?token=eyJjdHkiOiJKV1QiLCJlbmMiOiJBMjU2R0NNIiwiYWxnIjoiZGlyIn0..UJvZPNJuTOzbFvyT.nlT2sG-PdX78ZWOr2zz88RZgO9E8vNnGMplrislXJZn3zP7dnIEfPIot65oP867sJS7KRiN7CJf2q0s8DC-G_lq07m0VVyp7FoWYM1GYagcupWTqOSEjlcjSUGQ5UBofkj7kZgQHFfDjeHL-32UCOURAoZt1plCYbalLQ64rL-3ARWUC51NcpdHf3zw2SjezsTaMwjWq2uObU-I1uRD_VX7NMe7bSBf2-XW8hHxRyse0gZMpHQPjLv0rl-OlTKlKUKLN8KGwDSE__n6afXKOS2FHkenUTRAuEOxv1nVB2lUSajIZNGfGS3LQ8Qu70b8Du4X2xVWmZuoX0Pa_6kGMP6W0a5aY_RKn8__FV8JrT_DersJ59pdeRaz882KolLS3R6dPSPhz1h9kukukFLbVVm_jlGnGwS0lD3w7IfSEvUFLzohX3rmkpzxYs1WZDGfmvx8eDURDwrxaRWeRrl0KfObNKsDReSWjhmuBtjZSfbMoylh0VWSC8AQcEyp0cByTSVfhohf502zhQX75dxlNd7-TsguI1icGuqvsWvUUSkcNNWlLQZQ9I87mVMWyGwB5qn6_QcRCHP2xai-HKiqesHIwpQvJqHX7O3pd5Z14chiTlTNZ8N8-I_7VMjytUDLzOzHMdAL2aZL_SemNy-n6Vw.12ztQuTsdBJV45gNnk5CfA HTTP/1.1" 500 121 bytes "https://localhost:8181/archives" "Mozilla/5.0 (X11; Linux x86_64; rv:108.0) Gecko/20100101 Firefox/108.0"
```

6. Expand recording row on either table for this archived recording to generate the report and view it in an inline `iframe`. This works. Then retry the "three dots view report" menu item again, and this now works instantly.
## Anything else:
Only affects the 'fromPath", `jwt` versions of reports endpoints, which are used by the web-client for viewing reports in a new tab or download reports as files, not for the inline `iframe` views of reports embedded within recording tables.
|
1.0
|
[Bug] "View Report" failure/hang/timeout on archived recordings if not previously generated - ## Current Behavior:
Report generation fails, appears to hang and then time out on generation in some situations.
## Expected Behavior:
Report generation succeeds and the document is visible in a new tab.
## Steps To Reproduce:
1. `sh smoketest.sh`
2. Select a sample app and start a new recording (say 10 seconds, or any) with any template (I used Profiling). Leave "archive on stop" enabled.
3. Wait for recording to complete and be archived
4. Go to Archives > All Archives > sample target > recording > click three dots menu, "view report". Appears to hang processing and then time out and fail.
5. Go to Archives > All Targets > sample target > recording > three dots "view report" and observe similar behaviour. Logs show:
```
...
Jan 25, 2023 4:19:53 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData
WARNING: Thread Thread[vert.x-eventloop-thread-2,5,main] has been blocked for 27467 ms, time limit is 2000 ms
io.vertx.core.VertxException: Thread blocked
at java.base@17.0.5/jdk.internal.misc.Unsafe.park(Native Method)
at java.base@17.0.5/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252)
at java.base@17.0.5/java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1866)
at java.base@17.0.5/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3463)
at java.base@17.0.5/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3434)
at java.base@17.0.5/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1939)
at java.base@17.0.5/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095)
at app//io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99)
at app//io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58)
at app//io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148)
at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99)
at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72)
at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189)
at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48)
at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123)
at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65)
at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68)
at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37)
at app//io.cryostat.net.web.WebServer$$Lambda$842/0x00000008013b0238.handle(Unknown Source)
at app//io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168)
at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57)
at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30)
at app//io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55)
at app//io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158)
at app//io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145)
at app//io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157)
at app//io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93)
at app//io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at app//io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373)
at app//io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236)
at app//io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285)
at app//io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519)
at app//io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458)
at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at app//io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base@17.0.5/java.lang.Thread.run(Thread.java:833)
Jan 25, 2023 4:19:54 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData
WARNING: Thread Thread[vert.x-eventloop-thread-2,5,main] has been blocked for 28467 ms, time limit is 2000 ms
io.vertx.core.VertxException: Thread blocked
at java.base@17.0.5/jdk.internal.misc.Unsafe.park(Native Method)
at java.base@17.0.5/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252)
at java.base@17.0.5/java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1866)
at java.base@17.0.5/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3463)
at java.base@17.0.5/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3434)
at java.base@17.0.5/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1939)
at java.base@17.0.5/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095)
at app//io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99)
at app//io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58)
at app//io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148)
at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99)
at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72)
at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189)
at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48)
at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123)
at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65)
at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68)
at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37)
at app//io.cryostat.net.web.WebServer$$Lambda$842/0x00000008013b0238.handle(Unknown Source)
at app//io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168)
at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57)
at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30)
at app//io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55)
at app//io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158)
at app//io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145)
at app//io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157)
at app//io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93)
at app//io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at app//io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373)
at app//io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236)
at app//io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285)
at app//io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519)
at app//io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458)
at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at app//io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at app//io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at app//io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at app//io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base@17.0.5/java.lang.Thread.run(Thread.java:833)
Jan 25, 2023 4:19:55 PM io.cryostat.core.log.Logger error
SEVERE: Exception thrown
java.util.concurrent.TimeoutException
at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1960)
at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095)
at io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99)
at io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58)
at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148)
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99)
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72)
at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189)
at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48)
at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123)
at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65)
at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68)
at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37)
at io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168)
at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57)
at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30)
at io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55)
at io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158)
at io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145)
at io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157)
at io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93)
at io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373)
at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236)
at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285)
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519)
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:833)
Jan 25, 2023 4:19:55 PM io.cryostat.core.log.Logger error
SEVERE: HTTP 500: java.util.concurrent.TimeoutException
io.vertx.ext.web.handler.HttpException: Internal Server Error
Caused by: io.cryostat.net.web.http.api.v2.ApiException: java.util.concurrent.TimeoutException
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:119)
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72)
at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189)
at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48)
at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123)
at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65)
at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284)
at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173)
at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140)
at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68)
at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37)
at io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168)
at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57)
at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30)
at io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55)
at io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158)
at io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145)
at io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157)
at io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93)
at io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103)
at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373)
at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236)
at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285)
at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519)
at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458)
at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357)
at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379)
at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365)
at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919)
at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166)
at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788)
at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724)
at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562)
at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997)
at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
at java.base/java.lang.Thread.run(Thread.java:833)
Caused by: java.util.concurrent.ExecutionException: java.util.concurrent.TimeoutException
at java.base/java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:396)
at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2096)
at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:149)
at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99)
... 65 more
Caused by: java.util.concurrent.TimeoutException
at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1960)
at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095)
at io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99)
at io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58)
at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148)
... 66 more
Jan 25, 2023 4:19:55 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData
SEVERE: 10.0.2.100 - - [Wed, 25 Jan 2023 16:19:55 GMT] 29020ms "GET /api/beta/fs/reports/GAYHANDKNZVEEQKRIEZVO3SMMZTS24KEPJJDOS3BI5QU233TORRXOWKWPFFDE2KXKAZW6PI=/es-andrewazor-demo-Main_foo_20230125T161921Z.jfr/jwt?token=eyJjdHkiOiJKV1QiLCJlbmMiOiJBMjU2R0NNIiwiYWxnIjoiZGlyIn0..UJvZPNJuTOzbFvyT.nlT2sG-PdX78ZWOr2zz88RZgO9E8vNnGMplrislXJZn3zP7dnIEfPIot65oP867sJS7KRiN7CJf2q0s8DC-G_lq07m0VVyp7FoWYM1GYagcupWTqOSEjlcjSUGQ5UBofkj7kZgQHFfDjeHL-32UCOURAoZt1plCYbalLQ64rL-3ARWUC51NcpdHf3zw2SjezsTaMwjWq2uObU-I1uRD_VX7NMe7bSBf2-XW8hHxRyse0gZMpHQPjLv0rl-OlTKlKUKLN8KGwDSE__n6afXKOS2FHkenUTRAuEOxv1nVB2lUSajIZNGfGS3LQ8Qu70b8Du4X2xVWmZuoX0Pa_6kGMP6W0a5aY_RKn8__FV8JrT_DersJ59pdeRaz882KolLS3R6dPSPhz1h9kukukFLbVVm_jlGnGwS0lD3w7IfSEvUFLzohX3rmkpzxYs1WZDGfmvx8eDURDwrxaRWeRrl0KfObNKsDReSWjhmuBtjZSfbMoylh0VWSC8AQcEyp0cByTSVfhohf502zhQX75dxlNd7-TsguI1icGuqvsWvUUSkcNNWlLQZQ9I87mVMWyGwB5qn6_QcRCHP2xai-HKiqesHIwpQvJqHX7O3pd5Z14chiTlTNZ8N8-I_7VMjytUDLzOzHMdAL2aZL_SemNy-n6Vw.12ztQuTsdBJV45gNnk5CfA HTTP/1.1" 500 121 bytes "https://localhost:8181/archives" "Mozilla/5.0 (X11; Linux x86_64; rv:108.0) Gecko/20100101 Firefox/108.0"
```

6. Expand recording row on either table for this archived recording to generate the report and view it in an inline `iframe`. This works. Then retry the "three dots view report" menu item again, and this now works instantly.
## Anything else:
Only affects the 'fromPath", `jwt` versions of reports endpoints, which are used by the web-client for viewing reports in a new tab or download reports as files, not for the inline `iframe` views of reports embedded within recording tables.
|
non_process
|
view report failure hang timeout on archived recordings if not previously generated current behavior report generation fails appears to hang and then time out on generation in some situations expected behavior report generation succeeds and the document is visible in a new tab steps to reproduce sh smoketest sh select a sample app and start a new recording say seconds or any with any template i used profiling leave archive on stop enabled wait for recording to complete and be archived go to archives all archives sample target recording click three dots menu view report appears to hang processing and then time out and fail go to archives all targets sample target recording three dots view report and observe similar behaviour logs show jan pm org impl fillcallerdata warning thread thread has been blocked for ms time limit is ms io vertx core vertxexception thread blocked at java base jdk internal misc unsafe park native method at java base java util concurrent locks locksupport parknanos locksupport java at java base java util concurrent completablefuture signaller block completablefuture java at java base java util concurrent forkjoinpool unmanagedblock forkjoinpool java at java base java util concurrent forkjoinpool managedblock forkjoinpool java at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at app io cryostat net reports archivedrecordingreportcache getfrompath archivedrecordingreportcache java at app io cryostat net reports reportservice getfrompath reportservice java at app io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java at app io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at app io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at app io vertx ext web impl routestate handlecontext routestate java at app io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at app io vertx ext web impl routingcontextimpl next routingcontextimpl java at app io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at app io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at app io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at app io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at app io vertx ext web impl routestate handlecontext routestate java at app io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at app io vertx ext web impl routingcontextimpl next routingcontextimpl java at app io vertx ext web impl routerimpl handle routerimpl java at app io vertx ext web impl routerimpl handle routerimpl java at app io cryostat net web webserver lambda handle unknown source at app io cryostat net httpserver handlerdelegate handle httpserver java at app io vertx core http impl handle java at app io vertx core http impl handle java at app io vertx core impl eventloopcontext emit eventloopcontext java at app io vertx core impl duplicatedcontext emit duplicatedcontext java at app io vertx core http impl handlemessage java at app io vertx core net impl connectionbase read connectionbase java at app io vertx core net impl vertxhandler channelread vertxhandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at app io netty handler codec http websocketx extensions websocketserverextensionhandler channelread websocketserverextensionhandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at app io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler logging logginghandler channelread logginghandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler ssl sslhandler unwrap sslhandler java at app io netty handler ssl sslhandler decodejdkcompatible sslhandler java at app io netty handler ssl sslhandler decode sslhandler java at app io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at app io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at app io netty channel nio nioeventloop processselectedkey nioeventloop java at app io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at app io netty channel nio nioeventloop processselectedkeys nioeventloop java at app io netty channel nio nioeventloop run nioeventloop java at app io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at app io netty util internal threadexecutormap run threadexecutormap java at app io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java jan pm org impl fillcallerdata warning thread thread has been blocked for ms time limit is ms io vertx core vertxexception thread blocked at java base jdk internal misc unsafe park native method at java base java util concurrent locks locksupport parknanos locksupport java at java base java util concurrent completablefuture signaller block completablefuture java at java base java util concurrent forkjoinpool unmanagedblock forkjoinpool java at java base java util concurrent forkjoinpool managedblock forkjoinpool java at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at app io cryostat net reports archivedrecordingreportcache getfrompath archivedrecordingreportcache java at app io cryostat net reports reportservice getfrompath reportservice java at app io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java at app io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at app io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at app io vertx ext web impl routestate handlecontext routestate java at app io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at app io vertx ext web impl routingcontextimpl next routingcontextimpl java at app io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at app io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at app io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at app io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at app io vertx ext web impl routestate handlecontext routestate java at app io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at app io vertx ext web impl routingcontextimpl next routingcontextimpl java at app io vertx ext web impl routerimpl handle routerimpl java at app io vertx ext web impl routerimpl handle routerimpl java at app io cryostat net web webserver lambda handle unknown source at app io cryostat net httpserver handlerdelegate handle httpserver java at app io vertx core http impl handle java at app io vertx core http impl handle java at app io vertx core impl eventloopcontext emit eventloopcontext java at app io vertx core impl duplicatedcontext emit duplicatedcontext java at app io vertx core http impl handlemessage java at app io vertx core net impl connectionbase read connectionbase java at app io vertx core net impl vertxhandler channelread vertxhandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at app io netty handler codec http websocketx extensions websocketserverextensionhandler channelread websocketserverextensionhandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at app io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler logging logginghandler channelread logginghandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler ssl sslhandler unwrap sslhandler java at app io netty handler ssl sslhandler decodejdkcompatible sslhandler java at app io netty handler ssl sslhandler decode sslhandler java at app io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at app io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at app io netty channel nio nioeventloop processselectedkey nioeventloop java at app io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at app io netty channel nio nioeventloop processselectedkeys nioeventloop java at app io netty channel nio nioeventloop run nioeventloop java at app io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at app io netty util internal threadexecutormap run threadexecutormap java at app io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java jan pm io cryostat core log logger error severe exception thrown java util concurrent timeoutexception at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at io cryostat net reports archivedrecordingreportcache getfrompath archivedrecordingreportcache java at io cryostat net reports reportservice getfrompath reportservice java at io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at io vertx ext web impl routestate handlecontext routestate java at io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at io vertx ext web impl routingcontextimpl next routingcontextimpl java at io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at io vertx ext web impl routestate handlecontext routestate java at io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at io vertx ext web impl routingcontextimpl next routingcontextimpl java at io vertx ext web impl routerimpl handle routerimpl java at io vertx ext web impl routerimpl handle routerimpl java at io cryostat net httpserver handlerdelegate handle httpserver java at io vertx core http impl handle java at io vertx core http impl handle java at io vertx core impl eventloopcontext emit eventloopcontext java at io vertx core impl duplicatedcontext emit duplicatedcontext java at io vertx core http impl handlemessage java at io vertx core net impl connectionbase read connectionbase java at io vertx core net impl vertxhandler channelread vertxhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at io netty handler codec http websocketx extensions websocketserverextensionhandler channelread websocketserverextensionhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler logging logginghandler channelread logginghandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler ssl sslhandler unwrap sslhandler java at io netty handler ssl sslhandler decodejdkcompatible sslhandler java at io netty handler ssl sslhandler decode sslhandler java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java jan pm io cryostat core log logger error severe http java util concurrent timeoutexception io vertx ext web handler httpexception internal server error caused by io cryostat net web http api apiexception java util concurrent timeoutexception at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at io vertx ext web impl routestate handlecontext routestate java at io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at io vertx ext web impl routingcontextimpl next routingcontextimpl java at io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at io vertx ext web impl routestate handlecontext routestate java at io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at io vertx ext web impl routingcontextimpl next routingcontextimpl java at io vertx ext web impl routerimpl handle routerimpl java at io vertx ext web impl routerimpl handle routerimpl java at io cryostat net httpserver handlerdelegate handle httpserver java at io vertx core http impl handle java at io vertx core http impl handle java at io vertx core impl eventloopcontext emit eventloopcontext java at io vertx core impl duplicatedcontext emit duplicatedcontext java at io vertx core http impl handlemessage java at io vertx core net impl connectionbase read connectionbase java at io vertx core net impl vertxhandler channelread vertxhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at io netty handler codec http websocketx extensions websocketserverextensionhandler channelread websocketserverextensionhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler logging logginghandler channelread logginghandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler ssl sslhandler unwrap sslhandler java at io netty handler ssl sslhandler decodejdkcompatible sslhandler java at io netty handler ssl sslhandler decode sslhandler java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java caused by java util concurrent executionexception java util concurrent timeoutexception at java base java util concurrent completablefuture reportget completablefuture java at java base java util concurrent completablefuture get completablefuture java at io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java more caused by java util concurrent timeoutexception at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at io cryostat net reports archivedrecordingreportcache getfrompath archivedrecordingreportcache java at io cryostat net reports reportservice getfrompath reportservice java at io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java more jan pm org impl fillcallerdata severe get api beta fs reports es andrewazor demo main foo jfr jwt token ujvzpnjutozbfvyt g i semny http bytes mozilla linux rv gecko firefox expand recording row on either table for this archived recording to generate the report and view it in an inline iframe this works then retry the three dots view report menu item again and this now works instantly anything else only affects the frompath jwt versions of reports endpoints which are used by the web client for viewing reports in a new tab or download reports as files not for the inline iframe views of reports embedded within recording tables
| 0
|
4,273
| 3,793,636,933
|
IssuesEvent
|
2016-03-22 14:35:36
|
MISP/MISP
|
https://api.github.com/repos/MISP/MISP
|
closed
|
[UI] - Replace ShadowAttributes in the text to Proposal
|
usability
|
[UI] - Replace ShadowAttributes in the text to "Proposals" which much more easier for users to understand.
|
True
|
[UI] - Replace ShadowAttributes in the text to Proposal - [UI] - Replace ShadowAttributes in the text to "Proposals" which much more easier for users to understand.
|
non_process
|
replace shadowattributes in the text to proposal replace shadowattributes in the text to proposals which much more easier for users to understand
| 0
|
12,273
| 9,606,765,995
|
IssuesEvent
|
2019-05-11 13:25:02
|
microsoft/azure-pipelines-tasks
|
https://api.github.com/repos/microsoft/azure-pipelines-tasks
|
closed
|
Azure DevOps - Not able to run an inline Powershell in "Azure App Service Deploy's" post deployment action
|
Area: AzureAppService Area: Release question
|
I am looking to rename a file after deploying a package to the Azure Website. To do so, I understand that while deploying into an Azure Web app using the "Azure App Service Deploy" task, it has a Post Deployment action with which we can run an inline script that will execute on the app service itself - https://github.com/Microsoft/azure-pipelines-tasks/tree/master/Tasks/AzureRmWebAppDeploymentV4
I've tried to execute a simple copy powershell command and it throws an error
> [error]Error: EISDIR: illegal operation on a directory, read
Copy commands that I've tried :
```
Copy-Item ".\config\config.js" ".\config\<websiteName>.js" -Force
```
|
1.0
|
Azure DevOps - Not able to run an inline Powershell in "Azure App Service Deploy's" post deployment action - I am looking to rename a file after deploying a package to the Azure Website. To do so, I understand that while deploying into an Azure Web app using the "Azure App Service Deploy" task, it has a Post Deployment action with which we can run an inline script that will execute on the app service itself - https://github.com/Microsoft/azure-pipelines-tasks/tree/master/Tasks/AzureRmWebAppDeploymentV4
I've tried to execute a simple copy powershell command and it throws an error
> [error]Error: EISDIR: illegal operation on a directory, read
Copy commands that I've tried :
```
Copy-Item ".\config\config.js" ".\config\<websiteName>.js" -Force
```
|
non_process
|
azure devops not able to run an inline powershell in azure app service deploy s post deployment action i am looking to rename a file after deploying a package to the azure website to do so i understand that while deploying into an azure web app using the azure app service deploy task it has a post deployment action with which we can run an inline script that will execute on the app service itself i ve tried to execute a simple copy powershell command and it throws an error error eisdir illegal operation on a directory read copy commands that i ve tried copy item config config js config js force
| 0
|
10,333
| 13,163,425,367
|
IssuesEvent
|
2020-08-11 00:22:31
|
Obe101/eCommerceSite
|
https://api.github.com/repos/Obe101/eCommerceSite
|
closed
|
Add CI Pipeline
|
Developer process
|
Add continuous integration pipeline that makes sure code in pull request compiles successfully.
|
1.0
|
Add CI Pipeline - Add continuous integration pipeline that makes sure code in pull request compiles successfully.
|
process
|
add ci pipeline add continuous integration pipeline that makes sure code in pull request compiles successfully
| 1
|
95,172
| 16,074,076,474
|
IssuesEvent
|
2021-04-25 02:16:23
|
samq-ghdemo/JS-Demo
|
https://api.github.com/repos/samq-ghdemo/JS-Demo
|
opened
|
CVE-2017-1000048 (High) detected in multiple libraries
|
security vulnerability
|
## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>qs-0.6.6.tgz</b>, <b>qs-5.2.1.tgz</b>, <b>qs-6.2.1.tgz</b></p></summary>
<p>
<details><summary><b>qs-0.6.6.tgz</b></p></summary>
<p>querystring parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.6.6.tgz">https://registry.npmjs.org/qs/-/qs-0.6.6.tgz</a></p>
<p>Path to dependency file: JS-Demo/package.json</p>
<p>Path to vulnerable library: JS-Demo/node_modules/zaproxy/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- zaproxy-0.2.0.tgz (Root Library)
- request-2.36.0.tgz
- :x: **qs-0.6.6.tgz** (Vulnerable Library)
</details>
<details><summary><b>qs-5.2.1.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-5.2.1.tgz">https://registry.npmjs.org/qs/-/qs-5.2.1.tgz</a></p>
<p>Path to dependency file: JS-Demo/package.json</p>
<p>Path to vulnerable library: JS-Demo/node_modules/grunt-retire/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- grunt-retire-0.3.12.tgz (Root Library)
- request-2.67.0.tgz
- :x: **qs-5.2.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>qs-6.2.1.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.2.1.tgz">https://registry.npmjs.org/qs/-/qs-6.2.1.tgz</a></p>
<p>Path to dependency file: JS-Demo/package.json</p>
<p>Path to vulnerable library: JS-Demo/node_modules/npm/node_modules/request/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- request-2.75.0.tgz
- :x: **qs-6.2.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/JS-Demo/commit/46781df511f58c350408cb5158290290709b373c">46781df511f58c350408cb5158290290709b373c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.6.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;qs:0.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"},{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"5.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;qs:5.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"},{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"6.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;qs:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-1000048","vulnerabilityDetails":"the web framework using ljharb\u0027s qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2017-1000048 (High) detected in multiple libraries - ## CVE-2017-1000048 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>qs-0.6.6.tgz</b>, <b>qs-5.2.1.tgz</b>, <b>qs-6.2.1.tgz</b></p></summary>
<p>
<details><summary><b>qs-0.6.6.tgz</b></p></summary>
<p>querystring parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.6.6.tgz">https://registry.npmjs.org/qs/-/qs-0.6.6.tgz</a></p>
<p>Path to dependency file: JS-Demo/package.json</p>
<p>Path to vulnerable library: JS-Demo/node_modules/zaproxy/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- zaproxy-0.2.0.tgz (Root Library)
- request-2.36.0.tgz
- :x: **qs-0.6.6.tgz** (Vulnerable Library)
</details>
<details><summary><b>qs-5.2.1.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-5.2.1.tgz">https://registry.npmjs.org/qs/-/qs-5.2.1.tgz</a></p>
<p>Path to dependency file: JS-Demo/package.json</p>
<p>Path to vulnerable library: JS-Demo/node_modules/grunt-retire/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- grunt-retire-0.3.12.tgz (Root Library)
- request-2.67.0.tgz
- :x: **qs-5.2.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>qs-6.2.1.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.2.1.tgz">https://registry.npmjs.org/qs/-/qs-6.2.1.tgz</a></p>
<p>Path to dependency file: JS-Demo/package.json</p>
<p>Path to vulnerable library: JS-Demo/node_modules/npm/node_modules/request/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- request-2.75.0.tgz
- :x: **qs-6.2.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/JS-Demo/commit/46781df511f58c350408cb5158290290709b373c">46781df511f58c350408cb5158290290709b373c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.6.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;qs:0.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"},{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"5.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;qs:5.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"},{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"6.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;qs:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-1000048","vulnerabilityDetails":"the web framework using ljharb\u0027s qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries qs tgz qs tgz qs tgz qs tgz querystring parser library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules zaproxy node modules qs package json dependency hierarchy zaproxy tgz root library request tgz x qs tgz vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules grunt retire node modules qs package json dependency hierarchy grunt retire tgz root library request tgz x qs tgz vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules npm node modules request node modules qs package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz x qs tgz vulnerable library found in head commit a href found in base branch master vulnerability details the web framework using ljharb s qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree zaproxy request qs isminimumfixversionavailable true minimumfixversion qs packagetype javascript node js packagename qs packageversion packagefilepaths istransitivedependency true dependencytree grunt retire request qs isminimumfixversionavailable true minimumfixversion qs packagetype javascript node js packagename qs packageversion packagefilepaths istransitivedependency true dependencytree grunt npm install npm request qs isminimumfixversionavailable true minimumfixversion qs basebranches vulnerabilityidentifier cve vulnerabilitydetails the web framework using ljharb qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash vulnerabilityurl
| 0
|
76,169
| 26,272,511,279
|
IssuesEvent
|
2023-01-06 18:24:44
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
BUG: RegularGridInterpolator does not work on non-native byteorder anymore
|
defect scipy.interpolate
|
### Describe your issue.
With the most recent update, applying the `scipy.interpolate.REgularGridInterpolator` using non-native byteorder raises an exception.
This worked in 1.9.
### Reproducing Code Example
```python
import scipy
import numpy as np
from scipy.interpolate import RegularGridInterpolator
print(scipy.__version__)
x = np.linspace(0, 1, 10)
y = np.linspace(0, 1, 15)
xg, yg = np.meshgrid(x, y, indexing='ij', sparse=True)
# make values non-native byteorder
values = (2 * xg + yg).astype('>f8')
points = np.array([[0.5, 0.5]])
print(f'{points.dtype = }')
print(f'{values.dtype = }')
print(f'{x.dtype = }')
interpolator = RegularGridInterpolator((x, y), values)
print(interpolator(points))
```
### Error message
```shell
1.10.0
points.dtype = dtype('float64')
values.dtype = dtype('>f8')
x.dtype = dtype('float64')
Traceback (most recent call last):
File "/home/maxnoe/Uni/CTA/ctapipe/test_interp.py", line 22, in <module>
print(interpolator(points))
File "/usr/lib/python3.10/site-packages/scipy/interpolate/_rgi.py", line 336, in __call__
result = evaluate_linear_2d(self.values,
File "_rgi_cython.pyx", line 19, in scipy.interpolate._rgi_cython.evaluate_linear_2d
ValueError: Big-endian buffer not supported on little-endian compiler
```
### SciPy/NumPy/Python version information
1.10.0 1.23.5 sys.version_info(major=3, minor=10, micro=9, releaselevel='final', serial=0)
|
1.0
|
BUG: RegularGridInterpolator does not work on non-native byteorder anymore - ### Describe your issue.
With the most recent update, applying the `scipy.interpolate.REgularGridInterpolator` using non-native byteorder raises an exception.
This worked in 1.9.
### Reproducing Code Example
```python
import scipy
import numpy as np
from scipy.interpolate import RegularGridInterpolator
print(scipy.__version__)
x = np.linspace(0, 1, 10)
y = np.linspace(0, 1, 15)
xg, yg = np.meshgrid(x, y, indexing='ij', sparse=True)
# make values non-native byteorder
values = (2 * xg + yg).astype('>f8')
points = np.array([[0.5, 0.5]])
print(f'{points.dtype = }')
print(f'{values.dtype = }')
print(f'{x.dtype = }')
interpolator = RegularGridInterpolator((x, y), values)
print(interpolator(points))
```
### Error message
```shell
1.10.0
points.dtype = dtype('float64')
values.dtype = dtype('>f8')
x.dtype = dtype('float64')
Traceback (most recent call last):
File "/home/maxnoe/Uni/CTA/ctapipe/test_interp.py", line 22, in <module>
print(interpolator(points))
File "/usr/lib/python3.10/site-packages/scipy/interpolate/_rgi.py", line 336, in __call__
result = evaluate_linear_2d(self.values,
File "_rgi_cython.pyx", line 19, in scipy.interpolate._rgi_cython.evaluate_linear_2d
ValueError: Big-endian buffer not supported on little-endian compiler
```
### SciPy/NumPy/Python version information
1.10.0 1.23.5 sys.version_info(major=3, minor=10, micro=9, releaselevel='final', serial=0)
|
non_process
|
bug regulargridinterpolator does not work on non native byteorder anymore describe your issue with the most recent update applying the scipy interpolate regulargridinterpolator using non native byteorder raises an exception this worked in reproducing code example python import scipy import numpy as np from scipy interpolate import regulargridinterpolator print scipy version x np linspace y np linspace xg yg np meshgrid x y indexing ij sparse true make values non native byteorder values xg yg astype points np array print f points dtype print f values dtype print f x dtype interpolator regulargridinterpolator x y values print interpolator points error message shell points dtype dtype values dtype dtype x dtype dtype traceback most recent call last file home maxnoe uni cta ctapipe test interp py line in print interpolator points file usr lib site packages scipy interpolate rgi py line in call result evaluate linear self values file rgi cython pyx line in scipy interpolate rgi cython evaluate linear valueerror big endian buffer not supported on little endian compiler scipy numpy python version information sys version info major minor micro releaselevel final serial
| 0
|
540,696
| 15,815,892,993
|
IssuesEvent
|
2021-04-05 12:09:56
|
AY2021S2-CS2103T-T12-4/tp
|
https://api.github.com/repos/AY2021S2-CS2103T-T12-4/tp
|
closed
|
[PE-D] URL parameter for adding an API endpoint
|
priority.High severity.Low type.Bug
|
When user input url is in the format of www.example.org (without specifying the website protocol, a http protocol is assumed and automatically attached to the front of the user input url. However, it is possible that user does not want such an assumption.


<!--session: 1617429946124-258a37f6-cce5-4442-9226-410ccb423b7f-->
-------------
Labels: `severity.VeryLow` `type.FeatureFlaw`
original: Nanxi-Huang/ped#7
|
1.0
|
[PE-D] URL parameter for adding an API endpoint - When user input url is in the format of www.example.org (without specifying the website protocol, a http protocol is assumed and automatically attached to the front of the user input url. However, it is possible that user does not want such an assumption.


<!--session: 1617429946124-258a37f6-cce5-4442-9226-410ccb423b7f-->
-------------
Labels: `severity.VeryLow` `type.FeatureFlaw`
original: Nanxi-Huang/ped#7
|
non_process
|
url parameter for adding an api endpoint when user input url is in the format of without specifying the website protocol a http protocol is assumed and automatically attached to the front of the user input url however it is possible that user does not want such an assumption labels severity verylow type featureflaw original nanxi huang ped
| 0
|
21,207
| 28,262,769,798
|
IssuesEvent
|
2023-04-07 01:56:09
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] Exclude already sorted columns from `orderable_columns`
|
.Team/QueryProcessor :hammer_and_wrench:
|
We should exclude columns already used in order-by clauses from `metabase.lib.order_by/ordered_columns` output. More context [in Slack](https://metaboat.slack.com/archives/C04DN5VRQM6/p1680623669731539)
|
1.0
|
[MLv2] Exclude already sorted columns from `orderable_columns` - We should exclude columns already used in order-by clauses from `metabase.lib.order_by/ordered_columns` output. More context [in Slack](https://metaboat.slack.com/archives/C04DN5VRQM6/p1680623669731539)
|
process
|
exclude already sorted columns from orderable columns we should exclude columns already used in order by clauses from metabase lib order by ordered columns output more context
| 1
|
12,436
| 14,931,815,698
|
IssuesEvent
|
2021-01-25 06:35:53
|
pingcap/tidb
|
https://api.github.com/repos/pingcap/tidb
|
opened
|
Make stale read and history read compatible with coprocessor cache
|
component/coprocessor sig/execution sig/transaction type/enhancement
|
## Background
This is a subtask of #21094.
The coprocessor cache caches a response for each request on the TiDB side. If two coprocessor requests are equivalent, the second request may skip being processed on the TiKV side. However, the request is still sent to TiKV, because TiKV needs to judge whether the cached response is still valid.
For stale read and history read, the requested snapshot is older than the current. What concerns us is whether the coprocessor cache is still always valid for these kinds of reading.
## Analysis
The coprocessor cache is hit when (see `copIteratorWorker.handleTaskOnce` and `copIteratorWorker.handleCopResponse`):
- The number of requested ranges, the data size of the response, and the processing time of the request are all within the thresholds. That is, the response deserves to be cached.
- The region ID of the request and the cached response equals. That is, the region doesn't split or merge after it's cached.
- The start TS of the cached response is not greater than the start TS of the request. That is, the cached response won't be too new for the request.
- The applied index of this region doesn't change between the start TS of the cached response and the request. That is, the region is untouched after the response is cached. This is judged on the TiKV side.
Through the last two conditions, we can infer that as long as the start TS of the transactions of stale read and history read is set correctly, the cached response will be valid.
## Conclusion
We don't need to change anything.
|
1.0
|
Make stale read and history read compatible with coprocessor cache - ## Background
This is a subtask of #21094.
The coprocessor cache caches a response for each request on the TiDB side. If two coprocessor requests are equivalent, the second request may skip being processed on the TiKV side. However, the request is still sent to TiKV, because TiKV needs to judge whether the cached response is still valid.
For stale read and history read, the requested snapshot is older than the current. What concerns us is whether the coprocessor cache is still always valid for these kinds of reading.
## Analysis
The coprocessor cache is hit when (see `copIteratorWorker.handleTaskOnce` and `copIteratorWorker.handleCopResponse`):
- The number of requested ranges, the data size of the response, and the processing time of the request are all within the thresholds. That is, the response deserves to be cached.
- The region ID of the request and the cached response equals. That is, the region doesn't split or merge after it's cached.
- The start TS of the cached response is not greater than the start TS of the request. That is, the cached response won't be too new for the request.
- The applied index of this region doesn't change between the start TS of the cached response and the request. That is, the region is untouched after the response is cached. This is judged on the TiKV side.
Through the last two conditions, we can infer that as long as the start TS of the transactions of stale read and history read is set correctly, the cached response will be valid.
## Conclusion
We don't need to change anything.
|
process
|
make stale read and history read compatible with coprocessor cache background this is a subtask of the coprocessor cache caches a response for each request on the tidb side if two coprocessor requests are equivalent the second request may skip being processed on the tikv side however the request is still sent to tikv because tikv needs to judge whether the cached response is still valid for stale read and history read the requested snapshot is older than the current what concerns us is whether the coprocessor cache is still always valid for these kinds of reading analysis the coprocessor cache is hit when see copiteratorworker handletaskonce and copiteratorworker handlecopresponse the number of requested ranges the data size of the response and the processing time of the request are all within the thresholds that is the response deserves to be cached the region id of the request and the cached response equals that is the region doesn t split or merge after it s cached the start ts of the cached response is not greater than the start ts of the request that is the cached response won t be too new for the request the applied index of this region doesn t change between the start ts of the cached response and the request that is the region is untouched after the response is cached this is judged on the tikv side through the last two conditions we can infer that as long as the start ts of the transactions of stale read and history read is set correctly the cached response will be valid conclusion we don t need to change anything
| 1
|
11,099
| 13,941,131,628
|
IssuesEvent
|
2020-10-22 18:56:35
|
Arch666Angel/mods
|
https://api.github.com/repos/Arch666Angel/mods
|
closed
|
bob mods wood to coal recipe imbalanced with angels mods
|
Angels Bio Processing
|
bob mods adds a 2 wood -> 1 coal furnace recipe and it makes more energy than the wood -> wood bricks methods or the wood -> charcoal -> carbon route.
https://forums.factorio.com/viewtopic.php?f=185&t=25468&p=510592#p510592
the recipe is bob-coal-from-wood and its in the bob plates mod.
|
1.0
|
bob mods wood to coal recipe imbalanced with angels mods - bob mods adds a 2 wood -> 1 coal furnace recipe and it makes more energy than the wood -> wood bricks methods or the wood -> charcoal -> carbon route.
https://forums.factorio.com/viewtopic.php?f=185&t=25468&p=510592#p510592
the recipe is bob-coal-from-wood and its in the bob plates mod.
|
process
|
bob mods wood to coal recipe imbalanced with angels mods bob mods adds a wood coal furnace recipe and it makes more energy than the wood wood bricks methods or the wood charcoal carbon route the recipe is bob coal from wood and its in the bob plates mod
| 1
|
176,582
| 13,646,532,704
|
IssuesEvent
|
2020-09-25 23:27:49
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
DISABLED test_backward_ddp_outside (__main__.ProcessGroupDdpUnderDistAutogradTestWithSpawn)
|
high priority module: distributed module: rpc topic: flaky-tests triage review
|
## 🐛 test_backward_ddp_outside is flaky
See for ex. https://app.circleci.com/pipelines/github/pytorch/pytorch/183902/workflows/9bbc51db-4c8e-4d47-92ea-a43fe60ac4ce/jobs/5952141/steps
```
Jun 22 00:59:56 ======================================================================
Jun 22 00:59:56 ERROR [61.552s]: test_backward_ddp_outside (__main__.TestDdpUnderDistAutogradWrapper)
Jun 22 00:59:56 ----------------------------------------------------------------------
Jun 22 00:59:56 Traceback (most recent call last):
Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 204, in wrapper
Jun 22 00:59:56 self._join_processes(fn)
Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 306, in _join_processes
Jun 22 00:59:56 self._check_return_codes(elapsed_time)
Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 339, in _check_return_codes
Jun 22 00:59:56 raise RuntimeError(error)
Jun 22 00:59:56 RuntimeError: Processes 5 exited with error code 10
Jun 22 00:59:56
Jun 22 00:59:56 ----------------------------------------------------------------------
```
cc @ezyang @gchanan @zou3519 @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @xush6528 @osalpekar @jjlilley
|
1.0
|
DISABLED test_backward_ddp_outside (__main__.ProcessGroupDdpUnderDistAutogradTestWithSpawn) - ## 🐛 test_backward_ddp_outside is flaky
See for ex. https://app.circleci.com/pipelines/github/pytorch/pytorch/183902/workflows/9bbc51db-4c8e-4d47-92ea-a43fe60ac4ce/jobs/5952141/steps
```
Jun 22 00:59:56 ======================================================================
Jun 22 00:59:56 ERROR [61.552s]: test_backward_ddp_outside (__main__.TestDdpUnderDistAutogradWrapper)
Jun 22 00:59:56 ----------------------------------------------------------------------
Jun 22 00:59:56 Traceback (most recent call last):
Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 204, in wrapper
Jun 22 00:59:56 self._join_processes(fn)
Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 306, in _join_processes
Jun 22 00:59:56 self._check_return_codes(elapsed_time)
Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 339, in _check_return_codes
Jun 22 00:59:56 raise RuntimeError(error)
Jun 22 00:59:56 RuntimeError: Processes 5 exited with error code 10
Jun 22 00:59:56
Jun 22 00:59:56 ----------------------------------------------------------------------
```
cc @ezyang @gchanan @zou3519 @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @xush6528 @osalpekar @jjlilley
|
non_process
|
disabled test backward ddp outside main processgroupddpunderdistautogradtestwithspawn 🐛 test backward ddp outside is flaky see for ex jun jun error test backward ddp outside main testddpunderdistautogradwrapper jun jun traceback most recent call last jun file users distiller workspace lib site packages torch testing internal common distributed py line in wrapper jun self join processes fn jun file users distiller workspace lib site packages torch testing internal common distributed py line in join processes jun self check return codes elapsed time jun file users distiller workspace lib site packages torch testing internal common distributed py line in check return codes jun raise runtimeerror error jun runtimeerror processes exited with error code jun jun cc ezyang gchanan pietern mrshenli zhaojuanmao satgera rohan varma gqchen aazzolini osalpekar jjlilley
| 0
|
9,669
| 12,676,147,217
|
IssuesEvent
|
2020-06-19 04:09:34
|
medic/cht-core
|
https://api.github.com/repos/medic/cht-core
|
closed
|
Release 3.9.0
|
Type: Internal process
|
# Planning
- [x] Create an [organisation wide project](https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc) and add this issue to it. We use [semver](http://semver.org) so if there are breaking changes increment the major, otherwise if there are new features increment the minor, otherwise increment the service pack. Breaking changes in our case relate to updated software requirements (egs: CouchDB, node, minimum browser versions), broken backwards compatibility in an api, or a major visual update that requires user retraining.
- [x] Add all the issues to be worked on to the project. Ideally each minor release will have one or two features, a handful of improvements, and plenty of bugs.
# Development
When development is ready to begin one of the engineers should be nominated as a Release Manager. They will be responsible for making sure the following tasks are completed though not necessarily completing them.
- [x] Set the version number in `package.json` and `package-lock.json` and submit a PR. The easiest way to do this is to use `npm --no-git-tag-version version <major|minor>`.
- [x] Raise a new issue called `Update dependencies for <version>` with a description that links to [the documentation](https://github.com/medic/medic-docs/blob/master/development/update-dependencies.md). This should be done early in the release cycle so find a volunteer to take this on and assign it to them.
- [x] Write an update in the weekly Product Team call agenda summarising development and acceptance testing progress and identifying any blockers. The release manager is to update this every week until the version is released.
# Releasing
Once all issues have passed acceptance testing and have been merged into `master` release testing can begin.
- [x] Create a new release branch from `master` named `<major>.<minor>.x` in medic. Post a message to #development using this template:
```
@core_devs I've just created the `<major>.<minor>.x` release branch. Please be aware that any further changes intended for this release will have to be merged to `master` then backported. Thanks!
```
- [x] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing.
- [x] [Import translations keys](https://github.com/medic/medic-docs/blob/master/development/translations.md#adding-new-keys) into POE and notify the #translations Slack channel translate new and updated values, for example:
```
@channel I've just updated the translations in POE. These keys have been added: "<added-list>", and these keys have been updated: "<updated-list>"
```
- [x] Create a new document in the [release-notes folder](https://github.com/medic/medic/tree/master/release-notes) in `master`. Ensure all issues are in the GH Project, that they're correct labelled, and have human readable descriptions. Use [this script](https://github.com/medic/medic/blob/master/scripts/changelog-generator) to export the issues into our changelog format. Manually document any known migration steps and known issues. Provide description, screenshots, videos, and anything else to help communicate particularly important changes. Assign the PR to a) the Director of Technology, and b) an SRE to review and confirm the documentation on upgrade instructions and breaking changes is sufficient.
- [x] Create a Google Doc in the [blog posts folder](https://drive.google.com/drive/u/0/folders/0B2PTUNZFwxEvMHRWNTBjY2ZHNHc) with the draft of a blog post promoting the release based on the release notes above. Once it's ready ask Max and Kelly to review it.
- [x] Until release testing passes, make sure regressions are fixed in `master`, cherry-pick them into the release branch, and release another beta.
- [x] [Export the translations](https://github.com/medic/medic-docs/blob/master/development/translations.md#exporting-changes-from-poeditor-to-github), delete empty translation files and commit to `master`. Cherry-pick the commit into the release branch.
- [x] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/medic/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Link to the release notes in the description of the release.
- [x] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>`
- [x] Add the release to the [Supported versions](https://github.com/medic/medic-docs/blob/master/installation/supported-software.md#supported-versions) and update the EOL date and status of previous releases.
- [x] Announce the release in #products and #cht-contributors using this template:
```
@channel *We're excited to announce the release of {{version}}*
New features include {{key_features}}. We've also implemented loads of other improvements and fixed a heap of bugs.
Read the release notes for full details: {{url}}
Following our support policy, versions {{versions}} are no longer supported. Projects running these versions should start planning to upgrade in the near future. For more details read our software support documentation: https://github.com/medic/medic-docs/blob/master/installation/supported-software.md#supported-versions
To see what's scheduled for the next releases have a read of the product roadmap: https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc
```
- [x] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category. You can use the previous message and omit `@channel`.
- [x] Mark this issue "done" and close the project.
|
1.0
|
Release 3.9.0 - # Planning
- [x] Create an [organisation wide project](https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc) and add this issue to it. We use [semver](http://semver.org) so if there are breaking changes increment the major, otherwise if there are new features increment the minor, otherwise increment the service pack. Breaking changes in our case relate to updated software requirements (egs: CouchDB, node, minimum browser versions), broken backwards compatibility in an api, or a major visual update that requires user retraining.
- [x] Add all the issues to be worked on to the project. Ideally each minor release will have one or two features, a handful of improvements, and plenty of bugs.
# Development
When development is ready to begin one of the engineers should be nominated as a Release Manager. They will be responsible for making sure the following tasks are completed though not necessarily completing them.
- [x] Set the version number in `package.json` and `package-lock.json` and submit a PR. The easiest way to do this is to use `npm --no-git-tag-version version <major|minor>`.
- [x] Raise a new issue called `Update dependencies for <version>` with a description that links to [the documentation](https://github.com/medic/medic-docs/blob/master/development/update-dependencies.md). This should be done early in the release cycle so find a volunteer to take this on and assign it to them.
- [x] Write an update in the weekly Product Team call agenda summarising development and acceptance testing progress and identifying any blockers. The release manager is to update this every week until the version is released.
# Releasing
Once all issues have passed acceptance testing and have been merged into `master` release testing can begin.
- [x] Create a new release branch from `master` named `<major>.<minor>.x` in medic. Post a message to #development using this template:
```
@core_devs I've just created the `<major>.<minor>.x` release branch. Please be aware that any further changes intended for this release will have to be merged to `master` then backported. Thanks!
```
- [x] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing.
- [x] [Import translations keys](https://github.com/medic/medic-docs/blob/master/development/translations.md#adding-new-keys) into POE and notify the #translations Slack channel translate new and updated values, for example:
```
@channel I've just updated the translations in POE. These keys have been added: "<added-list>", and these keys have been updated: "<updated-list>"
```
- [x] Create a new document in the [release-notes folder](https://github.com/medic/medic/tree/master/release-notes) in `master`. Ensure all issues are in the GH Project, that they're correct labelled, and have human readable descriptions. Use [this script](https://github.com/medic/medic/blob/master/scripts/changelog-generator) to export the issues into our changelog format. Manually document any known migration steps and known issues. Provide description, screenshots, videos, and anything else to help communicate particularly important changes. Assign the PR to a) the Director of Technology, and b) an SRE to review and confirm the documentation on upgrade instructions and breaking changes is sufficient.
- [x] Create a Google Doc in the [blog posts folder](https://drive.google.com/drive/u/0/folders/0B2PTUNZFwxEvMHRWNTBjY2ZHNHc) with the draft of a blog post promoting the release based on the release notes above. Once it's ready ask Max and Kelly to review it.
- [x] Until release testing passes, make sure regressions are fixed in `master`, cherry-pick them into the release branch, and release another beta.
- [x] [Export the translations](https://github.com/medic/medic-docs/blob/master/development/translations.md#exporting-changes-from-poeditor-to-github), delete empty translation files and commit to `master`. Cherry-pick the commit into the release branch.
- [x] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/medic/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Link to the release notes in the description of the release.
- [x] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>`
- [x] Add the release to the [Supported versions](https://github.com/medic/medic-docs/blob/master/installation/supported-software.md#supported-versions) and update the EOL date and status of previous releases.
- [x] Announce the release in #products and #cht-contributors using this template:
```
@channel *We're excited to announce the release of {{version}}*
New features include {{key_features}}. We've also implemented loads of other improvements and fixed a heap of bugs.
Read the release notes for full details: {{url}}
Following our support policy, versions {{versions}} are no longer supported. Projects running these versions should start planning to upgrade in the near future. For more details read our software support documentation: https://github.com/medic/medic-docs/blob/master/installation/supported-software.md#supported-versions
To see what's scheduled for the next releases have a read of the product roadmap: https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc
```
- [x] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category. You can use the previous message and omit `@channel`.
- [x] Mark this issue "done" and close the project.
|
process
|
release planning create an and add this issue to it we use so if there are breaking changes increment the major otherwise if there are new features increment the minor otherwise increment the service pack breaking changes in our case relate to updated software requirements egs couchdb node minimum browser versions broken backwards compatibility in an api or a major visual update that requires user retraining add all the issues to be worked on to the project ideally each minor release will have one or two features a handful of improvements and plenty of bugs development when development is ready to begin one of the engineers should be nominated as a release manager they will be responsible for making sure the following tasks are completed though not necessarily completing them set the version number in package json and package lock json and submit a pr the easiest way to do this is to use npm no git tag version version raise a new issue called update dependencies for with a description that links to this should be done early in the release cycle so find a volunteer to take this on and assign it to them write an update in the weekly product team call agenda summarising development and acceptance testing progress and identifying any blockers the release manager is to update this every week until the version is released releasing once all issues have passed acceptance testing and have been merged into master release testing can begin create a new release branch from master named x in medic post a message to development using this template core devs i ve just created the x release branch please be aware that any further changes intended for this release will have to be merged to master then backported thanks build a beta named beta by pushing a git tag and when ci completes successfully notify the qa team that it s ready for release testing into poe and notify the translations slack channel translate new and updated values for example channel i ve just updated the translations in poe these keys have been added and these keys have been updated create a new document in the in master ensure all issues are in the gh project that they re correct labelled and have human readable descriptions use to export the issues into our changelog format manually document any known migration steps and known issues provide description screenshots videos and anything else to help communicate particularly important changes assign the pr to a the director of technology and b an sre to review and confirm the documentation on upgrade instructions and breaking changes is sufficient create a google doc in the with the draft of a blog post promoting the release based on the release notes above once it s ready ask max and kelly to review it until release testing passes make sure regressions are fixed in master cherry pick them into the release branch and release another beta delete empty translation files and commit to master cherry pick the commit into the release branch create a release in github from the release branch so it shows up under the with the naming convention this will create the git tag automatically link to the release notes in the description of the release confirm the release build completes successfully and the new release is available on the make sure that the document has new entry with id medic medic add the release to the and update the eol date and status of previous releases announce the release in products and cht contributors using this template channel we re excited to announce the release of version new features include key features we ve also implemented loads of other improvements and fixed a heap of bugs read the release notes for full details url following our support policy versions versions are no longer supported projects running these versions should start planning to upgrade in the near future for more details read our software support documentation to see what s scheduled for the next releases have a read of the product roadmap announce the release on the under the product releases category you can use the previous message and omit channel mark this issue done and close the project
| 1
|
437,282
| 30,594,328,138
|
IssuesEvent
|
2023-07-21 20:11:36
|
hirosystems/docs
|
https://api.github.com/repos/hirosystems/docs
|
closed
|
Registry sample app refactor
|
documentation help wanted
|
The registry sample app needs to be updated:
- the sample code on the docs page has new lines making the code unreadable
- the sample code isn't working anymore, as discussed here: https://github.com/blockstack/docs/issues/1046
|
1.0
|
Registry sample app refactor - The registry sample app needs to be updated:
- the sample code on the docs page has new lines making the code unreadable
- the sample code isn't working anymore, as discussed here: https://github.com/blockstack/docs/issues/1046
|
non_process
|
registry sample app refactor the registry sample app needs to be updated the sample code on the docs page has new lines making the code unreadable the sample code isn t working anymore as discussed here
| 0
|
158,538
| 13,735,832,608
|
IssuesEvent
|
2020-10-05 10:46:42
|
kevinpapst/kimai2
|
https://api.github.com/repos/kevinpapst/kimai2
|
closed
|
Token problem
|
documentation support waiting for feedback
|
Hello Kevin,
I agree on the choice to use tokens with a short duration to protect data security but, in my specific case, the duration of the token is too short and this involves a loss of time optimization because, both me and and my colleagues, we are forced several times a day to clean the cache and log in, especially for those who take a little longer to enter and confirm a registration.
Following what has been said, I ask you to kindly tell me, if possible, how to change the settings relating to the token in order to increase the duration of the token by at least 1 or 2 hours.
Thanks for the attention
Raffaele Colarusso
|
1.0
|
Token problem - Hello Kevin,
I agree on the choice to use tokens with a short duration to protect data security but, in my specific case, the duration of the token is too short and this involves a loss of time optimization because, both me and and my colleagues, we are forced several times a day to clean the cache and log in, especially for those who take a little longer to enter and confirm a registration.
Following what has been said, I ask you to kindly tell me, if possible, how to change the settings relating to the token in order to increase the duration of the token by at least 1 or 2 hours.
Thanks for the attention
Raffaele Colarusso
|
non_process
|
token problem hello kevin i agree on the choice to use tokens with a short duration to protect data security but in my specific case the duration of the token is too short and this involves a loss of time optimization because both me and and my colleagues we are forced several times a day to clean the cache and log in especially for those who take a little longer to enter and confirm a registration following what has been said i ask you to kindly tell me if possible how to change the settings relating to the token in order to increase the duration of the token by at least or hours thanks for the attention raffaele colarusso
| 0
|
196,564
| 22,442,145,139
|
IssuesEvent
|
2022-06-21 02:34:43
|
valdisiljuconoks/AlloyTech
|
https://api.github.com/repos/valdisiljuconoks/AlloyTech
|
closed
|
WS-2019-0047 (Medium) detected in tar-0.1.20.tgz - autoclosed
|
security vulnerability
|
## WS-2019-0047 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-0.1.20.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-0.1.20.tgz">https://registry.npmjs.org/tar/-/tar-0.1.20.tgz</a></p>
<p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_36a63706-cb38-4759-9162-9c49ba99f793/20190625175906_98125/git_depth_0/AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/package.json</p>
<p>Path to vulnerable library: /tmp/WhiteSource-ArchiveExtractor_36a63706-cb38-4759-9162-9c49ba99f793/20190625175906_98125/git_depth_0/AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- intern-geezer-2.2.3.tgz (Root Library)
- digdug-1.4.0.tgz
- decompress-0.2.3.tgz
- :x: **tar-0.1.20.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/valdisiljuconoks/AlloyTech/commits/365b59135886dba1f0332ba419b8693e0e68566f">365b59135886dba1f0332ba419b8693e0e68566f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of node-tar prior to 4.4.2 are vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink will overwrite the system's file with the contents of the extracted file.
<p>Publish Date: 2019-04-05
<p>URL: <a href=https://github.com/npm/node-tar/commit/b0c58433c22f5e7fe8b1c76373f27e3f81dcd4c8>WS-2019-0047</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/803">https://www.npmjs.com/advisories/803</a></p>
<p>Release Date: 2019-04-05</p>
<p>Fix Resolution: 4.4.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0047 (Medium) detected in tar-0.1.20.tgz - autoclosed - ## WS-2019-0047 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-0.1.20.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-0.1.20.tgz">https://registry.npmjs.org/tar/-/tar-0.1.20.tgz</a></p>
<p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_36a63706-cb38-4759-9162-9c49ba99f793/20190625175906_98125/git_depth_0/AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/package.json</p>
<p>Path to vulnerable library: /tmp/WhiteSource-ArchiveExtractor_36a63706-cb38-4759-9162-9c49ba99f793/20190625175906_98125/git_depth_0/AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- intern-geezer-2.2.3.tgz (Root Library)
- digdug-1.4.0.tgz
- decompress-0.2.3.tgz
- :x: **tar-0.1.20.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/valdisiljuconoks/AlloyTech/commits/365b59135886dba1f0332ba419b8693e0e68566f">365b59135886dba1f0332ba419b8693e0e68566f</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of node-tar prior to 4.4.2 are vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink will overwrite the system's file with the contents of the extracted file.
<p>Publish Date: 2019-04-05
<p>URL: <a href=https://github.com/npm/node-tar/commit/b0c58433c22f5e7fe8b1c76373f27e3f81dcd4c8>WS-2019-0047</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/803">https://www.npmjs.com/advisories/803</a></p>
<p>Release Date: 2019-04-05</p>
<p>Fix Resolution: 4.4.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws medium detected in tar tgz autoclosed ws medium severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file tmp whitesource archiveextractor git depth alloytech modules protected shell shell clientresources lib xstyle package json path to vulnerable library tmp whitesource archiveextractor git depth alloytech modules protected shell shell clientresources lib xstyle node modules tar package json dependency hierarchy intern geezer tgz root library digdug tgz decompress tgz x tar tgz vulnerable library found in head commit a href vulnerability details versions of node tar prior to are vulnerable to arbitrary file overwrite extracting tarballs containing a hardlink to a file that already exists in the system and a file that matches the hardlink will overwrite the system s file with the contents of the extracted file publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
59,566
| 14,422,008,211
|
IssuesEvent
|
2020-12-05 01:03:16
|
jgeraigery/blueocean-environments
|
https://api.github.com/repos/jgeraigery/blueocean-environments
|
opened
|
WS-2019-0427 (Medium) detected in elliptic-6.4.0.tgz
|
security vulnerability
|
## WS-2019-0427 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p>
<p>Path to dependency file: blueocean-environments/package.json</p>
<p>Path to vulnerable library: blueocean-environments/node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- browserify-sign-4.0.4.tgz (Root Library)
- :x: **elliptic-6.4.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2
<p>Publish Date: 2019-11-22
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0427</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a">https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a</a></p>
<p>Release Date: 2020-05-24</p>
<p>Fix Resolution: v6.5.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.4.0","isTransitiveDependency":true,"dependencyTree":"browserify-sign:4.0.4;elliptic:6.4.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v6.5.2"}],"vulnerabilityIdentifier":"WS-2019-0427","vulnerabilityDetails":"The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2","vulnerabilityUrl":"https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
|
True
|
WS-2019-0427 (Medium) detected in elliptic-6.4.0.tgz - ## WS-2019-0427 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p>
<p>Path to dependency file: blueocean-environments/package.json</p>
<p>Path to vulnerable library: blueocean-environments/node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- browserify-sign-4.0.4.tgz (Root Library)
- :x: **elliptic-6.4.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2
<p>Publish Date: 2019-11-22
<p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0427</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a">https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a</a></p>
<p>Release Date: 2020-05-24</p>
<p>Fix Resolution: v6.5.2</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.4.0","isTransitiveDependency":true,"dependencyTree":"browserify-sign:4.0.4;elliptic:6.4.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v6.5.2"}],"vulnerabilityIdentifier":"WS-2019-0427","vulnerabilityDetails":"The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2","vulnerabilityUrl":"https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
ws medium detected in elliptic tgz ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file blueocean environments package json path to vulnerable library blueocean environments node modules elliptic package json dependency hierarchy browserify sign tgz root library x elliptic tgz vulnerable library vulnerability details the function getnaf in elliptic library has information leakage this issue is mitigated in version publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails the function getnaf in elliptic library has information leakage this issue is mitigated in version vulnerabilityurl
| 0
|
6,043
| 2,806,340,278
|
IssuesEvent
|
2015-05-15 01:18:43
|
GoogleCloudPlatform/kubernetes
|
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
|
opened
|
kubectl e2e tests broken
|
component/CLI priority/P1 team/testing
|
Several of the kubectl e2e tests were broken today by #8249, which added a missing check of an err return value, all of which now fail when doing cleanup and calling `kubectl stop`.
Example:
```
kubectl guestbook should create and stop a working application
/go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/test/e2e/kubectl.go:125
Error running &{/jenkins-master-data/jobs/kubernetes-e2e-gce/workspace/kubernetes/hack/../cluster/../platforms/linux/amd64/kubectl [kubectl --server=https://104.197.99.179 --kubeconfig=/var/lib/jenkins/jobs/kubernetes-e2e-gce/workspace/.kube/config stop -f examples/guestbook] [] <nil> Error: [Resizing the controller failed with: replicationControllers "frontend" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-master" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-slave" not found; Current resource version Unknown]
[] <nil> 0xc2083e6860 exit status 1 <nil> true [0xc20803a3f8 0xc20803a438 0xc20803a470] [0xc20803a3f8 0xc20803a438 0xc20803a470] [0xc20803a430 0xc20803a468] [0x657730 0x657730] 0xc20841cc60}:
Command stdout:
stderr:
Error: [Resizing the controller failed with: replicationControllers "frontend" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-master" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-slave" not found; Current resource version Unknown]
```
I think this may have been originally broken in #6194. That change refactored the implementation of Stop, which started masking the error, but it also changed the order of code a bit.
Previously, the Stop implementation in `pkg/kubectl/stop.go` first queried for ReplicationControllers before creating the ResizeCondition; the first thing in the method was
```
rc := reaper.ReplicationControllers(namespace)
controller, err := rc.Get(name)
```
After the change, this is done following the ResizeCondition in `pkg/kubectl/resize.go`:
```
rc := &api.ReplicationController{ObjectMeta: api.ObjectMeta{Namespace: namespace, Name: name}}
```
Is it possible that we're trying to query for something that's already been deleted, and thus it will always fail?
cc @quinton-hoole @roberthbailey
|
1.0
|
kubectl e2e tests broken - Several of the kubectl e2e tests were broken today by #8249, which added a missing check of an err return value, all of which now fail when doing cleanup and calling `kubectl stop`.
Example:
```
kubectl guestbook should create and stop a working application
/go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/test/e2e/kubectl.go:125
Error running &{/jenkins-master-data/jobs/kubernetes-e2e-gce/workspace/kubernetes/hack/../cluster/../platforms/linux/amd64/kubectl [kubectl --server=https://104.197.99.179 --kubeconfig=/var/lib/jenkins/jobs/kubernetes-e2e-gce/workspace/.kube/config stop -f examples/guestbook] [] <nil> Error: [Resizing the controller failed with: replicationControllers "frontend" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-master" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-slave" not found; Current resource version Unknown]
[] <nil> 0xc2083e6860 exit status 1 <nil> true [0xc20803a3f8 0xc20803a438 0xc20803a470] [0xc20803a3f8 0xc20803a438 0xc20803a470] [0xc20803a430 0xc20803a468] [0x657730 0x657730] 0xc20841cc60}:
Command stdout:
stderr:
Error: [Resizing the controller failed with: replicationControllers "frontend" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-master" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-slave" not found; Current resource version Unknown]
```
I think this may have been originally broken in #6194. That change refactored the implementation of Stop, which started masking the error, but it also changed the order of code a bit.
Previously, the Stop implementation in `pkg/kubectl/stop.go` first queried for ReplicationControllers before creating the ResizeCondition; the first thing in the method was
```
rc := reaper.ReplicationControllers(namespace)
controller, err := rc.Get(name)
```
After the change, this is done following the ResizeCondition in `pkg/kubectl/resize.go`:
```
rc := &api.ReplicationController{ObjectMeta: api.ObjectMeta{Namespace: namespace, Name: name}}
```
Is it possible that we're trying to query for something that's already been deleted, and thus it will always fail?
cc @quinton-hoole @roberthbailey
|
non_process
|
kubectl tests broken several of the kubectl tests were broken today by which added a missing check of an err return value all of which now fail when doing cleanup and calling kubectl stop example kubectl guestbook should create and stop a working application go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes test kubectl go error running jenkins master data jobs kubernetes gce workspace kubernetes hack cluster platforms linux kubectl error exit status true command stdout stderr error i think this may have been originally broken in that change refactored the implementation of stop which started masking the error but it also changed the order of code a bit previously the stop implementation in pkg kubectl stop go first queried for replicationcontrollers before creating the resizecondition the first thing in the method was rc reaper replicationcontrollers namespace controller err rc get name after the change this is done following the resizecondition in pkg kubectl resize go rc api replicationcontroller objectmeta api objectmeta namespace namespace name name is it possible that we re trying to query for something that s already been deleted and thus it will always fail cc quinton hoole roberthbailey
| 0
|
46,549
| 19,293,888,270
|
IssuesEvent
|
2021-12-12 08:51:52
|
amirshnll/rtbf.ir
|
https://api.github.com/repos/amirshnll/rtbf.ir
|
closed
|
اکانتی شاپ
|
new website/service
|
"name" : "اکانتی شاپ"
"website": "https://accountishop.ir/"
"deleteurl": "#"
"info": "امکان حذف حساب کاربری وجود ندارد"
|
1.0
|
اکانتی شاپ - "name" : "اکانتی شاپ"
"website": "https://accountishop.ir/"
"deleteurl": "#"
"info": "امکان حذف حساب کاربری وجود ندارد"
|
non_process
|
اکانتی شاپ name اکانتی شاپ website deleteurl info امکان حذف حساب کاربری وجود ندارد
| 0
|
17,712
| 23,609,520,819
|
IssuesEvent
|
2022-08-24 11:10:30
|
nucleartide/Lets-Make-a-Roguelike-Deckbuilder
|
https://api.github.com/repos/nucleartide/Lets-Make-a-Roguelike-Deckbuilder
|
closed
|
Add a "share" step to process doc for an appropriate stopping point
|
process
|
See [Notion](https://www.notion.so/nucleartide/Game-Dev-1a8d6e5960b14adabc66e7b79d620b1a) for personal notes on action items
|
1.0
|
Add a "share" step to process doc for an appropriate stopping point - See [Notion](https://www.notion.so/nucleartide/Game-Dev-1a8d6e5960b14adabc66e7b79d620b1a) for personal notes on action items
|
process
|
add a share step to process doc for an appropriate stopping point see for personal notes on action items
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.