Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
15,185
18,955,159,595
IssuesEvent
2021-11-18 19:20:28
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
opened
unpin probot for owl-bot
type: process bot: owl-bot
#2912 added a pin on probot dependency to `12.1.0` because `12.1.1` gives me a compilation error.
1.0
unpin probot for owl-bot - #2912 added a pin on probot dependency to `12.1.0` because `12.1.1` gives me a compilation error.
process
unpin probot for owl bot added a pin on probot dependency to because gives me a compilation error
1
4,532
2,559,371,772
IssuesEvent
2015-02-05 00:11:17
tvkanters/Dopestreamer
https://api.github.com/repos/tvkanters/Dopestreamer
opened
Add viewer count
enhancement low-priority
There should be an viewer count in the main stream which sums the Hitbox, Twitch, Vacker live and Vacker restream viewcounts.
1.0
Add viewer count - There should be an viewer count in the main stream which sums the Hitbox, Twitch, Vacker live and Vacker restream viewcounts.
non_process
add viewer count there should be an viewer count in the main stream which sums the hitbox twitch vacker live and vacker restream viewcounts
0
6,765
9,888,280,189
IssuesEvent
2019-06-25 11:09:59
linnovate/root
https://api.github.com/repos/linnovate/root
closed
search office :can not click folder and signature
2.0.7 Fixed Process bug Search
go to offices open new office go to search and search for this office Click on the office, and then click Folder or Signature result : clicking on a signature or folder does not work go to offices open new office: ![image](https://user-images.githubusercontent.com/47353222/59250781-e6340b00-8c30-11e9-9fd2-9d2c7d3d8a9d.png) go to search and search for this office ![image](https://user-images.githubusercontent.com/47353222/59251014-75412300-8c31-11e9-92c2-8c6b35c82de7.png) Click on the office, and then click Folder or Signature ![image](https://user-images.githubusercontent.com/47353222/59250947-44f98480-8c31-11e9-87b7-9d0762aaf22a.png)
1.0
search office :can not click folder and signature - go to offices open new office go to search and search for this office Click on the office, and then click Folder or Signature result : clicking on a signature or folder does not work go to offices open new office: ![image](https://user-images.githubusercontent.com/47353222/59250781-e6340b00-8c30-11e9-9fd2-9d2c7d3d8a9d.png) go to search and search for this office ![image](https://user-images.githubusercontent.com/47353222/59251014-75412300-8c31-11e9-92c2-8c6b35c82de7.png) Click on the office, and then click Folder or Signature ![image](https://user-images.githubusercontent.com/47353222/59250947-44f98480-8c31-11e9-87b7-9d0762aaf22a.png)
process
search office can not click folder and signature go to offices open new office go to search and search for this office click on the office and then click folder or signature result clicking on a signature or folder does not work go to offices open new office go to search and search for this office click on the office and then click folder or signature
1
81,509
23,479,804,250
IssuesEvent
2022-08-17 09:33:20
godotengine/godot-proposals
https://api.github.com/repos/godotengine/godot-proposals
reopened
Tree-shaking compiler to reduce size and load time of exported games
archived topic:buildsystem
### Describe the project you are working on Many small HTML5 and iOS/Android games, where it's crucial to have fast download and load times. ### Describe the problem or limitation you are having in your project Godot has an ever-growing bundle of amazing functionality, which is hugely helpful for making games. However, most of this functionality is not used, leading to bloated game bundles that take longer for the player to download. For instance, even the simplest HTML5 games made with Godot need 15 MB for the engine, whereas the game logic and assets might only be 2-3 MB. These simple games can also take 10+ seconds to load on old devices because the device has to load a lot of unneeded functionality into memory. This will likely become an increasingly important issue as more functionality is added to Godot. ### Describe the feature / enhancement and how it helps to overcome the problem or limitation Add [tree-shaking](https://en.wikipedia.org/wiki/Tree_shaking) functionality to the compiler so only the part of the Godot engine that are actually used by the game are included. This would dramatically reduce size and load times. ### Describe how your proposal will work, with code, pseudo-code, mock-ups, and/or diagrams This would work similar to [webpack's tree-shaking](https://webpack.js.org/guides/tree-shaking/) in that Godot would identify which functions are being used and remove the rest from the exported bundle. This is similar to [disabling 3D mentioned in the docs](https://docs.godotengine.org/en/stable/development/compiling/optimizing_for_size.html#disabling-3d), but more refined and with much bigger benefit because it would disable lots of other unused functionality besides 3D. That being said, it might be cool to just do a simple initial version that just automatically detects if 3D is used and disables it otherwise. Implementing tree-shaking would likely take quite a bit of effort, but would improve the bundle size and load time of every single game exported with Godot (there’s no game that uses all the functionality). ### If this enhancement will not be used often, can it be worked around with a few lines of script? No. ### Is there a reason why this should be core and not an add-on in the asset library? AFAIK this would only be possible to do in core.
1.0
Tree-shaking compiler to reduce size and load time of exported games - ### Describe the project you are working on Many small HTML5 and iOS/Android games, where it's crucial to have fast download and load times. ### Describe the problem or limitation you are having in your project Godot has an ever-growing bundle of amazing functionality, which is hugely helpful for making games. However, most of this functionality is not used, leading to bloated game bundles that take longer for the player to download. For instance, even the simplest HTML5 games made with Godot need 15 MB for the engine, whereas the game logic and assets might only be 2-3 MB. These simple games can also take 10+ seconds to load on old devices because the device has to load a lot of unneeded functionality into memory. This will likely become an increasingly important issue as more functionality is added to Godot. ### Describe the feature / enhancement and how it helps to overcome the problem or limitation Add [tree-shaking](https://en.wikipedia.org/wiki/Tree_shaking) functionality to the compiler so only the part of the Godot engine that are actually used by the game are included. This would dramatically reduce size and load times. ### Describe how your proposal will work, with code, pseudo-code, mock-ups, and/or diagrams This would work similar to [webpack's tree-shaking](https://webpack.js.org/guides/tree-shaking/) in that Godot would identify which functions are being used and remove the rest from the exported bundle. This is similar to [disabling 3D mentioned in the docs](https://docs.godotengine.org/en/stable/development/compiling/optimizing_for_size.html#disabling-3d), but more refined and with much bigger benefit because it would disable lots of other unused functionality besides 3D. That being said, it might be cool to just do a simple initial version that just automatically detects if 3D is used and disables it otherwise. Implementing tree-shaking would likely take quite a bit of effort, but would improve the bundle size and load time of every single game exported with Godot (there’s no game that uses all the functionality). ### If this enhancement will not be used often, can it be worked around with a few lines of script? No. ### Is there a reason why this should be core and not an add-on in the asset library? AFAIK this would only be possible to do in core.
non_process
tree shaking compiler to reduce size and load time of exported games describe the project you are working on many small and ios android games where it s crucial to have fast download and load times describe the problem or limitation you are having in your project godot has an ever growing bundle of amazing functionality which is hugely helpful for making games however most of this functionality is not used leading to bloated game bundles that take longer for the player to download for instance even the simplest games made with godot need mb for the engine whereas the game logic and assets might only be mb these simple games can also take seconds to load on old devices because the device has to load a lot of unneeded functionality into memory this will likely become an increasingly important issue as more functionality is added to godot describe the feature enhancement and how it helps to overcome the problem or limitation add functionality to the compiler so only the part of the godot engine that are actually used by the game are included this would dramatically reduce size and load times describe how your proposal will work with code pseudo code mock ups and or diagrams this would work similar to in that godot would identify which functions are being used and remove the rest from the exported bundle this is similar to but more refined and with much bigger benefit because it would disable lots of other unused functionality besides that being said it might be cool to just do a simple initial version that just automatically detects if is used and disables it otherwise implementing tree shaking would likely take quite a bit of effort but would improve the bundle size and load time of every single game exported with godot there’s no game that uses all the functionality if this enhancement will not be used often can it be worked around with a few lines of script no is there a reason why this should be core and not an add on in the asset library afaik this would only be possible to do in core
0
649,766
21,320,384,885
IssuesEvent
2022-04-17 01:25:57
WeaponMechanics/MechanicsMain
https://api.github.com/repos/WeaponMechanics/MechanicsMain
closed
Add Vanilla Command Arguments/Validation
priority: low will add working on it
### Link to code https://github.com/WeaponMechanics/MechanicsMain/tree/master/MechanicsCore/src/main/java/me/deecaad/core/commands ### Related Issues _No response_ ### Improvements Lets say you are making a CSGO or Valorant or similar server. You need to give weapons to each member of a team. You cannot do this using `/wm give` since it doesn't accept command arguments like `@a[type=PLAYER,team=blue]`. However, in vanilla MC, commands handle this automatically. MC also shows validation through colors, making command usage more responsive on the fly. To handle this, you need an NMS based command api. Existing: * https://github.com/JorelAli/CommandAPI * https://github.com/lucko/commodore
1.0
Add Vanilla Command Arguments/Validation - ### Link to code https://github.com/WeaponMechanics/MechanicsMain/tree/master/MechanicsCore/src/main/java/me/deecaad/core/commands ### Related Issues _No response_ ### Improvements Lets say you are making a CSGO or Valorant or similar server. You need to give weapons to each member of a team. You cannot do this using `/wm give` since it doesn't accept command arguments like `@a[type=PLAYER,team=blue]`. However, in vanilla MC, commands handle this automatically. MC also shows validation through colors, making command usage more responsive on the fly. To handle this, you need an NMS based command api. Existing: * https://github.com/JorelAli/CommandAPI * https://github.com/lucko/commodore
non_process
add vanilla command arguments validation link to code related issues no response improvements lets say you are making a csgo or valorant or similar server you need to give weapons to each member of a team you cannot do this using wm give since it doesn t accept command arguments like a however in vanilla mc commands handle this automatically mc also shows validation through colors making command usage more responsive on the fly to handle this you need an nms based command api existing
0
296,191
25,535,616,349
IssuesEvent
2022-11-29 11:44:54
ToolJet/ToolJet
https://api.github.com/repos/ToolJet/ToolJet
closed
Add data-cy for table column edit options.
test cypress
### Specify the kind of test <!-- Provide the kind of test --> Cypress E2E ### Describe the test <!-- Provide a clear description of the test --> Add data -cy to help test cases for column edit options.
1.0
Add data-cy for table column edit options. - ### Specify the kind of test <!-- Provide the kind of test --> Cypress E2E ### Describe the test <!-- Provide a clear description of the test --> Add data -cy to help test cases for column edit options.
non_process
add data cy for table column edit options specify the kind of test provide the kind of test cypress describe the test provide a clear description of the test add data cy to help test cases for column edit options
0
7,139
10,281,285,005
IssuesEvent
2019-08-26 08:07:07
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Add the "-overwrite" flag to the "clip raster by mask" GDAL tool
Feature Request Processing
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When atempting to overwrite a previous raster (in this case) using clip raster by mask, it is not over written even when QGIS prompts you to overwrite. **How to Reproduce** Raster > Extraction > Clip Raster by Mask layer Save the file to a directory Go through the same process, save as same file name, and agree to overwrite the file. Run clipping process again. See error message below. <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error --> **QGIS and OS versions** QGIS: 3.4.10-Madeira OS: Ubuntu Xenial GDAL/OGR: 2.2.2 <!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here --> **Additional context** <!-- Add any other context about the problem here. --> Error Message: ``` Clip Raster by Mask Layer ERROR 1: Output dataset /filename_clipped.tif exists, but some command line options were provided indicating a new dataset should be created. Please delete existing dataset and run again. ``` I think here when running the tool the ```-overwrite``` needs to be specified?
1.0
Add the "-overwrite" flag to the "clip raster by mask" GDAL tool - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** <!-- A clear and concise description of what the bug is. --> When atempting to overwrite a previous raster (in this case) using clip raster by mask, it is not over written even when QGIS prompts you to overwrite. **How to Reproduce** Raster > Extraction > Clip Raster by Mask layer Save the file to a directory Go through the same process, save as same file name, and agree to overwrite the file. Run clipping process again. See error message below. <!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error --> **QGIS and OS versions** QGIS: 3.4.10-Madeira OS: Ubuntu Xenial GDAL/OGR: 2.2.2 <!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here --> **Additional context** <!-- Add any other context about the problem here. --> Error Message: ``` Clip Raster by Mask Layer ERROR 1: Output dataset /filename_clipped.tif exists, but some command line options were provided indicating a new dataset should be created. Please delete existing dataset and run again. ``` I think here when running the tool the ```-overwrite``` needs to be specified?
process
add the overwrite flag to the clip raster by mask gdal tool bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug when atempting to overwrite a previous raster in this case using clip raster by mask it is not over written even when qgis prompts you to overwrite how to reproduce raster extraction clip raster by mask layer save the file to a directory go through the same process save as same file name and agree to overwrite the file run clipping process again see error message below steps sample datasets and qgis project file to reproduce the behavior screencasts or screenshots welcome go to click on scroll down to see error qgis and os versions qgis madeira os ubuntu xenial gdal ogr additional context error message clip raster by mask layer error output dataset filename clipped tif exists but some command line options were provided indicating a new dataset should be created please delete existing dataset and run again i think here when running the tool the overwrite needs to be specified
1
10,810
13,609,288,898
IssuesEvent
2020-09-23 04:50:19
googleapis/java-accessapproval
https://api.github.com/repos/googleapis/java-accessapproval
closed
Dependency Dashboard
api: accessapproval type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-accessapproval-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-accessapproval to v1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-accessapproval-1.x -->chore(deps): update dependency com.google.cloud:google-cloud-accessapproval to v1 - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any chore deps update dependency com google cloud google cloud accessapproval to chore deps update dependency com google cloud libraries bom to check this box to trigger a request for renovate to run again on this repository
1
5,870
8,691,508,758
IssuesEvent
2018-12-04 01:40:39
knative/serving
https://api.github.com/repos/knative/serving
closed
We should only use knative-releases for public released artifacts
area/test-and-release kind/feature kind/process
<!-- Pro-tip: You can leave this block commented, and it still works! Select the appropriate areas for your issue: /area test-and-release Classify what kind of issue this is: /kind feature /kind process /assign @adrcunha --> ## Expected Behavior We should knative-releases bucket exclusively for public releases (vX.Y.Z releases) to avoid accident cleanups of nightly and other ephemeral releases. Nightly should be published to knative-nightly bucket. ## Actual Behavior Nightly is currently published to knative-releases GCS/GCR buckets.
1.0
We should only use knative-releases for public released artifacts - <!-- Pro-tip: You can leave this block commented, and it still works! Select the appropriate areas for your issue: /area test-and-release Classify what kind of issue this is: /kind feature /kind process /assign @adrcunha --> ## Expected Behavior We should knative-releases bucket exclusively for public releases (vX.Y.Z releases) to avoid accident cleanups of nightly and other ephemeral releases. Nightly should be published to knative-nightly bucket. ## Actual Behavior Nightly is currently published to knative-releases GCS/GCR buckets.
process
we should only use knative releases for public released artifacts pro tip you can leave this block commented and it still works select the appropriate areas for your issue area test and release classify what kind of issue this is kind feature kind process assign adrcunha expected behavior we should knative releases bucket exclusively for public releases vx y z releases to avoid accident cleanups of nightly and other ephemeral releases nightly should be published to knative nightly bucket actual behavior nightly is currently published to knative releases gcs gcr buckets
1
432,333
30,278,937,578
IssuesEvent
2023-07-07 23:12:07
houghj16/ShareX
https://api.github.com/repos/houghj16/ShareX
opened
Update the routing for the OneDrive
documentation
#3 #6 @houghj16 Description is very important ```[tasklist] ### Tasks - [ ] #7 ```
1.0
Update the routing for the OneDrive - #3 #6 @houghj16 Description is very important ```[tasklist] ### Tasks - [ ] #7 ```
non_process
update the routing for the onedrive description is very important tasks
0
173,030
14,399,550,475
IssuesEvent
2020-12-03 11:05:01
kubernetes-sigs/external-dns
https://api.github.com/repos/kubernetes-sigs/external-dns
closed
documentation for fast dns changed
kind/bug kind/documentation
<!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks! --> **What happened**: Documentation for dast dns changes as it become lagacy so old links in https://github.com/kubernetes-sigs/external-dns/blob/master/docs/tutorials/akamai-fastdns.md broken **What you expected to happen**: Fastdns documentation shoud redirect to correct documentation page
1.0
documentation for fast dns changed - <!-- Please use this template while reporting a bug and provide as much info as possible. Not doing so may result in your bug not being addressed in a timely manner. Thanks! --> **What happened**: Documentation for dast dns changes as it become lagacy so old links in https://github.com/kubernetes-sigs/external-dns/blob/master/docs/tutorials/akamai-fastdns.md broken **What you expected to happen**: Fastdns documentation shoud redirect to correct documentation page
non_process
documentation for fast dns changed please use this template while reporting a bug and provide as much info as possible not doing so may result in your bug not being addressed in a timely manner thanks what happened documentation for dast dns changes as it become lagacy so old links in broken what you expected to happen fastdns documentation shoud redirect to correct documentation page
0
302,930
26,174,182,533
IssuesEvent
2023-01-02 07:15:55
BoBAdministration/QA-Bug-Reports
https://api.github.com/repos/BoBAdministration/QA-Bug-Reports
closed
Tapping E makes you auto drink
Fixed-PendingTesting
**Describe the Bug** When you tap E (or use, whatever keybind you have) at a water source you can infinitely drink. **To Reproduce** 1. Logged onto a test server 2. Go on any creture 3. Go to any water source 4. Tap the button to drink, not hold **Expected behavior** Tapping the use button while trying to drink should make it that you only drink for a second **Actual behavior** Tapping the use button at water makes you drink forever **Screenshots & Video** Showed Pred on stream **Branch Version** Tester and Live **Additional Information** I first thought you could drain ponds with the bug but turns out it only takes water until you're full water/sat but you stay in the drinking animation, so not as huge of an issue in that case.
1.0
Tapping E makes you auto drink - **Describe the Bug** When you tap E (or use, whatever keybind you have) at a water source you can infinitely drink. **To Reproduce** 1. Logged onto a test server 2. Go on any creture 3. Go to any water source 4. Tap the button to drink, not hold **Expected behavior** Tapping the use button while trying to drink should make it that you only drink for a second **Actual behavior** Tapping the use button at water makes you drink forever **Screenshots & Video** Showed Pred on stream **Branch Version** Tester and Live **Additional Information** I first thought you could drain ponds with the bug but turns out it only takes water until you're full water/sat but you stay in the drinking animation, so not as huge of an issue in that case.
non_process
tapping e makes you auto drink describe the bug when you tap e or use whatever keybind you have at a water source you can infinitely drink to reproduce logged onto a test server go on any creture go to any water source tap the button to drink not hold expected behavior tapping the use button while trying to drink should make it that you only drink for a second actual behavior tapping the use button at water makes you drink forever screenshots video showed pred on stream branch version tester and live additional information i first thought you could drain ponds with the bug but turns out it only takes water until you re full water sat but you stay in the drinking animation so not as huge of an issue in that case
0
8,378
11,525,777,586
IssuesEvent
2020-02-15 10:55:57
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Make it possible to start a process suspened, and later resume it.
api-suggestion area-System.Diagnostics.Process
This would make it a lot easier to (for example) attach it to a job.
1.0
Make it possible to start a process suspened, and later resume it. - This would make it a lot easier to (for example) attach it to a job.
process
make it possible to start a process suspened and later resume it this would make it a lot easier to for example attach it to a job
1
80,989
15,613,862,545
IssuesEvent
2021-03-19 17:00:02
CliMA/RRTMGP.jl
https://api.github.com/repos/CliMA/RRTMGP.jl
closed
Add code coverage back in
code quality
It looks like code-coverage is no longer being reported, and we should add this back in.
1.0
Add code coverage back in - It looks like code-coverage is no longer being reported, and we should add this back in.
non_process
add code coverage back in it looks like code coverage is no longer being reported and we should add this back in
0
172,153
13,263,670,777
IssuesEvent
2020-08-21 01:16:58
omegaup/omegaup
https://api.github.com/repos/omegaup/omegaup
closed
[FEATURE] Hacer obligatorio el campo de lenguaje cuando creas un concurso
UX Task feature-request omegaUp for Contests
En https://omegaup.com/contest/new/: * Por default todos los lenguajes deben de estar seleccionados. * Si el usuario desmarca todos los lenguajes se le muestra un error.
1.0
[FEATURE] Hacer obligatorio el campo de lenguaje cuando creas un concurso - En https://omegaup.com/contest/new/: * Por default todos los lenguajes deben de estar seleccionados. * Si el usuario desmarca todos los lenguajes se le muestra un error.
non_process
hacer obligatorio el campo de lenguaje cuando creas un concurso en por default todos los lenguajes deben de estar seleccionados si el usuario desmarca todos los lenguajes se le muestra un error
0
506,413
14,664,483,700
IssuesEvent
2020-12-29 12:06:56
eXpandFramework/eXpand
https://api.github.com/repos/eXpandFramework/eXpand
closed
How can I debug this?
Priority Question ❤ Backer
From time to time, I found this exception in production environment (web app), any help to debug it? System.ArgumentException: Nombre de tipo duplicado en un ensamblado. en System.Reflection.Emit.ModuleBuilder.CheckTypeNameConflict(String strTypeName, Type enclosingType) en System.Reflection.Emit.AssemblyBuilderData.CheckTypeNameConflict(String strTypeName, TypeBuilder enclosingType) en System.Reflection.Emit.TypeBuilder.Init(String fullname, TypeAttributes attr, Type parent, Type[] interfaces, ModuleBuilder module, PackingSize iPackingSize, Int32 iTypeSize, TypeBuilder enclosingType) en System.Reflection.Emit.ModuleBuilder.DefineType(String name, TypeAttributes attr, Type parent) en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.NewControllerType[T](String id) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 170 en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.RegisterAction[TController,TAction](ApplicationModulesManager applicationModulesManager, String id, Func`2 actionBase) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 147 en Xpand.XAF.Modules.PositionInListview.SwapPositionInListViewService.<>c__DisplayClass4_0.b__0(String actionnId) en D:\a\1\s\src\Modules\PositionInListview\SwapPositionInListViewService.cs:línea 33 en System.Reactive.Linq.ObservableImpl.SelectMany`2.ObservableSelector._.OnNext(TSource value) WEB Void CheckTypeNameConflict(System.String, System.Type) en System.Reflection.Emit.ModuleBuilder.CheckTypeNameConflict(String strTypeName, Type enclosingType) en System.Reflection.Emit.AssemblyBuilderData.CheckTypeNameConflict(String strTypeName, TypeBuilder enclosingType) en System.Reflection.Emit.TypeBuilder.Init(String fullname, TypeAttributes attr, Type parent, Type[] interfaces, ModuleBuilder module, PackingSize iPackingSize, Int32 iTypeSize, TypeBuilder enclosingType) en System.Reflection.Emit.ModuleBuilder.DefineType(String name, TypeAttributes attr, Type parent) en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.NewControllerType[T](String id) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 170 en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.RegisterAction[TController,TAction](ApplicationModulesManager applicationModulesManager, String id, Func`2 actionBase) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 147 en Xpand.XAF.Modules.PositionInListview.SwapPositionInListViewService.<>c__DisplayClass4_0.b__0(String actionnId) en D:\a\1\s\src\Modules\PositionInListview\SwapPositionInListViewService.cs:línea 33 en System.Reactive.Linq.ObservableImpl.SelectMany`2.ObservableSelector._.OnNext(TSource value)
1.0
How can I debug this? - From time to time, I found this exception in production environment (web app), any help to debug it? System.ArgumentException: Nombre de tipo duplicado en un ensamblado. en System.Reflection.Emit.ModuleBuilder.CheckTypeNameConflict(String strTypeName, Type enclosingType) en System.Reflection.Emit.AssemblyBuilderData.CheckTypeNameConflict(String strTypeName, TypeBuilder enclosingType) en System.Reflection.Emit.TypeBuilder.Init(String fullname, TypeAttributes attr, Type parent, Type[] interfaces, ModuleBuilder module, PackingSize iPackingSize, Int32 iTypeSize, TypeBuilder enclosingType) en System.Reflection.Emit.ModuleBuilder.DefineType(String name, TypeAttributes attr, Type parent) en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.NewControllerType[T](String id) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 170 en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.RegisterAction[TController,TAction](ApplicationModulesManager applicationModulesManager, String id, Func`2 actionBase) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 147 en Xpand.XAF.Modules.PositionInListview.SwapPositionInListViewService.<>c__DisplayClass4_0.b__0(String actionnId) en D:\a\1\s\src\Modules\PositionInListview\SwapPositionInListViewService.cs:línea 33 en System.Reactive.Linq.ObservableImpl.SelectMany`2.ObservableSelector._.OnNext(TSource value) WEB Void CheckTypeNameConflict(System.String, System.Type) en System.Reflection.Emit.ModuleBuilder.CheckTypeNameConflict(String strTypeName, Type enclosingType) en System.Reflection.Emit.AssemblyBuilderData.CheckTypeNameConflict(String strTypeName, TypeBuilder enclosingType) en System.Reflection.Emit.TypeBuilder.Init(String fullname, TypeAttributes attr, Type parent, Type[] interfaces, ModuleBuilder module, PackingSize iPackingSize, Int32 iTypeSize, TypeBuilder enclosingType) en System.Reflection.Emit.ModuleBuilder.DefineType(String name, TypeAttributes attr, Type parent) en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.NewControllerType[T](String id) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 170 en Xpand.XAF.Modules.Reactive.Services.Actions.ActionsService.RegisterAction[TController,TAction](ApplicationModulesManager applicationModulesManager, String id, Func`2 actionBase) en D:\a\1\s\src\Modules\Reactive\Services\Actions\RegisterAction.cs:línea 147 en Xpand.XAF.Modules.PositionInListview.SwapPositionInListViewService.<>c__DisplayClass4_0.b__0(String actionnId) en D:\a\1\s\src\Modules\PositionInListview\SwapPositionInListViewService.cs:línea 33 en System.Reactive.Linq.ObservableImpl.SelectMany`2.ObservableSelector._.OnNext(TSource value)
non_process
how can i debug this from time to time i found this exception in production environment web app any help to debug it system argumentexception nombre de tipo duplicado en un ensamblado en system reflection emit modulebuilder checktypenameconflict string strtypename type enclosingtype en system reflection emit assemblybuilderdata checktypenameconflict string strtypename typebuilder enclosingtype en system reflection emit typebuilder init string fullname typeattributes attr type parent type interfaces modulebuilder module packingsize ipackingsize itypesize typebuilder enclosingtype en system reflection emit modulebuilder definetype string name typeattributes attr type parent en xpand xaf modules reactive services actions actionsservice newcontrollertype string id en d a s src modules reactive services actions registeraction cs línea en xpand xaf modules reactive services actions actionsservice registeraction applicationmodulesmanager applicationmodulesmanager string id func actionbase en d a s src modules reactive services actions registeraction cs línea en xpand xaf modules positioninlistview swappositioninlistviewservice c b string actionnid en d a s src modules positioninlistview swappositioninlistviewservice cs línea en system reactive linq observableimpl selectmany observableselector onnext tsource value web void checktypenameconflict system string system type en system reflection emit modulebuilder checktypenameconflict string strtypename type enclosingtype en system reflection emit assemblybuilderdata checktypenameconflict string strtypename typebuilder enclosingtype en system reflection emit typebuilder init string fullname typeattributes attr type parent type interfaces modulebuilder module packingsize ipackingsize itypesize typebuilder enclosingtype en system reflection emit modulebuilder definetype string name typeattributes attr type parent en xpand xaf modules reactive services actions actionsservice newcontrollertype string id en d a s src modules reactive services actions registeraction cs línea en xpand xaf modules reactive services actions actionsservice registeraction applicationmodulesmanager applicationmodulesmanager string id func actionbase en d a s src modules reactive services actions registeraction cs línea en xpand xaf modules positioninlistview swappositioninlistviewservice c b string actionnid en d a s src modules positioninlistview swappositioninlistviewservice cs línea en system reactive linq observableimpl selectmany observableselector onnext tsource value
0
12,550
14,976,333,471
IssuesEvent
2021-01-28 07:53:00
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Audit Logs] "studyVersion" is displayed null for the event in Response module
Bug P2 Process: Fixed Response datastore
Event: STUDY_METADATA_RECEIVED ![105966351-8a8aef80-60aa-11eb-8fd5-7bc18ddee4de](https://user-images.githubusercontent.com/60386291/105994561-c71c1280-60cd-11eb-8832-658967f60043.png)
1.0
[Audit Logs] "studyVersion" is displayed null for the event in Response module - Event: STUDY_METADATA_RECEIVED ![105966351-8a8aef80-60aa-11eb-8fd5-7bc18ddee4de](https://user-images.githubusercontent.com/60386291/105994561-c71c1280-60cd-11eb-8832-658967f60043.png)
process
studyversion is displayed null for the event in response module event study metadata received
1
11,640
14,496,618,192
IssuesEvent
2020-12-11 13:04:05
didi/mpx
https://api.github.com/repos/didi/mpx
closed
[Bug report]wx:model-value-path=""报错
processing
![image](https://user-images.githubusercontent.com/8046366/101130650-d12a1380-363e-11eb-9153-7cfa525c3019.png) ![image](https://user-images.githubusercontent.com/8046366/101130666-d8e9b800-363e-11eb-81dc-e9e8b566352d.png) ![image](https://user-images.githubusercontent.com/8046366/101130673-dc7d3f00-363e-11eb-9621-a96892ebaa26.png) 可以发现valuepath 这里为true 调用reduce就报错了,wx:model-value-path="[]" 没问题的 另外建议文档-api里 wx:model-value-path的链一下对应指南里的部分
1.0
[Bug report]wx:model-value-path=""报错 - ![image](https://user-images.githubusercontent.com/8046366/101130650-d12a1380-363e-11eb-9153-7cfa525c3019.png) ![image](https://user-images.githubusercontent.com/8046366/101130666-d8e9b800-363e-11eb-81dc-e9e8b566352d.png) ![image](https://user-images.githubusercontent.com/8046366/101130673-dc7d3f00-363e-11eb-9621-a96892ebaa26.png) 可以发现valuepath 这里为true 调用reduce就报错了,wx:model-value-path="[]" 没问题的 另外建议文档-api里 wx:model-value-path的链一下对应指南里的部分
process
wx model value path 报错 可以发现valuepath 这里为true 调用reduce就报错了 wx model value path 没问题的 另外建议文档 api里 wx model value path的链一下对应指南里的部分
1
287,675
8,818,181,568
IssuesEvent
2018-12-31 09:36:41
Veil-Project/veil
https://api.github.com/repos/Veil-Project/veil
opened
Daemon Crashes and corrupts wallet while mining.
core high priority wallet
ok you can get the debug log here https://veil.suprnova.cc/debug.log.2 and the wallet here https://veil.suprnova.cc/wallet.dat.2 just try to start the daemon with this wallet and it won't work it will go to 100% cpu usage and simply do nothing the only obvious errors i see in the logs are 2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error 2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error 2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error 2018-12-31T07:18:12Z UpdateTip: new best=f74862d1fbbd29087c606b7d7ea0333045d4015b52ab30cc2c7999c1f3a597cd height=9702 version=0x20000000 log2_work=60.34344 tx=16463 date='2018-12-31T06:34:48Z' progress=0.999463 cache=0.4MiB(2315txo) 2018-12-31T07:18:12Z ERROR: FindTx: txid mismatch 2018-12-31T07:18:12Z ERROR: FindTx: txid mismatch
1.0
Daemon Crashes and corrupts wallet while mining. - ok you can get the debug log here https://veil.suprnova.cc/debug.log.2 and the wallet here https://veil.suprnova.cc/wallet.dat.2 just try to start the daemon with this wallet and it won't work it will go to 100% cpu usage and simply do nothing the only obvious errors i see in the logs are 2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error 2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error 2018-12-31T07:18:12Z ERROR: FindTx: Deserialize or I/O error - ReadCompactSize(): size too large: iostream error 2018-12-31T07:18:12Z UpdateTip: new best=f74862d1fbbd29087c606b7d7ea0333045d4015b52ab30cc2c7999c1f3a597cd height=9702 version=0x20000000 log2_work=60.34344 tx=16463 date='2018-12-31T06:34:48Z' progress=0.999463 cache=0.4MiB(2315txo) 2018-12-31T07:18:12Z ERROR: FindTx: txid mismatch 2018-12-31T07:18:12Z ERROR: FindTx: txid mismatch
non_process
daemon crashes and corrupts wallet while mining ok you can get the debug log here and the wallet here just try to start the daemon with this wallet and it won t work it will go to cpu usage and simply do nothing the only obvious errors i see in the logs are error findtx deserialize or i o error readcompactsize size too large iostream error error findtx deserialize or i o error readcompactsize size too large iostream error error findtx deserialize or i o error readcompactsize size too large iostream error updatetip new best height version work tx date progress cache error findtx txid mismatch error findtx txid mismatch
0
12,273
3,061,890,203
IssuesEvent
2015-08-16 01:12:01
oppia/oppia
https://api.github.com/repos/oppia/oppia
closed
Bring the "featured exploration" flow within the site
feature: important ref: frontend/editor TODO: design doc
``` What steps will reproduce the problem? 1. Create a new exploration and publish it. 2. Click on "Nominate for featured status". What is the expected output? What do you see instead? A modal pops up and says "please write to this forum". This seems like too much hassle and is a bit of a weird flow. Instead it would be nicer for the nomination to be recorded in the moderator queue, and an email to be automatically sent to moderators, when someone clicks the button. There should also be a mechanism for the person who clicked the button to get a reply from the moderator. ``` Original issue reported on code.google.com by `s...@seanlip.org` on 1 Dec 2014 at 5:02
1.0
Bring the "featured exploration" flow within the site - ``` What steps will reproduce the problem? 1. Create a new exploration and publish it. 2. Click on "Nominate for featured status". What is the expected output? What do you see instead? A modal pops up and says "please write to this forum". This seems like too much hassle and is a bit of a weird flow. Instead it would be nicer for the nomination to be recorded in the moderator queue, and an email to be automatically sent to moderators, when someone clicks the button. There should also be a mechanism for the person who clicked the button to get a reply from the moderator. ``` Original issue reported on code.google.com by `s...@seanlip.org` on 1 Dec 2014 at 5:02
non_process
bring the featured exploration flow within the site what steps will reproduce the problem create a new exploration and publish it click on nominate for featured status what is the expected output what do you see instead a modal pops up and says please write to this forum this seems like too much hassle and is a bit of a weird flow instead it would be nicer for the nomination to be recorded in the moderator queue and an email to be automatically sent to moderators when someone clicks the button there should also be a mechanism for the person who clicked the button to get a reply from the moderator original issue reported on code google com by s seanlip org on dec at
0
7,370
10,512,610,703
IssuesEvent
2019-09-27 18:21:40
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Bug: Experience dates display incorrectly on application review page
Apply Process Bug State Dept.
Environment: Production Issue: Work experience start and end dates are displaying incorrectly on the application review page Steps to reproduce: Jot down work experience start and end dates in USAJOBS Apply for an internship in Open Opps Go to the review page of your application and the start and end dates display one month off Related ticket: 3947
1.0
Bug: Experience dates display incorrectly on application review page - Environment: Production Issue: Work experience start and end dates are displaying incorrectly on the application review page Steps to reproduce: Jot down work experience start and end dates in USAJOBS Apply for an internship in Open Opps Go to the review page of your application and the start and end dates display one month off Related ticket: 3947
process
bug experience dates display incorrectly on application review page environment production issue work experience start and end dates are displaying incorrectly on the application review page steps to reproduce jot down work experience start and end dates in usajobs apply for an internship in open opps go to the review page of your application and the start and end dates display one month off related ticket
1
5,341
8,167,601,147
IssuesEvent
2018-08-26 01:10:17
MobileOrg/mobileorg
https://api.github.com/repos/MobileOrg/mobileorg
opened
Fastlane tools
development process
Use fastlane.tools for build/release to automate more. Releasing to testflight should be fully automated. * https://fastlane.tools/
1.0
Fastlane tools - Use fastlane.tools for build/release to automate more. Releasing to testflight should be fully automated. * https://fastlane.tools/
process
fastlane tools use fastlane tools for build release to automate more releasing to testflight should be fully automated
1
5,444
8,306,330,490
IssuesEvent
2018-09-22 17:33:11
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
turn off undo for process buffer
process
I always get a warning when running loops that print something: `Warning (undo): Buffer `*julia*' undo info was 12560552 bytes long. The undo info was discarded because it exceeded `undo-outer-limit'.` I cannot think of a case where undoing something in the process buffer would be useful. So why not disable undo there completely? This also speeds things up if printing a lot into the REPL. This can easily be done by `(add-hook 'inferior-ess-mode-hook 'buffer-disable-undo)` in the `.emacs`, but I think, there should be an option/documentation in ESS to do this automatically for process buffers.
1.0
turn off undo for process buffer - I always get a warning when running loops that print something: `Warning (undo): Buffer `*julia*' undo info was 12560552 bytes long. The undo info was discarded because it exceeded `undo-outer-limit'.` I cannot think of a case where undoing something in the process buffer would be useful. So why not disable undo there completely? This also speeds things up if printing a lot into the REPL. This can easily be done by `(add-hook 'inferior-ess-mode-hook 'buffer-disable-undo)` in the `.emacs`, but I think, there should be an option/documentation in ESS to do this automatically for process buffers.
process
turn off undo for process buffer i always get a warning when running loops that print something warning undo buffer julia undo info was bytes long the undo info was discarded because it exceeded undo outer limit i cannot think of a case where undoing something in the process buffer would be useful so why not disable undo there completely this also speeds things up if printing a lot into the repl this can easily be done by add hook inferior ess mode hook buffer disable undo in the emacs but i think there should be an option documentation in ess to do this automatically for process buffers
1
1,184
3,687,011,074
IssuesEvent
2016-02-25 05:22:41
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Keyscope + branch filtering potential bug
bug needs-reproduction P2 preprocess/filtering
Related to this discussion I had no the DITA specs list: https://lists.oasis-open.org/archives/dita-comment/201602/msg00001.html So in one place of the DITA Map I have: <topichead navtitle="Test" collection-type="choice"> <ditavalref href="topics/getting_started/install_windows.ditaval"> <ditavalmeta> <dvrResourcePrefix>windows-</dvrResourcePrefix> <dvrKeyscopePrefix>windows_scope-</dvrKeyscopePrefix> </ditavalmeta> </ditavalref> <topicref href="topics/install.dita" keys="topicref_install" keyscope="scope_install"/> </topichead> and in some other place I have a keyref like: <topicref keyref="windows_scope-scope_install.topicref_install"/> From what I understand the keyref is invalid because the "windows_scope-" prefix must only be applied on the @keyscope attribute specified on the topichead in which the ditavalref is placed. But the DITA OT 2.x processes the keyref as valid and finds the target topic. As the @keyscope is missing on the topichead, according to the specs the @keyscope on the topichead will become "windows-" so the proper way to have the keyref would be: <topicref keyref="windows_scope-.scope_install_select_parallel.topicref_install_select_parallel"/> which currently does not work using DITA OT 2.x. What's your opinion on this?
1.0
Keyscope + branch filtering potential bug - Related to this discussion I had no the DITA specs list: https://lists.oasis-open.org/archives/dita-comment/201602/msg00001.html So in one place of the DITA Map I have: <topichead navtitle="Test" collection-type="choice"> <ditavalref href="topics/getting_started/install_windows.ditaval"> <ditavalmeta> <dvrResourcePrefix>windows-</dvrResourcePrefix> <dvrKeyscopePrefix>windows_scope-</dvrKeyscopePrefix> </ditavalmeta> </ditavalref> <topicref href="topics/install.dita" keys="topicref_install" keyscope="scope_install"/> </topichead> and in some other place I have a keyref like: <topicref keyref="windows_scope-scope_install.topicref_install"/> From what I understand the keyref is invalid because the "windows_scope-" prefix must only be applied on the @keyscope attribute specified on the topichead in which the ditavalref is placed. But the DITA OT 2.x processes the keyref as valid and finds the target topic. As the @keyscope is missing on the topichead, according to the specs the @keyscope on the topichead will become "windows-" so the proper way to have the keyref would be: <topicref keyref="windows_scope-.scope_install_select_parallel.topicref_install_select_parallel"/> which currently does not work using DITA OT 2.x. What's your opinion on this?
process
keyscope branch filtering potential bug related to this discussion i had no the dita specs list so in one place of the dita map i have windows windows scope topicref href topics install dita keys topicref install keyscope scope install and in some other place i have a keyref like topicref keyref windows scope scope install topicref install from what i understand the keyref is invalid because the windows scope prefix must only be applied on the keyscope attribute specified on the topichead in which the ditavalref is placed but the dita ot x processes the keyref as valid and finds the target topic as the keyscope is missing on the topichead according to the specs the keyscope on the topichead will become windows so the proper way to have the keyref would be topicref keyref windows scope scope install select parallel topicref install select parallel which currently does not work using dita ot x what s your opinion on this
1
180,557
13,937,210,133
IssuesEvent
2020-10-22 13:53:32
root-project/root
https://api.github.com/repos/root-project/root
closed
Unable to install pytest on MacOS with python2
bug in:Testing
We removed the pytest shipped with roottest because the source code was from 2014 and incompatible with py3.9 (see #6597). However, this poses now the issue on MacOs with python2 that we have to install pytest. Without a virtual environment, MacOS does not allow to pip packages. Since roottest fails on configuration level without pytest, roottest is currently broken in this configuration. @axel @oshadura What should we do? Our CI always runs roottest against python3, so we currently don't see the issue in our infrastructure. I see three options: 1. Ditch testing of python2 on MacOS and rely on the test coverage of other platforms (python2 is anyway dead) 2. Use a venv overlay in Jenkins for the MacOS nodes (haven't tested but it should work and is binary compatible with the system python) 3. We change the hard failure of roottest to a soft failure.
1.0
Unable to install pytest on MacOS with python2 - We removed the pytest shipped with roottest because the source code was from 2014 and incompatible with py3.9 (see #6597). However, this poses now the issue on MacOs with python2 that we have to install pytest. Without a virtual environment, MacOS does not allow to pip packages. Since roottest fails on configuration level without pytest, roottest is currently broken in this configuration. @axel @oshadura What should we do? Our CI always runs roottest against python3, so we currently don't see the issue in our infrastructure. I see three options: 1. Ditch testing of python2 on MacOS and rely on the test coverage of other platforms (python2 is anyway dead) 2. Use a venv overlay in Jenkins for the MacOS nodes (haven't tested but it should work and is binary compatible with the system python) 3. We change the hard failure of roottest to a soft failure.
non_process
unable to install pytest on macos with we removed the pytest shipped with roottest because the source code was from and incompatible with see however this poses now the issue on macos with that we have to install pytest without a virtual environment macos does not allow to pip packages since roottest fails on configuration level without pytest roottest is currently broken in this configuration axel oshadura what should we do our ci always runs roottest against so we currently don t see the issue in our infrastructure i see three options ditch testing of on macos and rely on the test coverage of other platforms is anyway dead use a venv overlay in jenkins for the macos nodes haven t tested but it should work and is binary compatible with the system python we change the hard failure of roottest to a soft failure
0
537,762
15,736,848,622
IssuesEvent
2021-03-30 01:34:35
musescore/MuseScore
https://api.github.com/repos/musescore/MuseScore
opened
[MU4 Issue] Scores should open on first page even if previous score was closed on a different page
Low Priority
**Describe the bug** If user closes score on the second or upper page and then open another score, it will be opened from those page which was opened on a previous closed score **To Reproduce** Steps to reproduce the behavior: 1. Create a score with at least 2 pages 2. Close the score on the 2-nd page or higher 3. Open new score with at least 2 pages > score will be opened on the 2-nd page **Expected behavior** Scores should open on first page even if previous score was closed on a different page **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** MacOS **Additional context** Add any other context about the problem here.
1.0
[MU4 Issue] Scores should open on first page even if previous score was closed on a different page - **Describe the bug** If user closes score on the second or upper page and then open another score, it will be opened from those page which was opened on a previous closed score **To Reproduce** Steps to reproduce the behavior: 1. Create a score with at least 2 pages 2. Close the score on the 2-nd page or higher 3. Open new score with at least 2 pages > score will be opened on the 2-nd page **Expected behavior** Scores should open on first page even if previous score was closed on a different page **Screenshots** If applicable, add screenshots to help explain your problem. **Desktop (please complete the following information):** MacOS **Additional context** Add any other context about the problem here.
non_process
scores should open on first page even if previous score was closed on a different page describe the bug if user closes score on the second or upper page and then open another score it will be opened from those page which was opened on a previous closed score to reproduce steps to reproduce the behavior create a score with at least pages close the score on the nd page or higher open new score with at least pages score will be opened on the nd page expected behavior scores should open on first page even if previous score was closed on a different page screenshots if applicable add screenshots to help explain your problem desktop please complete the following information macos additional context add any other context about the problem here
0
14,681
17,797,906,223
IssuesEvent
2021-09-01 02:01:23
Leviatan-Analytics/LA-data-processing
https://api.github.com/repos/Leviatan-Analytics/LA-data-processing
closed
Improve ward detection model [5]
Data Processing Week 4 Sprint 3
Re label images and find ways to improve ward detection model metrics: - Processing time - Model accuracy
1.0
Improve ward detection model [5] - Re label images and find ways to improve ward detection model metrics: - Processing time - Model accuracy
process
improve ward detection model re label images and find ways to improve ward detection model metrics processing time model accuracy
1
69,026
7,122,171,911
IssuesEvent
2018-01-19 10:50:38
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
rancher-cli install creates empty stack without services
area/cli status/resolved status/to-test
I've a local rancher installation. I've created an evironment and a API Key for it. I also have a custom catalog. If I now try to create a new stack using the following command line: `` rancher catalog install customcat/testitem:1.0 --name test `` the stack is successfully created, but does not contain any services. If I do the same using the GUI I get the expected services in my stack. Using --debug doesn't give any useful additional informations and the logs are empty either. Am I missing something? Rancher-CLI Version: 0.6.2 (current version offered as download) --- | Useful | Info | | :-- | :-- | |Versions|Rancher `v1.6.0` Cattle: `v0.179.7` UI: `v1.6.1` Rancher-CLI: `v0.6.2`| |Access|`ldap` `admin`| |Orchestration|`Cattle`| |Route|`stacks.index`|
1.0
rancher-cli install creates empty stack without services - I've a local rancher installation. I've created an evironment and a API Key for it. I also have a custom catalog. If I now try to create a new stack using the following command line: `` rancher catalog install customcat/testitem:1.0 --name test `` the stack is successfully created, but does not contain any services. If I do the same using the GUI I get the expected services in my stack. Using --debug doesn't give any useful additional informations and the logs are empty either. Am I missing something? Rancher-CLI Version: 0.6.2 (current version offered as download) --- | Useful | Info | | :-- | :-- | |Versions|Rancher `v1.6.0` Cattle: `v0.179.7` UI: `v1.6.1` Rancher-CLI: `v0.6.2`| |Access|`ldap` `admin`| |Orchestration|`Cattle`| |Route|`stacks.index`|
non_process
rancher cli install creates empty stack without services i ve a local rancher installation i ve created an evironment and a api key for it i also have a custom catalog if i now try to create a new stack using the following command line rancher catalog install customcat testitem name test the stack is successfully created but does not contain any services if i do the same using the gui i get the expected services in my stack using debug doesn t give any useful additional informations and the logs are empty either am i missing something rancher cli version current version offered as download useful info versions rancher cattle ui rancher cli access ldap admin orchestration cattle route stacks index
0
9,594
12,543,045,137
IssuesEvent
2020-06-05 14:56:21
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Improve validation message for @@unique
kind/improvement process/candidate topic: errors
```prisma // Specify a multi-field unique attribute that includes a relation field model Post { id Int @default(autoincrement()) author User @relation(fields: [authorId], references: [id]) authorId Int title String published Boolean @default(false) @@unique([author, title]) } model User { id Int @id @default(autoincrement()) email String @unique posts Post[] } ``` Current message for this schema `Error validating model "Post": The unique index definition refers to the relation fields author. Index definitions must reference only scalar fields.` Here the fix is to replace `@@unique([author, title])` by `@@unique([authorId, title])` so the message could mention it maybe? Discussion with @do4gr https://prisma-company.slack.com/archives/C5Z9TH6N9/p1591348814019400
1.0
Improve validation message for @@unique - ```prisma // Specify a multi-field unique attribute that includes a relation field model Post { id Int @default(autoincrement()) author User @relation(fields: [authorId], references: [id]) authorId Int title String published Boolean @default(false) @@unique([author, title]) } model User { id Int @id @default(autoincrement()) email String @unique posts Post[] } ``` Current message for this schema `Error validating model "Post": The unique index definition refers to the relation fields author. Index definitions must reference only scalar fields.` Here the fix is to replace `@@unique([author, title])` by `@@unique([authorId, title])` so the message could mention it maybe? Discussion with @do4gr https://prisma-company.slack.com/archives/C5Z9TH6N9/p1591348814019400
process
improve validation message for unique prisma specify a multi field unique attribute that includes a relation field model post id int default autoincrement author user relation fields references authorid int title string published boolean default false unique model user id int id default autoincrement email string unique posts post current message for this schema error validating model post the unique index definition refers to the relation fields author index definitions must reference only scalar fields here the fix is to replace unique by unique so the message could mention it maybe discussion with
1
89,670
18,019,568,097
IssuesEvent
2021-09-16 17:36:22
WordPress/openverse-frontend
https://api.github.com/repos/WordPress/openverse-frontend
closed
[Bug] Managing playback of multiple media files
🟧 priority: high 🛠 goal: fix 💻 aspect: code
## Description <!-- Concisely describe the bug. --> The current setup allows for multiple audio files to be played concurrently, which is a bad user experience. ## Reproduction <!-- Provide detailed steps to reproduce the bug. --> 1. View any page with multiple audio players 2. Press play on multiple audio players 3. Listen to the resulting 'chaos orchestra' ## Expectation <!-- Concisely describe what you expected to happen. --> When pressing 'play' on an audio file, if there is _already_ an active audio file it should be paused. ## Screenshots <!-- Add screenshots to show the problem; or delete the section entirely. --> ## Resolution <!-- Replace the [ ] with [x] to check the box. --> I have proposed a solution in #183. - [ ] 🙋 I would be interested in resolving this bug.
1.0
[Bug] Managing playback of multiple media files - ## Description <!-- Concisely describe the bug. --> The current setup allows for multiple audio files to be played concurrently, which is a bad user experience. ## Reproduction <!-- Provide detailed steps to reproduce the bug. --> 1. View any page with multiple audio players 2. Press play on multiple audio players 3. Listen to the resulting 'chaos orchestra' ## Expectation <!-- Concisely describe what you expected to happen. --> When pressing 'play' on an audio file, if there is _already_ an active audio file it should be paused. ## Screenshots <!-- Add screenshots to show the problem; or delete the section entirely. --> ## Resolution <!-- Replace the [ ] with [x] to check the box. --> I have proposed a solution in #183. - [ ] 🙋 I would be interested in resolving this bug.
non_process
managing playback of multiple media files description the current setup allows for multiple audio files to be played concurrently which is a bad user experience reproduction view any page with multiple audio players press play on multiple audio players listen to the resulting chaos orchestra expectation when pressing play on an audio file if there is already an active audio file it should be paused screenshots resolution i have proposed a solution in 🙋 i would be interested in resolving this bug
0
7,806
10,960,891,835
IssuesEvent
2019-11-27 14:25:50
codeuniversity/smag-mvp
https://api.github.com/repos/codeuniversity/smag-mvp
opened
Figure out favorite bands of users and find associated photos
Image Processing
As part of the interests, it would be nice to find musicians the user is interested and get representative pictures of them.
1.0
Figure out favorite bands of users and find associated photos - As part of the interests, it would be nice to find musicians the user is interested and get representative pictures of them.
process
figure out favorite bands of users and find associated photos as part of the interests it would be nice to find musicians the user is interested and get representative pictures of them
1
57,596
14,163,858,424
IssuesEvent
2020-11-12 03:26:55
woocommerce/woocommerce-admin
https://api.github.com/repos/woocommerce/woocommerce-admin
closed
e2e Testing: Set up Puppeteer master issue
Build [Type] Task [estimate] 13
End to end (e2e) testing automates user flows of navigating apps by simulating clicks and selections. Add Puppeteer infrastructure and tests so that we can check that reports, pages, and filtering are functioning as they should. Why Puppeteer? Core will eventually migrate in that direction (p7bje6-1ne-p2). ### Tasks - [x] Implement initial Puppeteer architecture #4343 - [x] Write a simple test to test Puppeteer config #4343 - [x] Integrate [Gutenberg’s WP util functions package](https://github.com/WordPress/gutenberg/tree/master/packages/e2e-test-utils) #4343 - [x] Write documentation on e2e test suite config #4343 - [ ] Create baseline tests for each report and page. Ensure elements are loading correctly. - [ ] Set up Testing in different browsers - [x] Integrate with Travis CI #4343 - [ ] Integrate with Slack/email/(something else?) to deliver a notice with screenshot of failed test - [ ] Identify and create tests for complex flows, ie filtering or settings manipulation.
1.0
e2e Testing: Set up Puppeteer master issue - End to end (e2e) testing automates user flows of navigating apps by simulating clicks and selections. Add Puppeteer infrastructure and tests so that we can check that reports, pages, and filtering are functioning as they should. Why Puppeteer? Core will eventually migrate in that direction (p7bje6-1ne-p2). ### Tasks - [x] Implement initial Puppeteer architecture #4343 - [x] Write a simple test to test Puppeteer config #4343 - [x] Integrate [Gutenberg’s WP util functions package](https://github.com/WordPress/gutenberg/tree/master/packages/e2e-test-utils) #4343 - [x] Write documentation on e2e test suite config #4343 - [ ] Create baseline tests for each report and page. Ensure elements are loading correctly. - [ ] Set up Testing in different browsers - [x] Integrate with Travis CI #4343 - [ ] Integrate with Slack/email/(something else?) to deliver a notice with screenshot of failed test - [ ] Identify and create tests for complex flows, ie filtering or settings manipulation.
non_process
testing set up puppeteer master issue end to end testing automates user flows of navigating apps by simulating clicks and selections add puppeteer infrastructure and tests so that we can check that reports pages and filtering are functioning as they should why puppeteer core will eventually migrate in that direction tasks implement initial puppeteer architecture write a simple test to test puppeteer config integrate write documentation on test suite config create baseline tests for each report and page ensure elements are loading correctly set up testing in different browsers integrate with travis ci integrate with slack email something else to deliver a notice with screenshot of failed test identify and create tests for complex flows ie filtering or settings manipulation
0
182,691
21,673,922,027
IssuesEvent
2022-05-08 12:05:29
turkdevops/vscode
https://api.github.com/repos/turkdevops/vscode
closed
WS-2018-0069 (High) detected in is-my-json-valid-2.16.1.tgz - autoclosed
security vulnerability
## WS-2018-0069 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.16.1.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.1.tgz</a></p> <p>Path to dependency file: /extensions/emmet/package.json</p> <p>Path to vulnerable library: /extensions/emmet/node_modules/is-my-json-valid/package.json,/extensions/vscode-api-tests/node_modules/is-my-json-valid/package.json,/extensions/vscode-colorize-tests/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - vscode-1.1.5.tgz (Root Library) - gulp-remote-src-0.4.3.tgz - request-2.79.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.16.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/vscode/commit/9fd6b056a06e14655f4f0b0f631d670b24878828">9fd6b056a06e14655f4f0b0f631d670b24878828</a></p> <p>Found in base branch: <b>webview-views</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function. <p>Publish Date: 2018-02-14 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976>WS-2018-0069</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/572">https://nodesecurity.io/advisories/572</a></p> <p>Release Date: 2018-02-14</p> <p>Fix Resolution (is-my-json-valid): 2.17.2</p> <p>Direct dependency fix Resolution (vscode): 1.1.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2018-0069 (High) detected in is-my-json-valid-2.16.1.tgz - autoclosed - ## WS-2018-0069 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.16.1.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.1.tgz</a></p> <p>Path to dependency file: /extensions/emmet/package.json</p> <p>Path to vulnerable library: /extensions/emmet/node_modules/is-my-json-valid/package.json,/extensions/vscode-api-tests/node_modules/is-my-json-valid/package.json,/extensions/vscode-colorize-tests/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - vscode-1.1.5.tgz (Root Library) - gulp-remote-src-0.4.3.tgz - request-2.79.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.16.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/turkdevops/vscode/commit/9fd6b056a06e14655f4f0b0f631d670b24878828">9fd6b056a06e14655f4f0b0f631d670b24878828</a></p> <p>Found in base branch: <b>webview-views</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function. <p>Publish Date: 2018-02-14 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976>WS-2018-0069</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/572">https://nodesecurity.io/advisories/572</a></p> <p>Release Date: 2018-02-14</p> <p>Fix Resolution (is-my-json-valid): 2.17.2</p> <p>Direct dependency fix Resolution (vscode): 1.1.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in is my json valid tgz autoclosed ws high severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file extensions emmet package json path to vulnerable library extensions emmet node modules is my json valid package json extensions vscode api tests node modules is my json valid package json extensions vscode colorize tests node modules is my json valid package json dependency hierarchy vscode tgz root library gulp remote src tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch webview views vulnerability details version of is my json valid before or are vulnerable to regular expression denial of service redos via the email validation function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid direct dependency fix resolution vscode step up your open source security game with whitesource
0
11,100
13,129,646,588
IssuesEvent
2020-08-06 14:12:45
sekiguchi-nagisa/ydsh
https://api.github.com/repos/sekiguchi-nagisa/ydsh
closed
prune redundant paths in source globbing
Core incompatible change
glob in source statement matches redundant paths when having ``..`` patterns as the following ```sh > echo /*r*/../r* /proc/../root /proc/../run /run/../root /run/../run /srv/../root /srv/../run /usr/../root /usr/../run /var/../root /var/../run ``` this is correct behavior in posix shell, but in source statement, only require unique match results.
True
prune redundant paths in source globbing - glob in source statement matches redundant paths when having ``..`` patterns as the following ```sh > echo /*r*/../r* /proc/../root /proc/../run /run/../root /run/../run /srv/../root /srv/../run /usr/../root /usr/../run /var/../root /var/../run ``` this is correct behavior in posix shell, but in source statement, only require unique match results.
non_process
prune redundant paths in source globbing glob in source statement matches redundant paths when having patterns as the following sh echo r r proc root proc run run root run run srv root srv run usr root usr run var root var run this is correct behavior in posix shell but in source statement only require unique match results
0
1,367
3,925,094,232
IssuesEvent
2016-04-22 17:37:49
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
opened
NTR: ‘(positive/negative) regulation of right ventricular blood pressure'
BHF-UCL miRNA New term request RNA processes
Dear Editors, I have recently been annotating a paper (PMID:22161164), which relied on the measurement of right systolic ventricular pressure (RSVP) as an indicator of the lung (pulmonary) blood pressure. Basic scientists as well as clinicians use the measurement of RVBP (e.g. in PMID:22161164) to estimate lung blood pressure. For instance, normal human RVBP and lung blood pressure are usually both close to 20/15 mmHg. However, although in normal circumstances RVBP and lung blood pressure will give the same numerical value, they are not the same pressures. RVBP is a function of the volume of the blood in the RV and the resistance of the ventricular walls, therefore as ventricles contract (systole), the pressure inside them will increase. However, until RVBP is sufficiently high to open the valve and allow the blood to start flowing into the pulmonary artery, the pressure in the artery will remain lower than in the RV. Whereas once the stroke of blood has been ejected into the pulmonary artery, the pressure in this artery (and then the lungs) will be higher than in the RV, which will have just emptied. In view of the above, we have had several group discussions about whether we should request a new term: ‘(positive/negative) regulation of right ventricular blood pressure (RVBP)’ as a child term of ‘regulation of blood pressure’. It would additionally be helpful to add a comment in QiuckGO to this ‘regulation of RVBP’ term to clarify that scientists and clinicians typically use RVSP as a measurement of the lung/pulmonary blood pressure, as the measurements of these pressures are typically the same. Nonetheless, regulation of RVSP is not synonymous with the regulation of lung/pulmonary blood pressure. The same comment could also be added to the already existing ‘regulation of lung blood pressure’ terms. Perhaps alternatively (or in addition), ‘regulation of RVBP’ could have a ‘part of’ relationship with ‘regulation of lung blood pressure’, as it is such a major factor influencing the latter? The blood pressure (positive and negative) regulation terms, which currently exist in GO include: Id: GO:0008217 Name: regulation of blood pressure Id: GO:0003073 Name: regulation of systemic arterial blood pressure Is_a: GO:0008217 regulation of blood pressure Id: GO:0014916 Name: regulation of lung blood pressure Is_a: GO:0008217 regulation of blood pressure Synonym: regulation of pulmonary blood pressure (exact) (Link to ancestor chart: https://www.ebi.ac.uk/QuickGO/GMultiTerm#a=64%2400m1020P03f4&tab=chart&c=&chartTab=settings) Unlike the systemic circulation, the pulmonary circulation is a low-pressure system, therefore there are no distinct differences between the right ventricular blood pressure (RVBP) and the pulmonary (lung) pressure in terms of measurements (both are ~20/15 mmHg in human). However, they are not synonymous. And additional factors, which do not affect RVBP (e.g. artery elasticity), will affect the lung blood pressure. Therefore, although the annotated paper (PMID:22161164) focused on pulmonary hypertension, we decided that ‘regulation of lung blood pressure’ was too far downstream to annotate to this term, based on the presented data and the measurements of RVBP. Instead, we annotated the entity rno-miR-17-5p to positive regulation of blood pressure (GO:0045777) IMP occurs_in UBERON:0002080 heart right ventricle. However, having RVBP regulation terms for this type of annotations would have been very helpful. I will look forward to hearing from you. Thank you, Barbara GOC:BHF_miRNA GOC:BHF GOC:bc @rachhuntley @RLovering
1.0
NTR: ‘(positive/negative) regulation of right ventricular blood pressure' - Dear Editors, I have recently been annotating a paper (PMID:22161164), which relied on the measurement of right systolic ventricular pressure (RSVP) as an indicator of the lung (pulmonary) blood pressure. Basic scientists as well as clinicians use the measurement of RVBP (e.g. in PMID:22161164) to estimate lung blood pressure. For instance, normal human RVBP and lung blood pressure are usually both close to 20/15 mmHg. However, although in normal circumstances RVBP and lung blood pressure will give the same numerical value, they are not the same pressures. RVBP is a function of the volume of the blood in the RV and the resistance of the ventricular walls, therefore as ventricles contract (systole), the pressure inside them will increase. However, until RVBP is sufficiently high to open the valve and allow the blood to start flowing into the pulmonary artery, the pressure in the artery will remain lower than in the RV. Whereas once the stroke of blood has been ejected into the pulmonary artery, the pressure in this artery (and then the lungs) will be higher than in the RV, which will have just emptied. In view of the above, we have had several group discussions about whether we should request a new term: ‘(positive/negative) regulation of right ventricular blood pressure (RVBP)’ as a child term of ‘regulation of blood pressure’. It would additionally be helpful to add a comment in QiuckGO to this ‘regulation of RVBP’ term to clarify that scientists and clinicians typically use RVSP as a measurement of the lung/pulmonary blood pressure, as the measurements of these pressures are typically the same. Nonetheless, regulation of RVSP is not synonymous with the regulation of lung/pulmonary blood pressure. The same comment could also be added to the already existing ‘regulation of lung blood pressure’ terms. Perhaps alternatively (or in addition), ‘regulation of RVBP’ could have a ‘part of’ relationship with ‘regulation of lung blood pressure’, as it is such a major factor influencing the latter? The blood pressure (positive and negative) regulation terms, which currently exist in GO include: Id: GO:0008217 Name: regulation of blood pressure Id: GO:0003073 Name: regulation of systemic arterial blood pressure Is_a: GO:0008217 regulation of blood pressure Id: GO:0014916 Name: regulation of lung blood pressure Is_a: GO:0008217 regulation of blood pressure Synonym: regulation of pulmonary blood pressure (exact) (Link to ancestor chart: https://www.ebi.ac.uk/QuickGO/GMultiTerm#a=64%2400m1020P03f4&tab=chart&c=&chartTab=settings) Unlike the systemic circulation, the pulmonary circulation is a low-pressure system, therefore there are no distinct differences between the right ventricular blood pressure (RVBP) and the pulmonary (lung) pressure in terms of measurements (both are ~20/15 mmHg in human). However, they are not synonymous. And additional factors, which do not affect RVBP (e.g. artery elasticity), will affect the lung blood pressure. Therefore, although the annotated paper (PMID:22161164) focused on pulmonary hypertension, we decided that ‘regulation of lung blood pressure’ was too far downstream to annotate to this term, based on the presented data and the measurements of RVBP. Instead, we annotated the entity rno-miR-17-5p to positive regulation of blood pressure (GO:0045777) IMP occurs_in UBERON:0002080 heart right ventricle. However, having RVBP regulation terms for this type of annotations would have been very helpful. I will look forward to hearing from you. Thank you, Barbara GOC:BHF_miRNA GOC:BHF GOC:bc @rachhuntley @RLovering
process
ntr ‘ positive negative regulation of right ventricular blood pressure dear editors i have recently been annotating a paper pmid which relied on the measurement of right systolic ventricular pressure rsvp as an indicator of the lung pulmonary blood pressure basic scientists as well as clinicians use the measurement of rvbp e g in pmid to estimate lung blood pressure for instance normal human rvbp and lung blood pressure are usually both close to mmhg however although in normal circumstances rvbp and lung blood pressure will give the same numerical value they are not the same pressures rvbp is a function of the volume of the blood in the rv and the resistance of the ventricular walls therefore as ventricles contract systole the pressure inside them will increase however until rvbp is sufficiently high to open the valve and allow the blood to start flowing into the pulmonary artery the pressure in the artery will remain lower than in the rv whereas once the stroke of blood has been ejected into the pulmonary artery the pressure in this artery and then the lungs will be higher than in the rv which will have just emptied in view of the above we have had several group discussions about whether we should request a new term ‘ positive negative regulation of right ventricular blood pressure rvbp ’ as a child term of ‘regulation of blood pressure’ it would additionally be helpful to add a comment in qiuckgo to this ‘regulation of rvbp’ term to clarify that scientists and clinicians typically use rvsp as a measurement of the lung pulmonary blood pressure as the measurements of these pressures are typically the same nonetheless regulation of rvsp is not synonymous with the regulation of lung pulmonary blood pressure the same comment could also be added to the already existing ‘regulation of lung blood pressure’ terms perhaps alternatively or in addition ‘regulation of rvbp’ could have a ‘part of’ relationship with ‘regulation of lung blood pressure’ as it is such a major factor influencing the latter the blood pressure positive and negative regulation terms which currently exist in go include id go name regulation of blood pressure id go name regulation of systemic arterial blood pressure is a go regulation of blood pressure id go name regulation of lung blood pressure is a go regulation of blood pressure synonym regulation of pulmonary blood pressure exact link to ancestor chart unlike the systemic circulation the pulmonary circulation is a low pressure system therefore there are no distinct differences between the right ventricular blood pressure rvbp and the pulmonary lung pressure in terms of measurements both are mmhg in human however they are not synonymous and additional factors which do not affect rvbp e g artery elasticity will affect the lung blood pressure therefore although the annotated paper pmid focused on pulmonary hypertension we decided that ‘regulation of lung blood pressure’ was too far downstream to annotate to this term based on the presented data and the measurements of rvbp instead we annotated the entity rno mir to positive regulation of blood pressure go imp occurs in uberon heart right ventricle however having rvbp regulation terms for this type of annotations would have been very helpful i will look forward to hearing from you thank you barbara goc bhf mirna goc bhf goc bc rachhuntley rlovering
1
12,065
14,739,732,895
IssuesEvent
2021-01-07 07:49:09
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Cerida Investment Corp. Invoice for account 118-SR0634
anc-process anp-important ant-bug has attachment
In GitLab by @kdjstudios on Sep 17, 2018, 09:12 **Submitted by:** "Amanda Jennings " <amanda.jennings@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-09-17-21134/conversation **Server:** internal **Client/Site:** Sarasota **Account:** 118-SR0634 **Issue:** Amanda wrote: > One of our customers got this message, I am unable to test this portion of SAB to see what I get. I have tried to pay this invoice on-line (three times!) but your web site is broken: ![image](/uploads/0dc227508a2a6639b281a091ea97ff2c/image.png) Please let me know when you get it fixed so I can pay with our company credit card through your web site.
1.0
Cerida Investment Corp. Invoice for account 118-SR0634 - In GitLab by @kdjstudios on Sep 17, 2018, 09:12 **Submitted by:** "Amanda Jennings " <amanda.jennings@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-09-17-21134/conversation **Server:** internal **Client/Site:** Sarasota **Account:** 118-SR0634 **Issue:** Amanda wrote: > One of our customers got this message, I am unable to test this portion of SAB to see what I get. I have tried to pay this invoice on-line (three times!) but your web site is broken: ![image](/uploads/0dc227508a2a6639b281a091ea97ff2c/image.png) Please let me know when you get it fixed so I can pay with our company credit card through your web site.
process
cerida investment corp invoice for account in gitlab by kdjstudios on sep submitted by amanda jennings helpdesk server internal client site sarasota account issue amanda wrote one of our customers got this message i am unable to test this portion of sab to see what i get i have tried to pay this invoice on line three times but your web site is broken uploads image png please let me know when you get it fixed so i can pay with our company credit card through your web site
1
387,565
26,727,075,465
IssuesEvent
2023-01-29 20:58:02
AntiMicroX/antimicrox
https://api.github.com/repos/AntiMicroX/antimicrox
closed
Add CONTRIBUTING.md file to repository
documentation
It could be done based on https://github.com/AntiMicroX/antimicrox/wiki/Coding-guidelines Iw may be very useful to do it as soon as possible because of Hacktoberfest https://hacktoberfest.com/participation/#maintainers
1.0
Add CONTRIBUTING.md file to repository - It could be done based on https://github.com/AntiMicroX/antimicrox/wiki/Coding-guidelines Iw may be very useful to do it as soon as possible because of Hacktoberfest https://hacktoberfest.com/participation/#maintainers
non_process
add contributing md file to repository it could be done based on iw may be very useful to do it as soon as possible because of hacktoberfest
0
10
2,496,235,514
IssuesEvent
2015-01-06 18:01:39
vivo-isf/vivo-isf-ontology
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
closed
learning or memory
biological_process imported
_From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 15:45:16_ GO:0007611 parent: neurological system process \<a href="http://purl.obolibrary.org/obo/GO_0050877" rel="nofollow">http://purl.obolibrary.org/obo/GO_0050877</a> _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=160_
1.0
learning or memory - _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 15:45:16_ GO:0007611 parent: neurological system process \<a href="http://purl.obolibrary.org/obo/GO_0050877" rel="nofollow">http://purl.obolibrary.org/obo/GO_0050877</a> _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=160_
process
learning or memory from on november go parent neurological system process original issue
1
16,724
3,352,844,664
IssuesEvent
2015-11-18 00:58:22
eloipuertas/ES2015A
https://api.github.com/repos/eloipuertas/ES2015A
closed
Placing trees. Part 1
design Group D
### DESCRIPTION: Map with trees. Six different models. ### OUTCOME EXPECTED / ACCEPTANCE CRITERIA: Trees in harmony with the environment. Part (1) ### Estimated time effort: 2h
1.0
Placing trees. Part 1 - ### DESCRIPTION: Map with trees. Six different models. ### OUTCOME EXPECTED / ACCEPTANCE CRITERIA: Trees in harmony with the environment. Part (1) ### Estimated time effort: 2h
non_process
placing trees part description map with trees six different models outcome expected acceptance criteria trees in harmony with the environment part estimated time effort
0
16,449
21,327,655,217
IssuesEvent
2022-04-18 02:27:50
emily-writes-poems/emily-writes-poems-processing
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-processing
closed
set current feature
processing
select from list of all features to set one as the current one. also option to deselect so there is no current feature.
1.0
set current feature - select from list of all features to set one as the current one. also option to deselect so there is no current feature.
process
set current feature select from list of all features to set one as the current one also option to deselect so there is no current feature
1
289,010
24,950,216,732
IssuesEvent
2022-11-01 06:17:49
redhat-developer/vscode-openshift-tools
https://api.github.com/repos/redhat-developer/vscode-openshift-tools
closed
Application Explorer options (buttons) are missing and only "loaded-context" item is present when no kubeconfig is set
priority/blocker kind/bug odov3-test
In a situation when I do not have any kubeconfig on ~/.kube/config or env. var set. the extension shows only "loaded-context" tree item in application explorer instead of Login, Create component, etc.
1.0
Application Explorer options (buttons) are missing and only "loaded-context" item is present when no kubeconfig is set - In a situation when I do not have any kubeconfig on ~/.kube/config or env. var set. the extension shows only "loaded-context" tree item in application explorer instead of Login, Create component, etc.
non_process
application explorer options buttons are missing and only loaded context item is present when no kubeconfig is set in a situation when i do not have any kubeconfig on kube config or env var set the extension shows only loaded context tree item in application explorer instead of login create component etc
0
292,681
22,034,308,335
IssuesEvent
2022-05-28 10:18:24
KorBasilion/OSS-Basic-Project
https://api.github.com/repos/KorBasilion/OSS-Basic-Project
closed
현재의 가계부 삭제 로직의 재구성 예정
documentation
![image](https://user-images.githubusercontent.com/34836246/170805754-2774637f-3c62-4447-8b56-a4b7e5f16404.png) 현재 가계부 삭제방식은 사용자가 직접 추가할 때 처럼 내용을 모두 입력받아 해당내용과 일치한 행에 대한 정보들을 삭제하는 로직으로 구성되어있으나 앞선 이슈에서 언급했듯이 이 방법은 다소 불편할 것으로 예상되어 행 더블클릭시 삭제 창에 해당 행의 정보를 자동으로 입력시킨 뒤 사용자로부터 삭제 명령만 받을 수 있게끔 구현 예정입니다. 관련 내용을 찾아 본 결과 https://blog.naver.com/happyyhj/221849262839 https://stackoverflow.com/questions/61404261/tkinter-selecting-an-item-from-a-treeview-using-single-click-instead-of-double 위 링크에 내용이 있어 참고 차 첨부드립니다.
1.0
현재의 가계부 삭제 로직의 재구성 예정 - ![image](https://user-images.githubusercontent.com/34836246/170805754-2774637f-3c62-4447-8b56-a4b7e5f16404.png) 현재 가계부 삭제방식은 사용자가 직접 추가할 때 처럼 내용을 모두 입력받아 해당내용과 일치한 행에 대한 정보들을 삭제하는 로직으로 구성되어있으나 앞선 이슈에서 언급했듯이 이 방법은 다소 불편할 것으로 예상되어 행 더블클릭시 삭제 창에 해당 행의 정보를 자동으로 입력시킨 뒤 사용자로부터 삭제 명령만 받을 수 있게끔 구현 예정입니다. 관련 내용을 찾아 본 결과 https://blog.naver.com/happyyhj/221849262839 https://stackoverflow.com/questions/61404261/tkinter-selecting-an-item-from-a-treeview-using-single-click-instead-of-double 위 링크에 내용이 있어 참고 차 첨부드립니다.
non_process
현재의 가계부 삭제 로직의 재구성 예정 현재 가계부 삭제방식은 사용자가 직접 추가할 때 처럼 내용을 모두 입력받아 해당내용과 일치한 행에 대한 정보들을 삭제하는 로직으로 구성되어있으나 앞선 이슈에서 언급했듯이 이 방법은 다소 불편할 것으로 예상되어 행 더블클릭시 삭제 창에 해당 행의 정보를 자동으로 입력시킨 뒤 사용자로부터 삭제 명령만 받을 수 있게끔 구현 예정입니다 관련 내용을 찾아 본 결과 위 링크에 내용이 있어 참고 차 첨부드립니다
0
243,355
20,381,284,490
IssuesEvent
2022-02-21 22:15:22
Qiskit/platypus
https://api.github.com/repos/Qiskit/platypus
opened
Validate Grader Integration Feature with Professors
User Testing
Success Metric: - Number of problem sets or exams created - Number of scores returned 1. Reach out to 3 - 5 Professors 2. Ask to integrate with ongoing course 3. Measure success metrics as described above
1.0
Validate Grader Integration Feature with Professors - Success Metric: - Number of problem sets or exams created - Number of scores returned 1. Reach out to 3 - 5 Professors 2. Ask to integrate with ongoing course 3. Measure success metrics as described above
non_process
validate grader integration feature with professors success metric number of problem sets or exams created number of scores returned reach out to professors ask to integrate with ongoing course measure success metrics as described above
0
10,485
13,252,927,498
IssuesEvent
2020-08-20 06:35:02
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
Introduce Arena / Bump Allocators for Expression Evaluation
PCP-S1 difficulty/hard sig/coprocessor status/help-wanted
## Description In Coprocessor evaluation framework, each column occupies a `Vec`. Column is the source, the immediate container and the destination of all evaluation. For example, consider expression `a+b+c`, a vector `d` will be created to store results of `a+b` and then a vector `e` will be created to store results of `d+c`. It would be efficient to use a single memory pool to store these vectors so that there can be no allocation cost and deallocation cost, which can improve performance. The memory pool could be an arena allocator or a bump allocator. ## Difficulty * Hard ## Score * 3000 ## Mentor(s) - @breeswish - @sticnarf - @andylokandy ## Recommended Skills - Unsafe Rust programming
1.0
Introduce Arena / Bump Allocators for Expression Evaluation - ## Description In Coprocessor evaluation framework, each column occupies a `Vec`. Column is the source, the immediate container and the destination of all evaluation. For example, consider expression `a+b+c`, a vector `d` will be created to store results of `a+b` and then a vector `e` will be created to store results of `d+c`. It would be efficient to use a single memory pool to store these vectors so that there can be no allocation cost and deallocation cost, which can improve performance. The memory pool could be an arena allocator or a bump allocator. ## Difficulty * Hard ## Score * 3000 ## Mentor(s) - @breeswish - @sticnarf - @andylokandy ## Recommended Skills - Unsafe Rust programming
process
introduce arena bump allocators for expression evaluation description in coprocessor evaluation framework each column occupies a vec column is the source the immediate container and the destination of all evaluation for example consider expression a b c a vector d will be created to store results of a b and then a vector e will be created to store results of d c it would be efficient to use a single memory pool to store these vectors so that there can be no allocation cost and deallocation cost which can improve performance the memory pool could be an arena allocator or a bump allocator difficulty hard score mentor s breeswish sticnarf andylokandy recommended skills unsafe rust programming
1
828,467
31,830,425,731
IssuesEvent
2023-09-14 10:12:50
MuntashirAkon/AppManager
https://api.github.com/repos/MuntashirAkon/AppManager
closed
WakeLock under-locked
Bug Severity: 2 Priority: 1 Status: Accepted
### Please check before submitting an issue - [X] I know what my device, OS and App Manager versions are - [X] I know how to take logs - [ ] I know how to reproduce the issue which may not be specific to my device ### Describe the bug AppManager crashed while performing 1-click ops immediately after restarting the device. (1 time crash) ### To Reproduce 1. Restart the device 2. Immediately open AM > 1-Click Ops > Clear data from uninstall apps & Trim caches in all apps 3. See error ### Expected behavior _No response_ ### Screenshots _No response_ ### Logs ``` java.lang.RuntimeException: WakeLock under-locked AppManager::1-click_ops at android.os.PowerManager$WakeLock.release(PowerManager.java:2899) at android.os.PowerManager$WakeLock.release(PowerManager.java:2861) at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity.clearData(OneClickOpsActivity.java:404) at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity.$r8$lambda$PAAA15TnXXKPI_Vc49jE5M68oeE(Unknown Source:0) at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity$$ExternalSyntheticLambda15.onChanged(Unknown Source:4) at androidx.lifecycle.LiveData.considerNotify(LiveData.java:133) at androidx.lifecycle.LiveData.dispatchingValue(LiveData.java:151) at androidx.lifecycle.LiveData.setValue(LiveData.java:309) at androidx.lifecycle.MutableLiveData.setValue(MutableLiveData.java:50) at androidx.lifecycle.LiveData$1.run(LiveData.java:93) at android.os.Handler.handleCallback(Handler.java:978) at android.os.Handler.dispatchMessage(Handler.java:104) at android.os.Looper.loopOnce(Looper.java:238) at android.os.Looper.loop(Looper.java:357) at android.app.ActivityThread.main(ActivityThread.java:8090) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:548) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1026) ``` ### Device info Device Info: App version: 4.0.0-alpha01-DEBUG App version code: 434 Android build version: 8049b-fa5cf Android release version: 12 Android SDK version: 31 Android build ID: S2RUBS32.51-15-9-7 Device brand: motorola Device manufacturer: motorola Device name: corfur Device model: moto g71 5G Device product name: corfur_g Device hardware name: qcom ABIs: [arm64-v8a, armeabi-v7a, armeabi] ABIs (32bit): [armeabi-v7a, armeabi] ABIs (64bit): [arm64-v8a] System language: en-IN In-App Language: auto Mode: root Inferred Mode: Root ### Additional context _No response_
1.0
WakeLock under-locked - ### Please check before submitting an issue - [X] I know what my device, OS and App Manager versions are - [X] I know how to take logs - [ ] I know how to reproduce the issue which may not be specific to my device ### Describe the bug AppManager crashed while performing 1-click ops immediately after restarting the device. (1 time crash) ### To Reproduce 1. Restart the device 2. Immediately open AM > 1-Click Ops > Clear data from uninstall apps & Trim caches in all apps 3. See error ### Expected behavior _No response_ ### Screenshots _No response_ ### Logs ``` java.lang.RuntimeException: WakeLock under-locked AppManager::1-click_ops at android.os.PowerManager$WakeLock.release(PowerManager.java:2899) at android.os.PowerManager$WakeLock.release(PowerManager.java:2861) at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity.clearData(OneClickOpsActivity.java:404) at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity.$r8$lambda$PAAA15TnXXKPI_Vc49jE5M68oeE(Unknown Source:0) at io.github.muntashirakon.AppManager.oneclickops.OneClickOpsActivity$$ExternalSyntheticLambda15.onChanged(Unknown Source:4) at androidx.lifecycle.LiveData.considerNotify(LiveData.java:133) at androidx.lifecycle.LiveData.dispatchingValue(LiveData.java:151) at androidx.lifecycle.LiveData.setValue(LiveData.java:309) at androidx.lifecycle.MutableLiveData.setValue(MutableLiveData.java:50) at androidx.lifecycle.LiveData$1.run(LiveData.java:93) at android.os.Handler.handleCallback(Handler.java:978) at android.os.Handler.dispatchMessage(Handler.java:104) at android.os.Looper.loopOnce(Looper.java:238) at android.os.Looper.loop(Looper.java:357) at android.app.ActivityThread.main(ActivityThread.java:8090) at java.lang.reflect.Method.invoke(Native Method) at com.android.internal.os.RuntimeInit$MethodAndArgsCaller.run(RuntimeInit.java:548) at com.android.internal.os.ZygoteInit.main(ZygoteInit.java:1026) ``` ### Device info Device Info: App version: 4.0.0-alpha01-DEBUG App version code: 434 Android build version: 8049b-fa5cf Android release version: 12 Android SDK version: 31 Android build ID: S2RUBS32.51-15-9-7 Device brand: motorola Device manufacturer: motorola Device name: corfur Device model: moto g71 5G Device product name: corfur_g Device hardware name: qcom ABIs: [arm64-v8a, armeabi-v7a, armeabi] ABIs (32bit): [armeabi-v7a, armeabi] ABIs (64bit): [arm64-v8a] System language: en-IN In-App Language: auto Mode: root Inferred Mode: Root ### Additional context _No response_
non_process
wakelock under locked please check before submitting an issue i know what my device os and app manager versions are i know how to take logs i know how to reproduce the issue which may not be specific to my device describe the bug appmanager crashed while performing click ops immediately after restarting the device time crash to reproduce restart the device immediately open am click ops clear data from uninstall apps trim caches in all apps see error expected behavior no response screenshots no response logs java lang runtimeexception wakelock under locked appmanager click ops at android os powermanager wakelock release powermanager java at android os powermanager wakelock release powermanager java at io github muntashirakon appmanager oneclickops oneclickopsactivity cleardata oneclickopsactivity java at io github muntashirakon appmanager oneclickops oneclickopsactivity lambda unknown source at io github muntashirakon appmanager oneclickops oneclickopsactivity onchanged unknown source at androidx lifecycle livedata considernotify livedata java at androidx lifecycle livedata dispatchingvalue livedata java at androidx lifecycle livedata setvalue livedata java at androidx lifecycle mutablelivedata setvalue mutablelivedata java at androidx lifecycle livedata run livedata java at android os handler handlecallback handler java at android os handler dispatchmessage handler java at android os looper looponce looper java at android os looper loop looper java at android app activitythread main activitythread java at java lang reflect method invoke native method at com android internal os runtimeinit methodandargscaller run runtimeinit java at com android internal os zygoteinit main zygoteinit java device info device info app version debug app version code android build version android release version android sdk version android build id device brand motorola device manufacturer motorola device name corfur device model moto device product name corfur g device hardware name qcom abis abis abis system language en in in app language auto mode root inferred mode root additional context no response
0
23,943
2,664,916,681
IssuesEvent
2015-03-20 17:14:08
QuiteRSS/quiterss
https://api.github.com/repos/QuiteRSS/quiterss
opened
Не показывать уведомления в полноэкранном режиме
2–5 stars bug imported OpSys-Linux Priority-Low
_From [raistlin...@gmail.com](https://code.google.com/u/109057101288473400433/) on January 24, 2013 04:23:53_ Хочется такую опцию, ибо при просмотре фильма это окошко - не к месту. _Original issue: http://code.google.com/p/quite-rss/issues/detail?id=106_
1.0
Не показывать уведомления в полноэкранном режиме - _From [raistlin...@gmail.com](https://code.google.com/u/109057101288473400433/) on January 24, 2013 04:23:53_ Хочется такую опцию, ибо при просмотре фильма это окошко - не к месту. _Original issue: http://code.google.com/p/quite-rss/issues/detail?id=106_
non_process
не показывать уведомления в полноэкранном режиме from on january хочется такую опцию ибо при просмотре фильма это окошко не к месту original issue
0
18,891
24,833,454,031
IssuesEvent
2022-10-26 06:48:34
didi/mpx
https://api.github.com/repos/didi/mpx
closed
改进:在写小程序分包时,编译后页面使用到的组件引用路径优化
processing
**问题描述** 我司现有原生语法写的微信小程序,近期需要升级新增页面,由于大小限制,我便使用了分包机制,我正在使用mpx写小程序的几个新加页面,并且将其编译文件生成的目录由原来的dist改到了另外一个小程序的分包路径下了,按照mpx的单文件形式进行了编译处理,生成的页面所引用的组件路径都是以 / 开头的(例如:/components/chat5bda63f4/lib/iview/input/index),但实际呢该组件存在的路径是 /我的分包名/components/chat5bda63f4/lib/iview/input/index 这样,我就要手动将分包下生成的components目录复制到 根路径下 或者 将生成的页面的包引用路径 手动更改为合适的位置;这样问题就来了,现有的小程序空间已经无法装下那些生成组件了,必须放到子包路径下,所以,第一种方法不可行,那么第二种方法,我就要不断的修改生成页面的组件引用路径,源码每次修改自动编译后都会重置,这显然是不现实的,希望大佬们可以增加配置项,将生成页面的组件引用路径可以改为相对路径的形式,这样,我在子包里使用框架生成的页面就毫无问题了。
1.0
改进:在写小程序分包时,编译后页面使用到的组件引用路径优化 - **问题描述** 我司现有原生语法写的微信小程序,近期需要升级新增页面,由于大小限制,我便使用了分包机制,我正在使用mpx写小程序的几个新加页面,并且将其编译文件生成的目录由原来的dist改到了另外一个小程序的分包路径下了,按照mpx的单文件形式进行了编译处理,生成的页面所引用的组件路径都是以 / 开头的(例如:/components/chat5bda63f4/lib/iview/input/index),但实际呢该组件存在的路径是 /我的分包名/components/chat5bda63f4/lib/iview/input/index 这样,我就要手动将分包下生成的components目录复制到 根路径下 或者 将生成的页面的包引用路径 手动更改为合适的位置;这样问题就来了,现有的小程序空间已经无法装下那些生成组件了,必须放到子包路径下,所以,第一种方法不可行,那么第二种方法,我就要不断的修改生成页面的组件引用路径,源码每次修改自动编译后都会重置,这显然是不现实的,希望大佬们可以增加配置项,将生成页面的组件引用路径可以改为相对路径的形式,这样,我在子包里使用框架生成的页面就毫无问题了。
process
改进:在写小程序分包时,编译后页面使用到的组件引用路径优化 问题描述 我司现有原生语法写的微信小程序,近期需要升级新增页面,由于大小限制,我便使用了分包机制,我正在使用mpx写小程序的几个新加页面,并且将其编译文件生成的目录由原来的dist改到了另外一个小程序的分包路径下了,按照mpx的单文件形式进行了编译处理,生成的页面所引用的组件路径都是以 开头的(例如: components lib iview input index),但实际呢该组件存在的路径是 我的分包名 components lib iview input index 这样,我就要手动将分包下生成的components目录复制到 根路径下 或者 将生成的页面的包引用路径 手动更改为合适的位置;这样问题就来了,现有的小程序空间已经无法装下那些生成组件了,必须放到子包路径下,所以,第一种方法不可行,那么第二种方法,我就要不断的修改生成页面的组件引用路径,源码每次修改自动编译后都会重置,这显然是不现实的,希望大佬们可以增加配置项,将生成页面的组件引用路径可以改为相对路径的形式,这样,我在子包里使用框架生成的页面就毫无问题了。
1
531,080
15,440,189,323
IssuesEvent
2021-03-08 02:37:41
AY2021S2-CS2103-W16-3/tp
https://api.github.com/repos/AY2021S2-CS2103-W16-3/tp
closed
💡 Resize GUI textbox to fit messages
priority::3 type::enhancement
### Describe the issue *What is the problem, if any? What is the desired behavior?* Textbox in HippoCampus GUI currently too small to fit help messages. It is currently scrollable, but preferably resized to show more. Or at least allow it to be resized with window. ________________________________________________________________ ### Additional context *Describe possible and/or alternative solutions, to highlight required scope of work.*
1.0
💡 Resize GUI textbox to fit messages - ### Describe the issue *What is the problem, if any? What is the desired behavior?* Textbox in HippoCampus GUI currently too small to fit help messages. It is currently scrollable, but preferably resized to show more. Or at least allow it to be resized with window. ________________________________________________________________ ### Additional context *Describe possible and/or alternative solutions, to highlight required scope of work.*
non_process
💡 resize gui textbox to fit messages describe the issue what is the problem if any what is the desired behavior textbox in hippocampus gui currently too small to fit help messages it is currently scrollable but preferably resized to show more or at least allow it to be resized with window additional context describe possible and or alternative solutions to highlight required scope of work
0
51,336
13,635,095,445
IssuesEvent
2020-09-25 01:52:47
nasifimtiazohi/openmrs-core-2.3.0
https://api.github.com/repos/nasifimtiazohi/openmrs-core-2.3.0
opened
CVE-2019-17571 (High) detected in log4j-1.2.17.jar
security vulnerability
## CVE-2019-17571 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to dependency file: openmrs-core-2.3.0/web/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.17.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-core-2.3.0/commit/51205552c13cbb164119762e0de32dd074f8ba6d">51205552c13cbb164119762e0de32dd074f8ba6d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17. <p>Publish Date: 2019-12-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571</a></p> <p>Release Date: 2019-12-20</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0-alpha1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-17571 (High) detected in log4j-1.2.17.jar - ## CVE-2019-17571 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to dependency file: openmrs-core-2.3.0/web/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.17.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nasifimtiazohi/openmrs-core-2.3.0/commit/51205552c13cbb164119762e0de32dd074f8ba6d">51205552c13cbb164119762e0de32dd074f8ba6d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17. <p>Publish Date: 2019-12-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17571</a></p> <p>Release Date: 2019-12-20</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0-alpha1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jar cve high severity vulnerability vulnerable library jar apache path to dependency file openmrs core web pom xml path to vulnerable library home wss scanner repository jar canner repository jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core step up your open source security game with whitesource
0
15,884
20,072,970,399
IssuesEvent
2022-02-04 09:28:34
amor71/LiuAlgoTrader
https://api.github.com/repos/amor71/LiuAlgoTrader
closed
migrate the momentum-scanner to support Alpaca
enhancement in-process
**Is your feature request related to a problem? Please describe.** The current `momentum-scanner` is using Polygon, and the older way the platform was loading data. Need to migrate the scanner to support both Polygon and Alpaca
1.0
migrate the momentum-scanner to support Alpaca - **Is your feature request related to a problem? Please describe.** The current `momentum-scanner` is using Polygon, and the older way the platform was loading data. Need to migrate the scanner to support both Polygon and Alpaca
process
migrate the momentum scanner to support alpaca is your feature request related to a problem please describe the current momentum scanner is using polygon and the older way the platform was loading data need to migrate the scanner to support both polygon and alpaca
1
15,537
19,703,299,750
IssuesEvent
2022-01-12 18:54:31
googleapis/python-translate
https://api.github.com/repos/googleapis/python-translate
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'translation' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * api_shortname 'translation' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 api shortname translation invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
633,353
20,252,751,757
IssuesEvent
2022-02-14 19:37:19
wasmerio/wasmer
https://api.github.com/repos/wasmerio/wasmer
closed
OOB table access with singlepass backend
🐞 bug priority-high
<!-- Thanks for the bug report! --> ### Describe the bug singlepass backend somehow emits instructions that can cause unexpected OOB in table access. llvm and cranelift backend doesn't reproduce this issue. Also this issue has been introduced since 2.2.0-rc1 ```console $ echo "`wasmer -V` | `rustc -V` | `uname -m`" wasmer 2.2.0-rc1 | rustc 1.56.1 (59eed8a2a 2021-11-01) | x86_64 ``` ### Steps to reproduce [VarArgs.swift.wasm.zip](https://github.com/wasmerio/wasmer/files/8011496/VarArgs.swift.wasm.zip) ```console $ wasmer run --singlepass VarArgs.swift.wasm ``` ### Expected behavior ```console $ wasmer run --singlepass VarArgs.swift.wasm The answer to life and everything is 42, 42, -42, 3.141593 dig it: 0 0 -1 1 -2 2 -3 3 -4 4 -5 5 -6 6 -7 7 -8 8 -9 9 -10 10 -11 11 pointers: '0x12345670' '0x12345671' '0x12345672' '0x12345673' '0x12345674' a 1.1 1 2.2 1 3.3 1 a b 1 1.1 1 2.2 1 3.3 1 b c 1 2 1 3 1 4444444444444444 1 c d 1 1 2 1 3 1 4444444444444444 1 d e 10 20 10 30 10 4040404040404040 10 e f 10 10 20 10 30 10 4040404040404040 10 f rdar-32547102: 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 10.0 11.0 a 1.1 1 2.2 1 4.5 1 1.1 a b 1 1.1 1 2.2 1 4.5 1 1.1 b done. ``` ### Actual behavior ```console $ wasmer run --singlepass VarArgs.swift.wasm error: failed to run `VarArgs.swift.wasm` │ 1: RuntimeError: undefined element: out of bounds table access ╰─▶ 2: table_get_oob ``` ### Additional context This issue was found in SwiftWasm's CI: https://github.com/swiftwasm/swift/pull/4203
1.0
OOB table access with singlepass backend - <!-- Thanks for the bug report! --> ### Describe the bug singlepass backend somehow emits instructions that can cause unexpected OOB in table access. llvm and cranelift backend doesn't reproduce this issue. Also this issue has been introduced since 2.2.0-rc1 ```console $ echo "`wasmer -V` | `rustc -V` | `uname -m`" wasmer 2.2.0-rc1 | rustc 1.56.1 (59eed8a2a 2021-11-01) | x86_64 ``` ### Steps to reproduce [VarArgs.swift.wasm.zip](https://github.com/wasmerio/wasmer/files/8011496/VarArgs.swift.wasm.zip) ```console $ wasmer run --singlepass VarArgs.swift.wasm ``` ### Expected behavior ```console $ wasmer run --singlepass VarArgs.swift.wasm The answer to life and everything is 42, 42, -42, 3.141593 dig it: 0 0 -1 1 -2 2 -3 3 -4 4 -5 5 -6 6 -7 7 -8 8 -9 9 -10 10 -11 11 pointers: '0x12345670' '0x12345671' '0x12345672' '0x12345673' '0x12345674' a 1.1 1 2.2 1 3.3 1 a b 1 1.1 1 2.2 1 3.3 1 b c 1 2 1 3 1 4444444444444444 1 c d 1 1 2 1 3 1 4444444444444444 1 d e 10 20 10 30 10 4040404040404040 10 e f 10 10 20 10 30 10 4040404040404040 10 f rdar-32547102: 0.0 1.0 2.0 3.0 4.0 5.0 6.0 7.0 8.0 9.0 10.0 11.0 a 1.1 1 2.2 1 4.5 1 1.1 a b 1 1.1 1 2.2 1 4.5 1 1.1 b done. ``` ### Actual behavior ```console $ wasmer run --singlepass VarArgs.swift.wasm error: failed to run `VarArgs.swift.wasm` │ 1: RuntimeError: undefined element: out of bounds table access ╰─▶ 2: table_get_oob ``` ### Additional context This issue was found in SwiftWasm's CI: https://github.com/swiftwasm/swift/pull/4203
non_process
oob table access with singlepass backend describe the bug singlepass backend somehow emits instructions that can cause unexpected oob in table access llvm and cranelift backend doesn t reproduce this issue also this issue has been introduced since console echo wasmer v rustc v uname m wasmer rustc steps to reproduce console wasmer run singlepass varargs swift wasm expected behavior console wasmer run singlepass varargs swift wasm the answer to life and everything is dig it pointers a a b b c c d d e e f f rdar a a b b done actual behavior console wasmer run singlepass varargs swift wasm error failed to run varargs swift wasm │ runtimeerror undefined element out of bounds table access ╰─▶ table get oob additional context this issue was found in swiftwasm s ci
0
19,896
26,347,038,309
IssuesEvent
2023-01-10 23:18:07
googleapis/nodejs-compute
https://api.github.com/repos/googleapis/nodejs-compute
closed
test(flakiness): reenable "instances with custom machine type tests" tests
priority: p2 type: process api: compute samples
The `instances with custom machine type tests` have been turned off due to flakiness. Rather than fixing integration tests, it would be worth investigating using mocking library (sinon or nock) to test library behavior (_reducing the number of integration tests_). Refs: #791
1.0
test(flakiness): reenable "instances with custom machine type tests" tests - The `instances with custom machine type tests` have been turned off due to flakiness. Rather than fixing integration tests, it would be worth investigating using mocking library (sinon or nock) to test library behavior (_reducing the number of integration tests_). Refs: #791
process
test flakiness reenable instances with custom machine type tests tests the instances with custom machine type tests have been turned off due to flakiness rather than fixing integration tests it would be worth investigating using mocking library sinon or nock to test library behavior reducing the number of integration tests refs
1
14,240
10,721,167,915
IssuesEvent
2019-10-26 23:34:45
anikau31/systemc-clang
https://api.github.com/repos/anikau31/systemc-clang
closed
catch2 on mac
testing infrastructure
Now, there are additional errors being generated from the catch header when compiling on a mac with clang 11.0.0. The prior one was "raise" on line 10630 being undeclared. The new errors after the last push are line 10639 signalstack being undeclared, 10645, 10657, 10660 sigaction having the wrong number of arguments (3, when 1 is required).
1.0
catch2 on mac - Now, there are additional errors being generated from the catch header when compiling on a mac with clang 11.0.0. The prior one was "raise" on line 10630 being undeclared. The new errors after the last push are line 10639 signalstack being undeclared, 10645, 10657, 10660 sigaction having the wrong number of arguments (3, when 1 is required).
non_process
on mac now there are additional errors being generated from the catch header when compiling on a mac with clang the prior one was raise on line being undeclared the new errors after the last push are line signalstack being undeclared sigaction having the wrong number of arguments when is required
0
2,213
5,051,502,272
IssuesEvent
2016-12-20 22:05:12
cfpb/design-manual
https://api.github.com/repos/cfpb/design-manual
closed
include links to incomplete issues on the manual site
1 - Working process and planning
cfpb.github.io/design-manual/ is how most non-employees are viewing our design manual work, but there's so much more work that hasn't been published yet, living in our issues. I propose two updates: 1. Include links to the related github issues on each page of the design manual site 2. Create pages on the design manual site for incomplete issues. This will make it easier for the public to know what we're working on and join our conversations
1.0
include links to incomplete issues on the manual site - cfpb.github.io/design-manual/ is how most non-employees are viewing our design manual work, but there's so much more work that hasn't been published yet, living in our issues. I propose two updates: 1. Include links to the related github issues on each page of the design manual site 2. Create pages on the design manual site for incomplete issues. This will make it easier for the public to know what we're working on and join our conversations
process
include links to incomplete issues on the manual site cfpb github io design manual is how most non employees are viewing our design manual work but there s so much more work that hasn t been published yet living in our issues i propose two updates include links to the related github issues on each page of the design manual site create pages on the design manual site for incomplete issues this will make it easier for the public to know what we re working on and join our conversations
1
9,214
12,247,259,933
IssuesEvent
2020-05-05 15:37:42
labnote-ant/labnote
https://api.github.com/repos/labnote-ant/labnote
closed
Set ingredient's volume, weight, pressure to null for Mix
process-view
The ingredients' volume, weight, and pressure in the output of Mix process are set to null. It will be changed in the major updates.
1.0
Set ingredient's volume, weight, pressure to null for Mix - The ingredients' volume, weight, and pressure in the output of Mix process are set to null. It will be changed in the major updates.
process
set ingredient s volume weight pressure to null for mix the ingredients volume weight and pressure in the output of mix process are set to null it will be changed in the major updates
1
11,183
13,957,696,324
IssuesEvent
2020-10-24 08:11:49
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
PT: Harvesting
Geoportal Harvesting process PT - Portugal
Geoportal team, The harvesting to the Portuguese catalogue is still not running daily... It&acute;s taking at least 1 week for us to check the results. Can you please start a harvesting do the Portuguese catalogue? Thank you! Best regards, Vanda Marcos
1.0
PT: Harvesting - Geoportal team, The harvesting to the Portuguese catalogue is still not running daily... It&acute;s taking at least 1 week for us to check the results. Can you please start a harvesting do the Portuguese catalogue? Thank you! Best regards, Vanda Marcos
process
pt harvesting geoportal team the harvesting to the portuguese catalogue is still not running daily it acute s taking at least week for us to check the results can you please start a harvesting do the portuguese catalogue thank you best regards vanda marcos
1
530,561
15,434,031,600
IssuesEvent
2021-03-07 00:55:42
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
opened
[Coverity CID :219573] Untrusted value as argument in samples/net/sockets/coap_client/src/coap-client.c
Coverity bug priority: low
Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/samples/net/sockets/coap_client/src/coap-client.c Category: Insecure data handling Function: `process_simple_coap_reply` Component: Samples CID: [219573](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219573) Details: https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/samples/net/sockets/coap_client/src/coap-client.c#L124 Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
1.0
[Coverity CID :219573] Untrusted value as argument in samples/net/sockets/coap_client/src/coap-client.c - Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/samples/net/sockets/coap_client/src/coap-client.c Category: Insecure data handling Function: `process_simple_coap_reply` Component: Samples CID: [219573](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219573) Details: https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/samples/net/sockets/coap_client/src/coap-client.c#L124 Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
non_process
untrusted value as argument in samples net sockets coap client src coap client c static code scan issues found in file category insecure data handling function process simple coap reply component samples cid details please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
0
78,211
9,670,301,658
IssuesEvent
2019-05-21 19:33:53
apollographql/apollo-client
https://api.github.com/repos/apollographql/apollo-client
closed
Apollo-client 2.5 state management resolvers not firing correctly
:wilted_flower: needs-reproduction ✍️ working-as-designed 🏓 waiting-response 🛬 fixed-in-prerelease
**Intended outcome:** I am working towards replacing our app's state management with apollo-client 2.5's local state management (because graphql is awesome). I have the following query ``` query GetClient { client(id: "0y9FVRmIiDIIhjr1czVi", live: true) @client { id name queueCount updated tickets @client { id deleted comments @client { id author_email public } } } } ``` Here is what my resolver looks like: ``` Query: { client: async (parent, { id, live }, context, info) => { // DO SOME STUFF var data = {} // DATA is actually returned after some stuff is done return data; } }, Client: { tickets: async (client, { live }, context, info) => { // DO SOME STUFF var data = {} // DATA is actually returned after some stuff is done return data; } }, Ticket: { comments: async (parent, { live }, context, info) => { // DO SOME STUFF var data = {} // DATA is actually returned after some stuff is done return data; } } ``` What I expect for the query to do is to start in the Query --> client section, and then after I return the data from there, to go to the Client --> tickets section of the resolver to run the logic there, and once each ticket is retrieved, it goes down to the Ticket --> comments section. **Actual outcome:** The Query --> client section fires, but neither the Client --> tickets nor Ticket --> comments section fire :(. I have tried removing the @clients decorators on the tickets and comments parts of the query, still no luck. **How to reproduce the issue:** See above code **Versions** System: OS: macOS 10.14.4 Binaries: Node: 10.15.3 - ~/.nvm/versions/node/v10.15.3/bin/node npm: 6.4.1 - ~/.nvm/versions/node/v10.15.3/bin/npm Browsers: Chrome: 74.0.3729.157 Firefox: 62.0 Safari: 12.1 npmPackages: apollo-boost: ^0.3.1 => 0.3.1 apollo-client: ^2.5.1 => 2.5.1 react-apollo: ^2.5.5 => 2.5.5
1.0
Apollo-client 2.5 state management resolvers not firing correctly - **Intended outcome:** I am working towards replacing our app's state management with apollo-client 2.5's local state management (because graphql is awesome). I have the following query ``` query GetClient { client(id: "0y9FVRmIiDIIhjr1czVi", live: true) @client { id name queueCount updated tickets @client { id deleted comments @client { id author_email public } } } } ``` Here is what my resolver looks like: ``` Query: { client: async (parent, { id, live }, context, info) => { // DO SOME STUFF var data = {} // DATA is actually returned after some stuff is done return data; } }, Client: { tickets: async (client, { live }, context, info) => { // DO SOME STUFF var data = {} // DATA is actually returned after some stuff is done return data; } }, Ticket: { comments: async (parent, { live }, context, info) => { // DO SOME STUFF var data = {} // DATA is actually returned after some stuff is done return data; } } ``` What I expect for the query to do is to start in the Query --> client section, and then after I return the data from there, to go to the Client --> tickets section of the resolver to run the logic there, and once each ticket is retrieved, it goes down to the Ticket --> comments section. **Actual outcome:** The Query --> client section fires, but neither the Client --> tickets nor Ticket --> comments section fire :(. I have tried removing the @clients decorators on the tickets and comments parts of the query, still no luck. **How to reproduce the issue:** See above code **Versions** System: OS: macOS 10.14.4 Binaries: Node: 10.15.3 - ~/.nvm/versions/node/v10.15.3/bin/node npm: 6.4.1 - ~/.nvm/versions/node/v10.15.3/bin/npm Browsers: Chrome: 74.0.3729.157 Firefox: 62.0 Safari: 12.1 npmPackages: apollo-boost: ^0.3.1 => 0.3.1 apollo-client: ^2.5.1 => 2.5.1 react-apollo: ^2.5.5 => 2.5.5
non_process
apollo client state management resolvers not firing correctly intended outcome i am working towards replacing our app s state management with apollo client s local state management because graphql is awesome i have the following query query getclient client id live true client id name queuecount updated tickets client id deleted comments client id author email public here is what my resolver looks like query client async parent id live context info do some stuff var data data is actually returned after some stuff is done return data client tickets async client live context info do some stuff var data data is actually returned after some stuff is done return data ticket comments async parent live context info do some stuff var data data is actually returned after some stuff is done return data what i expect for the query to do is to start in the query client section and then after i return the data from there to go to the client tickets section of the resolver to run the logic there and once each ticket is retrieved it goes down to the ticket comments section actual outcome the query client section fires but neither the client tickets nor ticket comments section fire i have tried removing the clients decorators on the tickets and comments parts of the query still no luck how to reproduce the issue see above code versions system os macos binaries node nvm versions node bin node npm nvm versions node bin npm browsers chrome firefox safari npmpackages apollo boost apollo client react apollo
0
60,127
6,671,984,739
IssuesEvent
2017-10-04 09:41:24
zalando/zalenium
https://api.github.com/repos/zalando/zalenium
closed
Ability to group videos
waiting-retest
Is there a way at the moment to group videos by a constant? like for example, the date, or the build server build number?
1.0
Ability to group videos - Is there a way at the moment to group videos by a constant? like for example, the date, or the build server build number?
non_process
ability to group videos is there a way at the moment to group videos by a constant like for example the date or the build server build number
0
12,447
14,934,725,065
IssuesEvent
2021-01-25 10:55:17
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Open study > Enrollment status is incorrectly displayed as 'Not eligible' in participant details page
Bug P2 Participant manager datastore Process: Fixed Process: Tested dev
**Steps:** 1. Enroll into any open study successfully 2. WIthdraw from the study 3. Fail the eligibility test for the same study 4. Observe the Enrollment status value in participant details page **Actual:** Enrollment status is incorrect displayed as 'Not eligible' in participant details page **Expected:** Enrollment status should displayt 'Withdrawn' in participant details page Note: Issue not observed in Enrollment registry page ![Screenshot_2](https://user-images.githubusercontent.com/60386291/105465274-b5470380-5cb8-11eb-860f-caa948324401.png)
2.0
[PM] Open study > Enrollment status is incorrectly displayed as 'Not eligible' in participant details page - **Steps:** 1. Enroll into any open study successfully 2. WIthdraw from the study 3. Fail the eligibility test for the same study 4. Observe the Enrollment status value in participant details page **Actual:** Enrollment status is incorrect displayed as 'Not eligible' in participant details page **Expected:** Enrollment status should displayt 'Withdrawn' in participant details page Note: Issue not observed in Enrollment registry page ![Screenshot_2](https://user-images.githubusercontent.com/60386291/105465274-b5470380-5cb8-11eb-860f-caa948324401.png)
process
open study enrollment status is incorrectly displayed as not eligible in participant details page steps enroll into any open study successfully withdraw from the study fail the eligibility test for the same study observe the enrollment status value in participant details page actual enrollment status is incorrect displayed as not eligible in participant details page expected enrollment status should displayt withdrawn in participant details page note issue not observed in enrollment registry page
1
6,184
9,101,753,499
IssuesEvent
2019-02-20 11:58:45
FACK1/ReservationSystem
https://api.github.com/repos/FACK1/ReservationSystem
opened
Get the events of the requested service ( front-end )
inProcess technical
* When an `service card` in the services page is clicked: - [ ] Make a request from the front side that should return the `service details` in the bookevent page, and details view page
1.0
Get the events of the requested service ( front-end ) - * When an `service card` in the services page is clicked: - [ ] Make a request from the front side that should return the `service details` in the bookevent page, and details view page
process
get the events of the requested service front end when an service card in the services page is clicked make a request from the front side that should return the service details in the bookevent page and details view page
1
16,111
11,840,366,991
IssuesEvent
2020-03-23 18:44:27
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Set minimum CMake version to 3.14.5
area-Infrastructure
See discussion in https://github.com/dotnet/runtime/pull/33959/commits/16e1d309221e70d4f3d484e8c8a4fb2a4347f83c#r396389297 The documented minimum CMake version is 3.15.5 already but a few places still check for 3.14: ```bash $ git grep "_minimum_required\|3\.14\.2" # search for _minimum_required OR 3.14.2 docs/workflow/requirements/linux-requirements.md:Building the repo requires CMake 3.14.2 or newer on Linux. Add Kitware's APT feed to your configuration for a newer version of CMake. See their instructions at <https://apt.kitware.com/>. Also, add LLVM/s APT feed to your configuration for a newer version of CMake. See their instructions as <http://apt.llvm.org/>. eng/native/build-commons.sh: if [[ "$(version "$cmake_version")" -lt "$(version 3.14.2)" ]]; then eng/native/build-commons.sh: echo "Please install CMake 3.14.2 or newer from http://www.cmake.org/download/ or https://apt.kitware.com and ensure it is on your path."; exit 1; global.json: "cmake": "3.14.2", src/coreclr/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/coreclr/src/pal/tests/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/coreclr/tests/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/installer/corehost/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/installer/corehost/cli/test/mockhostfxr/CMakeLists.txt:cmake_minimum_required (VERSION 2.6) src/libraries/Native/Unix/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/libraries/Native/Windows/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/libraries/Native/Windows/clrcompression/zlib/CMakeLists.txt:cmake_minimum_required(VERSION 2.4.4) src/mono/CMakeLists.txt:cmake_minimum_required (VERSION 3.10) src/mono/mono/btls/CMakeLists.txt:cmake_minimum_required (VERSION 2.8.10) src/mono/netcore/sample/iOS/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.5) ```
1.0
Set minimum CMake version to 3.14.5 - See discussion in https://github.com/dotnet/runtime/pull/33959/commits/16e1d309221e70d4f3d484e8c8a4fb2a4347f83c#r396389297 The documented minimum CMake version is 3.15.5 already but a few places still check for 3.14: ```bash $ git grep "_minimum_required\|3\.14\.2" # search for _minimum_required OR 3.14.2 docs/workflow/requirements/linux-requirements.md:Building the repo requires CMake 3.14.2 or newer on Linux. Add Kitware's APT feed to your configuration for a newer version of CMake. See their instructions at <https://apt.kitware.com/>. Also, add LLVM/s APT feed to your configuration for a newer version of CMake. See their instructions as <http://apt.llvm.org/>. eng/native/build-commons.sh: if [[ "$(version "$cmake_version")" -lt "$(version 3.14.2)" ]]; then eng/native/build-commons.sh: echo "Please install CMake 3.14.2 or newer from http://www.cmake.org/download/ or https://apt.kitware.com and ensure it is on your path."; exit 1; global.json: "cmake": "3.14.2", src/coreclr/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/coreclr/src/pal/tests/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/coreclr/tests/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/installer/corehost/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/installer/corehost/cli/test/mockhostfxr/CMakeLists.txt:cmake_minimum_required (VERSION 2.6) src/libraries/Native/Unix/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/libraries/Native/Windows/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.2) src/libraries/Native/Windows/clrcompression/zlib/CMakeLists.txt:cmake_minimum_required(VERSION 2.4.4) src/mono/CMakeLists.txt:cmake_minimum_required (VERSION 3.10) src/mono/mono/btls/CMakeLists.txt:cmake_minimum_required (VERSION 2.8.10) src/mono/netcore/sample/iOS/CMakeLists.txt:cmake_minimum_required(VERSION 3.14.5) ```
non_process
set minimum cmake version to see discussion in the documented minimum cmake version is already but a few places still check for bash git grep minimum required search for minimum required or docs workflow requirements linux requirements md building the repo requires cmake or newer on linux add kitware s apt feed to your configuration for a newer version of cmake see their instructions at also add llvm s apt feed to your configuration for a newer version of cmake see their instructions as eng native build commons sh if then eng native build commons sh echo please install cmake or newer from or and ensure it is on your path exit global json cmake src coreclr cmakelists txt cmake minimum required version src coreclr src pal tests cmakelists txt cmake minimum required version src coreclr tests cmakelists txt cmake minimum required version src installer corehost cmakelists txt cmake minimum required version src installer corehost cli test mockhostfxr cmakelists txt cmake minimum required version src libraries native unix cmakelists txt cmake minimum required version src libraries native windows cmakelists txt cmake minimum required version src libraries native windows clrcompression zlib cmakelists txt cmake minimum required version src mono cmakelists txt cmake minimum required version src mono mono btls cmakelists txt cmake minimum required version src mono netcore sample ios cmakelists txt cmake minimum required version
0
799
3,276,146,500
IssuesEvent
2015-10-26 18:13:36
hammerlab/pileup.js
https://api.github.com/repos/hammerlab/pileup.js
closed
Code coverage doesn't match source
bug process
See, for example, coverage for [`RemoteFile.js`][2]. The lines covered don't make logical sense. This was broken by my recent [switch to babelify][1]. The issue is that the coverage is being generated for the babel-ified source, which doesn't necessarily match up 1-1 line for line with the original source. [1]: https://github.com/hammerlab/pileup.js/pull/316 [2]: https://coveralls.io/builds/3932994/source?filename=src%2Fmain%2FRemoteFile.js
1.0
Code coverage doesn't match source - See, for example, coverage for [`RemoteFile.js`][2]. The lines covered don't make logical sense. This was broken by my recent [switch to babelify][1]. The issue is that the coverage is being generated for the babel-ified source, which doesn't necessarily match up 1-1 line for line with the original source. [1]: https://github.com/hammerlab/pileup.js/pull/316 [2]: https://coveralls.io/builds/3932994/source?filename=src%2Fmain%2FRemoteFile.js
process
code coverage doesn t match source see for example coverage for the lines covered don t make logical sense this was broken by my recent the issue is that the coverage is being generated for the babel ified source which doesn t necessarily match up line for line with the original source
1
19,682
26,033,039,027
IssuesEvent
2022-12-22 00:04:28
nkdAgility/azure-devops-migration-tools
https://api.github.com/repos/nkdAgility/azure-devops-migration-tools
closed
Edited discussion comments can be lost when migrating work items
question no-issue-activity Refactor WorkItem Processor
## Describe your issue: When migrating work items between Azure DevOps Services instances using "ReplayRevisions": true there is a situation where a work item discussion comment can be lost. To replicate: - Add a discussion comment - Add another discussion comment - Edit the first discussion comment When the work item is migrated the edited comment will not exist in the target. ## Source Details - **Source Version:** Azure DevOps Services - **Source Language:** Eng ## Target Details - **Target Version:** Azure DevOps Services - **Target Language:** Eng
1.0
Edited discussion comments can be lost when migrating work items - ## Describe your issue: When migrating work items between Azure DevOps Services instances using "ReplayRevisions": true there is a situation where a work item discussion comment can be lost. To replicate: - Add a discussion comment - Add another discussion comment - Edit the first discussion comment When the work item is migrated the edited comment will not exist in the target. ## Source Details - **Source Version:** Azure DevOps Services - **Source Language:** Eng ## Target Details - **Target Version:** Azure DevOps Services - **Target Language:** Eng
process
edited discussion comments can be lost when migrating work items describe your issue when migrating work items between azure devops services instances using replayrevisions true there is a situation where a work item discussion comment can be lost to replicate add a discussion comment add another discussion comment edit the first discussion comment when the work item is migrated the edited comment will not exist in the target source details source version azure devops services source language eng target details target version azure devops services target language eng
1
10,050
13,044,161,659
IssuesEvent
2020-07-29 03:47:25
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `SubDateDatetimeString` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `SubDateDatetimeString` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `SubDateDatetimeString` from TiDB - ## Description Port the scalar function `SubDateDatetimeString` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function subdatedatetimestring from tidb description port the scalar function subdatedatetimestring from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
1
193,326
6,884,032,685
IssuesEvent
2017-11-21 11:30:15
EyeSeeTea/dhis2-android-dashboard
https://api.github.com/repos/EyeSeeTea/dhis2-android-dashboard
closed
On login, toast are crazy
priority - critical
- [ ] When no connexion, it doesn't say a thing - [ ] When URL is introduced wrong, it says you have no connexion
1.0
On login, toast are crazy - - [ ] When no connexion, it doesn't say a thing - [ ] When URL is introduced wrong, it says you have no connexion
non_process
on login toast are crazy when no connexion it doesn t say a thing when url is introduced wrong it says you have no connexion
0
38,086
8,408,655,828
IssuesEvent
2018-10-12 02:55:47
fga-eps-mds/2018.2-NaturalSearch
https://api.github.com/repos/fga-eps-mds/2018.2-NaturalSearch
closed
US10-Eu, como desenvolvedor, desejo remover informações desnecessárias de proponentes que serão utilizadas.
Code User Story eps mds
Critérios de aceitação: - [ ] tratar os dados de proponentes para que sejam obtidos apenas os valores desejados. - [ ] visualizar os dados tratados.
1.0
US10-Eu, como desenvolvedor, desejo remover informações desnecessárias de proponentes que serão utilizadas. - Critérios de aceitação: - [ ] tratar os dados de proponentes para que sejam obtidos apenas os valores desejados. - [ ] visualizar os dados tratados.
non_process
eu como desenvolvedor desejo remover informações desnecessárias de proponentes que serão utilizadas critérios de aceitação tratar os dados de proponentes para que sejam obtidos apenas os valores desejados visualizar os dados tratados
0
216,815
16,820,001,766
IssuesEvent
2021-06-17 12:02:12
ARMmbed/mbed-os
https://api.github.com/repos/ARMmbed/mbed-os
opened
Findout a place to keep mbed-os/UNITTESTS/target_h
UNITTESTS component: core component: test
<!-- ************************************** WARNING ************************************** The ciarcom bot parses this header automatically. Any deviation from the template may cause the bot to automatically correct this header or may result in a warning message, requesting updates. Please ensure all sections of the template below are filled in and no changes are made to the template format. Only bugs should be raised here as issues. Questions or enhancements should instead be raised on our forums: https://forums.mbed.com/ . ************************************************************************************* --> ### Description of defect Findout a place to keep mbed-os/UNITTESTS/target_h <!-- Add detailed description of what you are reporting. Good example: https://os.mbed.com/docs/mbed-os/latest/contributing/workflow.html --> #### Target(s) affected by this defect ? N/A #### Toolchain(s) (name and version) displaying this defect ? N/A #### What version of Mbed-os are you using (tag or sha) ? <!-- For a released version please provide the release tag (this can be found as per the instructions below) mbed-os version can be found in /platform/mbed_version.h. The tag can be reconstructed as follows: mbed-os-MBED_MAJOR_VERSION.MBED_MINOR_VERSION.MBED_PATCH_VERSION Master branch is indicated by 'mbed-os-99.99.99 For an issue found on Master please provide the sha being used. --> Latest #### What version(s) of tools are you using. List all that apply (E.g. mbed-cli) N/A #### How is this defect reproduced ? N/A <!-- For connectivity issues it is recommended to enable mbed trace and attach the traces generated by the logger to the issue. -->
2.0
Findout a place to keep mbed-os/UNITTESTS/target_h - <!-- ************************************** WARNING ************************************** The ciarcom bot parses this header automatically. Any deviation from the template may cause the bot to automatically correct this header or may result in a warning message, requesting updates. Please ensure all sections of the template below are filled in and no changes are made to the template format. Only bugs should be raised here as issues. Questions or enhancements should instead be raised on our forums: https://forums.mbed.com/ . ************************************************************************************* --> ### Description of defect Findout a place to keep mbed-os/UNITTESTS/target_h <!-- Add detailed description of what you are reporting. Good example: https://os.mbed.com/docs/mbed-os/latest/contributing/workflow.html --> #### Target(s) affected by this defect ? N/A #### Toolchain(s) (name and version) displaying this defect ? N/A #### What version of Mbed-os are you using (tag or sha) ? <!-- For a released version please provide the release tag (this can be found as per the instructions below) mbed-os version can be found in /platform/mbed_version.h. The tag can be reconstructed as follows: mbed-os-MBED_MAJOR_VERSION.MBED_MINOR_VERSION.MBED_PATCH_VERSION Master branch is indicated by 'mbed-os-99.99.99 For an issue found on Master please provide the sha being used. --> Latest #### What version(s) of tools are you using. List all that apply (E.g. mbed-cli) N/A #### How is this defect reproduced ? N/A <!-- For connectivity issues it is recommended to enable mbed trace and attach the traces generated by the logger to the issue. -->
non_process
findout a place to keep mbed os unittests target h warning the ciarcom bot parses this header automatically any deviation from the template may cause the bot to automatically correct this header or may result in a warning message requesting updates please ensure all sections of the template below are filled in and no changes are made to the template format only bugs should be raised here as issues questions or enhancements should instead be raised on our forums description of defect findout a place to keep mbed os unittests target h add detailed description of what you are reporting good example target s affected by this defect n a toolchain s name and version displaying this defect n a what version of mbed os are you using tag or sha for a released version please provide the release tag this can be found as per the instructions below mbed os version can be found in platform mbed version h the tag can be reconstructed as follows mbed os mbed major version mbed minor version mbed patch version master branch is indicated by mbed os for an issue found on master please provide the sha being used latest what version s of tools are you using list all that apply e g mbed cli n a how is this defect reproduced n a for connectivity issues it is recommended to enable mbed trace and attach the traces generated by the logger to the issue
0
892
3,354,875,271
IssuesEvent
2015-11-18 14:22:03
kerubistan/kerub
https://api.github.com/repos/kerubistan/kerub
closed
generalize 'host assignment'
cleanup component:data processing priority: normal
*Host* assignment should be just *assignment* * modify data structure * all references should be called *assignment*
1.0
generalize 'host assignment' - *Host* assignment should be just *assignment* * modify data structure * all references should be called *assignment*
process
generalize host assignment host assignment should be just assignment modify data structure all references should be called assignment
1
7,000
10,145,505,138
IssuesEvent
2019-08-05 04:43:11
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
joins across multiple tables
.Proposal Query Language (MBQL) Query Processor
one of the biggest feature requests here is the ability to join more than one hop from the starting table. usually in filters, but sometimes in group-bys too. i know this is a nontrivial feature, but i can imagine it would be at least somewhat doable, UI wise, by just including connections in the filter column list, and collapsing when necessary if you keep following foreign keys. (and when rendering the filter "chip.") background: i've been doing "metabase office hours" here for the last couple days, and i've collected a handful of bug reports/feature requests. i'll try to search first to make sure i don't dupe existing issues, but apologies in advance if i do. people are loving metabase here, btw! here are a few messages i've gotten on slack, verbatim: > hey! everyone is raving about metabase, but my invite is not working. i want in on the new hotness! any chance i can get another invite? > oh wow nice! this thing is impressive. i _love_ that you can do raw sql! > holy crap metabase is amazing :rainbow: :sparkles: ⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
1.0
joins across multiple tables - one of the biggest feature requests here is the ability to join more than one hop from the starting table. usually in filters, but sometimes in group-bys too. i know this is a nontrivial feature, but i can imagine it would be at least somewhat doable, UI wise, by just including connections in the filter column list, and collapsing when necessary if you keep following foreign keys. (and when rendering the filter "chip.") background: i've been doing "metabase office hours" here for the last couple days, and i've collected a handful of bug reports/feature requests. i'll try to search first to make sure i don't dupe existing issues, but apologies in advance if i do. people are loving metabase here, btw! here are a few messages i've gotten on slack, verbatim: > hey! everyone is raving about metabase, but my invite is not working. i want in on the new hotness! any chance i can get another invite? > oh wow nice! this thing is impressive. i _love_ that you can do raw sql! > holy crap metabase is amazing :rainbow: :sparkles: ⬇️ **Please click the 👍 reaction instead of leaving a `+1` or 👍 comment**
process
joins across multiple tables one of the biggest feature requests here is the ability to join more than one hop from the starting table usually in filters but sometimes in group bys too i know this is a nontrivial feature but i can imagine it would be at least somewhat doable ui wise by just including connections in the filter column list and collapsing when necessary if you keep following foreign keys and when rendering the filter chip background i ve been doing metabase office hours here for the last couple days and i ve collected a handful of bug reports feature requests i ll try to search first to make sure i don t dupe existing issues but apologies in advance if i do people are loving metabase here btw here are a few messages i ve gotten on slack verbatim hey everyone is raving about metabase but my invite is not working i want in on the new hotness any chance i can get another invite oh wow nice this thing is impressive i love that you can do raw sql holy crap metabase is amazing rainbow sparkles ⬇️ please click the 👍 reaction instead of leaving a or 👍 comment
1
161,269
13,821,975,254
IssuesEvent
2020-10-13 03:51:40
Hbdkfk/Automated-Prostate-Cancer-Detection-using-MRI
https://api.github.com/repos/Hbdkfk/Automated-Prostate-Cancer-Detection-using-MRI
opened
EDA and further researches
documentation
-[ ] check number of patient -[ ] distribution of series (not all the patient has same amount of series, select patient ids that have all series), zones, benign/cancerous findings -[ ] search for the ROI sizes used -[ ] see if other teams have dealt with the edge detection issues for CNN
1.0
EDA and further researches - -[ ] check number of patient -[ ] distribution of series (not all the patient has same amount of series, select patient ids that have all series), zones, benign/cancerous findings -[ ] search for the ROI sizes used -[ ] see if other teams have dealt with the edge detection issues for CNN
non_process
eda and further researches check number of patient distribution of series not all the patient has same amount of series select patient ids that have all series zones benign cancerous findings search for the roi sizes used see if other teams have dealt with the edge detection issues for cnn
0
8,669
11,802,675,811
IssuesEvent
2020-03-18 22:06:07
phokz/mod-auth-external
https://api.github.com/repos/phokz/mod-auth-external
opened
Now that we have official Windows builds, we should update INSTALL
process-item
Now that we have official Windows builds, we should update INSTALL with example steps and requirements. The current one claims it may not work. :) The Wiki was [previously updated](https://github.com/phokz/mod-auth-external/wiki/Installation), but could use more details (example, requirements) and maybe reorganization would be good.
1.0
Now that we have official Windows builds, we should update INSTALL - Now that we have official Windows builds, we should update INSTALL with example steps and requirements. The current one claims it may not work. :) The Wiki was [previously updated](https://github.com/phokz/mod-auth-external/wiki/Installation), but could use more details (example, requirements) and maybe reorganization would be good.
process
now that we have official windows builds we should update install now that we have official windows builds we should update install with example steps and requirements the current one claims it may not work the wiki was but could use more details example requirements and maybe reorganization would be good
1
11,904
14,698,310,116
IssuesEvent
2021-01-04 06:08:23
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
"Extract by location" atributes bug when two predicates are selected
Bug Processing
<!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** When using "Extract by location" from _Processings_, choosing two predicates as seen below causes duplicates rows ... when object intersects and is within. Geometries are unique but I get duplicates rows in attributes. **How to Reproduce** 1. Open Processings toolbox 2. Click on "Extract by location" 3. Choose any input layer : here "polygons-2" 4. Check predicates : 'intersects' and 'are within' 5. Choose second layer `By comparing features from` "polygon" ![image](https://user-images.githubusercontent.com/154323/68209241-355d4c80-ffd3-11e9-82d1-646ef4af075d.png) 6. Open attributes table from created layer 7. Observe duplicated rows in attributes when object match both 'intersects' and 'within' conditions ! ![image](https://user-images.githubusercontent.com/154323/68212517-f67ec500-ffd9-11e9-8689-c16cf8022fe1.png) **QGIS and OS versions** Version de QGIS | 3.10.0-A Coruña | Révision du code | 6c816b4204 -- | -- | -- | -- Compilé avec Qt | 5.9.5 | Utilisant Qt | 5.9.5 Compilé avec GDAL/OGR | 2.2.3 | Utilisé avec GDAL/OGR | 2.2.3 Compilé avec GEOS | 3.6.2-CAPI-1.10.2 | Utilisé avec GEOS | 3.6.2-CAPI-1.10.2 4d2925d6 Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0 Version du client PostgreSQL | 10.10 (Ubuntu 10.10-0ubuntu0.18.04.1) | Version de SpatiaLite | 4.3.0a Version de QWT | 6.1.3 | Version de QScintilla2 | 2.10.2 Version de PROJ.4 : | 493 OS Version | Ubuntu 18.04.3 LTS Extensions Python actives | VectorBender; AnotherDXF2Shape; mask; Qgis2threejs; GeoCoding; vector_tiles_reader; qconsolidate3; shapetools; Hqgis; DataPlotly; french_locator_filter; SRTM-Downloader; Beeline; OSMtools; QNEAT3; loadthemall; db_manager; MetaSearch; processing **Additional context** <!-- Add any other context about the problem here. -->
1.0
"Extract by location" atributes bug when two predicates are selected - <!-- Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone. If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix Checklist before submitting - [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists - [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles). - [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). --> **Describe the bug** When using "Extract by location" from _Processings_, choosing two predicates as seen below causes duplicates rows ... when object intersects and is within. Geometries are unique but I get duplicates rows in attributes. **How to Reproduce** 1. Open Processings toolbox 2. Click on "Extract by location" 3. Choose any input layer : here "polygons-2" 4. Check predicates : 'intersects' and 'are within' 5. Choose second layer `By comparing features from` "polygon" ![image](https://user-images.githubusercontent.com/154323/68209241-355d4c80-ffd3-11e9-82d1-646ef4af075d.png) 6. Open attributes table from created layer 7. Observe duplicated rows in attributes when object match both 'intersects' and 'within' conditions ! ![image](https://user-images.githubusercontent.com/154323/68212517-f67ec500-ffd9-11e9-8689-c16cf8022fe1.png) **QGIS and OS versions** Version de QGIS | 3.10.0-A Coruña | Révision du code | 6c816b4204 -- | -- | -- | -- Compilé avec Qt | 5.9.5 | Utilisant Qt | 5.9.5 Compilé avec GDAL/OGR | 2.2.3 | Utilisé avec GDAL/OGR | 2.2.3 Compilé avec GEOS | 3.6.2-CAPI-1.10.2 | Utilisé avec GEOS | 3.6.2-CAPI-1.10.2 4d2925d6 Compiled against SQLite | 3.22.0 | Running against SQLite | 3.22.0 Version du client PostgreSQL | 10.10 (Ubuntu 10.10-0ubuntu0.18.04.1) | Version de SpatiaLite | 4.3.0a Version de QWT | 6.1.3 | Version de QScintilla2 | 2.10.2 Version de PROJ.4 : | 493 OS Version | Ubuntu 18.04.3 LTS Extensions Python actives | VectorBender; AnotherDXF2Shape; mask; Qgis2threejs; GeoCoding; vector_tiles_reader; qconsolidate3; shapetools; Hqgis; DataPlotly; french_locator_filter; SRTM-Downloader; Beeline; OSMtools; QNEAT3; loadthemall; db_manager; MetaSearch; processing **Additional context** <!-- Add any other context about the problem here. -->
process
extract by location atributes bug when two predicates are selected bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug when using extract by location from processings choosing two predicates as seen below causes duplicates rows when object intersects and is within geometries are unique but i get duplicates rows in attributes how to reproduce open processings toolbox click on extract by location choose any input layer here polygons check predicates intersects and are within choose second layer by comparing features from polygon open attributes table from created layer observe duplicated rows in attributes when object match both intersects and within conditions qgis and os versions version de qgis a coruña révision du code compilé avec qt utilisant qt compilé avec gdal ogr utilisé avec gdal ogr compilé avec geos capi utilisé avec geos capi compiled against sqlite running against sqlite version du client postgresql ubuntu version de spatialite version de qwt version de version de proj os version ubuntu lts extensions python actives vectorbender mask geocoding vector tiles reader shapetools hqgis dataplotly french locator filter srtm downloader beeline osmtools loadthemall db manager metasearch processing additional context
1
525,381
15,252,099,894
IssuesEvent
2021-02-20 01:31:30
rubyforgood/casa
https://api.github.com/repos/rubyforgood/casa
closed
Do not show inactive assignment or inactive case to volunteers. Do not include in supervisor weekly digest.
:clipboard: Supervisor :email: Email :no_good_man: User Permissions :raised_hands: Volunteer Priority: Medium
**What is the problem, and what should happen instead?** function `case_assignments_with_cases` contains `case_assignments.includes(:casa_case)` called by `views/volunteers/edit.html.erb` on line 80 and `views/supervisor_mailer/weekly_digest.html.erb `on line 15 **Acceptance criteria:** - [ ] Volunteer Edit should NOT include casa cases from which the volunteer has been unassigned (inactive assignment) - [ ] Write a test - [ ] Volunteer Edit should NOT include **inactive** casa cases (inactive case) - [ ] Write a test - [ ] Supervisor email should NOT include cases which are inactive at the time that the mailer is generated (inactive case) - [ ] Write a test - [ ] Supervisor email should NOT include cases which have been unassigned from volunteers (inactive assignment) - [ ] Write a test - [ ] If a volunteer performs actions - - and is unassigned from a supervisor in the same week - - and **not** assigned to another supervisor then the supervisor should see the activity in the email and the email should indicate that the volunteer is unassigned - [ ] If a volunteer performs actions - - and is unassigned from a supervisor in the same week - - and **is** assigned to another supervisor, then the volunteer's old supervisor should not see the volunteer's activity in the email - - and the volunteer's new supervisor should see the volunteer's activity in their email Thanks @FireLemons for finding this! https://rubyforgood.slack.com/archives/CVB0QJGVD/p1607574310223400
1.0
Do not show inactive assignment or inactive case to volunteers. Do not include in supervisor weekly digest. - **What is the problem, and what should happen instead?** function `case_assignments_with_cases` contains `case_assignments.includes(:casa_case)` called by `views/volunteers/edit.html.erb` on line 80 and `views/supervisor_mailer/weekly_digest.html.erb `on line 15 **Acceptance criteria:** - [ ] Volunteer Edit should NOT include casa cases from which the volunteer has been unassigned (inactive assignment) - [ ] Write a test - [ ] Volunteer Edit should NOT include **inactive** casa cases (inactive case) - [ ] Write a test - [ ] Supervisor email should NOT include cases which are inactive at the time that the mailer is generated (inactive case) - [ ] Write a test - [ ] Supervisor email should NOT include cases which have been unassigned from volunteers (inactive assignment) - [ ] Write a test - [ ] If a volunteer performs actions - - and is unassigned from a supervisor in the same week - - and **not** assigned to another supervisor then the supervisor should see the activity in the email and the email should indicate that the volunteer is unassigned - [ ] If a volunteer performs actions - - and is unassigned from a supervisor in the same week - - and **is** assigned to another supervisor, then the volunteer's old supervisor should not see the volunteer's activity in the email - - and the volunteer's new supervisor should see the volunteer's activity in their email Thanks @FireLemons for finding this! https://rubyforgood.slack.com/archives/CVB0QJGVD/p1607574310223400
non_process
do not show inactive assignment or inactive case to volunteers do not include in supervisor weekly digest what is the problem and what should happen instead function case assignments with cases contains case assignments includes casa case called by views volunteers edit html erb on line and views supervisor mailer weekly digest html erb on line acceptance criteria volunteer edit should not include casa cases from which the volunteer has been unassigned inactive assignment write a test volunteer edit should not include inactive casa cases inactive case write a test supervisor email should not include cases which are inactive at the time that the mailer is generated inactive case write a test supervisor email should not include cases which have been unassigned from volunteers inactive assignment write a test if a volunteer performs actions and is unassigned from a supervisor in the same week and not assigned to another supervisor then the supervisor should see the activity in the email and the email should indicate that the volunteer is unassigned if a volunteer performs actions and is unassigned from a supervisor in the same week and is assigned to another supervisor then the volunteer s old supervisor should not see the volunteer s activity in the email and the volunteer s new supervisor should see the volunteer s activity in their email thanks firelemons for finding this
0
256,496
27,561,680,730
IssuesEvent
2023-03-07 22:39:42
samqws-marketing/electronicarts_ava-capture
https://api.github.com/repos/samqws-marketing/electronicarts_ava-capture
closed
CVE-2021-44906 (High) detected in minimist-1.2.5.tgz - autoclosed
Mend: dependency security vulnerability
## CVE-2021-44906 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p> <p>Path to dependency file: /website-frontend/package.json</p> <p>Path to vulnerable library: /website-frontend/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - awesome-typescript-loader-5.2.1.tgz (Root Library) - mkdirp-0.5.5.tgz - :x: **minimist-1.2.5.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95). <p>Publish Date: 2022-03-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-44906>CVE-2021-44906</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-03-17</p> <p>Fix Resolution: minimist - 1.2.6</p> </p> </details> <p></p>
True
CVE-2021-44906 (High) detected in minimist-1.2.5.tgz - autoclosed - ## CVE-2021-44906 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimist-1.2.5.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.5.tgz</a></p> <p>Path to dependency file: /website-frontend/package.json</p> <p>Path to vulnerable library: /website-frontend/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - awesome-typescript-loader-5.2.1.tgz (Root Library) - mkdirp-0.5.5.tgz - :x: **minimist-1.2.5.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Minimist <=1.2.5 is vulnerable to Prototype Pollution via file index.js, function setKey() (lines 69-95). <p>Publish Date: 2022-03-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-44906>CVE-2021-44906</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-03-17</p> <p>Fix Resolution: minimist - 1.2.6</p> </p> </details> <p></p>
non_process
cve high detected in minimist tgz autoclosed cve high severity vulnerability vulnerable library minimist tgz parse argument options library home page a href path to dependency file website frontend package json path to vulnerable library website frontend node modules minimist package json dependency hierarchy awesome typescript loader tgz root library mkdirp tgz x minimist tgz vulnerable library found in base branch master vulnerability details minimist is vulnerable to prototype pollution via file index js function setkey lines publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimist
0
26,900
4,826,052,555
IssuesEvent
2016-11-07 08:49:09
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
opened
Build: Module 'XXX' has no exported member 'YYY'.
defect
http://forums.bridge.net/forum/bridge-net-pro/bugs/2976-build-module-xxx-has-no-exported-member-yyy ### Expected ```js module DictionaryMap$2 ``` ### Actual ```js export interface DictionaryMap$2<T1,T2> { getForward(): Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T1,T2>; setForward(value: Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T1,T2>): void; getReverse(): Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T2,T1>; setReverse(value: Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T2,T1>): void; add(t1: T1, t2: T2): void; } export function DictionaryMap$2<T1,T2>(T1: {prototype: T1}, T2: {prototype: T2}): { prototype: DictionaryMap$2<T1,T2>; ctor: { new (): DictionaryMap$2<T1,T2> }; $ctor1: { new (initialValues: System.Collections.Generic.KeyValuePair$2<T1,T2>[]): DictionaryMap$2<T1,T2> }; } module DictionaryMap { export interface Indexer$2<T1,T2,T3,T4> { getItem(index: T3): T4; setItem(index: T3, value: T4): void; containsKey(index: T3): boolean; } export function Indexer$2<T1,T2,T3,T4>(T1: {prototype: T1}, T2: {prototype: T2}, T3: {prototype: T3}, T4: {prototype: T4}): { prototype: Indexer$2<T1,T2,T3,T4>; new (dictionary: System.Collections.Generic.Dictionary$2<T3,T4>): Indexer$2<T1,T2,T3,T4>; } } ``` ### Steps To Reproduce [Deck](http://deck.net/REPLACE-WITH-YOUR-DECK-LINK) ```cs public class DictionaryMap<T1, T2> { private Dictionary<T1, T2> _forward = new Dictionary<T1, T2>(); private Dictionary<T2, T1> _reverse = new Dictionary<T2, T1>(); public DictionaryMap() { this.Forward = new Indexer<T1, T2>(_forward); this.Reverse = new Indexer<T2, T1>(_reverse); } public DictionaryMap(params KeyValuePair<T1, T2>[] initialValues): this() { foreach (KeyValuePair<T1, T2> value in initialValues) Add(value.Key, value.Value); } public class Indexer<T3, T4> { private Dictionary<T3, T4> _dictionary; public Indexer(Dictionary<T3, T4> dictionary) { _dictionary = dictionary; } public T4 this[T3 index] { get { return _dictionary[index]; } set { _dictionary[index] = value; } } public bool ContainsKey(T3 index) { return _dictionary.ContainsKey(index); } } public void Add(T1 t1, T2 t2) { _forward.Add(t1, t2); _reverse.Add(t2, t1); } public Indexer<T1, T2> Forward { get; private set; } public Indexer<T2, T1> Reverse { get; private set; } } ```
1.0
Build: Module 'XXX' has no exported member 'YYY'. - http://forums.bridge.net/forum/bridge-net-pro/bugs/2976-build-module-xxx-has-no-exported-member-yyy ### Expected ```js module DictionaryMap$2 ``` ### Actual ```js export interface DictionaryMap$2<T1,T2> { getForward(): Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T1,T2>; setForward(value: Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T1,T2>): void; getReverse(): Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T2,T1>; setReverse(value: Mosaic.Infrastructure.General.DictionaryMap$2.Indexer$2<T1,T2,T2,T1>): void; add(t1: T1, t2: T2): void; } export function DictionaryMap$2<T1,T2>(T1: {prototype: T1}, T2: {prototype: T2}): { prototype: DictionaryMap$2<T1,T2>; ctor: { new (): DictionaryMap$2<T1,T2> }; $ctor1: { new (initialValues: System.Collections.Generic.KeyValuePair$2<T1,T2>[]): DictionaryMap$2<T1,T2> }; } module DictionaryMap { export interface Indexer$2<T1,T2,T3,T4> { getItem(index: T3): T4; setItem(index: T3, value: T4): void; containsKey(index: T3): boolean; } export function Indexer$2<T1,T2,T3,T4>(T1: {prototype: T1}, T2: {prototype: T2}, T3: {prototype: T3}, T4: {prototype: T4}): { prototype: Indexer$2<T1,T2,T3,T4>; new (dictionary: System.Collections.Generic.Dictionary$2<T3,T4>): Indexer$2<T1,T2,T3,T4>; } } ``` ### Steps To Reproduce [Deck](http://deck.net/REPLACE-WITH-YOUR-DECK-LINK) ```cs public class DictionaryMap<T1, T2> { private Dictionary<T1, T2> _forward = new Dictionary<T1, T2>(); private Dictionary<T2, T1> _reverse = new Dictionary<T2, T1>(); public DictionaryMap() { this.Forward = new Indexer<T1, T2>(_forward); this.Reverse = new Indexer<T2, T1>(_reverse); } public DictionaryMap(params KeyValuePair<T1, T2>[] initialValues): this() { foreach (KeyValuePair<T1, T2> value in initialValues) Add(value.Key, value.Value); } public class Indexer<T3, T4> { private Dictionary<T3, T4> _dictionary; public Indexer(Dictionary<T3, T4> dictionary) { _dictionary = dictionary; } public T4 this[T3 index] { get { return _dictionary[index]; } set { _dictionary[index] = value; } } public bool ContainsKey(T3 index) { return _dictionary.ContainsKey(index); } } public void Add(T1 t1, T2 t2) { _forward.Add(t1, t2); _reverse.Add(t2, t1); } public Indexer<T1, T2> Forward { get; private set; } public Indexer<T2, T1> Reverse { get; private set; } } ```
non_process
build module xxx has no exported member yyy expected js module dictionarymap actual js export interface dictionarymap getforward mosaic infrastructure general dictionarymap indexer setforward value mosaic infrastructure general dictionarymap indexer void getreverse mosaic infrastructure general dictionarymap indexer setreverse value mosaic infrastructure general dictionarymap indexer void add void export function dictionarymap prototype prototype prototype dictionarymap ctor new dictionarymap new initialvalues system collections generic keyvaluepair dictionarymap module dictionarymap export interface indexer getitem index setitem index value void containskey index boolean export function indexer prototype prototype prototype prototype prototype indexer new dictionary system collections generic dictionary indexer steps to reproduce cs public class dictionarymap private dictionary forward new dictionary private dictionary reverse new dictionary public dictionarymap this forward new indexer forward this reverse new indexer reverse public dictionarymap params keyvaluepair initialvalues this foreach keyvaluepair value in initialvalues add value key value value public class indexer private dictionary dictionary public indexer dictionary dictionary dictionary dictionary public this get return dictionary set dictionary value public bool containskey index return dictionary containskey index public void add forward add reverse add public indexer forward get private set public indexer reverse get private set
0
19,960
26,441,240,973
IssuesEvent
2023-01-16 00:28:48
hsmusic/hsmusic-wiki
https://api.github.com/repos/hsmusic/hsmusic-wiki
closed
Totally empty documents (end of file only?) break YAML processing
type: bug scope: data processing
I.e, `---` at the end of an album yaml apparently breaks stuff. Wow yikes!! Thanks to koba and Niklink for spotting this.
1.0
Totally empty documents (end of file only?) break YAML processing - I.e, `---` at the end of an album yaml apparently breaks stuff. Wow yikes!! Thanks to koba and Niklink for spotting this.
process
totally empty documents end of file only break yaml processing i e at the end of an album yaml apparently breaks stuff wow yikes thanks to koba and niklink for spotting this
1
19,152
11,151,025,416
IssuesEvent
2019-12-24 02:01:42
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
opened
C/C++ IntelliSense "expression must have a pointer type" is incorrectly shown on a pointer type (wrong location)
Language Service Visual Studio bug
Use ``` class c2 { public: int i; }; class c1 { public: c2 func() { return c2(); } }; void main(void) { std::shared_ptr<c1> ptr = std::make_shared<c1>(); ptr->func()->i; } ``` Bug: "expression must have a pointer type" is shown on the "ptr", which is the wrong location and confusing. It should be shown after the func() call, which is the expression that doesn't have the pointer type, which is where the cl.exe compiler errors at. Reported at https://developercommunity.visualstudio.com/content/problem/866095/cc-intellisense-expression-must-have-a-pointer-typ.html .
1.0
C/C++ IntelliSense "expression must have a pointer type" is incorrectly shown on a pointer type (wrong location) - Use ``` class c2 { public: int i; }; class c1 { public: c2 func() { return c2(); } }; void main(void) { std::shared_ptr<c1> ptr = std::make_shared<c1>(); ptr->func()->i; } ``` Bug: "expression must have a pointer type" is shown on the "ptr", which is the wrong location and confusing. It should be shown after the func() call, which is the expression that doesn't have the pointer type, which is where the cl.exe compiler errors at. Reported at https://developercommunity.visualstudio.com/content/problem/866095/cc-intellisense-expression-must-have-a-pointer-typ.html .
non_process
c c intellisense expression must have a pointer type is incorrectly shown on a pointer type wrong location use class public int i class public func return void main void std shared ptr ptr std make shared ptr func i bug expression must have a pointer type is shown on the ptr which is the wrong location and confusing it should be shown after the func call which is the expression that doesn t have the pointer type which is where the cl exe compiler errors at reported at
0
290,313
25,047,544,576
IssuesEvent
2022-11-05 13:01:54
dotnet/maui
https://api.github.com/repos/dotnet/maui
closed
[Critical] Command + V and more crashes the application (macOS 13)
t/bug platform/macOS 🍏 area/controls 🎮 control-entry s/needs-attention s/try-latest-version
### Description When developing my own application, I noticed a very serious bug: when I tried to copy and paste certain text, the application stalled and then crashed. I can note that when testing on Windows (version 11) everything worked fine. ### Steps to Reproduce 1. Create a File > New .NET MAUI App 2. Add a `Entry` like so: `<Entry Text="this is a bug" />` 3. Click the added entry, press in keyboard Command + V, Command + C, Keyboard arrows left or right (macOS system) and observe the bug 🐞 Expected result: The text insertion will be successful. Actual result: the application is slowing down/crashing. There were no errors when trying to run this bug on Windows. ### Link to public reproduction project repository none ### Version with bug 6.0.400 ### Last version that worked well Unknown/Other ### Affected platforms macOS ### Affected platform versions macOS 13.0 Ventura Beta (22A5373b) ### Did you find any workaround? Sadly, no. ### Relevant log output ```shell 2022-10-22 22:39:46.959 BugMauiApp[29575:379247] Warning: observer object was not disposed manually with Dispose() 2022-10-22 22:39:46.959 BugMauiApp[29575:379247] Warning: observer object was not disposed manually with Dispose() 2022-10-22 22:39:51.040 BugMauiApp[29575:379239] *** Terminating app due to uncaught exception 'NSInvalidArgumentException', reason: '-[<_UIObscurableTextContentStorage: 0x600001445950> offsetFromLocation:toLocation:] received invalid locations 0 (null)' *** First throw call stack: ( 0 CoreFoundation 0x00007ff80bb31cd3 __exceptionPreprocess + 242 1 libobjc.A.dylib 0x00007ff80b68210a objc_exception_throw + 48 2 CoreFoundation 0x00007ff80bb31b38 +[NSException raise:format:] + 206 3 UIFoundation 0x00007ff80fb8fbc0 -[NSTextContentStorage offsetFromLocation:toLocation:] + 290 4 UIKitCore 0x00007ff919906647 -[_UITextKit2LayoutController offsetFromPosition:toPosition:] + 234 5 UIKitCore 0x00007ff91a0f6430 -[UITextInputController offsetFromPosition:toPosition:] + 85 6 UIKitCore 0x00007ff91a104fea __91-[_UITextInputControllerTokenizer _closestTokenSubrangeForPosition:granularity:downstream:]_block_invoke + 175 7 UIFoundation 0x00007ff80fb67b5c -[NSTextStorage coordinateReading:] + 42 8 UIKitCore 0x00007ff91a104cf0 -[_UITextInputControllerTokenizer _closestTokenSubrangeForPosition:granularity:downstream:] + 412 9 UIKitCore 0x00007ff919d2864b -[UITextInputStringTokenizer positionFromPosition:toBoundary:inDirection:] + 100 10 UIKitCore 0x00007ff919c984c6 -[UIKBRTIPartner _queryUIKitDocumentRequest:completion:] + 1329 11 UIKitCore 0x00007ff919c96aee -[UIKBRTIPartner _queryDocumentRequest:completion:] + 159 12 UIKitCore 0x00007ff919c90f13 -[UIKBRTIPartner _updateRTIStateWithCompletion:] + 559 13 UIKitCore 0x00007ff919c8f912 -[UIKBRTIPartner updateStateWithCompletion:] + 65 14 UIKitCore 0x00007ff919c8ba37 -[UIKBRTIPartner documentStateChanged:] + 126 15 UIKitCore 0x00007ff919be8200 -[UIKeyboardImpl setDocumentState:] + 263 16 UIKitCore 0x00007ff919c03cbf -[UIKeyboardImpl updateKeyboardStateForDeletion] + 337 17 UIKitCore 0x00007ff9198c31b3 -[UIKBInputDelegateManager _deleteBackwardAndNotify:reinsertText:] + 651 18 UIKitCore 0x00007ff919c03b5d -[UIKeyboardImpl deleteBackwardAndNotify:reinsertText:] + 64 19 UIKitCore 0x00007ff919be7865 -[UIKeyboardImpl performTextOperationActionSelector:] + 1272 20 UIKitCore 0x00007ff919c95477 -[UIKBRTIPartner _queued_performTextOperations:resultHandler:] + 1408 21 libdispatch.dylib 0x00007ff80b85e7fb _dispatch_call_block_and_release + 12 22 libdispatch.dylib 0x00007ff80b85fa44 _dispatch_client_callout + 8 23 libdispatch.dylib 0x00007ff80b86c7b9 _dispatch_main_queue_drain + 952 24 libdispatch.dylib 0x00007ff80b86c3f3 _dispatch_main_queue_callback_4CF + 31 25 CoreFoundation 0x00007ff80baf844d __CFRUNLOOP_IS_SERVICING_THE_MAIN_DISPATCH_QUEUE__ + 9 26 CoreFoundation 0x00007ff80bab8edb __CFRunLoopRun + 2498 27 CoreFoundation 0x00007ff80bab7e9f CFRunLoopRunSpecific + 560 28 HIToolbox 0x00007ff8158eabd6 RunCurrentEventLoopInMode + 292 29 HIToolbox 0x00007ff8158ea806 ReceiveNextEventCommon + 199 30 HIToolbox 0x00007ff8158ea723 _BlockUntilNextEventMatchingListInModeWithFilter + 70 31 AppKit 0x00007ff80eb19b87 _DPSNextEvent + 909 32 AppKit 0x00007ff80eb18a08 -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:] + 1219 33 AppKit 0x00007ff80eb0b043 -[NSApplication run] + 586 34 AppKit 0x00007ff80eadf0a8 NSApplicationMain + 817 35 AppKit 0x00007ff80ed9f6d7 _NSApplicationMainWithInfoDictionary + 16 36 UIKitMacHelper 0x00007ff82300ce97 UINSApplicationMain + 1413 37 UIKitCore 0x00007ff918f615b9 UIApplicationMain + 144 38 libmonosgen-2.0.dylib 0x000000010e6d43b0 do_icall + 288 39 libmonosgen-2.0.dylib 0x000000010e6d3388 do_icall_wrapper + 248 40 libmonosgen-2.0.dylib 0x000000010e6c54d1 interp_exec_method + 3393 41 libmonosgen-2.0.dylib 0x000000010e6c35ff interp_runtime_invoke + 239 42 libmonosgen-2.0.dylib 0x000000010e5c20c4 mono_jit_runtime_invoke + 1220 43 libmonosgen-2.0.dylib 0x000000010e4de2b8 mono_runtime_invoke_checked + 136 44 libmonosgen-2.0.dylib 0x000000010e4e620c do_exec_main_checked + 92 45 libmonosgen-2.0.dylib 0x000000010e61ea92 mono_jit_exec + 354 46 libxamarin-dotnet-debug.dylib 0x000000010dea4b99 xamarin_main + 809 47 BugMauiApp 0x0000000104d99ba4 main + 52 48 dyld 0x00000002054fa310 start + 2432 ) libc++abi: terminating with uncaught exception of type NSException ================================================================= Native Crash Reporting ================================================================= Got a SIGABRT while executing native code. This usually indicates a fatal error in the mono runtime or one of the native libraries used by your application. ================================================================= ================================================================= Native stacktrace: ================================================================= 0x10e6c1e85 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_dump_native_crash_info 0x10e663d6e - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_handle_native_crash 0x10e6c1782 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : sigabrt_signal_handler 0x7ff80ba09c1d - /usr/lib/system/libsystem_platform.dylib : _sigtramp 0x0 - Unknown 0x7ff80b928ca5 - /usr/lib/system/libsystem_c.dylib : abort 0x7ff80b999082 - /usr/lib/libc++abi.dylib : abort_message 0x7ff80b98a23d - /usr/lib/libc++abi.dylib : _ZL29demangling_unexpected_handlerv 0x7ff80b689250 - /usr/lib/libobjc.A.dylib : _ZL15_objc_terminatev 0x7ff80b9984a5 - /usr/lib/libc++abi.dylib : _ZSt11__terminatePFvvE 0x7ff80b998456 - /usr/lib/libc++abi.dylib : _ZSt9terminatev 0x7ff80b85fa58 - /usr/lib/system/libdispatch.dylib : _dispatch_client_callout 0x7ff80b86c7b9 - /usr/lib/system/libdispatch.dylib : _dispatch_main_queue_drain 0x7ff80b86c3f3 - /usr/lib/system/libdispatch.dylib : _dispatch_main_queue_callback_4CF 0x7ff80baf844d - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : __CFRUNLOOP_IS_SERVICING_THE_MAIN_DISPATCH_QUEUE__ 0x7ff80bab8edb - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : __CFRunLoopRun 0x7ff80bab7e9f - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : CFRunLoopRunSpecific 0x7ff8158eabd6 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : RunCurrentEventLoopInMode 0x7ff8158ea806 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : ReceiveNextEventCommon 0x7ff8158ea723 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : _BlockUntilNextEventMatchingListInModeWithFilter 0x7ff80eb19b87 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : _DPSNextEvent 0x7ff80eb18a08 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:] 0x7ff80eb0b043 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : -[NSApplication run] 0x7ff80eadf0a8 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : NSApplicationMain 0x7ff80ed9f6d7 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : _NSApplicationMainWithInfoDictionary 0x7ff82300ce97 - /System/Library/PrivateFrameworks/UIKitMacHelper.framework/Versions/A/UIKitMacHelper : UINSApplicationMain 0x7ff918f615b9 - /System/iOSSupport/System/Library/PrivateFrameworks/UIKitCore.framework/Versions/A/UIKitCore : UIApplicationMain 0x10e6d43b0 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_icall 0x10e6d3388 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_icall_wrapper 0x10e6c54d1 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : interp_exec_method 0x10e6c35ff - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : interp_runtime_invoke 0x10e5c20c4 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_jit_runtime_invoke 0x10e4de2b8 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_runtime_invoke_checked 0x10e4e620c - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_exec_main_checked 0x10e61ea92 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_jit_exec 0x10dea4b99 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libxamarin-dotnet-debug.dylib : xamarin_main 0x104d99ba4 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MacOS/BugMauiApp : main 0x2054fa310 - Unknown ================================================================= Telemetry Dumper: ================================================================= Pkilling 0x13128519680x from 0x8679686848x Pkilling 0x13126410240x from 0x8679686848x Could not exec mono-hang-watchdog, expected on path '/Users/runner/work/1/s/artifacts/obj/mono/MacCatalyst.x64.Release/out/etc/../bin/mono-hang-watchdog' (errno 2) Entering thread summarizer pause from 0x8679686848x Finished thread summarizer pause from 0x8679686848x. Failed to create breadcrumb file (null)/crash_hash_0x29e26bbc7a Waiting for dumping threads to resume ================================================================= Basic Fault Address Reporting ================================================================= Memory around native instruction pointer (0x7ff80b9a730e):0x7ff80b9a72fe ff ff c3 90 90 90 b8 48 01 00 02 49 89 ca 0f 05 .......H...I.... 0x7ff80b9a730e 73 08 48 89 c7 e9 aa 9a ff ff c3 90 90 90 b8 53 s.H............S 0x7ff80b9a731e 00 00 02 49 89 ca 0f 05 73 08 48 89 c7 e9 92 9a ...I....s.H..... 0x7ff80b9a732e ff ff c3 90 90 90 b8 83 01 00 02 49 89 ca 0f 05 ...........I.... ================================================================= Managed Stacktrace: ================================================================= at <unknown> <0xffffffff> at UIKit.UIApplication:UIApplicationMain <0x0019e> at UIKit.UIApplication:Main <0x00106> at BugMauiApp.Program:Main <0x00016> at <Module>:runtime_invoke_direct_void_string[] <0x00074> at <unknown> <0x00000> ================================================================= ```
1.0
[Critical] Command + V and more crashes the application (macOS 13) - ### Description When developing my own application, I noticed a very serious bug: when I tried to copy and paste certain text, the application stalled and then crashed. I can note that when testing on Windows (version 11) everything worked fine. ### Steps to Reproduce 1. Create a File > New .NET MAUI App 2. Add a `Entry` like so: `<Entry Text="this is a bug" />` 3. Click the added entry, press in keyboard Command + V, Command + C, Keyboard arrows left or right (macOS system) and observe the bug 🐞 Expected result: The text insertion will be successful. Actual result: the application is slowing down/crashing. There were no errors when trying to run this bug on Windows. ### Link to public reproduction project repository none ### Version with bug 6.0.400 ### Last version that worked well Unknown/Other ### Affected platforms macOS ### Affected platform versions macOS 13.0 Ventura Beta (22A5373b) ### Did you find any workaround? Sadly, no. ### Relevant log output ```shell 2022-10-22 22:39:46.959 BugMauiApp[29575:379247] Warning: observer object was not disposed manually with Dispose() 2022-10-22 22:39:46.959 BugMauiApp[29575:379247] Warning: observer object was not disposed manually with Dispose() 2022-10-22 22:39:51.040 BugMauiApp[29575:379239] *** Terminating app due to uncaught exception 'NSInvalidArgumentException', reason: '-[<_UIObscurableTextContentStorage: 0x600001445950> offsetFromLocation:toLocation:] received invalid locations 0 (null)' *** First throw call stack: ( 0 CoreFoundation 0x00007ff80bb31cd3 __exceptionPreprocess + 242 1 libobjc.A.dylib 0x00007ff80b68210a objc_exception_throw + 48 2 CoreFoundation 0x00007ff80bb31b38 +[NSException raise:format:] + 206 3 UIFoundation 0x00007ff80fb8fbc0 -[NSTextContentStorage offsetFromLocation:toLocation:] + 290 4 UIKitCore 0x00007ff919906647 -[_UITextKit2LayoutController offsetFromPosition:toPosition:] + 234 5 UIKitCore 0x00007ff91a0f6430 -[UITextInputController offsetFromPosition:toPosition:] + 85 6 UIKitCore 0x00007ff91a104fea __91-[_UITextInputControllerTokenizer _closestTokenSubrangeForPosition:granularity:downstream:]_block_invoke + 175 7 UIFoundation 0x00007ff80fb67b5c -[NSTextStorage coordinateReading:] + 42 8 UIKitCore 0x00007ff91a104cf0 -[_UITextInputControllerTokenizer _closestTokenSubrangeForPosition:granularity:downstream:] + 412 9 UIKitCore 0x00007ff919d2864b -[UITextInputStringTokenizer positionFromPosition:toBoundary:inDirection:] + 100 10 UIKitCore 0x00007ff919c984c6 -[UIKBRTIPartner _queryUIKitDocumentRequest:completion:] + 1329 11 UIKitCore 0x00007ff919c96aee -[UIKBRTIPartner _queryDocumentRequest:completion:] + 159 12 UIKitCore 0x00007ff919c90f13 -[UIKBRTIPartner _updateRTIStateWithCompletion:] + 559 13 UIKitCore 0x00007ff919c8f912 -[UIKBRTIPartner updateStateWithCompletion:] + 65 14 UIKitCore 0x00007ff919c8ba37 -[UIKBRTIPartner documentStateChanged:] + 126 15 UIKitCore 0x00007ff919be8200 -[UIKeyboardImpl setDocumentState:] + 263 16 UIKitCore 0x00007ff919c03cbf -[UIKeyboardImpl updateKeyboardStateForDeletion] + 337 17 UIKitCore 0x00007ff9198c31b3 -[UIKBInputDelegateManager _deleteBackwardAndNotify:reinsertText:] + 651 18 UIKitCore 0x00007ff919c03b5d -[UIKeyboardImpl deleteBackwardAndNotify:reinsertText:] + 64 19 UIKitCore 0x00007ff919be7865 -[UIKeyboardImpl performTextOperationActionSelector:] + 1272 20 UIKitCore 0x00007ff919c95477 -[UIKBRTIPartner _queued_performTextOperations:resultHandler:] + 1408 21 libdispatch.dylib 0x00007ff80b85e7fb _dispatch_call_block_and_release + 12 22 libdispatch.dylib 0x00007ff80b85fa44 _dispatch_client_callout + 8 23 libdispatch.dylib 0x00007ff80b86c7b9 _dispatch_main_queue_drain + 952 24 libdispatch.dylib 0x00007ff80b86c3f3 _dispatch_main_queue_callback_4CF + 31 25 CoreFoundation 0x00007ff80baf844d __CFRUNLOOP_IS_SERVICING_THE_MAIN_DISPATCH_QUEUE__ + 9 26 CoreFoundation 0x00007ff80bab8edb __CFRunLoopRun + 2498 27 CoreFoundation 0x00007ff80bab7e9f CFRunLoopRunSpecific + 560 28 HIToolbox 0x00007ff8158eabd6 RunCurrentEventLoopInMode + 292 29 HIToolbox 0x00007ff8158ea806 ReceiveNextEventCommon + 199 30 HIToolbox 0x00007ff8158ea723 _BlockUntilNextEventMatchingListInModeWithFilter + 70 31 AppKit 0x00007ff80eb19b87 _DPSNextEvent + 909 32 AppKit 0x00007ff80eb18a08 -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:] + 1219 33 AppKit 0x00007ff80eb0b043 -[NSApplication run] + 586 34 AppKit 0x00007ff80eadf0a8 NSApplicationMain + 817 35 AppKit 0x00007ff80ed9f6d7 _NSApplicationMainWithInfoDictionary + 16 36 UIKitMacHelper 0x00007ff82300ce97 UINSApplicationMain + 1413 37 UIKitCore 0x00007ff918f615b9 UIApplicationMain + 144 38 libmonosgen-2.0.dylib 0x000000010e6d43b0 do_icall + 288 39 libmonosgen-2.0.dylib 0x000000010e6d3388 do_icall_wrapper + 248 40 libmonosgen-2.0.dylib 0x000000010e6c54d1 interp_exec_method + 3393 41 libmonosgen-2.0.dylib 0x000000010e6c35ff interp_runtime_invoke + 239 42 libmonosgen-2.0.dylib 0x000000010e5c20c4 mono_jit_runtime_invoke + 1220 43 libmonosgen-2.0.dylib 0x000000010e4de2b8 mono_runtime_invoke_checked + 136 44 libmonosgen-2.0.dylib 0x000000010e4e620c do_exec_main_checked + 92 45 libmonosgen-2.0.dylib 0x000000010e61ea92 mono_jit_exec + 354 46 libxamarin-dotnet-debug.dylib 0x000000010dea4b99 xamarin_main + 809 47 BugMauiApp 0x0000000104d99ba4 main + 52 48 dyld 0x00000002054fa310 start + 2432 ) libc++abi: terminating with uncaught exception of type NSException ================================================================= Native Crash Reporting ================================================================= Got a SIGABRT while executing native code. This usually indicates a fatal error in the mono runtime or one of the native libraries used by your application. ================================================================= ================================================================= Native stacktrace: ================================================================= 0x10e6c1e85 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_dump_native_crash_info 0x10e663d6e - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_handle_native_crash 0x10e6c1782 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : sigabrt_signal_handler 0x7ff80ba09c1d - /usr/lib/system/libsystem_platform.dylib : _sigtramp 0x0 - Unknown 0x7ff80b928ca5 - /usr/lib/system/libsystem_c.dylib : abort 0x7ff80b999082 - /usr/lib/libc++abi.dylib : abort_message 0x7ff80b98a23d - /usr/lib/libc++abi.dylib : _ZL29demangling_unexpected_handlerv 0x7ff80b689250 - /usr/lib/libobjc.A.dylib : _ZL15_objc_terminatev 0x7ff80b9984a5 - /usr/lib/libc++abi.dylib : _ZSt11__terminatePFvvE 0x7ff80b998456 - /usr/lib/libc++abi.dylib : _ZSt9terminatev 0x7ff80b85fa58 - /usr/lib/system/libdispatch.dylib : _dispatch_client_callout 0x7ff80b86c7b9 - /usr/lib/system/libdispatch.dylib : _dispatch_main_queue_drain 0x7ff80b86c3f3 - /usr/lib/system/libdispatch.dylib : _dispatch_main_queue_callback_4CF 0x7ff80baf844d - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : __CFRUNLOOP_IS_SERVICING_THE_MAIN_DISPATCH_QUEUE__ 0x7ff80bab8edb - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : __CFRunLoopRun 0x7ff80bab7e9f - /System/Library/Frameworks/CoreFoundation.framework/Versions/A/CoreFoundation : CFRunLoopRunSpecific 0x7ff8158eabd6 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : RunCurrentEventLoopInMode 0x7ff8158ea806 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : ReceiveNextEventCommon 0x7ff8158ea723 - /System/Library/Frameworks/Carbon.framework/Versions/A/Frameworks/HIToolbox.framework/Versions/A/HIToolbox : _BlockUntilNextEventMatchingListInModeWithFilter 0x7ff80eb19b87 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : _DPSNextEvent 0x7ff80eb18a08 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : -[NSApplication(NSEvent) _nextEventMatchingEventMask:untilDate:inMode:dequeue:] 0x7ff80eb0b043 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : -[NSApplication run] 0x7ff80eadf0a8 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : NSApplicationMain 0x7ff80ed9f6d7 - /System/Library/Frameworks/AppKit.framework/Versions/C/AppKit : _NSApplicationMainWithInfoDictionary 0x7ff82300ce97 - /System/Library/PrivateFrameworks/UIKitMacHelper.framework/Versions/A/UIKitMacHelper : UINSApplicationMain 0x7ff918f615b9 - /System/iOSSupport/System/Library/PrivateFrameworks/UIKitCore.framework/Versions/A/UIKitCore : UIApplicationMain 0x10e6d43b0 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_icall 0x10e6d3388 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_icall_wrapper 0x10e6c54d1 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : interp_exec_method 0x10e6c35ff - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : interp_runtime_invoke 0x10e5c20c4 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_jit_runtime_invoke 0x10e4de2b8 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_runtime_invoke_checked 0x10e4e620c - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : do_exec_main_checked 0x10e61ea92 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libmonosgen-2.0.dylib : mono_jit_exec 0x10dea4b99 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MonoBundle/libxamarin-dotnet-debug.dylib : xamarin_main 0x104d99ba4 - /Users/nik-801/Developer/Rider/BugMauiApp/BugMauiApp/bin/Debug/net6.0-maccatalyst/maccatalyst-x64/BugMauiApp.app/Contents/MacOS/BugMauiApp : main 0x2054fa310 - Unknown ================================================================= Telemetry Dumper: ================================================================= Pkilling 0x13128519680x from 0x8679686848x Pkilling 0x13126410240x from 0x8679686848x Could not exec mono-hang-watchdog, expected on path '/Users/runner/work/1/s/artifacts/obj/mono/MacCatalyst.x64.Release/out/etc/../bin/mono-hang-watchdog' (errno 2) Entering thread summarizer pause from 0x8679686848x Finished thread summarizer pause from 0x8679686848x. Failed to create breadcrumb file (null)/crash_hash_0x29e26bbc7a Waiting for dumping threads to resume ================================================================= Basic Fault Address Reporting ================================================================= Memory around native instruction pointer (0x7ff80b9a730e):0x7ff80b9a72fe ff ff c3 90 90 90 b8 48 01 00 02 49 89 ca 0f 05 .......H...I.... 0x7ff80b9a730e 73 08 48 89 c7 e9 aa 9a ff ff c3 90 90 90 b8 53 s.H............S 0x7ff80b9a731e 00 00 02 49 89 ca 0f 05 73 08 48 89 c7 e9 92 9a ...I....s.H..... 0x7ff80b9a732e ff ff c3 90 90 90 b8 83 01 00 02 49 89 ca 0f 05 ...........I.... ================================================================= Managed Stacktrace: ================================================================= at <unknown> <0xffffffff> at UIKit.UIApplication:UIApplicationMain <0x0019e> at UIKit.UIApplication:Main <0x00106> at BugMauiApp.Program:Main <0x00016> at <Module>:runtime_invoke_direct_void_string[] <0x00074> at <unknown> <0x00000> ================================================================= ```
non_process
command v and more crashes the application macos description when developing my own application i noticed a very serious bug when i tried to copy and paste certain text the application stalled and then crashed i can note that when testing on windows version everything worked fine steps to reproduce create a file new net maui app add a entry like so click the added entry press in keyboard command v command c keyboard arrows left or right macos system and observe the bug 🐞 expected result the text insertion will be successful actual result the application is slowing down crashing there were no errors when trying to run this bug on windows link to public reproduction project repository none version with bug last version that worked well unknown other affected platforms macos affected platform versions macos ventura beta did you find any workaround sadly no relevant log output shell bugmauiapp warning observer object was not disposed manually with dispose bugmauiapp warning observer object was not disposed manually with dispose bugmauiapp terminating app due to uncaught exception nsinvalidargumentexception reason received invalid locations null first throw call stack corefoundation exceptionpreprocess libobjc a dylib objc exception throw corefoundation uifoundation uikitcore uikitcore uikitcore block invoke uifoundation uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore uikitcore libdispatch dylib dispatch call block and release libdispatch dylib dispatch client callout libdispatch dylib dispatch main queue drain libdispatch dylib dispatch main queue callback corefoundation cfrunloop is servicing the main dispatch queue corefoundation cfrunlooprun corefoundation cfrunlooprunspecific hitoolbox runcurrenteventloopinmode hitoolbox receivenexteventcommon hitoolbox blockuntilnexteventmatchinglistinmodewithfilter appkit dpsnextevent appkit appkit appkit nsapplicationmain appkit nsapplicationmainwithinfodictionary uikitmachelper uinsapplicationmain uikitcore uiapplicationmain libmonosgen dylib do icall libmonosgen dylib do icall wrapper libmonosgen dylib interp exec method libmonosgen dylib interp runtime invoke libmonosgen dylib mono jit runtime invoke libmonosgen dylib mono runtime invoke checked libmonosgen dylib do exec main checked libmonosgen dylib mono jit exec libxamarin dotnet debug dylib xamarin main bugmauiapp main dyld start libc abi terminating with uncaught exception of type nsexception native crash reporting got a sigabrt while executing native code this usually indicates a fatal error in the mono runtime or one of the native libraries used by your application native stacktrace users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono dump native crash info users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono handle native crash users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib sigabrt signal handler usr lib system libsystem platform dylib sigtramp unknown usr lib system libsystem c dylib abort usr lib libc abi dylib abort message usr lib libc abi dylib unexpected handlerv usr lib libobjc a dylib objc terminatev usr lib libc abi dylib terminatepfvve usr lib libc abi dylib usr lib system libdispatch dylib dispatch client callout usr lib system libdispatch dylib dispatch main queue drain usr lib system libdispatch dylib dispatch main queue callback system library frameworks corefoundation framework versions a corefoundation cfrunloop is servicing the main dispatch queue system library frameworks corefoundation framework versions a corefoundation cfrunlooprun system library frameworks corefoundation framework versions a corefoundation cfrunlooprunspecific system library frameworks carbon framework versions a frameworks hitoolbox framework versions a hitoolbox runcurrenteventloopinmode system library frameworks carbon framework versions a frameworks hitoolbox framework versions a hitoolbox receivenexteventcommon system library frameworks carbon framework versions a frameworks hitoolbox framework versions a hitoolbox blockuntilnexteventmatchinglistinmodewithfilter system library frameworks appkit framework versions c appkit dpsnextevent system library frameworks appkit framework versions c appkit system library frameworks appkit framework versions c appkit system library frameworks appkit framework versions c appkit nsapplicationmain system library frameworks appkit framework versions c appkit nsapplicationmainwithinfodictionary system library privateframeworks uikitmachelper framework versions a uikitmachelper uinsapplicationmain system iossupport system library privateframeworks uikitcore framework versions a uikitcore uiapplicationmain users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib do icall users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib do icall wrapper users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib interp exec method users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib interp runtime invoke users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono jit runtime invoke users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono runtime invoke checked users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib do exec main checked users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libmonosgen dylib mono jit exec users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents monobundle libxamarin dotnet debug dylib xamarin main users nik developer rider bugmauiapp bugmauiapp bin debug maccatalyst maccatalyst bugmauiapp app contents macos bugmauiapp main unknown telemetry dumper pkilling from pkilling from could not exec mono hang watchdog expected on path users runner work s artifacts obj mono maccatalyst release out etc bin mono hang watchdog errno entering thread summarizer pause from finished thread summarizer pause from failed to create breadcrumb file null crash hash waiting for dumping threads to resume basic fault address reporting memory around native instruction pointer ff ff ca h i aa ff ff s h s ca i s h ff ff ca i managed stacktrace at at uikit uiapplication uiapplicationmain at uikit uiapplication main at bugmauiapp program main at runtime invoke direct void string at
0
20,992
27,855,441,283
IssuesEvent
2023-03-20 22:22:19
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Enable remote terminals to be run in systemd, where available
feature-request terminal-process
<!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> On remote Linux machines, we'd like to be able to run the VS Code terminals within a systemd slice. This enables us to be able to set resource usage limits and direct OOM kills to a particular slice so we can manage processes in our infra. Everything created under a slice parents up to the main slice so the sub-processes (such as large build jobs) can be managed separately while having vscode-server run within a different slice to enable protection for vscode-server and avoid possible connectivity loss.
1.0
Enable remote terminals to be run in systemd, where available - <!-- ⚠️⚠️ Do Not Delete This! feature_request_template ⚠️⚠️ --> <!-- Please read our Rules of Conduct: https://opensource.microsoft.com/codeofconduct/ --> <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> On remote Linux machines, we'd like to be able to run the VS Code terminals within a systemd slice. This enables us to be able to set resource usage limits and direct OOM kills to a particular slice so we can manage processes in our infra. Everything created under a slice parents up to the main slice so the sub-processes (such as large build jobs) can be managed separately while having vscode-server run within a different slice to enable protection for vscode-server and avoid possible connectivity loss.
process
enable remote terminals to be run in systemd where available on remote linux machines we d like to be able to run the vs code terminals within a systemd slice this enables us to be able to set resource usage limits and direct oom kills to a particular slice so we can manage processes in our infra everything created under a slice parents up to the main slice so the sub processes such as large build jobs can be managed separately while having vscode server run within a different slice to enable protection for vscode server and avoid possible connectivity loss
1
761,012
26,663,328,869
IssuesEvent
2023-01-25 23:39:25
cryostatio/cryostat
https://api.github.com/repos/cryostatio/cryostat
closed
[Bug] "View Report" failure/hang/timeout on archived recordings if not previously generated
bug high-priority
## Current Behavior: Report generation fails, appears to hang and then time out on generation in some situations. ## Expected Behavior: Report generation succeeds and the document is visible in a new tab. ## Steps To Reproduce: 1. `sh smoketest.sh` 2. Select a sample app and start a new recording (say 10 seconds, or any) with any template (I used Profiling). Leave "archive on stop" enabled. 3. Wait for recording to complete and be archived 4. Go to Archives > All Archives > sample target > recording > click three dots menu, "view report". Appears to hang processing and then time out and fail. 5. Go to Archives > All Targets > sample target > recording > three dots "view report" and observe similar behaviour. Logs show: ``` ... Jan 25, 2023 4:19:53 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData WARNING: Thread Thread[vert.x-eventloop-thread-2,5,main] has been blocked for 27467 ms, time limit is 2000 ms io.vertx.core.VertxException: Thread blocked at java.base@17.0.5/jdk.internal.misc.Unsafe.park(Native Method) at java.base@17.0.5/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) at java.base@17.0.5/java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1866) at java.base@17.0.5/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3463) at java.base@17.0.5/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3434) at java.base@17.0.5/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1939) at java.base@17.0.5/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095) at app//io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99) at app//io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58) at app//io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148) at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99) at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72) at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189) at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48) at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123) at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65) at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68) at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37) at app//io.cryostat.net.web.WebServer$$Lambda$842/0x00000008013b0238.handle(Unknown Source) at app//io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168) at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57) at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30) at app//io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55) at app//io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158) at app//io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145) at app//io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157) at app//io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93) at app//io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) at app//io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336) at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373) at app//io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236) at app//io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285) at app//io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519) at app//io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458) at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at app//io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at app//io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at app//io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at app//io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base@17.0.5/java.lang.Thread.run(Thread.java:833) Jan 25, 2023 4:19:54 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData WARNING: Thread Thread[vert.x-eventloop-thread-2,5,main] has been blocked for 28467 ms, time limit is 2000 ms io.vertx.core.VertxException: Thread blocked at java.base@17.0.5/jdk.internal.misc.Unsafe.park(Native Method) at java.base@17.0.5/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) at java.base@17.0.5/java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1866) at java.base@17.0.5/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3463) at java.base@17.0.5/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3434) at java.base@17.0.5/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1939) at java.base@17.0.5/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095) at app//io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99) at app//io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58) at app//io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148) at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99) at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72) at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189) at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48) at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123) at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65) at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68) at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37) at app//io.cryostat.net.web.WebServer$$Lambda$842/0x00000008013b0238.handle(Unknown Source) at app//io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168) at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57) at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30) at app//io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55) at app//io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158) at app//io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145) at app//io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157) at app//io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93) at app//io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) at app//io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336) at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373) at app//io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236) at app//io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285) at app//io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519) at app//io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458) at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at app//io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at app//io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at app//io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at app//io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base@17.0.5/java.lang.Thread.run(Thread.java:833) Jan 25, 2023 4:19:55 PM io.cryostat.core.log.Logger error SEVERE: Exception thrown java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1960) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095) at io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99) at io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58) at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148) at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99) at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72) at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189) at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48) at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123) at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65) at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68) at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37) at io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168) at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57) at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30) at io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55) at io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158) at io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145) at io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157) at io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93) at io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373) at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236) at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:833) Jan 25, 2023 4:19:55 PM io.cryostat.core.log.Logger error SEVERE: HTTP 500: java.util.concurrent.TimeoutException io.vertx.ext.web.handler.HttpException: Internal Server Error Caused by: io.cryostat.net.web.http.api.v2.ApiException: java.util.concurrent.TimeoutException at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:119) at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72) at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189) at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48) at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123) at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65) at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68) at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37) at io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168) at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57) at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30) at io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55) at io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158) at io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145) at io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157) at io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93) at io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373) at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236) at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.util.concurrent.ExecutionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:396) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2096) at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:149) at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99) ... 65 more Caused by: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1960) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095) at io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99) at io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58) at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148) ... 66 more Jan 25, 2023 4:19:55 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData SEVERE: 10.0.2.100 - - [Wed, 25 Jan 2023 16:19:55 GMT] 29020ms "GET /api/beta/fs/reports/GAYHANDKNZVEEQKRIEZVO3SMMZTS24KEPJJDOS3BI5QU233TORRXOWKWPFFDE2KXKAZW6PI=/es-andrewazor-demo-Main_foo_20230125T161921Z.jfr/jwt?token=eyJjdHkiOiJKV1QiLCJlbmMiOiJBMjU2R0NNIiwiYWxnIjoiZGlyIn0..UJvZPNJuTOzbFvyT.nlT2sG-PdX78ZWOr2zz88RZgO9E8vNnGMplrislXJZn3zP7dnIEfPIot65oP867sJS7KRiN7CJf2q0s8DC-G_lq07m0VVyp7FoWYM1GYagcupWTqOSEjlcjSUGQ5UBofkj7kZgQHFfDjeHL-32UCOURAoZt1plCYbalLQ64rL-3ARWUC51NcpdHf3zw2SjezsTaMwjWq2uObU-I1uRD_VX7NMe7bSBf2-XW8hHxRyse0gZMpHQPjLv0rl-OlTKlKUKLN8KGwDSE__n6afXKOS2FHkenUTRAuEOxv1nVB2lUSajIZNGfGS3LQ8Qu70b8Du4X2xVWmZuoX0Pa_6kGMP6W0a5aY_RKn8__FV8JrT_DersJ59pdeRaz882KolLS3R6dPSPhz1h9kukukFLbVVm_jlGnGwS0lD3w7IfSEvUFLzohX3rmkpzxYs1WZDGfmvx8eDURDwrxaRWeRrl0KfObNKsDReSWjhmuBtjZSfbMoylh0VWSC8AQcEyp0cByTSVfhohf502zhQX75dxlNd7-TsguI1icGuqvsWvUUSkcNNWlLQZQ9I87mVMWyGwB5qn6_QcRCHP2xai-HKiqesHIwpQvJqHX7O3pd5Z14chiTlTNZ8N8-I_7VMjytUDLzOzHMdAL2aZL_SemNy-n6Vw.12ztQuTsdBJV45gNnk5CfA HTTP/1.1" 500 121 bytes "https://localhost:8181/archives" "Mozilla/5.0 (X11; Linux x86_64; rv:108.0) Gecko/20100101 Firefox/108.0" ``` ![image](https://user-images.githubusercontent.com/3787464/214618186-3aa555fd-639c-4eb3-b899-b95bfc7fece1.png) 6. Expand recording row on either table for this archived recording to generate the report and view it in an inline `iframe`. This works. Then retry the "three dots view report" menu item again, and this now works instantly. ## Anything else: Only affects the 'fromPath", `jwt` versions of reports endpoints, which are used by the web-client for viewing reports in a new tab or download reports as files, not for the inline `iframe` views of reports embedded within recording tables.
1.0
[Bug] "View Report" failure/hang/timeout on archived recordings if not previously generated - ## Current Behavior: Report generation fails, appears to hang and then time out on generation in some situations. ## Expected Behavior: Report generation succeeds and the document is visible in a new tab. ## Steps To Reproduce: 1. `sh smoketest.sh` 2. Select a sample app and start a new recording (say 10 seconds, or any) with any template (I used Profiling). Leave "archive on stop" enabled. 3. Wait for recording to complete and be archived 4. Go to Archives > All Archives > sample target > recording > click three dots menu, "view report". Appears to hang processing and then time out and fail. 5. Go to Archives > All Targets > sample target > recording > three dots "view report" and observe similar behaviour. Logs show: ``` ... Jan 25, 2023 4:19:53 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData WARNING: Thread Thread[vert.x-eventloop-thread-2,5,main] has been blocked for 27467 ms, time limit is 2000 ms io.vertx.core.VertxException: Thread blocked at java.base@17.0.5/jdk.internal.misc.Unsafe.park(Native Method) at java.base@17.0.5/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) at java.base@17.0.5/java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1866) at java.base@17.0.5/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3463) at java.base@17.0.5/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3434) at java.base@17.0.5/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1939) at java.base@17.0.5/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095) at app//io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99) at app//io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58) at app//io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148) at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99) at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72) at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189) at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48) at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123) at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65) at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68) at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37) at app//io.cryostat.net.web.WebServer$$Lambda$842/0x00000008013b0238.handle(Unknown Source) at app//io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168) at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57) at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30) at app//io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55) at app//io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158) at app//io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145) at app//io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157) at app//io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93) at app//io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) at app//io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336) at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373) at app//io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236) at app//io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285) at app//io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519) at app//io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458) at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at app//io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at app//io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at app//io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at app//io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base@17.0.5/java.lang.Thread.run(Thread.java:833) Jan 25, 2023 4:19:54 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData WARNING: Thread Thread[vert.x-eventloop-thread-2,5,main] has been blocked for 28467 ms, time limit is 2000 ms io.vertx.core.VertxException: Thread blocked at java.base@17.0.5/jdk.internal.misc.Unsafe.park(Native Method) at java.base@17.0.5/java.util.concurrent.locks.LockSupport.parkNanos(LockSupport.java:252) at java.base@17.0.5/java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1866) at java.base@17.0.5/java.util.concurrent.ForkJoinPool.unmanagedBlock(ForkJoinPool.java:3463) at java.base@17.0.5/java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3434) at java.base@17.0.5/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1939) at java.base@17.0.5/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095) at app//io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99) at app//io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58) at app//io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148) at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99) at app//io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72) at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189) at app//io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48) at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123) at app//io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65) at app//io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at app//io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at app//io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68) at app//io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37) at app//io.cryostat.net.web.WebServer$$Lambda$842/0x00000008013b0238.handle(Unknown Source) at app//io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168) at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57) at app//io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30) at app//io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55) at app//io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158) at app//io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145) at app//io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157) at app//io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93) at app//io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) at app//io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336) at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373) at app//io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236) at app//io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285) at app//io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519) at app//io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458) at app//io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at app//io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at app//io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at app//io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at app//io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at app//io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at app//io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at app//io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at app//io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at app//io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at app//io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at app//io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base@17.0.5/java.lang.Thread.run(Thread.java:833) Jan 25, 2023 4:19:55 PM io.cryostat.core.log.Logger error SEVERE: Exception thrown java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1960) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095) at io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99) at io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58) at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148) at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99) at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72) at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189) at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48) at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123) at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65) at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68) at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37) at io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168) at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57) at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30) at io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55) at io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158) at io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145) at io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157) at io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93) at io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373) at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236) at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:833) Jan 25, 2023 4:19:55 PM io.cryostat.core.log.Logger error SEVERE: HTTP 500: java.util.concurrent.TimeoutException io.vertx.ext.web.handler.HttpException: Internal Server Error Caused by: io.cryostat.net.web.http.api.v2.ApiException: java.util.concurrent.TimeoutException at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:119) at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:72) at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:189) at io.vertx.ext.web.handler.impl.LoggerHandlerImpl.handle(LoggerHandlerImpl.java:48) at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:123) at io.cryostat.net.web.http.generic.RequestLoggingHandler.handle(RequestLoggingHandler.java:65) at io.vertx.ext.web.impl.RouteState.handleContext(RouteState.java:1284) at io.vertx.ext.web.impl.RoutingContextImplBase.iterateNext(RoutingContextImplBase.java:173) at io.vertx.ext.web.impl.RoutingContextImpl.next(RoutingContextImpl.java:140) at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:68) at io.vertx.ext.web.impl.RouterImpl.handle(RouterImpl.java:37) at io.cryostat.net.HttpServer$HandlerDelegate.handle(HttpServer.java:168) at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:57) at io.vertx.core.http.impl.Http1xServerRequestHandler.handle(Http1xServerRequestHandler.java:30) at io.vertx.core.impl.EventLoopContext.emit(EventLoopContext.java:55) at io.vertx.core.impl.DuplicatedContext.emit(DuplicatedContext.java:158) at io.vertx.core.http.impl.Http1xServerConnection.handleMessage(Http1xServerConnection.java:145) at io.vertx.core.net.impl.ConnectionBase.read(ConnectionBase.java:157) at io.vertx.core.net.impl.VertxHandler.channelRead(VertxHandler.java:153) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.ChannelInboundHandlerAdapter.channelRead(ChannelInboundHandlerAdapter.java:93) at io.netty.handler.codec.http.websocketx.extensions.WebSocketServerExtensionHandler.channelRead(WebSocketServerExtensionHandler.java:99) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.MessageToMessageDecoder.channelRead(MessageToMessageDecoder.java:103) at io.netty.handler.codec.MessageToMessageCodec.channelRead(MessageToMessageCodec.java:111) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:336) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:308) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:280) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.handler.ssl.SslHandler.unwrap(SslHandler.java:1373) at io.netty.handler.ssl.SslHandler.decodeJdkCompatible(SslHandler.java:1236) at io.netty.handler.ssl.SslHandler.decode(SslHandler.java:1285) at io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:519) at io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:458) at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:280) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:357) at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1410) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:379) at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:365) at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:919) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:166) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:788) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:724) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:650) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:562) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:997) at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.base/java.lang.Thread.run(Thread.java:833) Caused by: java.util.concurrent.ExecutionException: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:396) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2096) at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:149) at io.cryostat.net.web.http.api.v2.AbstractAssetJwtConsumingHandler.handle(AbstractAssetJwtConsumingHandler.java:99) ... 65 more Caused by: java.util.concurrent.TimeoutException at java.base/java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1960) at java.base/java.util.concurrent.CompletableFuture.get(CompletableFuture.java:2095) at io.cryostat.net.reports.ArchivedRecordingReportCache.getFromPath(ArchivedRecordingReportCache.java:99) at io.cryostat.net.reports.ReportService.getFromPath(ReportService.java:58) at io.cryostat.net.web.http.api.beta.ReportGetFromPathWithJwtHandler.handleWithValidJwt(ReportGetFromPathWithJwtHandler.java:148) ... 66 more Jan 25, 2023 4:19:55 PM org.slf4j.impl.JDK14LoggerAdapter fillCallerData SEVERE: 10.0.2.100 - - [Wed, 25 Jan 2023 16:19:55 GMT] 29020ms "GET /api/beta/fs/reports/GAYHANDKNZVEEQKRIEZVO3SMMZTS24KEPJJDOS3BI5QU233TORRXOWKWPFFDE2KXKAZW6PI=/es-andrewazor-demo-Main_foo_20230125T161921Z.jfr/jwt?token=eyJjdHkiOiJKV1QiLCJlbmMiOiJBMjU2R0NNIiwiYWxnIjoiZGlyIn0..UJvZPNJuTOzbFvyT.nlT2sG-PdX78ZWOr2zz88RZgO9E8vNnGMplrislXJZn3zP7dnIEfPIot65oP867sJS7KRiN7CJf2q0s8DC-G_lq07m0VVyp7FoWYM1GYagcupWTqOSEjlcjSUGQ5UBofkj7kZgQHFfDjeHL-32UCOURAoZt1plCYbalLQ64rL-3ARWUC51NcpdHf3zw2SjezsTaMwjWq2uObU-I1uRD_VX7NMe7bSBf2-XW8hHxRyse0gZMpHQPjLv0rl-OlTKlKUKLN8KGwDSE__n6afXKOS2FHkenUTRAuEOxv1nVB2lUSajIZNGfGS3LQ8Qu70b8Du4X2xVWmZuoX0Pa_6kGMP6W0a5aY_RKn8__FV8JrT_DersJ59pdeRaz882KolLS3R6dPSPhz1h9kukukFLbVVm_jlGnGwS0lD3w7IfSEvUFLzohX3rmkpzxYs1WZDGfmvx8eDURDwrxaRWeRrl0KfObNKsDReSWjhmuBtjZSfbMoylh0VWSC8AQcEyp0cByTSVfhohf502zhQX75dxlNd7-TsguI1icGuqvsWvUUSkcNNWlLQZQ9I87mVMWyGwB5qn6_QcRCHP2xai-HKiqesHIwpQvJqHX7O3pd5Z14chiTlTNZ8N8-I_7VMjytUDLzOzHMdAL2aZL_SemNy-n6Vw.12ztQuTsdBJV45gNnk5CfA HTTP/1.1" 500 121 bytes "https://localhost:8181/archives" "Mozilla/5.0 (X11; Linux x86_64; rv:108.0) Gecko/20100101 Firefox/108.0" ``` ![image](https://user-images.githubusercontent.com/3787464/214618186-3aa555fd-639c-4eb3-b899-b95bfc7fece1.png) 6. Expand recording row on either table for this archived recording to generate the report and view it in an inline `iframe`. This works. Then retry the "three dots view report" menu item again, and this now works instantly. ## Anything else: Only affects the 'fromPath", `jwt` versions of reports endpoints, which are used by the web-client for viewing reports in a new tab or download reports as files, not for the inline `iframe` views of reports embedded within recording tables.
non_process
view report failure hang timeout on archived recordings if not previously generated current behavior report generation fails appears to hang and then time out on generation in some situations expected behavior report generation succeeds and the document is visible in a new tab steps to reproduce sh smoketest sh select a sample app and start a new recording say seconds or any with any template i used profiling leave archive on stop enabled wait for recording to complete and be archived go to archives all archives sample target recording click three dots menu view report appears to hang processing and then time out and fail go to archives all targets sample target recording three dots view report and observe similar behaviour logs show jan pm org impl fillcallerdata warning thread thread has been blocked for ms time limit is ms io vertx core vertxexception thread blocked at java base jdk internal misc unsafe park native method at java base java util concurrent locks locksupport parknanos locksupport java at java base java util concurrent completablefuture signaller block completablefuture java at java base java util concurrent forkjoinpool unmanagedblock forkjoinpool java at java base java util concurrent forkjoinpool managedblock forkjoinpool java at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at app io cryostat net reports archivedrecordingreportcache getfrompath archivedrecordingreportcache java at app io cryostat net reports reportservice getfrompath reportservice java at app io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java at app io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at app io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at app io vertx ext web impl routestate handlecontext routestate java at app io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at app io vertx ext web impl routingcontextimpl next routingcontextimpl java at app io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at app io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at app io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at app io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at app io vertx ext web impl routestate handlecontext routestate java at app io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at app io vertx ext web impl routingcontextimpl next routingcontextimpl java at app io vertx ext web impl routerimpl handle routerimpl java at app io vertx ext web impl routerimpl handle routerimpl java at app io cryostat net web webserver lambda handle unknown source at app io cryostat net httpserver handlerdelegate handle httpserver java at app io vertx core http impl handle java at app io vertx core http impl handle java at app io vertx core impl eventloopcontext emit eventloopcontext java at app io vertx core impl duplicatedcontext emit duplicatedcontext java at app io vertx core http impl handlemessage java at app io vertx core net impl connectionbase read connectionbase java at app io vertx core net impl vertxhandler channelread vertxhandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at app io netty handler codec http websocketx extensions websocketserverextensionhandler channelread websocketserverextensionhandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at app io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler logging logginghandler channelread logginghandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler ssl sslhandler unwrap sslhandler java at app io netty handler ssl sslhandler decodejdkcompatible sslhandler java at app io netty handler ssl sslhandler decode sslhandler java at app io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at app io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at app io netty channel nio nioeventloop processselectedkey nioeventloop java at app io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at app io netty channel nio nioeventloop processselectedkeys nioeventloop java at app io netty channel nio nioeventloop run nioeventloop java at app io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at app io netty util internal threadexecutormap run threadexecutormap java at app io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java jan pm org impl fillcallerdata warning thread thread has been blocked for ms time limit is ms io vertx core vertxexception thread blocked at java base jdk internal misc unsafe park native method at java base java util concurrent locks locksupport parknanos locksupport java at java base java util concurrent completablefuture signaller block completablefuture java at java base java util concurrent forkjoinpool unmanagedblock forkjoinpool java at java base java util concurrent forkjoinpool managedblock forkjoinpool java at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at app io cryostat net reports archivedrecordingreportcache getfrompath archivedrecordingreportcache java at app io cryostat net reports reportservice getfrompath reportservice java at app io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java at app io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at app io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at app io vertx ext web impl routestate handlecontext routestate java at app io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at app io vertx ext web impl routingcontextimpl next routingcontextimpl java at app io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at app io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at app io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at app io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at app io vertx ext web impl routestate handlecontext routestate java at app io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at app io vertx ext web impl routingcontextimpl next routingcontextimpl java at app io vertx ext web impl routerimpl handle routerimpl java at app io vertx ext web impl routerimpl handle routerimpl java at app io cryostat net web webserver lambda handle unknown source at app io cryostat net httpserver handlerdelegate handle httpserver java at app io vertx core http impl handle java at app io vertx core http impl handle java at app io vertx core impl eventloopcontext emit eventloopcontext java at app io vertx core impl duplicatedcontext emit duplicatedcontext java at app io vertx core http impl handlemessage java at app io vertx core net impl connectionbase read connectionbase java at app io vertx core net impl vertxhandler channelread vertxhandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at app io netty handler codec http websocketx extensions websocketserverextensionhandler channelread websocketserverextensionhandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at app io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler logging logginghandler channelread logginghandler java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty handler ssl sslhandler unwrap sslhandler java at app io netty handler ssl sslhandler decodejdkcompatible sslhandler java at app io netty handler ssl sslhandler decode sslhandler java at app io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at app io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at app io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at app io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at app io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at app io netty channel nio nioeventloop processselectedkey nioeventloop java at app io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at app io netty channel nio nioeventloop processselectedkeys nioeventloop java at app io netty channel nio nioeventloop run nioeventloop java at app io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at app io netty util internal threadexecutormap run threadexecutormap java at app io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java jan pm io cryostat core log logger error severe exception thrown java util concurrent timeoutexception at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at io cryostat net reports archivedrecordingreportcache getfrompath archivedrecordingreportcache java at io cryostat net reports reportservice getfrompath reportservice java at io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at io vertx ext web impl routestate handlecontext routestate java at io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at io vertx ext web impl routingcontextimpl next routingcontextimpl java at io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at io vertx ext web impl routestate handlecontext routestate java at io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at io vertx ext web impl routingcontextimpl next routingcontextimpl java at io vertx ext web impl routerimpl handle routerimpl java at io vertx ext web impl routerimpl handle routerimpl java at io cryostat net httpserver handlerdelegate handle httpserver java at io vertx core http impl handle java at io vertx core http impl handle java at io vertx core impl eventloopcontext emit eventloopcontext java at io vertx core impl duplicatedcontext emit duplicatedcontext java at io vertx core http impl handlemessage java at io vertx core net impl connectionbase read connectionbase java at io vertx core net impl vertxhandler channelread vertxhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at io netty handler codec http websocketx extensions websocketserverextensionhandler channelread websocketserverextensionhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler logging logginghandler channelread logginghandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler ssl sslhandler unwrap sslhandler java at io netty handler ssl sslhandler decodejdkcompatible sslhandler java at io netty handler ssl sslhandler decode sslhandler java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java jan pm io cryostat core log logger error severe http java util concurrent timeoutexception io vertx ext web handler httpexception internal server error caused by io cryostat net web http api apiexception java util concurrent timeoutexception at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java at io vertx ext web impl routestate handlecontext routestate java at io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at io vertx ext web impl routingcontextimpl next routingcontextimpl java at io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at io vertx ext web handler impl loggerhandlerimpl handle loggerhandlerimpl java at io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at io cryostat net web http generic requestlogginghandler handle requestlogginghandler java at io vertx ext web impl routestate handlecontext routestate java at io vertx ext web impl routingcontextimplbase iteratenext routingcontextimplbase java at io vertx ext web impl routingcontextimpl next routingcontextimpl java at io vertx ext web impl routerimpl handle routerimpl java at io vertx ext web impl routerimpl handle routerimpl java at io cryostat net httpserver handlerdelegate handle httpserver java at io vertx core http impl handle java at io vertx core http impl handle java at io vertx core impl eventloopcontext emit eventloopcontext java at io vertx core impl duplicatedcontext emit duplicatedcontext java at io vertx core http impl handlemessage java at io vertx core net impl connectionbase read connectionbase java at io vertx core net impl vertxhandler channelread vertxhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel channelinboundhandleradapter channelread channelinboundhandleradapter java at io netty handler codec http websocketx extensions websocketserverextensionhandler channelread websocketserverextensionhandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec messagetomessagedecoder channelread messagetomessagedecoder java at io netty handler codec messagetomessagecodec channelread messagetomessagecodec java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler logging logginghandler channelread logginghandler java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty handler ssl sslhandler unwrap sslhandler java at io netty handler ssl sslhandler decodejdkcompatible sslhandler java at io netty handler ssl sslhandler decode sslhandler java at io netty handler codec bytetomessagedecoder decoderemovalreentryprotection bytetomessagedecoder java at io netty handler codec bytetomessagedecoder calldecode bytetomessagedecoder java at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java at io netty channel nio nioeventloop processselectedkey nioeventloop java at io netty channel nio nioeventloop processselectedkeysoptimized nioeventloop java at io netty channel nio nioeventloop processselectedkeys nioeventloop java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java caused by java util concurrent executionexception java util concurrent timeoutexception at java base java util concurrent completablefuture reportget completablefuture java at java base java util concurrent completablefuture get completablefuture java at io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java at io cryostat net web http api abstractassetjwtconsuminghandler handle abstractassetjwtconsuminghandler java more caused by java util concurrent timeoutexception at java base java util concurrent completablefuture timedget completablefuture java at java base java util concurrent completablefuture get completablefuture java at io cryostat net reports archivedrecordingreportcache getfrompath archivedrecordingreportcache java at io cryostat net reports reportservice getfrompath reportservice java at io cryostat net web http api beta reportgetfrompathwithjwthandler handlewithvalidjwt reportgetfrompathwithjwthandler java more jan pm org impl fillcallerdata severe get api beta fs reports es andrewazor demo main foo jfr jwt token ujvzpnjutozbfvyt g i semny http bytes mozilla linux rv gecko firefox expand recording row on either table for this archived recording to generate the report and view it in an inline iframe this works then retry the three dots view report menu item again and this now works instantly anything else only affects the frompath jwt versions of reports endpoints which are used by the web client for viewing reports in a new tab or download reports as files not for the inline iframe views of reports embedded within recording tables
0
4,273
3,793,636,933
IssuesEvent
2016-03-22 14:35:36
MISP/MISP
https://api.github.com/repos/MISP/MISP
closed
[UI] - Replace ShadowAttributes in the text to Proposal
usability
[UI] - Replace ShadowAttributes in the text to "Proposals" which much more easier for users to understand.
True
[UI] - Replace ShadowAttributes in the text to Proposal - [UI] - Replace ShadowAttributes in the text to "Proposals" which much more easier for users to understand.
non_process
replace shadowattributes in the text to proposal replace shadowattributes in the text to proposals which much more easier for users to understand
0
12,273
9,606,765,995
IssuesEvent
2019-05-11 13:25:02
microsoft/azure-pipelines-tasks
https://api.github.com/repos/microsoft/azure-pipelines-tasks
closed
Azure DevOps - Not able to run an inline Powershell in "Azure App Service Deploy's" post deployment action
Area: AzureAppService Area: Release question
I am looking to rename a file after deploying a package to the Azure Website. To do so, I understand that while deploying into an Azure Web app using the "Azure App Service Deploy" task, it has a Post Deployment action with which we can run an inline script that will execute on the app service itself - https://github.com/Microsoft/azure-pipelines-tasks/tree/master/Tasks/AzureRmWebAppDeploymentV4 I've tried to execute a simple copy powershell command and it throws an error > [error]Error: EISDIR: illegal operation on a directory, read Copy commands that I've tried : ``` Copy-Item ".\config\config.js" ".\config\<websiteName>.js" -Force ```
1.0
Azure DevOps - Not able to run an inline Powershell in "Azure App Service Deploy's" post deployment action - I am looking to rename a file after deploying a package to the Azure Website. To do so, I understand that while deploying into an Azure Web app using the "Azure App Service Deploy" task, it has a Post Deployment action with which we can run an inline script that will execute on the app service itself - https://github.com/Microsoft/azure-pipelines-tasks/tree/master/Tasks/AzureRmWebAppDeploymentV4 I've tried to execute a simple copy powershell command and it throws an error > [error]Error: EISDIR: illegal operation on a directory, read Copy commands that I've tried : ``` Copy-Item ".\config\config.js" ".\config\<websiteName>.js" -Force ```
non_process
azure devops not able to run an inline powershell in azure app service deploy s post deployment action i am looking to rename a file after deploying a package to the azure website to do so i understand that while deploying into an azure web app using the azure app service deploy task it has a post deployment action with which we can run an inline script that will execute on the app service itself i ve tried to execute a simple copy powershell command and it throws an error error eisdir illegal operation on a directory read copy commands that i ve tried copy item config config js config js force
0
10,333
13,163,425,367
IssuesEvent
2020-08-11 00:22:31
Obe101/eCommerceSite
https://api.github.com/repos/Obe101/eCommerceSite
closed
Add CI Pipeline
Developer process
Add continuous integration pipeline that makes sure code in pull request compiles successfully.
1.0
Add CI Pipeline - Add continuous integration pipeline that makes sure code in pull request compiles successfully.
process
add ci pipeline add continuous integration pipeline that makes sure code in pull request compiles successfully
1
95,172
16,074,076,474
IssuesEvent
2021-04-25 02:16:23
samq-ghdemo/JS-Demo
https://api.github.com/repos/samq-ghdemo/JS-Demo
opened
CVE-2017-1000048 (High) detected in multiple libraries
security vulnerability
## CVE-2017-1000048 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>qs-0.6.6.tgz</b>, <b>qs-5.2.1.tgz</b>, <b>qs-6.2.1.tgz</b></p></summary> <p> <details><summary><b>qs-0.6.6.tgz</b></p></summary> <p>querystring parser</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.6.6.tgz">https://registry.npmjs.org/qs/-/qs-0.6.6.tgz</a></p> <p>Path to dependency file: JS-Demo/package.json</p> <p>Path to vulnerable library: JS-Demo/node_modules/zaproxy/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - zaproxy-0.2.0.tgz (Root Library) - request-2.36.0.tgz - :x: **qs-0.6.6.tgz** (Vulnerable Library) </details> <details><summary><b>qs-5.2.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-5.2.1.tgz">https://registry.npmjs.org/qs/-/qs-5.2.1.tgz</a></p> <p>Path to dependency file: JS-Demo/package.json</p> <p>Path to vulnerable library: JS-Demo/node_modules/grunt-retire/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - grunt-retire-0.3.12.tgz (Root Library) - request-2.67.0.tgz - :x: **qs-5.2.1.tgz** (Vulnerable Library) </details> <details><summary><b>qs-6.2.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.2.1.tgz">https://registry.npmjs.org/qs/-/qs-6.2.1.tgz</a></p> <p>Path to dependency file: JS-Demo/package.json</p> <p>Path to vulnerable library: JS-Demo/node_modules/npm/node_modules/request/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - :x: **qs-6.2.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/JS-Demo/commit/46781df511f58c350408cb5158290290709b373c">46781df511f58c350408cb5158290290709b373c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash. <p>Publish Date: 2017-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p> <p>Release Date: 2017-07-17</p> <p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.6.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;qs:0.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"},{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"5.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;qs:5.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"},{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"6.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;qs:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-1000048","vulnerabilityDetails":"the web framework using ljharb\u0027s qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-1000048 (High) detected in multiple libraries - ## CVE-2017-1000048 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>qs-0.6.6.tgz</b>, <b>qs-5.2.1.tgz</b>, <b>qs-6.2.1.tgz</b></p></summary> <p> <details><summary><b>qs-0.6.6.tgz</b></p></summary> <p>querystring parser</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-0.6.6.tgz">https://registry.npmjs.org/qs/-/qs-0.6.6.tgz</a></p> <p>Path to dependency file: JS-Demo/package.json</p> <p>Path to vulnerable library: JS-Demo/node_modules/zaproxy/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - zaproxy-0.2.0.tgz (Root Library) - request-2.36.0.tgz - :x: **qs-0.6.6.tgz** (Vulnerable Library) </details> <details><summary><b>qs-5.2.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-5.2.1.tgz">https://registry.npmjs.org/qs/-/qs-5.2.1.tgz</a></p> <p>Path to dependency file: JS-Demo/package.json</p> <p>Path to vulnerable library: JS-Demo/node_modules/grunt-retire/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - grunt-retire-0.3.12.tgz (Root Library) - request-2.67.0.tgz - :x: **qs-5.2.1.tgz** (Vulnerable Library) </details> <details><summary><b>qs-6.2.1.tgz</b></p></summary> <p>A querystring parser that supports nesting and arrays, with a depth limit</p> <p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.2.1.tgz">https://registry.npmjs.org/qs/-/qs-6.2.1.tgz</a></p> <p>Path to dependency file: JS-Demo/package.json</p> <p>Path to vulnerable library: JS-Demo/node_modules/npm/node_modules/request/node_modules/qs/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - :x: **qs-6.2.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/JS-Demo/commit/46781df511f58c350408cb5158290290709b373c">46781df511f58c350408cb5158290290709b373c</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> the web framework using ljharb's qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash. <p>Publish Date: 2017-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048>CVE-2017-1000048</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-1000048</a></p> <p>Release Date: 2017-07-17</p> <p>Fix Resolution: qs - 6.0.4,6.1.2,6.2.3,6.3.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"0.6.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"zaproxy:0.2.0;request:2.36.0;qs:0.6.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"},{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"5.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;qs:5.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"},{"packageType":"javascript/Node.js","packageName":"qs","packageVersion":"6.2.1","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;qs:6.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"qs - 6.0.4,6.1.2,6.2.3,6.3.2"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-1000048","vulnerabilityDetails":"the web framework using ljharb\u0027s qs module older than v6.3.2, v6.2.3, v6.1.2, and v6.0.4 is vulnerable to a DoS. A malicious user can send a evil request to cause the web framework crash.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000048","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries qs tgz qs tgz qs tgz qs tgz querystring parser library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules zaproxy node modules qs package json dependency hierarchy zaproxy tgz root library request tgz x qs tgz vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules grunt retire node modules qs package json dependency hierarchy grunt retire tgz root library request tgz x qs tgz vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules npm node modules request node modules qs package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz x qs tgz vulnerable library found in head commit a href found in base branch master vulnerability details the web framework using ljharb s qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree zaproxy request qs isminimumfixversionavailable true minimumfixversion qs packagetype javascript node js packagename qs packageversion packagefilepaths istransitivedependency true dependencytree grunt retire request qs isminimumfixversionavailable true minimumfixversion qs packagetype javascript node js packagename qs packageversion packagefilepaths istransitivedependency true dependencytree grunt npm install npm request qs isminimumfixversionavailable true minimumfixversion qs basebranches vulnerabilityidentifier cve vulnerabilitydetails the web framework using ljharb qs module older than and is vulnerable to a dos a malicious user can send a evil request to cause the web framework crash vulnerabilityurl
0
76,169
26,272,511,279
IssuesEvent
2023-01-06 18:24:44
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: RegularGridInterpolator does not work on non-native byteorder anymore
defect scipy.interpolate
### Describe your issue. With the most recent update, applying the `scipy.interpolate.REgularGridInterpolator` using non-native byteorder raises an exception. This worked in 1.9. ### Reproducing Code Example ```python import scipy import numpy as np from scipy.interpolate import RegularGridInterpolator print(scipy.__version__) x = np.linspace(0, 1, 10) y = np.linspace(0, 1, 15) xg, yg = np.meshgrid(x, y, indexing='ij', sparse=True) # make values non-native byteorder values = (2 * xg + yg).astype('>f8') points = np.array([[0.5, 0.5]]) print(f'{points.dtype = }') print(f'{values.dtype = }') print(f'{x.dtype = }') interpolator = RegularGridInterpolator((x, y), values) print(interpolator(points)) ``` ### Error message ```shell 1.10.0 points.dtype = dtype('float64') values.dtype = dtype('>f8') x.dtype = dtype('float64') Traceback (most recent call last): File "/home/maxnoe/Uni/CTA/ctapipe/test_interp.py", line 22, in <module> print(interpolator(points)) File "/usr/lib/python3.10/site-packages/scipy/interpolate/_rgi.py", line 336, in __call__ result = evaluate_linear_2d(self.values, File "_rgi_cython.pyx", line 19, in scipy.interpolate._rgi_cython.evaluate_linear_2d ValueError: Big-endian buffer not supported on little-endian compiler ``` ### SciPy/NumPy/Python version information 1.10.0 1.23.5 sys.version_info(major=3, minor=10, micro=9, releaselevel='final', serial=0)
1.0
BUG: RegularGridInterpolator does not work on non-native byteorder anymore - ### Describe your issue. With the most recent update, applying the `scipy.interpolate.REgularGridInterpolator` using non-native byteorder raises an exception. This worked in 1.9. ### Reproducing Code Example ```python import scipy import numpy as np from scipy.interpolate import RegularGridInterpolator print(scipy.__version__) x = np.linspace(0, 1, 10) y = np.linspace(0, 1, 15) xg, yg = np.meshgrid(x, y, indexing='ij', sparse=True) # make values non-native byteorder values = (2 * xg + yg).astype('>f8') points = np.array([[0.5, 0.5]]) print(f'{points.dtype = }') print(f'{values.dtype = }') print(f'{x.dtype = }') interpolator = RegularGridInterpolator((x, y), values) print(interpolator(points)) ``` ### Error message ```shell 1.10.0 points.dtype = dtype('float64') values.dtype = dtype('>f8') x.dtype = dtype('float64') Traceback (most recent call last): File "/home/maxnoe/Uni/CTA/ctapipe/test_interp.py", line 22, in <module> print(interpolator(points)) File "/usr/lib/python3.10/site-packages/scipy/interpolate/_rgi.py", line 336, in __call__ result = evaluate_linear_2d(self.values, File "_rgi_cython.pyx", line 19, in scipy.interpolate._rgi_cython.evaluate_linear_2d ValueError: Big-endian buffer not supported on little-endian compiler ``` ### SciPy/NumPy/Python version information 1.10.0 1.23.5 sys.version_info(major=3, minor=10, micro=9, releaselevel='final', serial=0)
non_process
bug regulargridinterpolator does not work on non native byteorder anymore describe your issue with the most recent update applying the scipy interpolate regulargridinterpolator using non native byteorder raises an exception this worked in reproducing code example python import scipy import numpy as np from scipy interpolate import regulargridinterpolator print scipy version x np linspace y np linspace xg yg np meshgrid x y indexing ij sparse true make values non native byteorder values xg yg astype points np array print f points dtype print f values dtype print f x dtype interpolator regulargridinterpolator x y values print interpolator points error message shell points dtype dtype values dtype dtype x dtype dtype traceback most recent call last file home maxnoe uni cta ctapipe test interp py line in print interpolator points file usr lib site packages scipy interpolate rgi py line in call result evaluate linear self values file rgi cython pyx line in scipy interpolate rgi cython evaluate linear valueerror big endian buffer not supported on little endian compiler scipy numpy python version information sys version info major minor micro releaselevel final serial
0
540,696
15,815,892,993
IssuesEvent
2021-04-05 12:09:56
AY2021S2-CS2103T-T12-4/tp
https://api.github.com/repos/AY2021S2-CS2103T-T12-4/tp
closed
[PE-D] URL parameter for adding an API endpoint
priority.High severity.Low type.Bug
When user input url is in the format of www.example.org (without specifying the website protocol, a http protocol is assumed and automatically attached to the front of the user input url. However, it is possible that user does not want such an assumption. ![image.png](https://raw.githubusercontent.com/Nanxi-Huang/ped/main/files/edb74c83-8e63-4c0c-ba32-7319a782b0d7.png) ![image.png](https://raw.githubusercontent.com/Nanxi-Huang/ped/main/files/1d82327c-9dcd-4c05-b0fc-875786b2cf81.png) <!--session: 1617429946124-258a37f6-cce5-4442-9226-410ccb423b7f--> ------------- Labels: `severity.VeryLow` `type.FeatureFlaw` original: Nanxi-Huang/ped#7
1.0
[PE-D] URL parameter for adding an API endpoint - When user input url is in the format of www.example.org (without specifying the website protocol, a http protocol is assumed and automatically attached to the front of the user input url. However, it is possible that user does not want such an assumption. ![image.png](https://raw.githubusercontent.com/Nanxi-Huang/ped/main/files/edb74c83-8e63-4c0c-ba32-7319a782b0d7.png) ![image.png](https://raw.githubusercontent.com/Nanxi-Huang/ped/main/files/1d82327c-9dcd-4c05-b0fc-875786b2cf81.png) <!--session: 1617429946124-258a37f6-cce5-4442-9226-410ccb423b7f--> ------------- Labels: `severity.VeryLow` `type.FeatureFlaw` original: Nanxi-Huang/ped#7
non_process
url parameter for adding an api endpoint when user input url is in the format of without specifying the website protocol a http protocol is assumed and automatically attached to the front of the user input url however it is possible that user does not want such an assumption labels severity verylow type featureflaw original nanxi huang ped
0
21,207
28,262,769,798
IssuesEvent
2023-04-07 01:56:09
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] Exclude already sorted columns from `orderable_columns`
.Team/QueryProcessor :hammer_and_wrench:
We should exclude columns already used in order-by clauses from `metabase.lib.order_by/ordered_columns` output. More context [in Slack](https://metaboat.slack.com/archives/C04DN5VRQM6/p1680623669731539)
1.0
[MLv2] Exclude already sorted columns from `orderable_columns` - We should exclude columns already used in order-by clauses from `metabase.lib.order_by/ordered_columns` output. More context [in Slack](https://metaboat.slack.com/archives/C04DN5VRQM6/p1680623669731539)
process
exclude already sorted columns from orderable columns we should exclude columns already used in order by clauses from metabase lib order by ordered columns output more context
1
12,436
14,931,815,698
IssuesEvent
2021-01-25 06:35:53
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
opened
Make stale read and history read compatible with coprocessor cache
component/coprocessor sig/execution sig/transaction type/enhancement
## Background This is a subtask of #21094. The coprocessor cache caches a response for each request on the TiDB side. If two coprocessor requests are equivalent, the second request may skip being processed on the TiKV side. However, the request is still sent to TiKV, because TiKV needs to judge whether the cached response is still valid. For stale read and history read, the requested snapshot is older than the current. What concerns us is whether the coprocessor cache is still always valid for these kinds of reading. ## Analysis The coprocessor cache is hit when (see `copIteratorWorker.handleTaskOnce` and `copIteratorWorker.handleCopResponse`): - The number of requested ranges, the data size of the response, and the processing time of the request are all within the thresholds. That is, the response deserves to be cached. - The region ID of the request and the cached response equals. That is, the region doesn't split or merge after it's cached. - The start TS of the cached response is not greater than the start TS of the request. That is, the cached response won't be too new for the request. - The applied index of this region doesn't change between the start TS of the cached response and the request. That is, the region is untouched after the response is cached. This is judged on the TiKV side. Through the last two conditions, we can infer that as long as the start TS of the transactions of stale read and history read is set correctly, the cached response will be valid. ## Conclusion We don't need to change anything.
1.0
Make stale read and history read compatible with coprocessor cache - ## Background This is a subtask of #21094. The coprocessor cache caches a response for each request on the TiDB side. If two coprocessor requests are equivalent, the second request may skip being processed on the TiKV side. However, the request is still sent to TiKV, because TiKV needs to judge whether the cached response is still valid. For stale read and history read, the requested snapshot is older than the current. What concerns us is whether the coprocessor cache is still always valid for these kinds of reading. ## Analysis The coprocessor cache is hit when (see `copIteratorWorker.handleTaskOnce` and `copIteratorWorker.handleCopResponse`): - The number of requested ranges, the data size of the response, and the processing time of the request are all within the thresholds. That is, the response deserves to be cached. - The region ID of the request and the cached response equals. That is, the region doesn't split or merge after it's cached. - The start TS of the cached response is not greater than the start TS of the request. That is, the cached response won't be too new for the request. - The applied index of this region doesn't change between the start TS of the cached response and the request. That is, the region is untouched after the response is cached. This is judged on the TiKV side. Through the last two conditions, we can infer that as long as the start TS of the transactions of stale read and history read is set correctly, the cached response will be valid. ## Conclusion We don't need to change anything.
process
make stale read and history read compatible with coprocessor cache background this is a subtask of the coprocessor cache caches a response for each request on the tidb side if two coprocessor requests are equivalent the second request may skip being processed on the tikv side however the request is still sent to tikv because tikv needs to judge whether the cached response is still valid for stale read and history read the requested snapshot is older than the current what concerns us is whether the coprocessor cache is still always valid for these kinds of reading analysis the coprocessor cache is hit when see copiteratorworker handletaskonce and copiteratorworker handlecopresponse the number of requested ranges the data size of the response and the processing time of the request are all within the thresholds that is the response deserves to be cached the region id of the request and the cached response equals that is the region doesn t split or merge after it s cached the start ts of the cached response is not greater than the start ts of the request that is the cached response won t be too new for the request the applied index of this region doesn t change between the start ts of the cached response and the request that is the region is untouched after the response is cached this is judged on the tikv side through the last two conditions we can infer that as long as the start ts of the transactions of stale read and history read is set correctly the cached response will be valid conclusion we don t need to change anything
1
11,099
13,941,131,628
IssuesEvent
2020-10-22 18:56:35
Arch666Angel/mods
https://api.github.com/repos/Arch666Angel/mods
closed
bob mods wood to coal recipe imbalanced with angels mods
Angels Bio Processing
bob mods adds a 2 wood -> 1 coal furnace recipe and it makes more energy than the wood -> wood bricks methods or the wood -> charcoal -> carbon route. https://forums.factorio.com/viewtopic.php?f=185&t=25468&p=510592#p510592 the recipe is bob-coal-from-wood and its in the bob plates mod.
1.0
bob mods wood to coal recipe imbalanced with angels mods - bob mods adds a 2 wood -> 1 coal furnace recipe and it makes more energy than the wood -> wood bricks methods or the wood -> charcoal -> carbon route. https://forums.factorio.com/viewtopic.php?f=185&t=25468&p=510592#p510592 the recipe is bob-coal-from-wood and its in the bob plates mod.
process
bob mods wood to coal recipe imbalanced with angels mods bob mods adds a wood coal furnace recipe and it makes more energy than the wood wood bricks methods or the wood charcoal carbon route the recipe is bob coal from wood and its in the bob plates mod
1
176,582
13,646,532,704
IssuesEvent
2020-09-25 23:27:49
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_backward_ddp_outside (__main__.ProcessGroupDdpUnderDistAutogradTestWithSpawn)
high priority module: distributed module: rpc topic: flaky-tests triage review
## 🐛 test_backward_ddp_outside is flaky See for ex. https://app.circleci.com/pipelines/github/pytorch/pytorch/183902/workflows/9bbc51db-4c8e-4d47-92ea-a43fe60ac4ce/jobs/5952141/steps ``` Jun 22 00:59:56 ====================================================================== Jun 22 00:59:56 ERROR [61.552s]: test_backward_ddp_outside (__main__.TestDdpUnderDistAutogradWrapper) Jun 22 00:59:56 ---------------------------------------------------------------------- Jun 22 00:59:56 Traceback (most recent call last): Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 204, in wrapper Jun 22 00:59:56 self._join_processes(fn) Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 306, in _join_processes Jun 22 00:59:56 self._check_return_codes(elapsed_time) Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 339, in _check_return_codes Jun 22 00:59:56 raise RuntimeError(error) Jun 22 00:59:56 RuntimeError: Processes 5 exited with error code 10 Jun 22 00:59:56 Jun 22 00:59:56 ---------------------------------------------------------------------- ``` cc @ezyang @gchanan @zou3519 @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @xush6528 @osalpekar @jjlilley
1.0
DISABLED test_backward_ddp_outside (__main__.ProcessGroupDdpUnderDistAutogradTestWithSpawn) - ## 🐛 test_backward_ddp_outside is flaky See for ex. https://app.circleci.com/pipelines/github/pytorch/pytorch/183902/workflows/9bbc51db-4c8e-4d47-92ea-a43fe60ac4ce/jobs/5952141/steps ``` Jun 22 00:59:56 ====================================================================== Jun 22 00:59:56 ERROR [61.552s]: test_backward_ddp_outside (__main__.TestDdpUnderDistAutogradWrapper) Jun 22 00:59:56 ---------------------------------------------------------------------- Jun 22 00:59:56 Traceback (most recent call last): Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 204, in wrapper Jun 22 00:59:56 self._join_processes(fn) Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 306, in _join_processes Jun 22 00:59:56 self._check_return_codes(elapsed_time) Jun 22 00:59:56 File "/Users/distiller/workspace/miniconda3/lib/python3.7/site-packages/torch/testing/_internal/common_distributed.py", line 339, in _check_return_codes Jun 22 00:59:56 raise RuntimeError(error) Jun 22 00:59:56 RuntimeError: Processes 5 exited with error code 10 Jun 22 00:59:56 Jun 22 00:59:56 ---------------------------------------------------------------------- ``` cc @ezyang @gchanan @zou3519 @pietern @mrshenli @pritamdamania87 @zhaojuanmao @satgera @rohan-varma @gqchen @aazzolini @xush6528 @osalpekar @jjlilley
non_process
disabled test backward ddp outside main processgroupddpunderdistautogradtestwithspawn 🐛 test backward ddp outside is flaky see for ex jun jun error test backward ddp outside main testddpunderdistautogradwrapper jun jun traceback most recent call last jun file users distiller workspace lib site packages torch testing internal common distributed py line in wrapper jun self join processes fn jun file users distiller workspace lib site packages torch testing internal common distributed py line in join processes jun self check return codes elapsed time jun file users distiller workspace lib site packages torch testing internal common distributed py line in check return codes jun raise runtimeerror error jun runtimeerror processes exited with error code jun jun cc ezyang gchanan pietern mrshenli zhaojuanmao satgera rohan varma gqchen aazzolini osalpekar jjlilley
0
9,669
12,676,147,217
IssuesEvent
2020-06-19 04:09:34
medic/cht-core
https://api.github.com/repos/medic/cht-core
closed
Release 3.9.0
Type: Internal process
# Planning - [x] Create an [organisation wide project](https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc) and add this issue to it. We use [semver](http://semver.org) so if there are breaking changes increment the major, otherwise if there are new features increment the minor, otherwise increment the service pack. Breaking changes in our case relate to updated software requirements (egs: CouchDB, node, minimum browser versions), broken backwards compatibility in an api, or a major visual update that requires user retraining. - [x] Add all the issues to be worked on to the project. Ideally each minor release will have one or two features, a handful of improvements, and plenty of bugs. # Development When development is ready to begin one of the engineers should be nominated as a Release Manager. They will be responsible for making sure the following tasks are completed though not necessarily completing them. - [x] Set the version number in `package.json` and `package-lock.json` and submit a PR. The easiest way to do this is to use `npm --no-git-tag-version version <major|minor>`. - [x] Raise a new issue called `Update dependencies for <version>` with a description that links to [the documentation](https://github.com/medic/medic-docs/blob/master/development/update-dependencies.md). This should be done early in the release cycle so find a volunteer to take this on and assign it to them. - [x] Write an update in the weekly Product Team call agenda summarising development and acceptance testing progress and identifying any blockers. The release manager is to update this every week until the version is released. # Releasing Once all issues have passed acceptance testing and have been merged into `master` release testing can begin. - [x] Create a new release branch from `master` named `<major>.<minor>.x` in medic. Post a message to #development using this template: ``` @core_devs I've just created the `<major>.<minor>.x` release branch. Please be aware that any further changes intended for this release will have to be merged to `master` then backported. Thanks! ``` - [x] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing. - [x] [Import translations keys](https://github.com/medic/medic-docs/blob/master/development/translations.md#adding-new-keys) into POE and notify the #translations Slack channel translate new and updated values, for example: ``` @channel I've just updated the translations in POE. These keys have been added: "<added-list>", and these keys have been updated: "<updated-list>" ``` - [x] Create a new document in the [release-notes folder](https://github.com/medic/medic/tree/master/release-notes) in `master`. Ensure all issues are in the GH Project, that they're correct labelled, and have human readable descriptions. Use [this script](https://github.com/medic/medic/blob/master/scripts/changelog-generator) to export the issues into our changelog format. Manually document any known migration steps and known issues. Provide description, screenshots, videos, and anything else to help communicate particularly important changes. Assign the PR to a) the Director of Technology, and b) an SRE to review and confirm the documentation on upgrade instructions and breaking changes is sufficient. - [x] Create a Google Doc in the [blog posts folder](https://drive.google.com/drive/u/0/folders/0B2PTUNZFwxEvMHRWNTBjY2ZHNHc) with the draft of a blog post promoting the release based on the release notes above. Once it's ready ask Max and Kelly to review it. - [x] Until release testing passes, make sure regressions are fixed in `master`, cherry-pick them into the release branch, and release another beta. - [x] [Export the translations](https://github.com/medic/medic-docs/blob/master/development/translations.md#exporting-changes-from-poeditor-to-github), delete empty translation files and commit to `master`. Cherry-pick the commit into the release branch. - [x] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/medic/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Link to the release notes in the description of the release. - [x] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>` - [x] Add the release to the [Supported versions](https://github.com/medic/medic-docs/blob/master/installation/supported-software.md#supported-versions) and update the EOL date and status of previous releases. - [x] Announce the release in #products and #cht-contributors using this template: ``` @channel *We're excited to announce the release of {{version}}* New features include {{key_features}}. We've also implemented loads of other improvements and fixed a heap of bugs. Read the release notes for full details: {{url}} Following our support policy, versions {{versions}} are no longer supported. Projects running these versions should start planning to upgrade in the near future. For more details read our software support documentation: https://github.com/medic/medic-docs/blob/master/installation/supported-software.md#supported-versions To see what's scheduled for the next releases have a read of the product roadmap: https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc ``` - [x] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category. You can use the previous message and omit `@channel`. - [x] Mark this issue "done" and close the project.
1.0
Release 3.9.0 - # Planning - [x] Create an [organisation wide project](https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc) and add this issue to it. We use [semver](http://semver.org) so if there are breaking changes increment the major, otherwise if there are new features increment the minor, otherwise increment the service pack. Breaking changes in our case relate to updated software requirements (egs: CouchDB, node, minimum browser versions), broken backwards compatibility in an api, or a major visual update that requires user retraining. - [x] Add all the issues to be worked on to the project. Ideally each minor release will have one or two features, a handful of improvements, and plenty of bugs. # Development When development is ready to begin one of the engineers should be nominated as a Release Manager. They will be responsible for making sure the following tasks are completed though not necessarily completing them. - [x] Set the version number in `package.json` and `package-lock.json` and submit a PR. The easiest way to do this is to use `npm --no-git-tag-version version <major|minor>`. - [x] Raise a new issue called `Update dependencies for <version>` with a description that links to [the documentation](https://github.com/medic/medic-docs/blob/master/development/update-dependencies.md). This should be done early in the release cycle so find a volunteer to take this on and assign it to them. - [x] Write an update in the weekly Product Team call agenda summarising development and acceptance testing progress and identifying any blockers. The release manager is to update this every week until the version is released. # Releasing Once all issues have passed acceptance testing and have been merged into `master` release testing can begin. - [x] Create a new release branch from `master` named `<major>.<minor>.x` in medic. Post a message to #development using this template: ``` @core_devs I've just created the `<major>.<minor>.x` release branch. Please be aware that any further changes intended for this release will have to be merged to `master` then backported. Thanks! ``` - [x] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing. - [x] [Import translations keys](https://github.com/medic/medic-docs/blob/master/development/translations.md#adding-new-keys) into POE and notify the #translations Slack channel translate new and updated values, for example: ``` @channel I've just updated the translations in POE. These keys have been added: "<added-list>", and these keys have been updated: "<updated-list>" ``` - [x] Create a new document in the [release-notes folder](https://github.com/medic/medic/tree/master/release-notes) in `master`. Ensure all issues are in the GH Project, that they're correct labelled, and have human readable descriptions. Use [this script](https://github.com/medic/medic/blob/master/scripts/changelog-generator) to export the issues into our changelog format. Manually document any known migration steps and known issues. Provide description, screenshots, videos, and anything else to help communicate particularly important changes. Assign the PR to a) the Director of Technology, and b) an SRE to review and confirm the documentation on upgrade instructions and breaking changes is sufficient. - [x] Create a Google Doc in the [blog posts folder](https://drive.google.com/drive/u/0/folders/0B2PTUNZFwxEvMHRWNTBjY2ZHNHc) with the draft of a blog post promoting the release based on the release notes above. Once it's ready ask Max and Kelly to review it. - [x] Until release testing passes, make sure regressions are fixed in `master`, cherry-pick them into the release branch, and release another beta. - [x] [Export the translations](https://github.com/medic/medic-docs/blob/master/development/translations.md#exporting-changes-from-poeditor-to-github), delete empty translation files and commit to `master`. Cherry-pick the commit into the release branch. - [x] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/medic/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Link to the release notes in the description of the release. - [x] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>` - [x] Add the release to the [Supported versions](https://github.com/medic/medic-docs/blob/master/installation/supported-software.md#supported-versions) and update the EOL date and status of previous releases. - [x] Announce the release in #products and #cht-contributors using this template: ``` @channel *We're excited to announce the release of {{version}}* New features include {{key_features}}. We've also implemented loads of other improvements and fixed a heap of bugs. Read the release notes for full details: {{url}} Following our support policy, versions {{versions}} are no longer supported. Projects running these versions should start planning to upgrade in the near future. For more details read our software support documentation: https://github.com/medic/medic-docs/blob/master/installation/supported-software.md#supported-versions To see what's scheduled for the next releases have a read of the product roadmap: https://github.com/orgs/medic/projects?query=is%3Aopen+sort%3Aname-asc ``` - [x] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category. You can use the previous message and omit `@channel`. - [x] Mark this issue "done" and close the project.
process
release planning create an and add this issue to it we use so if there are breaking changes increment the major otherwise if there are new features increment the minor otherwise increment the service pack breaking changes in our case relate to updated software requirements egs couchdb node minimum browser versions broken backwards compatibility in an api or a major visual update that requires user retraining add all the issues to be worked on to the project ideally each minor release will have one or two features a handful of improvements and plenty of bugs development when development is ready to begin one of the engineers should be nominated as a release manager they will be responsible for making sure the following tasks are completed though not necessarily completing them set the version number in package json and package lock json and submit a pr the easiest way to do this is to use npm no git tag version version raise a new issue called update dependencies for with a description that links to this should be done early in the release cycle so find a volunteer to take this on and assign it to them write an update in the weekly product team call agenda summarising development and acceptance testing progress and identifying any blockers the release manager is to update this every week until the version is released releasing once all issues have passed acceptance testing and have been merged into master release testing can begin create a new release branch from master named x in medic post a message to development using this template core devs i ve just created the x release branch please be aware that any further changes intended for this release will have to be merged to master then backported thanks build a beta named beta by pushing a git tag and when ci completes successfully notify the qa team that it s ready for release testing into poe and notify the translations slack channel translate new and updated values for example channel i ve just updated the translations in poe these keys have been added and these keys have been updated create a new document in the in master ensure all issues are in the gh project that they re correct labelled and have human readable descriptions use to export the issues into our changelog format manually document any known migration steps and known issues provide description screenshots videos and anything else to help communicate particularly important changes assign the pr to a the director of technology and b an sre to review and confirm the documentation on upgrade instructions and breaking changes is sufficient create a google doc in the with the draft of a blog post promoting the release based on the release notes above once it s ready ask max and kelly to review it until release testing passes make sure regressions are fixed in master cherry pick them into the release branch and release another beta delete empty translation files and commit to master cherry pick the commit into the release branch create a release in github from the release branch so it shows up under the with the naming convention this will create the git tag automatically link to the release notes in the description of the release confirm the release build completes successfully and the new release is available on the make sure that the document has new entry with id medic medic add the release to the and update the eol date and status of previous releases announce the release in products and cht contributors using this template channel we re excited to announce the release of version new features include key features we ve also implemented loads of other improvements and fixed a heap of bugs read the release notes for full details url following our support policy versions versions are no longer supported projects running these versions should start planning to upgrade in the near future for more details read our software support documentation to see what s scheduled for the next releases have a read of the product roadmap announce the release on the under the product releases category you can use the previous message and omit channel mark this issue done and close the project
1
437,282
30,594,328,138
IssuesEvent
2023-07-21 20:11:36
hirosystems/docs
https://api.github.com/repos/hirosystems/docs
closed
Registry sample app refactor
documentation help wanted
The registry sample app needs to be updated: - the sample code on the docs page has new lines making the code unreadable - the sample code isn't working anymore, as discussed here: https://github.com/blockstack/docs/issues/1046
1.0
Registry sample app refactor - The registry sample app needs to be updated: - the sample code on the docs page has new lines making the code unreadable - the sample code isn't working anymore, as discussed here: https://github.com/blockstack/docs/issues/1046
non_process
registry sample app refactor the registry sample app needs to be updated the sample code on the docs page has new lines making the code unreadable the sample code isn t working anymore as discussed here
0
158,538
13,735,832,608
IssuesEvent
2020-10-05 10:46:42
kevinpapst/kimai2
https://api.github.com/repos/kevinpapst/kimai2
closed
Token problem
documentation support waiting for feedback
Hello Kevin, I agree on the choice to use tokens with a short duration to protect data security but, in my specific case, the duration of the token is too short and this involves a loss of time optimization because, both me and and my colleagues, we are forced several times a day to clean the cache and log in, especially for those who take a little longer to enter and confirm a registration. Following what has been said, I ask you to kindly tell me, if possible, how to change the settings relating to the token in order to increase the duration of the token by at least 1 or 2 hours. Thanks for the attention Raffaele Colarusso
1.0
Token problem - Hello Kevin, I agree on the choice to use tokens with a short duration to protect data security but, in my specific case, the duration of the token is too short and this involves a loss of time optimization because, both me and and my colleagues, we are forced several times a day to clean the cache and log in, especially for those who take a little longer to enter and confirm a registration. Following what has been said, I ask you to kindly tell me, if possible, how to change the settings relating to the token in order to increase the duration of the token by at least 1 or 2 hours. Thanks for the attention Raffaele Colarusso
non_process
token problem hello kevin i agree on the choice to use tokens with a short duration to protect data security but in my specific case the duration of the token is too short and this involves a loss of time optimization because both me and and my colleagues we are forced several times a day to clean the cache and log in especially for those who take a little longer to enter and confirm a registration following what has been said i ask you to kindly tell me if possible how to change the settings relating to the token in order to increase the duration of the token by at least or hours thanks for the attention raffaele colarusso
0
196,564
22,442,145,139
IssuesEvent
2022-06-21 02:34:43
valdisiljuconoks/AlloyTech
https://api.github.com/repos/valdisiljuconoks/AlloyTech
closed
WS-2019-0047 (Medium) detected in tar-0.1.20.tgz - autoclosed
security vulnerability
## WS-2019-0047 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-0.1.20.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-0.1.20.tgz">https://registry.npmjs.org/tar/-/tar-0.1.20.tgz</a></p> <p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_36a63706-cb38-4759-9162-9c49ba99f793/20190625175906_98125/git_depth_0/AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: /tmp/WhiteSource-ArchiveExtractor_36a63706-cb38-4759-9162-9c49ba99f793/20190625175906_98125/git_depth_0/AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - digdug-1.4.0.tgz - decompress-0.2.3.tgz - :x: **tar-0.1.20.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/valdisiljuconoks/AlloyTech/commits/365b59135886dba1f0332ba419b8693e0e68566f">365b59135886dba1f0332ba419b8693e0e68566f</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of node-tar prior to 4.4.2 are vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink will overwrite the system's file with the contents of the extracted file. <p>Publish Date: 2019-04-05 <p>URL: <a href=https://github.com/npm/node-tar/commit/b0c58433c22f5e7fe8b1c76373f27e3f81dcd4c8>WS-2019-0047</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/803">https://www.npmjs.com/advisories/803</a></p> <p>Release Date: 2019-04-05</p> <p>Fix Resolution: 4.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0047 (Medium) detected in tar-0.1.20.tgz - autoclosed - ## WS-2019-0047 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-0.1.20.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-0.1.20.tgz">https://registry.npmjs.org/tar/-/tar-0.1.20.tgz</a></p> <p>Path to dependency file: /tmp/WhiteSource-ArchiveExtractor_36a63706-cb38-4759-9162-9c49ba99f793/20190625175906_98125/git_depth_0/AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: /tmp/WhiteSource-ArchiveExtractor_36a63706-cb38-4759-9162-9c49ba99f793/20190625175906_98125/git_depth_0/AlloyTech/AlloyTechEpi10/modules/_protected/Shell/Shell/10.1.0.0/ClientResources/lib/xstyle/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - digdug-1.4.0.tgz - decompress-0.2.3.tgz - :x: **tar-0.1.20.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/valdisiljuconoks/AlloyTech/commits/365b59135886dba1f0332ba419b8693e0e68566f">365b59135886dba1f0332ba419b8693e0e68566f</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of node-tar prior to 4.4.2 are vulnerable to Arbitrary File Overwrite. Extracting tarballs containing a hardlink to a file that already exists in the system, and a file that matches the hardlink will overwrite the system's file with the contents of the extracted file. <p>Publish Date: 2019-04-05 <p>URL: <a href=https://github.com/npm/node-tar/commit/b0c58433c22f5e7fe8b1c76373f27e3f81dcd4c8>WS-2019-0047</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/803">https://www.npmjs.com/advisories/803</a></p> <p>Release Date: 2019-04-05</p> <p>Fix Resolution: 4.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in tar tgz autoclosed ws medium severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file tmp whitesource archiveextractor git depth alloytech modules protected shell shell clientresources lib xstyle package json path to vulnerable library tmp whitesource archiveextractor git depth alloytech modules protected shell shell clientresources lib xstyle node modules tar package json dependency hierarchy intern geezer tgz root library digdug tgz decompress tgz x tar tgz vulnerable library found in head commit a href vulnerability details versions of node tar prior to are vulnerable to arbitrary file overwrite extracting tarballs containing a hardlink to a file that already exists in the system and a file that matches the hardlink will overwrite the system s file with the contents of the extracted file publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
59,566
14,422,008,211
IssuesEvent
2020-12-05 01:03:16
jgeraigery/blueocean-environments
https://api.github.com/repos/jgeraigery/blueocean-environments
opened
WS-2019-0427 (Medium) detected in elliptic-6.4.0.tgz
security vulnerability
## WS-2019-0427 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p> <p>Path to dependency file: blueocean-environments/package.json</p> <p>Path to vulnerable library: blueocean-environments/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - browserify-sign-4.0.4.tgz (Root Library) - :x: **elliptic-6.4.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2 <p>Publish Date: 2019-11-22 <p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0427</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a">https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a</a></p> <p>Release Date: 2020-05-24</p> <p>Fix Resolution: v6.5.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.4.0","isTransitiveDependency":true,"dependencyTree":"browserify-sign:4.0.4;elliptic:6.4.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v6.5.2"}],"vulnerabilityIdentifier":"WS-2019-0427","vulnerabilityDetails":"The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2","vulnerabilityUrl":"https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0427 (Medium) detected in elliptic-6.4.0.tgz - ## WS-2019-0427 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p> <p>Path to dependency file: blueocean-environments/package.json</p> <p>Path to vulnerable library: blueocean-environments/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - browserify-sign-4.0.4.tgz (Root Library) - :x: **elliptic-6.4.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2 <p>Publish Date: 2019-11-22 <p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0427</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a">https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a</a></p> <p>Release Date: 2020-05-24</p> <p>Fix Resolution: v6.5.2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.4.0","isTransitiveDependency":true,"dependencyTree":"browserify-sign:4.0.4;elliptic:6.4.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v6.5.2"}],"vulnerabilityIdentifier":"WS-2019-0427","vulnerabilityDetails":"The function getNAF() in elliptic library has information leakage. This issue is mitigated in version 6.5.2","vulnerabilityUrl":"https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_process
ws medium detected in elliptic tgz ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file blueocean environments package json path to vulnerable library blueocean environments node modules elliptic package json dependency hierarchy browserify sign tgz root library x elliptic tgz vulnerable library vulnerability details the function getnaf in elliptic library has information leakage this issue is mitigated in version publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails the function getnaf in elliptic library has information leakage this issue is mitigated in version vulnerabilityurl
0
6,043
2,806,340,278
IssuesEvent
2015-05-15 01:18:43
GoogleCloudPlatform/kubernetes
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
opened
kubectl e2e tests broken
component/CLI priority/P1 team/testing
Several of the kubectl e2e tests were broken today by #8249, which added a missing check of an err return value, all of which now fail when doing cleanup and calling `kubectl stop`. Example: ``` kubectl guestbook should create and stop a working application /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/test/e2e/kubectl.go:125 Error running &{/jenkins-master-data/jobs/kubernetes-e2e-gce/workspace/kubernetes/hack/../cluster/../platforms/linux/amd64/kubectl [kubectl --server=https://104.197.99.179 --kubeconfig=/var/lib/jenkins/jobs/kubernetes-e2e-gce/workspace/.kube/config stop -f examples/guestbook] [] <nil> Error: [Resizing the controller failed with: replicationControllers "frontend" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-master" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-slave" not found; Current resource version Unknown] [] <nil> 0xc2083e6860 exit status 1 <nil> true [0xc20803a3f8 0xc20803a438 0xc20803a470] [0xc20803a3f8 0xc20803a438 0xc20803a470] [0xc20803a430 0xc20803a468] [0x657730 0x657730] 0xc20841cc60}: Command stdout: stderr: Error: [Resizing the controller failed with: replicationControllers "frontend" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-master" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-slave" not found; Current resource version Unknown] ``` I think this may have been originally broken in #6194. That change refactored the implementation of Stop, which started masking the error, but it also changed the order of code a bit. Previously, the Stop implementation in `pkg/kubectl/stop.go` first queried for ReplicationControllers before creating the ResizeCondition; the first thing in the method was ``` rc := reaper.ReplicationControllers(namespace) controller, err := rc.Get(name) ``` After the change, this is done following the ResizeCondition in `pkg/kubectl/resize.go`: ``` rc := &api.ReplicationController{ObjectMeta: api.ObjectMeta{Namespace: namespace, Name: name}} ``` Is it possible that we're trying to query for something that's already been deleted, and thus it will always fail? cc @quinton-hoole @roberthbailey
1.0
kubectl e2e tests broken - Several of the kubectl e2e tests were broken today by #8249, which added a missing check of an err return value, all of which now fail when doing cleanup and calling `kubectl stop`. Example: ``` kubectl guestbook should create and stop a working application /go/src/github.com/GoogleCloudPlatform/kubernetes/_output/dockerized/go/src/github.com/GoogleCloudPlatform/kubernetes/test/e2e/kubectl.go:125 Error running &{/jenkins-master-data/jobs/kubernetes-e2e-gce/workspace/kubernetes/hack/../cluster/../platforms/linux/amd64/kubectl [kubectl --server=https://104.197.99.179 --kubeconfig=/var/lib/jenkins/jobs/kubernetes-e2e-gce/workspace/.kube/config stop -f examples/guestbook] [] <nil> Error: [Resizing the controller failed with: replicationControllers "frontend" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-master" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-slave" not found; Current resource version Unknown] [] <nil> 0xc2083e6860 exit status 1 <nil> true [0xc20803a3f8 0xc20803a438 0xc20803a470] [0xc20803a3f8 0xc20803a438 0xc20803a470] [0xc20803a430 0xc20803a468] [0x657730 0x657730] 0xc20841cc60}: Command stdout: stderr: Error: [Resizing the controller failed with: replicationControllers "frontend" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-master" not found; Current resource version Unknown, Resizing the controller failed with: replicationControllers "redis-slave" not found; Current resource version Unknown] ``` I think this may have been originally broken in #6194. That change refactored the implementation of Stop, which started masking the error, but it also changed the order of code a bit. Previously, the Stop implementation in `pkg/kubectl/stop.go` first queried for ReplicationControllers before creating the ResizeCondition; the first thing in the method was ``` rc := reaper.ReplicationControllers(namespace) controller, err := rc.Get(name) ``` After the change, this is done following the ResizeCondition in `pkg/kubectl/resize.go`: ``` rc := &api.ReplicationController{ObjectMeta: api.ObjectMeta{Namespace: namespace, Name: name}} ``` Is it possible that we're trying to query for something that's already been deleted, and thus it will always fail? cc @quinton-hoole @roberthbailey
non_process
kubectl tests broken several of the kubectl tests were broken today by which added a missing check of an err return value all of which now fail when doing cleanup and calling kubectl stop example kubectl guestbook should create and stop a working application go src github com googlecloudplatform kubernetes output dockerized go src github com googlecloudplatform kubernetes test kubectl go error running jenkins master data jobs kubernetes gce workspace kubernetes hack cluster platforms linux kubectl error exit status true command stdout stderr error i think this may have been originally broken in that change refactored the implementation of stop which started masking the error but it also changed the order of code a bit previously the stop implementation in pkg kubectl stop go first queried for replicationcontrollers before creating the resizecondition the first thing in the method was rc reaper replicationcontrollers namespace controller err rc get name after the change this is done following the resizecondition in pkg kubectl resize go rc api replicationcontroller objectmeta api objectmeta namespace namespace name name is it possible that we re trying to query for something that s already been deleted and thus it will always fail cc quinton hoole roberthbailey
0
46,549
19,293,888,270
IssuesEvent
2021-12-12 08:51:52
amirshnll/rtbf.ir
https://api.github.com/repos/amirshnll/rtbf.ir
closed
اکانتی شاپ
new website/service
"name" : "اکانتی شاپ" "website": "https://accountishop.ir/" "deleteurl": "#" "info": "امکان حذف حساب کاربری وجود ندارد"
1.0
اکانتی شاپ - "name" : "اکانتی شاپ" "website": "https://accountishop.ir/" "deleteurl": "#" "info": "امکان حذف حساب کاربری وجود ندارد"
non_process
اکانتی شاپ name اکانتی شاپ website deleteurl info امکان حذف حساب کاربری وجود ندارد
0
17,712
23,609,520,819
IssuesEvent
2022-08-24 11:10:30
nucleartide/Lets-Make-a-Roguelike-Deckbuilder
https://api.github.com/repos/nucleartide/Lets-Make-a-Roguelike-Deckbuilder
closed
Add a "share" step to process doc for an appropriate stopping point
process
See [Notion](https://www.notion.so/nucleartide/Game-Dev-1a8d6e5960b14adabc66e7b79d620b1a) for personal notes on action items
1.0
Add a "share" step to process doc for an appropriate stopping point - See [Notion](https://www.notion.so/nucleartide/Game-Dev-1a8d6e5960b14adabc66e7b79d620b1a) for personal notes on action items
process
add a share step to process doc for an appropriate stopping point see for personal notes on action items
1