Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
4,066
6,996,181,679
IssuesEvent
2017-12-15 22:50:28
P0cL4bs/WiFi-Pumpkin
https://api.github.com/repos/P0cL4bs/WiFi-Pumpkin
closed
Redirect Traffic from all domain doesn't working
enhancement in process priority solved
## Redirect Traffic from all domain doesn't working ? Hi, i use DNS Spoofer with Windows Update Attack Generator. In DNS Spoofer i activate "Redirect traffic from all domains" but when on AP Wifi i go on any website i have no connection screen but if i disable "Redirect traffic from all domains" and add manualy "google.com" its working on this and only on this. Thanks for helping me (screen of my tool config) ![screenshot from 2017-11-02 21-00-40](https://user-images.githubusercontent.com/20757385/32347547-53adceac-c011-11e7-9cf5-bb61732182c0.png) #### * Card wireless adapters name : Don't know the name but working good. * Version used tool: Last Update * Virtual Machine : No * Operating System and version: Kali Linux 2017.2
1.0
Redirect Traffic from all domain doesn't working - ## Redirect Traffic from all domain doesn't working ? Hi, i use DNS Spoofer with Windows Update Attack Generator. In DNS Spoofer i activate "Redirect traffic from all domains" but when on AP Wifi i go on any website i have no connection screen but if i disable "Redirect traffic from all domains" and add manualy "google.com" its working on this and only on this. Thanks for helping me (screen of my tool config) ![screenshot from 2017-11-02 21-00-40](https://user-images.githubusercontent.com/20757385/32347547-53adceac-c011-11e7-9cf5-bb61732182c0.png) #### * Card wireless adapters name : Don't know the name but working good. * Version used tool: Last Update * Virtual Machine : No * Operating System and version: Kali Linux 2017.2
process
redirect traffic from all domain doesn t working redirect traffic from all domain doesn t working hi i use dns spoofer with windows update attack generator in dns spoofer i activate redirect traffic from all domains but when on ap wifi i go on any website i have no connection screen but if i disable redirect traffic from all domains and add manualy google com its working on this and only on this thanks for helping me screen of my tool config card wireless adapters name don t know the name but working good version used tool last update virtual machine no operating system and version kali linux
1
397,227
27,155,924,072
IssuesEvent
2023-02-17 07:48:04
WordPress/hosting-handbook
https://api.github.com/repos/WordPress/hosting-handbook
closed
Security page changes (6): Cache (Opcache)
documentation WCEU
Some changes, but more or less the same. ACTUAL TEXT: ### OpCache Security PHP opcode caching can significantly improve the performance of PHP processing for WordPress websites, as outlined in the Performance section of the WordPress Hosting Handbook. However, when improperly configured PHP opcode caching can enable users to access other users' PHP files without authorization. There are important PHP configuration options for opcode caching that mitigate vulnerabilities such as accessing files without authorization. #### Validate permission The following setting makes PHP check that the current user has the necessary permissions to access the cached file. It should be enabled at the root php.ini configuration level to prevent users from accessing other users cached files. `opcache.validate_permission = On` This setting is not enabled by default. It is also only available as of PHP 7.0.14. #### Validate root The following setting prevents PHP users from accessing files outside of the chroot'd directory to which they normally would not have access. It should also be added to the root php.ini configuration level to prevent unauthorized access to files. `opcache.validate_root = On` This setting is not enabled by default. It is also only available as of PHP 7.0.14. #### Restrict API Normally, any PHP user can access the opcache API for viewing the currently cached files and for managing the PHP opcode cache. With some PHP configurations, however, the PHP opcode cache shares the same memory between all users on the server. Sharing the PHP opcode cache between all users means all users can view and access the PHP opcode cache and can access other users' cached PHP files. Restricting the Opcache API prevents PHP scripts run in unauthorized directories from viewing cached files and interacting with the PHP opcode cache manually from within PHP scripts. The following setting defines the directory path PHP scripts must start with to be able to access the Opcache API. `opcache.restrict_api = '/some/folder/path'` The default value for the setting is `''`, which means there are no restrictions on which PHP scripts can access the Opcache API. This setting should be defined in the root php.ini for your PHP configuration in order to prevent users from overriding it. NEW PROPOSAL: ### Opcache Security The PHP Opcode can significantly improve the performance of PHP processing, however, when misconfigured it can allow users to access other users' PHP files without authorization. There are important PHP configuration options for opcode caching that mitigate vulnerabilities such as unauthorized file access. #### Access validation The following configuration causes PHP to check that the current user has the necessary permissions to access the cache file. It must be enabled at the root level of php.ini to prevent users from accessing other users' cached files. ``` opcache.validate_permission = on ``` This setting is not activated by default. Available as of PHP 7.0.14. #### Root validation The following configuration prevents PHP users from accessing files outside the chrooted directory that they would not normally have access to. It should also be added to the root level of php.ini to prevent unauthorized access to files. ``` opcache.validate_root = on ``` This setting is not activated by default. Available as of PHP 7.0.14. #### API Restriction Normally any PHP user can access the Opcache API to view the currently cached files and to manage the PHP opcode cache. However, with some PHP configurations, the PHP opcode cache shares the same memory among all users on the server. Restricting the Opcache API prevents PHP scripts from running in directories that are not authorized to view cached files and interact with the PHP opcode cache manually from within the PHP scripts. The following configuration defines the directory path with which PHP scripts must start in order to access the opcache API. ``` opcache.restrict_api = '/some/folder/path ``` The default value for the configuration is `"` (*nothing*), which means that there are no restrictions on which PHP scripts can access the Opcache API. This setting must be defined in the php.ini root of your PHP configuration to prevent users from overriding it.
1.0
Security page changes (6): Cache (Opcache) - Some changes, but more or less the same. ACTUAL TEXT: ### OpCache Security PHP opcode caching can significantly improve the performance of PHP processing for WordPress websites, as outlined in the Performance section of the WordPress Hosting Handbook. However, when improperly configured PHP opcode caching can enable users to access other users' PHP files without authorization. There are important PHP configuration options for opcode caching that mitigate vulnerabilities such as accessing files without authorization. #### Validate permission The following setting makes PHP check that the current user has the necessary permissions to access the cached file. It should be enabled at the root php.ini configuration level to prevent users from accessing other users cached files. `opcache.validate_permission = On` This setting is not enabled by default. It is also only available as of PHP 7.0.14. #### Validate root The following setting prevents PHP users from accessing files outside of the chroot'd directory to which they normally would not have access. It should also be added to the root php.ini configuration level to prevent unauthorized access to files. `opcache.validate_root = On` This setting is not enabled by default. It is also only available as of PHP 7.0.14. #### Restrict API Normally, any PHP user can access the opcache API for viewing the currently cached files and for managing the PHP opcode cache. With some PHP configurations, however, the PHP opcode cache shares the same memory between all users on the server. Sharing the PHP opcode cache between all users means all users can view and access the PHP opcode cache and can access other users' cached PHP files. Restricting the Opcache API prevents PHP scripts run in unauthorized directories from viewing cached files and interacting with the PHP opcode cache manually from within PHP scripts. The following setting defines the directory path PHP scripts must start with to be able to access the Opcache API. `opcache.restrict_api = '/some/folder/path'` The default value for the setting is `''`, which means there are no restrictions on which PHP scripts can access the Opcache API. This setting should be defined in the root php.ini for your PHP configuration in order to prevent users from overriding it. NEW PROPOSAL: ### Opcache Security The PHP Opcode can significantly improve the performance of PHP processing, however, when misconfigured it can allow users to access other users' PHP files without authorization. There are important PHP configuration options for opcode caching that mitigate vulnerabilities such as unauthorized file access. #### Access validation The following configuration causes PHP to check that the current user has the necessary permissions to access the cache file. It must be enabled at the root level of php.ini to prevent users from accessing other users' cached files. ``` opcache.validate_permission = on ``` This setting is not activated by default. Available as of PHP 7.0.14. #### Root validation The following configuration prevents PHP users from accessing files outside the chrooted directory that they would not normally have access to. It should also be added to the root level of php.ini to prevent unauthorized access to files. ``` opcache.validate_root = on ``` This setting is not activated by default. Available as of PHP 7.0.14. #### API Restriction Normally any PHP user can access the Opcache API to view the currently cached files and to manage the PHP opcode cache. However, with some PHP configurations, the PHP opcode cache shares the same memory among all users on the server. Restricting the Opcache API prevents PHP scripts from running in directories that are not authorized to view cached files and interact with the PHP opcode cache manually from within the PHP scripts. The following configuration defines the directory path with which PHP scripts must start in order to access the opcache API. ``` opcache.restrict_api = '/some/folder/path ``` The default value for the configuration is `"` (*nothing*), which means that there are no restrictions on which PHP scripts can access the Opcache API. This setting must be defined in the php.ini root of your PHP configuration to prevent users from overriding it.
non_process
security page changes cache opcache some changes but more or less the same actual text opcache security php opcode caching can significantly improve the performance of php processing for wordpress websites as outlined in the performance section of the wordpress hosting handbook however when improperly configured php opcode caching can enable users to access other users php files without authorization there are important php configuration options for opcode caching that mitigate vulnerabilities such as accessing files without authorization validate permission the following setting makes php check that the current user has the necessary permissions to access the cached file it should be enabled at the root php ini configuration level to prevent users from accessing other users cached files opcache validate permission on this setting is not enabled by default it is also only available as of php validate root the following setting prevents php users from accessing files outside of the chroot d directory to which they normally would not have access it should also be added to the root php ini configuration level to prevent unauthorized access to files opcache validate root on this setting is not enabled by default it is also only available as of php restrict api normally any php user can access the opcache api for viewing the currently cached files and for managing the php opcode cache with some php configurations however the php opcode cache shares the same memory between all users on the server sharing the php opcode cache between all users means all users can view and access the php opcode cache and can access other users cached php files restricting the opcache api prevents php scripts run in unauthorized directories from viewing cached files and interacting with the php opcode cache manually from within php scripts the following setting defines the directory path php scripts must start with to be able to access the opcache api opcache restrict api some folder path the default value for the setting is which means there are no restrictions on which php scripts can access the opcache api this setting should be defined in the root php ini for your php configuration in order to prevent users from overriding it new proposal opcache security the php opcode can significantly improve the performance of php processing however when misconfigured it can allow users to access other users php files without authorization there are important php configuration options for opcode caching that mitigate vulnerabilities such as unauthorized file access access validation the following configuration causes php to check that the current user has the necessary permissions to access the cache file it must be enabled at the root level of php ini to prevent users from accessing other users cached files opcache validate permission on this setting is not activated by default available as of php root validation the following configuration prevents php users from accessing files outside the chrooted directory that they would not normally have access to it should also be added to the root level of php ini to prevent unauthorized access to files opcache validate root on this setting is not activated by default available as of php api restriction normally any php user can access the opcache api to view the currently cached files and to manage the php opcode cache however with some php configurations the php opcode cache shares the same memory among all users on the server restricting the opcache api prevents php scripts from running in directories that are not authorized to view cached files and interact with the php opcode cache manually from within the php scripts the following configuration defines the directory path with which php scripts must start in order to access the opcache api opcache restrict api some folder path the default value for the configuration is nothing which means that there are no restrictions on which php scripts can access the opcache api this setting must be defined in the php ini root of your php configuration to prevent users from overriding it
0
2,065
4,876,031,751
IssuesEvent
2016-11-16 11:27:52
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Consider setting transaction_persistent=1 by default.
ADMIN CONNECTION POOL QUERY PROCESSOR ROUTING
I've seen comments which explains why `transaction_persistent=0` by default .(https://github.com/sysown/proxysql/issues/653) This issue here is to discuss this, after reading the comment, I still think `transaction_persistent=1` should be default. ### why `0` is good Yes, I understand that it is possible that some connectors, applications, frameworks and many other tools disable `autocommit=0` by default or wrap everything inside a transaction. so yes, having `transaction_persistent=1` by default could indeed lead the query routing behaviour to not work and feel not flexible as an unknowing inexperienced human would expect when (s)he's just adding some users or just copy pastes something from a blogpost found on the internet and then figures out that that `proxysql` tool did not send those darn selects to a replica. ### why `1` is good However, that exact same inexperienced person who is a dba/sysadmin might actually create some major problems in an application with the current default behaviour of `transaction_persistent=0`. Suddenly after setting up proxysql, `select` statements within the transaction go to a slave and that inexperienced person might become very happy after checking his statistics seeing that the query went to the slave. But in turn, several of these `select`s actually start to show different results as it would have when executed within the transaction, as maybe the `select` read some data that is being changed in that transaction, and it might make wrong decisions for the application to see what to perform as `insert` as next query. This problem is similar to sending all/ a lot of reads to asynchronous slaves. Even in this case, I am very careful before I tell my customers in for example healthcare and financial sectors to just send those reads to slaves. ### plea for `1` default! To summarise, I think that, because we are dealing with data, big and small, we should aim at keeping safety/expected behavior first, only when we really want it, we can have that flexibility and set `transaction_persistent=0` when it is really required. Let's make ProxySQL proper-expected-transactional-behavior Default (and Great) Again!!
1.0
Consider setting transaction_persistent=1 by default. - I've seen comments which explains why `transaction_persistent=0` by default .(https://github.com/sysown/proxysql/issues/653) This issue here is to discuss this, after reading the comment, I still think `transaction_persistent=1` should be default. ### why `0` is good Yes, I understand that it is possible that some connectors, applications, frameworks and many other tools disable `autocommit=0` by default or wrap everything inside a transaction. so yes, having `transaction_persistent=1` by default could indeed lead the query routing behaviour to not work and feel not flexible as an unknowing inexperienced human would expect when (s)he's just adding some users or just copy pastes something from a blogpost found on the internet and then figures out that that `proxysql` tool did not send those darn selects to a replica. ### why `1` is good However, that exact same inexperienced person who is a dba/sysadmin might actually create some major problems in an application with the current default behaviour of `transaction_persistent=0`. Suddenly after setting up proxysql, `select` statements within the transaction go to a slave and that inexperienced person might become very happy after checking his statistics seeing that the query went to the slave. But in turn, several of these `select`s actually start to show different results as it would have when executed within the transaction, as maybe the `select` read some data that is being changed in that transaction, and it might make wrong decisions for the application to see what to perform as `insert` as next query. This problem is similar to sending all/ a lot of reads to asynchronous slaves. Even in this case, I am very careful before I tell my customers in for example healthcare and financial sectors to just send those reads to slaves. ### plea for `1` default! To summarise, I think that, because we are dealing with data, big and small, we should aim at keeping safety/expected behavior first, only when we really want it, we can have that flexibility and set `transaction_persistent=0` when it is really required. Let's make ProxySQL proper-expected-transactional-behavior Default (and Great) Again!!
process
consider setting transaction persistent by default i ve seen comments which explains why transaction persistent by default this issue here is to discuss this after reading the comment i still think transaction persistent should be default why is good yes i understand that it is possible that some connectors applications frameworks and many other tools disable autocommit by default or wrap everything inside a transaction so yes having transaction persistent by default could indeed lead the query routing behaviour to not work and feel not flexible as an unknowing inexperienced human would expect when s he s just adding some users or just copy pastes something from a blogpost found on the internet and then figures out that that proxysql tool did not send those darn selects to a replica why is good however that exact same inexperienced person who is a dba sysadmin might actually create some major problems in an application with the current default behaviour of transaction persistent suddenly after setting up proxysql select statements within the transaction go to a slave and that inexperienced person might become very happy after checking his statistics seeing that the query went to the slave but in turn several of these select s actually start to show different results as it would have when executed within the transaction as maybe the select read some data that is being changed in that transaction and it might make wrong decisions for the application to see what to perform as insert as next query this problem is similar to sending all a lot of reads to asynchronous slaves even in this case i am very careful before i tell my customers in for example healthcare and financial sectors to just send those reads to slaves plea for default to summarise i think that because we are dealing with data big and small we should aim at keeping safety expected behavior first only when we really want it we can have that flexibility and set transaction persistent when it is really required let s make proxysql proper expected transactional behavior default and great again
1
16,543
21,568,582,415
IssuesEvent
2022-05-02 04:16:06
jmacost5/CPP-528-Project
https://api.github.com/repos/jmacost5/CPP-528-Project
closed
Add file names and file types to the .gitignore file to prevent them from syncing to the team repository
TEAM PROCESS
Use the R specific .gitignore file courtesy of GitHub
1.0
Add file names and file types to the .gitignore file to prevent them from syncing to the team repository - Use the R specific .gitignore file courtesy of GitHub
process
add file names and file types to the gitignore file to prevent them from syncing to the team repository use the r specific gitignore file courtesy of github
1
682
3,163,302,786
IssuesEvent
2015-09-20 05:03:39
spootTheLousy/saguaro
https://api.github.com/repos/spootTheLousy/saguaro
closed
Captcha doesn't disable for admin form
Admin Function Bug: Minor Post/text processing
Probably an issue with regist. Should be a simple fix using valid()
1.0
Captcha doesn't disable for admin form - Probably an issue with regist. Should be a simple fix using valid()
process
captcha doesn t disable for admin form probably an issue with regist should be a simple fix using valid
1
2,782
5,716,439,801
IssuesEvent
2017-04-19 15:08:54
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Filter request url from into log file
log-processing question
Hi I want to know if it's possible to filter the request URL from a log file that's means If I could Know what functionality the user use of my application since the url?
1.0
Filter request url from into log file - Hi I want to know if it's possible to filter the request URL from a log file that's means If I could Know what functionality the user use of my application since the url?
process
filter request url from into log file hi i want to know if it s possible to filter the request url from a log file that s means if i could know what functionality the user use of my application since the url
1
14,862
18,267,411,270
IssuesEvent
2021-10-04 10:03:29
googleapis/java-logging
https://api.github.com/repos/googleapis/java-logging
closed
blocking batch flow control belies async logging
type: process api: logging lang: java
google-cloud-logging enables the blocking flow controller in the GAX batching library, going to lengths to modify the [default gapic behavior](https://github.com/googleapis/googleapis/blob/ee3c7eb3401f8de84348046b71fba7b1ac2215cf/google/logging/v2/logging_gapic.yaml#L22): https://github.com/googleapis/java-logging/blob/7b3928e1df1986bfd6fdd1197e22bcc00ce96a6d/google-cloud-logging/src/main/java/com/google/cloud/logging/spi/v2/GrpcLoggingRpc.java#L168-L181 The trouble with the blocking batch flow controller is that it can block `LoggingImpl.writeAsync` and thus any application code that happens to log after the flow controller is full. (That can happen either because of too much application logging or because of a slow-down in the logging service.) Preferable behavior would be dropping the log messages that cannot be uploaded and leaving a note about dropped messages in the next batch or stderr. (This issue is notwithstanding a recent regression described in #632. That regression occludes the problematic behavior described here by removing the GAX batcher.)
1.0
blocking batch flow control belies async logging - google-cloud-logging enables the blocking flow controller in the GAX batching library, going to lengths to modify the [default gapic behavior](https://github.com/googleapis/googleapis/blob/ee3c7eb3401f8de84348046b71fba7b1ac2215cf/google/logging/v2/logging_gapic.yaml#L22): https://github.com/googleapis/java-logging/blob/7b3928e1df1986bfd6fdd1197e22bcc00ce96a6d/google-cloud-logging/src/main/java/com/google/cloud/logging/spi/v2/GrpcLoggingRpc.java#L168-L181 The trouble with the blocking batch flow controller is that it can block `LoggingImpl.writeAsync` and thus any application code that happens to log after the flow controller is full. (That can happen either because of too much application logging or because of a slow-down in the logging service.) Preferable behavior would be dropping the log messages that cannot be uploaded and leaving a note about dropped messages in the next batch or stderr. (This issue is notwithstanding a recent regression described in #632. That regression occludes the problematic behavior described here by removing the GAX batcher.)
process
blocking batch flow control belies async logging google cloud logging enables the blocking flow controller in the gax batching library going to lengths to modify the the trouble with the blocking batch flow controller is that it can block loggingimpl writeasync and thus any application code that happens to log after the flow controller is full that can happen either because of too much application logging or because of a slow down in the logging service preferable behavior would be dropping the log messages that cannot be uploaded and leaving a note about dropped messages in the next batch or stderr this issue is notwithstanding a recent regression described in that regression occludes the problematic behavior described here by removing the gax batcher
1
609
3,078,014,780
IssuesEvent
2015-08-21 07:05:10
maraujop/django-crispy-forms
https://api.github.com/repos/maraujop/django-crispy-forms
closed
"Not a registered tag library" against Django Pre-1.9 master
Testing/Process
Getting an error running against latest Django master: TemplateSyntaxError: 'crispy_forms_tags' is not a registered tag library. Must be one of: (Note there's literally nothing in the output after the `:` there — i.e. it looks like a list is expected but missing.) [See here for an example build](https://travis-ci.org/maraujop/django-crispy-forms/builds/75753379)
1.0
"Not a registered tag library" against Django Pre-1.9 master - Getting an error running against latest Django master: TemplateSyntaxError: 'crispy_forms_tags' is not a registered tag library. Must be one of: (Note there's literally nothing in the output after the `:` there — i.e. it looks like a list is expected but missing.) [See here for an example build](https://travis-ci.org/maraujop/django-crispy-forms/builds/75753379)
process
not a registered tag library against django pre master getting an error running against latest django master templatesyntaxerror crispy forms tags is not a registered tag library must be one of note there s literally nothing in the output after the there — i e it looks like a list is expected but missing
1
17,081
22,586,510,273
IssuesEvent
2022-06-28 15:40:13
metallb/metallb
https://api.github.com/repos/metallb/metallb
closed
Issues with helm release process
process helm
I just published v0.10.0, which is our first release since including the new helm chart. Unfortunately the helm release portion of the release process didn't work correctly. https://github.com/metallb/metallb/runs/2767270351?check_suite_focus=true We have the chart-releaser github action set to run against tags, with the idea that we would publish new helm chart releases in parallel with metallb releases. The chart-releaser action determined that no chart release was necessary, though. Based on the [docs of the action](https://github.com/helm/chart-releaser-action/), it sounds like running against the `main` branch is the common way this action is used. We either need to figure out what changes are necessary to make this work against tags, or adjust our process to release the helm chart from `main`. If the helm chart is *only* released from main when we update the `charts/metallb/Chart.yaml` file with a new version, then that is probably OK.
1.0
Issues with helm release process - I just published v0.10.0, which is our first release since including the new helm chart. Unfortunately the helm release portion of the release process didn't work correctly. https://github.com/metallb/metallb/runs/2767270351?check_suite_focus=true We have the chart-releaser github action set to run against tags, with the idea that we would publish new helm chart releases in parallel with metallb releases. The chart-releaser action determined that no chart release was necessary, though. Based on the [docs of the action](https://github.com/helm/chart-releaser-action/), it sounds like running against the `main` branch is the common way this action is used. We either need to figure out what changes are necessary to make this work against tags, or adjust our process to release the helm chart from `main`. If the helm chart is *only* released from main when we update the `charts/metallb/Chart.yaml` file with a new version, then that is probably OK.
process
issues with helm release process i just published which is our first release since including the new helm chart unfortunately the helm release portion of the release process didn t work correctly we have the chart releaser github action set to run against tags with the idea that we would publish new helm chart releases in parallel with metallb releases the chart releaser action determined that no chart release was necessary though based on the it sounds like running against the main branch is the common way this action is used we either need to figure out what changes are necessary to make this work against tags or adjust our process to release the helm chart from main if the helm chart is only released from main when we update the charts metallb chart yaml file with a new version then that is probably ok
1
3,003
5,997,235,639
IssuesEvent
2017-06-03 21:51:32
alexrj/Slic3r
https://api.github.com/repos/alexrj/Slic3r
closed
Perform Background Processing also while the preset editor dialog is open
Background Processing Feature request
Fresh build on April 2, 2017: 1.2.9-901-g90623c4 OS: Debian Stable Description: Background Processing is enabled, but changes in the Print Settings window do not update until the Print Settings window is closed.
1.0
Perform Background Processing also while the preset editor dialog is open - Fresh build on April 2, 2017: 1.2.9-901-g90623c4 OS: Debian Stable Description: Background Processing is enabled, but changes in the Print Settings window do not update until the Print Settings window is closed.
process
perform background processing also while the preset editor dialog is open fresh build on april os debian stable description background processing is enabled but changes in the print settings window do not update until the print settings window is closed
1
21,060
28,007,608,601
IssuesEvent
2023-03-27 16:13:17
varabyte/kobweb
https://api.github.com/repos/varabyte/kobweb
opened
Lighthouse investigation
process
Do some [lighthouse analysis](https://github.com/GoogleChrome/lighthouse/blob/HEAD/docs/user-flows.md) across several pages that were rendering with Kobweb, make sure what's being generated is generally healthy. This bug can be used to document the list of things worth visiting to improve lighthouse scorse. For example, here's the current score for https://bitspittle.dev at the time of creating this issue: ![welcome-to-bitspittle-lighthouse](https://user-images.githubusercontent.com/43705986/228000391-e40561a8-a73c-4111-88d0-1d21e099ce46.png) a11y looks like it needs some digging into. Not sure yet if that's a Kobweb issue OR an issue with how I made that page.
1.0
Lighthouse investigation - Do some [lighthouse analysis](https://github.com/GoogleChrome/lighthouse/blob/HEAD/docs/user-flows.md) across several pages that were rendering with Kobweb, make sure what's being generated is generally healthy. This bug can be used to document the list of things worth visiting to improve lighthouse scorse. For example, here's the current score for https://bitspittle.dev at the time of creating this issue: ![welcome-to-bitspittle-lighthouse](https://user-images.githubusercontent.com/43705986/228000391-e40561a8-a73c-4111-88d0-1d21e099ce46.png) a11y looks like it needs some digging into. Not sure yet if that's a Kobweb issue OR an issue with how I made that page.
process
lighthouse investigation do some across several pages that were rendering with kobweb make sure what s being generated is generally healthy this bug can be used to document the list of things worth visiting to improve lighthouse scorse for example here s the current score for at the time of creating this issue looks like it needs some digging into not sure yet if that s a kobweb issue or an issue with how i made that page
1
11,382
17,019,102,129
IssuesEvent
2021-07-02 16:00:52
samoody2/wiki
https://api.github.com/repos/samoody2/wiki
closed
search suggestion
project requirement
If the query does not match the name of an encyclopedia entry, the user should instead be taken to a search results page that displays a list of all encyclopedia entries that have the query as a substring. For example, if the search query were ytho, then Python should appear in the search results.
1.0
search suggestion - If the query does not match the name of an encyclopedia entry, the user should instead be taken to a search results page that displays a list of all encyclopedia entries that have the query as a substring. For example, if the search query were ytho, then Python should appear in the search results.
non_process
search suggestion if the query does not match the name of an encyclopedia entry the user should instead be taken to a search results page that displays a list of all encyclopedia entries that have the query as a substring for example if the search query were ytho then python should appear in the search results
0
547,246
16,040,703,826
IssuesEvent
2021-04-22 07:30:03
googleapis/java-aiplatform
https://api.github.com/repos/googleapis/java-aiplatform
reopened
aiplatform.CreateBatchPredictionJobTextEntityExtractionSampleTest: testCreateBatchPredictionJobTextEntityExtractionSample failed
api: aiplatform flakybot: flaky flakybot: issue priority: p1 type: bug
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 8cfdab67f373cbd7aa3f139a82c1d63502a360fa buildURL: [Build Status](https://source.cloud.google.com/results/invocations/109b6dd4-4ffe-453b-8e53-5e15c4a70504), [Sponge](http://sponge2/109b6dd4-4ffe-453b-8e53-5e15c4a70504) status: failed <details><summary>Test output</summary><br><pre>java.lang.ArrayIndexOutOfBoundsException: 1 at aiplatform.CreateBatchPredictionJobTextEntityExtractionSampleTest.tearDown(CreateBatchPredictionJobTextEntityExtractionSampleTest.java:71) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) </pre></details>
1.0
aiplatform.CreateBatchPredictionJobTextEntityExtractionSampleTest: testCreateBatchPredictionJobTextEntityExtractionSample failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 8cfdab67f373cbd7aa3f139a82c1d63502a360fa buildURL: [Build Status](https://source.cloud.google.com/results/invocations/109b6dd4-4ffe-453b-8e53-5e15c4a70504), [Sponge](http://sponge2/109b6dd4-4ffe-453b-8e53-5e15c4a70504) status: failed <details><summary>Test output</summary><br><pre>java.lang.ArrayIndexOutOfBoundsException: 1 at aiplatform.CreateBatchPredictionJobTextEntityExtractionSampleTest.tearDown(CreateBatchPredictionJobTextEntityExtractionSampleTest.java:71) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunAfters.invokeMethod(RunAfters.java:46) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:33) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) </pre></details>
non_process
aiplatform createbatchpredictionjobtextentityextractionsampletest testcreatebatchpredictionjobtextentityextractionsample failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output java lang arrayindexoutofboundsexception at aiplatform createbatchpredictionjobtextentityextractionsampletest teardown createbatchpredictionjobtextentityextractionsampletest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements runafters invokemethod runafters java at org junit internal runners statements runafters evaluate runafters java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit internal runners statements runbefores evaluate runbefores java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java
0
183
2,588,444,240
IssuesEvent
2015-02-18 01:24:19
GsDevKit/gsDevKitHome
https://api.github.com/repos/GsDevKit/gsDevKitHome
closed
`Do not clone as root` ...
in process
It is worth mentioning that when the initial clone is made there is no need to run the clone as root and that it should be cloned such that the normal user has write permissions ... `sudo` is used where and when `root` operations are performed
1.0
`Do not clone as root` ... - It is worth mentioning that when the initial clone is made there is no need to run the clone as root and that it should be cloned such that the normal user has write permissions ... `sudo` is used where and when `root` operations are performed
process
do not clone as root it is worth mentioning that when the initial clone is made there is no need to run the clone as root and that it should be cloned such that the normal user has write permissions sudo is used where and when root operations are performed
1
29,945
8,444,315,766
IssuesEvent
2018-10-18 18:04:55
gimli-org/gimli
https://api.github.com/repos/gimli-org/gimli
closed
Fix install script for Mac OS
building and distribution
We should test the curl installer and give additional instructions on the website for manual compilation on MacOS. Any experiences & help from Mac users is welcome.
1.0
Fix install script for Mac OS - We should test the curl installer and give additional instructions on the website for manual compilation on MacOS. Any experiences & help from Mac users is welcome.
non_process
fix install script for mac os we should test the curl installer and give additional instructions on the website for manual compilation on macos any experiences help from mac users is welcome
0
261,468
8,233,709,340
IssuesEvent
2018-09-08 04:35:30
Sakuten/backend
https://api.github.com/repos/Sakuten/backend
opened
Use `app.logger.warning` instead of `app.logger.warn`
low priority refactoring small
<!-- あくまでテンプレートなので必ずしもすべての項目を埋めなくてよい --> Step 1: 目的 ============ * `The 'warn' method is deprecated, use 'warning' instead` Step 2: 概要 ============ * Use `app.logger.warning` instead of `app.logger.warn`
1.0
Use `app.logger.warning` instead of `app.logger.warn` - <!-- あくまでテンプレートなので必ずしもすべての項目を埋めなくてよい --> Step 1: 目的 ============ * `The 'warn' method is deprecated, use 'warning' instead` Step 2: 概要 ============ * Use `app.logger.warning` instead of `app.logger.warn`
non_process
use app logger warning instead of app logger warn step 目的 the warn method is deprecated use warning instead step 概要 use app logger warning instead of app logger warn
0
60,924
14,596,421,154
IssuesEvent
2020-12-20 15:46:59
billmcchesney1/superagent
https://api.github.com/repos/billmcchesney1/superagent
opened
CVE-2020-11023 (Medium) detected in jquery-1.7.1.min.js
security vulnerability
## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: superagent/node_modules/watchify/node_modules/vm-browserify/example/run/index.html</p> <p>Path to vulnerable library: superagent/node_modules/watchify/node_modules/vm-browserify/example/run/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/superagent/commit/77fefdaffd4ef3cef2e5b252e165b5f40fae61d5">77fefdaffd4ef3cef2e5b252e165b5f40fae61d5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 3.5.0"}],"vulnerabilityIdentifier":"CVE-2020-11023","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing \u003coption\u003e elements from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11023 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p> <p>Path to dependency file: superagent/node_modules/watchify/node_modules/vm-browserify/example/run/index.html</p> <p>Path to vulnerable library: superagent/node_modules/watchify/node_modules/vm-browserify/example/run/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/superagent/commit/77fefdaffd4ef3cef2e5b252e165b5f40fae61d5">77fefdaffd4ef3cef2e5b252e165b5f40fae61d5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.7.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.7.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 3.5.0"}],"vulnerabilityIdentifier":"CVE-2020-11023","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing \u003coption\u003e elements from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file superagent node modules watchify node modules vm browserify example run index html path to vulnerable library superagent node modules watchify node modules vm browserify example run index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery vulnerabilityurl
0
1,822
4,579,701,103
IssuesEvent
2016-09-18 10:40:18
edovio/OpenStudentBook
https://api.github.com/repos/edovio/OpenStudentBook
closed
Task 2: Identify all the Faculties
Help wanted - Bisogno di Aiuto Processing - In Lavorazione
Identify all the Faculties to insert in this guide Identificare tutte le Facoltà o indirizzi di studio da inserire nella guida
1.0
Task 2: Identify all the Faculties - Identify all the Faculties to insert in this guide Identificare tutte le Facoltà o indirizzi di studio da inserire nella guida
process
task identify all the faculties identify all the faculties to insert in this guide identificare tutte le facoltà o indirizzi di studio da inserire nella guida
1
22,011
30,515,200,812
IssuesEvent
2023-07-19 02:00:10
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Wed, 19 Jul 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Online Self-Supervised Thermal Water Segmentation for Aerial Vehicles - **Authors:** Connor Lee, Jonathan Gustafsson Frennert, Lu Gan, Matthew Anderson, Soon-Jo Chung - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO) - **Arxiv link:** https://arxiv.org/abs/2307.09027 - **Pdf link:** https://arxiv.org/pdf/2307.09027 - **Abstract** We present a new method to adapt an RGB-trained water segmentation network to target-domain aerial thermal imagery using online self-supervision by leveraging texture and motion cues as supervisory signals. This new thermal capability enables current autonomous aerial robots operating in near-shore environments to perform tasks such as visual navigation, bathymetry, and flow tracking at night. Our method overcomes the problem of scarce and difficult-to-obtain near-shore thermal data that prevents the application of conventional supervised and unsupervised methods. In this work, we curate the first aerial thermal near-shore dataset, show that our approach outperforms fully-supervised segmentation models trained on limited target-domain thermal data, and demonstrate real-time capabilities onboard an Nvidia Jetson embedded computing platform. Code and datasets used in this work will be available at: https://github.com/connorlee77/uav-thermal-water-segmentation. ### Plug the Leaks: Advancing Audio-driven Talking Face Generation by Preventing Unintended Information Flow - **Authors:** Dogucan Yaman, Fevziye Irem Eyiokur, Leonard Bärmann, Hazim Kemal Ekenel, Alexander Waibel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09368 - **Pdf link:** https://arxiv.org/pdf/2307.09368 - **Abstract** Audio-driven talking face generation is the task of creating a lip-synchronized, realistic face video from given audio and reference frames. This involves two major challenges: overall visual quality of generated images on the one hand, and audio-visual synchronization of the mouth part on the other hand. In this paper, we start by identifying several problematic aspects of synchronization methods in recent audio-driven talking face generation approaches. Specifically, this involves unintended flow of lip and pose information from the reference to the generated image, as well as instabilities during model training. Subsequently, we propose various techniques for obviating these issues: First, a silent-lip reference image generator prevents leaking of lips from the reference to the generated image. Second, an adaptive triplet loss handles the pose leaking problem. Finally, we propose a stabilized formulation of synchronization loss, circumventing aforementioned training instabilities while additionally further alleviating the lip leaking issue. Combining the individual improvements, we present state-of-the art performance on LRS2 and LRW in both synchronization and visual quality. We further validate our design in various ablation experiments, confirming the individual contributions as well as their complementary effects. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### EVIL: Evidential Inference Learning for Trustworthy Semi-supervised Medical Image Segmentation - **Authors:** Yingyu Chen, Ziyuan Yang, Chenyu Shen, Zhiwen Wang, Yang Qin, Yi Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2307.08988 - **Pdf link:** https://arxiv.org/pdf/2307.08988 - **Abstract** Recently, uncertainty-aware methods have attracted increasing attention in semi-supervised medical image segmentation. However, current methods usually suffer from the drawback that it is difficult to balance the computational cost, estimation accuracy, and theoretical support in a unified framework. To alleviate this problem, we introduce the Dempster-Shafer Theory of Evidence (DST) into semi-supervised medical image segmentation, dubbed Evidential Inference Learning (EVIL). EVIL provides a theoretically guaranteed solution to infer accurate uncertainty quantification in a single forward pass. Trustworthy pseudo labels on unlabeled data are generated after uncertainty estimation. The recently proposed consistency regularization-based training paradigm is adopted in our framework, which enforces the consistency on the perturbed predictions to enhance the generalization with few labeled data. Experimental results show that EVIL achieves competitive performance in comparison with several state-of-the-art methods on the public dataset. ## Keyword: ISP ### RepViT: Revisiting Mobile CNN From ViT Perspective - **Authors:** Ao Wang, Hui Chen, Zijia Lin, Hengjun Pu, Guiguang Ding - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09283 - **Pdf link:** https://arxiv.org/pdf/2307.09283 - **Abstract** Recently, lightweight Vision Transformers (ViTs) demonstrate superior performance and lower latency compared with lightweight Convolutional Neural Networks (CNNs) on resource-constrained mobile devices. This improvement is usually attributed to the multi-head self-attention module, which enables the model to learn global representations. However, the architectural disparities between lightweight ViTs and lightweight CNNs have not been adequately examined. In this study, we revisit the efficient design of lightweight CNNs and emphasize their potential for mobile devices. We incrementally enhance the mobile-friendliness of a standard lightweight CNN, specifically MobileNetV3, by integrating the efficient architectural choices of lightweight ViTs. This ends up with a new family of pure lightweight CNNs, namely RepViT. Extensive experiments show that RepViT outperforms existing state-of-the-art lightweight ViTs and exhibits favorable latency in various vision tasks. On ImageNet, RepViT achieves over 80\% top-1 accuracy with nearly 1ms latency on an iPhone 12, which is the first time for a lightweight model, to the best of our knowledge. Our largest model, RepViT-M3, obtains 81.4\% accuracy with only 1.3ms latency. The code and trained models are available at \url{https://github.com/jameslahm/RepViT}. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### UPSCALE: Unconstrained Channel Pruning - **Authors:** Alvin Wan, Hanxiang Hao, Kaushik Patnaik, Yueyang Xu, Omer Hadad, David Güera, Zhile Ren, Qi Shan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.08771 - **Pdf link:** https://arxiv.org/pdf/2307.08771 - **Abstract** As neural networks grow in size and complexity, inference speeds decline. To combat this, one of the most effective compression techniques -- channel pruning -- removes channels from weights. However, for multi-branch segments of a model, channel removal can introduce inference-time memory copies. In turn, these copies increase inference latency -- so much so that the pruned model can be slower than the unpruned model. As a workaround, pruners conventionally constrain certain channels to be pruned together. This fully eliminates memory copies but, as we show, significantly impairs accuracy. We now have a dilemma: Remove constraints but increase latency, or add constraints and impair accuracy. In response, our insight is to reorder channels at export time, (1) reducing latency by reducing memory copies and (2) improving accuracy by removing constraints. Using this insight, we design a generic algorithm UPSCALE to prune models with any pruning pattern. By removing constraints from existing pruners, we improve ImageNet accuracy for post-training pruned models by 2.1 points on average -- benefiting DenseNet (+16.9), EfficientNetV2 (+7.9), and ResNet (+6.2). Furthermore, by reordering channels, UPSCALE improves inference speeds by up to 2x over a baseline export. ### In Defense of Clip-based Video Relation Detection - **Authors:** Meng Wei, Long Chen, Wei Ji, Xiaoyu Yue, Roger Zimmermann - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.08984 - **Pdf link:** https://arxiv.org/pdf/2307.08984 - **Abstract** Video Visual Relation Detection (VidVRD) aims to detect visual relationship triplets in videos using spatial bounding boxes and temporal boundaries. Existing VidVRD methods can be broadly categorized into bottom-up and top-down paradigms, depending on their approach to classifying relations. Bottom-up methods follow a clip-based approach where they classify relations of short clip tubelet pairs and then merge them into long video relations. On the other hand, top-down methods directly classify long video tubelet pairs. While recent video-based methods utilizing video tubelets have shown promising results, we argue that the effective modeling of spatial and temporal context plays a more significant role than the choice between clip tubelets and video tubelets. This motivates us to revisit the clip-based paradigm and explore the key success factors in VidVRD. In this paper, we propose a Hierarchical Context Model (HCM) that enriches the object-based spatial context and relation-based temporal context based on clips. We demonstrate that using clip tubelets can achieve superior performance compared to most video-based methods. Additionally, using clip tubelets offers more flexibility in model designs and helps alleviate the limitations associated with video tubelets, such as the challenging long-term object tracking problem and the loss of temporal information in long-term tubelet feature compression. Extensive experiments conducted on two challenging VidVRD benchmarks validate that our HCM achieves a new state-of-the-art performance, highlighting the effectiveness of incorporating advanced spatial and temporal context modeling within the clip-based paradigm. ### Knowledge Distillation for Object Detection: from generic to remote sensing datasets - **Authors:** Hoàng-Ân Lê, Minh-Tan Pham - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09264 - **Pdf link:** https://arxiv.org/pdf/2307.09264 - **Abstract** Knowledge distillation, a well-known model compression technique, is an active research area in both computer vision and remote sensing communities. In this paper, we evaluate in a remote sensing context various off-the-shelf object detection knowledge distillation methods which have been originally developed on generic computer vision datasets such as Pascal VOC. In particular, methods covering both logit mimicking and feature imitation approaches are applied for vehicle detection using the well-known benchmarks such as xView and VEDAI datasets. Extensive experiments are performed to compare the relative performance and interrelationships of the methods. Experimental results show high variations and confirm the importance of result aggregation and cross validation on remote sensing datasets. ## Keyword: RAW ### EVIL: Evidential Inference Learning for Trustworthy Semi-supervised Medical Image Segmentation - **Authors:** Yingyu Chen, Ziyuan Yang, Chenyu Shen, Zhiwen Wang, Yang Qin, Yi Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2307.08988 - **Pdf link:** https://arxiv.org/pdf/2307.08988 - **Abstract** Recently, uncertainty-aware methods have attracted increasing attention in semi-supervised medical image segmentation. However, current methods usually suffer from the drawback that it is difficult to balance the computational cost, estimation accuracy, and theoretical support in a unified framework. To alleviate this problem, we introduce the Dempster-Shafer Theory of Evidence (DST) into semi-supervised medical image segmentation, dubbed Evidential Inference Learning (EVIL). EVIL provides a theoretically guaranteed solution to infer accurate uncertainty quantification in a single forward pass. Trustworthy pseudo labels on unlabeled data are generated after uncertainty estimation. The recently proposed consistency regularization-based training paradigm is adopted in our framework, which enforces the consistency on the perturbed predictions to enhance the generalization with few labeled data. Experimental results show that EVIL achieves competitive performance in comparison with several state-of-the-art methods on the public dataset. ### Visual Validation versus Visual Estimation: A Study on the Average Value in Scatterplots - **Authors:** Daniel Braun, Ashley Suh, Remco Chang, Michael Gleicher, Tatiana von Landesberger - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR) - **Arxiv link:** https://arxiv.org/abs/2307.09330 - **Pdf link:** https://arxiv.org/pdf/2307.09330 - **Abstract** We investigate the ability of individuals to visually validate statistical models in terms of their fit to the data. While visual model estimation has been studied extensively, visual model validation remains under-investigated. It is unknown how well people are able to visually validate models, and how their performance compares to visual and computational estimation. As a starting point, we conducted a study across two populations (crowdsourced and volunteers). Participants had to both visually estimate (i.e, draw) and visually validate (i.e., accept or reject) the frequently studied model of averages. Across both populations, the level of accuracy of the models that were considered valid was lower than the accuracy of the estimated models. We find that participants' validation and estimation were unbiased. Moreover, their natural critical point between accepting and rejecting a given mean value is close to the boundary of its 95% confidence interval, indicating that the visually perceived confidence interval corresponds to a common statistical standard. Our work contributes to the understanding of visual model validation and opens new research opportunities. ## Keyword: raw image There is no result
2.0
New submissions for Wed, 19 Jul 23 - ## Keyword: events ### Online Self-Supervised Thermal Water Segmentation for Aerial Vehicles - **Authors:** Connor Lee, Jonathan Gustafsson Frennert, Lu Gan, Matthew Anderson, Soon-Jo Chung - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Robotics (cs.RO) - **Arxiv link:** https://arxiv.org/abs/2307.09027 - **Pdf link:** https://arxiv.org/pdf/2307.09027 - **Abstract** We present a new method to adapt an RGB-trained water segmentation network to target-domain aerial thermal imagery using online self-supervision by leveraging texture and motion cues as supervisory signals. This new thermal capability enables current autonomous aerial robots operating in near-shore environments to perform tasks such as visual navigation, bathymetry, and flow tracking at night. Our method overcomes the problem of scarce and difficult-to-obtain near-shore thermal data that prevents the application of conventional supervised and unsupervised methods. In this work, we curate the first aerial thermal near-shore dataset, show that our approach outperforms fully-supervised segmentation models trained on limited target-domain thermal data, and demonstrate real-time capabilities onboard an Nvidia Jetson embedded computing platform. Code and datasets used in this work will be available at: https://github.com/connorlee77/uav-thermal-water-segmentation. ### Plug the Leaks: Advancing Audio-driven Talking Face Generation by Preventing Unintended Information Flow - **Authors:** Dogucan Yaman, Fevziye Irem Eyiokur, Leonard Bärmann, Hazim Kemal Ekenel, Alexander Waibel - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09368 - **Pdf link:** https://arxiv.org/pdf/2307.09368 - **Abstract** Audio-driven talking face generation is the task of creating a lip-synchronized, realistic face video from given audio and reference frames. This involves two major challenges: overall visual quality of generated images on the one hand, and audio-visual synchronization of the mouth part on the other hand. In this paper, we start by identifying several problematic aspects of synchronization methods in recent audio-driven talking face generation approaches. Specifically, this involves unintended flow of lip and pose information from the reference to the generated image, as well as instabilities during model training. Subsequently, we propose various techniques for obviating these issues: First, a silent-lip reference image generator prevents leaking of lips from the reference to the generated image. Second, an adaptive triplet loss handles the pose leaking problem. Finally, we propose a stabilized formulation of synchronization loss, circumventing aforementioned training instabilities while additionally further alleviating the lip leaking issue. Combining the individual improvements, we present state-of-the art performance on LRS2 and LRW in both synchronization and visual quality. We further validate our design in various ablation experiments, confirming the individual contributions as well as their complementary effects. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### EVIL: Evidential Inference Learning for Trustworthy Semi-supervised Medical Image Segmentation - **Authors:** Yingyu Chen, Ziyuan Yang, Chenyu Shen, Zhiwen Wang, Yang Qin, Yi Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2307.08988 - **Pdf link:** https://arxiv.org/pdf/2307.08988 - **Abstract** Recently, uncertainty-aware methods have attracted increasing attention in semi-supervised medical image segmentation. However, current methods usually suffer from the drawback that it is difficult to balance the computational cost, estimation accuracy, and theoretical support in a unified framework. To alleviate this problem, we introduce the Dempster-Shafer Theory of Evidence (DST) into semi-supervised medical image segmentation, dubbed Evidential Inference Learning (EVIL). EVIL provides a theoretically guaranteed solution to infer accurate uncertainty quantification in a single forward pass. Trustworthy pseudo labels on unlabeled data are generated after uncertainty estimation. The recently proposed consistency regularization-based training paradigm is adopted in our framework, which enforces the consistency on the perturbed predictions to enhance the generalization with few labeled data. Experimental results show that EVIL achieves competitive performance in comparison with several state-of-the-art methods on the public dataset. ## Keyword: ISP ### RepViT: Revisiting Mobile CNN From ViT Perspective - **Authors:** Ao Wang, Hui Chen, Zijia Lin, Hengjun Pu, Guiguang Ding - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09283 - **Pdf link:** https://arxiv.org/pdf/2307.09283 - **Abstract** Recently, lightweight Vision Transformers (ViTs) demonstrate superior performance and lower latency compared with lightweight Convolutional Neural Networks (CNNs) on resource-constrained mobile devices. This improvement is usually attributed to the multi-head self-attention module, which enables the model to learn global representations. However, the architectural disparities between lightweight ViTs and lightweight CNNs have not been adequately examined. In this study, we revisit the efficient design of lightweight CNNs and emphasize their potential for mobile devices. We incrementally enhance the mobile-friendliness of a standard lightweight CNN, specifically MobileNetV3, by integrating the efficient architectural choices of lightweight ViTs. This ends up with a new family of pure lightweight CNNs, namely RepViT. Extensive experiments show that RepViT outperforms existing state-of-the-art lightweight ViTs and exhibits favorable latency in various vision tasks. On ImageNet, RepViT achieves over 80\% top-1 accuracy with nearly 1ms latency on an iPhone 12, which is the first time for a lightweight model, to the best of our knowledge. Our largest model, RepViT-M3, obtains 81.4\% accuracy with only 1.3ms latency. The code and trained models are available at \url{https://github.com/jameslahm/RepViT}. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### UPSCALE: Unconstrained Channel Pruning - **Authors:** Alvin Wan, Hanxiang Hao, Kaushik Patnaik, Yueyang Xu, Omer Hadad, David Güera, Zhile Ren, Qi Shan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.08771 - **Pdf link:** https://arxiv.org/pdf/2307.08771 - **Abstract** As neural networks grow in size and complexity, inference speeds decline. To combat this, one of the most effective compression techniques -- channel pruning -- removes channels from weights. However, for multi-branch segments of a model, channel removal can introduce inference-time memory copies. In turn, these copies increase inference latency -- so much so that the pruned model can be slower than the unpruned model. As a workaround, pruners conventionally constrain certain channels to be pruned together. This fully eliminates memory copies but, as we show, significantly impairs accuracy. We now have a dilemma: Remove constraints but increase latency, or add constraints and impair accuracy. In response, our insight is to reorder channels at export time, (1) reducing latency by reducing memory copies and (2) improving accuracy by removing constraints. Using this insight, we design a generic algorithm UPSCALE to prune models with any pruning pattern. By removing constraints from existing pruners, we improve ImageNet accuracy for post-training pruned models by 2.1 points on average -- benefiting DenseNet (+16.9), EfficientNetV2 (+7.9), and ResNet (+6.2). Furthermore, by reordering channels, UPSCALE improves inference speeds by up to 2x over a baseline export. ### In Defense of Clip-based Video Relation Detection - **Authors:** Meng Wei, Long Chen, Wei Ji, Xiaoyu Yue, Roger Zimmermann - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.08984 - **Pdf link:** https://arxiv.org/pdf/2307.08984 - **Abstract** Video Visual Relation Detection (VidVRD) aims to detect visual relationship triplets in videos using spatial bounding boxes and temporal boundaries. Existing VidVRD methods can be broadly categorized into bottom-up and top-down paradigms, depending on their approach to classifying relations. Bottom-up methods follow a clip-based approach where they classify relations of short clip tubelet pairs and then merge them into long video relations. On the other hand, top-down methods directly classify long video tubelet pairs. While recent video-based methods utilizing video tubelets have shown promising results, we argue that the effective modeling of spatial and temporal context plays a more significant role than the choice between clip tubelets and video tubelets. This motivates us to revisit the clip-based paradigm and explore the key success factors in VidVRD. In this paper, we propose a Hierarchical Context Model (HCM) that enriches the object-based spatial context and relation-based temporal context based on clips. We demonstrate that using clip tubelets can achieve superior performance compared to most video-based methods. Additionally, using clip tubelets offers more flexibility in model designs and helps alleviate the limitations associated with video tubelets, such as the challenging long-term object tracking problem and the loss of temporal information in long-term tubelet feature compression. Extensive experiments conducted on two challenging VidVRD benchmarks validate that our HCM achieves a new state-of-the-art performance, highlighting the effectiveness of incorporating advanced spatial and temporal context modeling within the clip-based paradigm. ### Knowledge Distillation for Object Detection: from generic to remote sensing datasets - **Authors:** Hoàng-Ân Lê, Minh-Tan Pham - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2307.09264 - **Pdf link:** https://arxiv.org/pdf/2307.09264 - **Abstract** Knowledge distillation, a well-known model compression technique, is an active research area in both computer vision and remote sensing communities. In this paper, we evaluate in a remote sensing context various off-the-shelf object detection knowledge distillation methods which have been originally developed on generic computer vision datasets such as Pascal VOC. In particular, methods covering both logit mimicking and feature imitation approaches are applied for vehicle detection using the well-known benchmarks such as xView and VEDAI datasets. Extensive experiments are performed to compare the relative performance and interrelationships of the methods. Experimental results show high variations and confirm the importance of result aggregation and cross validation on remote sensing datasets. ## Keyword: RAW ### EVIL: Evidential Inference Learning for Trustworthy Semi-supervised Medical Image Segmentation - **Authors:** Yingyu Chen, Ziyuan Yang, Chenyu Shen, Zhiwen Wang, Yang Qin, Yi Zhang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2307.08988 - **Pdf link:** https://arxiv.org/pdf/2307.08988 - **Abstract** Recently, uncertainty-aware methods have attracted increasing attention in semi-supervised medical image segmentation. However, current methods usually suffer from the drawback that it is difficult to balance the computational cost, estimation accuracy, and theoretical support in a unified framework. To alleviate this problem, we introduce the Dempster-Shafer Theory of Evidence (DST) into semi-supervised medical image segmentation, dubbed Evidential Inference Learning (EVIL). EVIL provides a theoretically guaranteed solution to infer accurate uncertainty quantification in a single forward pass. Trustworthy pseudo labels on unlabeled data are generated after uncertainty estimation. The recently proposed consistency regularization-based training paradigm is adopted in our framework, which enforces the consistency on the perturbed predictions to enhance the generalization with few labeled data. Experimental results show that EVIL achieves competitive performance in comparison with several state-of-the-art methods on the public dataset. ### Visual Validation versus Visual Estimation: A Study on the Average Value in Scatterplots - **Authors:** Daniel Braun, Ashley Suh, Remco Chang, Michael Gleicher, Tatiana von Landesberger - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Graphics (cs.GR) - **Arxiv link:** https://arxiv.org/abs/2307.09330 - **Pdf link:** https://arxiv.org/pdf/2307.09330 - **Abstract** We investigate the ability of individuals to visually validate statistical models in terms of their fit to the data. While visual model estimation has been studied extensively, visual model validation remains under-investigated. It is unknown how well people are able to visually validate models, and how their performance compares to visual and computational estimation. As a starting point, we conducted a study across two populations (crowdsourced and volunteers). Participants had to both visually estimate (i.e, draw) and visually validate (i.e., accept or reject) the frequently studied model of averages. Across both populations, the level of accuracy of the models that were considered valid was lower than the accuracy of the estimated models. We find that participants' validation and estimation were unbiased. Moreover, their natural critical point between accepting and rejecting a given mean value is close to the boundary of its 95% confidence interval, indicating that the visually perceived confidence interval corresponds to a common statistical standard. Our work contributes to the understanding of visual model validation and opens new research opportunities. ## Keyword: raw image There is no result
process
new submissions for wed jul keyword events online self supervised thermal water segmentation for aerial vehicles authors connor lee jonathan gustafsson frennert lu gan matthew anderson soon jo chung subjects computer vision and pattern recognition cs cv robotics cs ro arxiv link pdf link abstract we present a new method to adapt an rgb trained water segmentation network to target domain aerial thermal imagery using online self supervision by leveraging texture and motion cues as supervisory signals this new thermal capability enables current autonomous aerial robots operating in near shore environments to perform tasks such as visual navigation bathymetry and flow tracking at night our method overcomes the problem of scarce and difficult to obtain near shore thermal data that prevents the application of conventional supervised and unsupervised methods in this work we curate the first aerial thermal near shore dataset show that our approach outperforms fully supervised segmentation models trained on limited target domain thermal data and demonstrate real time capabilities onboard an nvidia jetson embedded computing platform code and datasets used in this work will be available at plug the leaks advancing audio driven talking face generation by preventing unintended information flow authors dogucan yaman fevziye irem eyiokur leonard bärmann hazim kemal ekenel alexander waibel subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract audio driven talking face generation is the task of creating a lip synchronized realistic face video from given audio and reference frames this involves two major challenges overall visual quality of generated images on the one hand and audio visual synchronization of the mouth part on the other hand in this paper we start by identifying several problematic aspects of synchronization methods in recent audio driven talking face generation approaches specifically this involves unintended flow of lip and pose information from the reference to the generated image as well as instabilities during model training subsequently we propose various techniques for obviating these issues first a silent lip reference image generator prevents leaking of lips from the reference to the generated image second an adaptive triplet loss handles the pose leaking problem finally we propose a stabilized formulation of synchronization loss circumventing aforementioned training instabilities while additionally further alleviating the lip leaking issue combining the individual improvements we present state of the art performance on and lrw in both synchronization and visual quality we further validate our design in various ablation experiments confirming the individual contributions as well as their complementary effects keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb evil evidential inference learning for trustworthy semi supervised medical image segmentation authors yingyu chen ziyuan yang chenyu shen zhiwen wang yang qin yi zhang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract recently uncertainty aware methods have attracted increasing attention in semi supervised medical image segmentation however current methods usually suffer from the drawback that it is difficult to balance the computational cost estimation accuracy and theoretical support in a unified framework to alleviate this problem we introduce the dempster shafer theory of evidence dst into semi supervised medical image segmentation dubbed evidential inference learning evil evil provides a theoretically guaranteed solution to infer accurate uncertainty quantification in a single forward pass trustworthy pseudo labels on unlabeled data are generated after uncertainty estimation the recently proposed consistency regularization based training paradigm is adopted in our framework which enforces the consistency on the perturbed predictions to enhance the generalization with few labeled data experimental results show that evil achieves competitive performance in comparison with several state of the art methods on the public dataset keyword isp repvit revisiting mobile cnn from vit perspective authors ao wang hui chen zijia lin hengjun pu guiguang ding subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract recently lightweight vision transformers vits demonstrate superior performance and lower latency compared with lightweight convolutional neural networks cnns on resource constrained mobile devices this improvement is usually attributed to the multi head self attention module which enables the model to learn global representations however the architectural disparities between lightweight vits and lightweight cnns have not been adequately examined in this study we revisit the efficient design of lightweight cnns and emphasize their potential for mobile devices we incrementally enhance the mobile friendliness of a standard lightweight cnn specifically by integrating the efficient architectural choices of lightweight vits this ends up with a new family of pure lightweight cnns namely repvit extensive experiments show that repvit outperforms existing state of the art lightweight vits and exhibits favorable latency in various vision tasks on imagenet repvit achieves over top accuracy with nearly latency on an iphone which is the first time for a lightweight model to the best of our knowledge our largest model repvit obtains accuracy with only latency the code and trained models are available at url keyword image signal processing there is no result keyword image signal process there is no result keyword compression upscale unconstrained channel pruning authors alvin wan hanxiang hao kaushik patnaik yueyang xu omer hadad david güera zhile ren qi shan subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract as neural networks grow in size and complexity inference speeds decline to combat this one of the most effective compression techniques channel pruning removes channels from weights however for multi branch segments of a model channel removal can introduce inference time memory copies in turn these copies increase inference latency so much so that the pruned model can be slower than the unpruned model as a workaround pruners conventionally constrain certain channels to be pruned together this fully eliminates memory copies but as we show significantly impairs accuracy we now have a dilemma remove constraints but increase latency or add constraints and impair accuracy in response our insight is to reorder channels at export time reducing latency by reducing memory copies and improving accuracy by removing constraints using this insight we design a generic algorithm upscale to prune models with any pruning pattern by removing constraints from existing pruners we improve imagenet accuracy for post training pruned models by points on average benefiting densenet and resnet furthermore by reordering channels upscale improves inference speeds by up to over a baseline export in defense of clip based video relation detection authors meng wei long chen wei ji xiaoyu yue roger zimmermann subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract video visual relation detection vidvrd aims to detect visual relationship triplets in videos using spatial bounding boxes and temporal boundaries existing vidvrd methods can be broadly categorized into bottom up and top down paradigms depending on their approach to classifying relations bottom up methods follow a clip based approach where they classify relations of short clip tubelet pairs and then merge them into long video relations on the other hand top down methods directly classify long video tubelet pairs while recent video based methods utilizing video tubelets have shown promising results we argue that the effective modeling of spatial and temporal context plays a more significant role than the choice between clip tubelets and video tubelets this motivates us to revisit the clip based paradigm and explore the key success factors in vidvrd in this paper we propose a hierarchical context model hcm that enriches the object based spatial context and relation based temporal context based on clips we demonstrate that using clip tubelets can achieve superior performance compared to most video based methods additionally using clip tubelets offers more flexibility in model designs and helps alleviate the limitations associated with video tubelets such as the challenging long term object tracking problem and the loss of temporal information in long term tubelet feature compression extensive experiments conducted on two challenging vidvrd benchmarks validate that our hcm achieves a new state of the art performance highlighting the effectiveness of incorporating advanced spatial and temporal context modeling within the clip based paradigm knowledge distillation for object detection from generic to remote sensing datasets authors hoàng ân lê minh tan pham subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract knowledge distillation a well known model compression technique is an active research area in both computer vision and remote sensing communities in this paper we evaluate in a remote sensing context various off the shelf object detection knowledge distillation methods which have been originally developed on generic computer vision datasets such as pascal voc in particular methods covering both logit mimicking and feature imitation approaches are applied for vehicle detection using the well known benchmarks such as xview and vedai datasets extensive experiments are performed to compare the relative performance and interrelationships of the methods experimental results show high variations and confirm the importance of result aggregation and cross validation on remote sensing datasets keyword raw evil evidential inference learning for trustworthy semi supervised medical image segmentation authors yingyu chen ziyuan yang chenyu shen zhiwen wang yang qin yi zhang subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract recently uncertainty aware methods have attracted increasing attention in semi supervised medical image segmentation however current methods usually suffer from the drawback that it is difficult to balance the computational cost estimation accuracy and theoretical support in a unified framework to alleviate this problem we introduce the dempster shafer theory of evidence dst into semi supervised medical image segmentation dubbed evidential inference learning evil evil provides a theoretically guaranteed solution to infer accurate uncertainty quantification in a single forward pass trustworthy pseudo labels on unlabeled data are generated after uncertainty estimation the recently proposed consistency regularization based training paradigm is adopted in our framework which enforces the consistency on the perturbed predictions to enhance the generalization with few labeled data experimental results show that evil achieves competitive performance in comparison with several state of the art methods on the public dataset visual validation versus visual estimation a study on the average value in scatterplots authors daniel braun ashley suh remco chang michael gleicher tatiana von landesberger subjects computer vision and pattern recognition cs cv graphics cs gr arxiv link pdf link abstract we investigate the ability of individuals to visually validate statistical models in terms of their fit to the data while visual model estimation has been studied extensively visual model validation remains under investigated it is unknown how well people are able to visually validate models and how their performance compares to visual and computational estimation as a starting point we conducted a study across two populations crowdsourced and volunteers participants had to both visually estimate i e draw and visually validate i e accept or reject the frequently studied model of averages across both populations the level of accuracy of the models that were considered valid was lower than the accuracy of the estimated models we find that participants validation and estimation were unbiased moreover their natural critical point between accepting and rejecting a given mean value is close to the boundary of its confidence interval indicating that the visually perceived confidence interval corresponds to a common statistical standard our work contributes to the understanding of visual model validation and opens new research opportunities keyword raw image there is no result
1
70,973
23,392,770,631
IssuesEvent
2022-08-11 19:35:03
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Verification modal shows red shield while info panel shows grey shield
T-Defect
### Steps to reproduce Alice is using some device that does not support cross-signing. Alice has 3 sessions. Bob verifies Alice's first session. In the info panel, Alices 3 devices shields are now shown to Bob: Green, Grey, Grey. Bob clicks on one of Alice's grey sessions. ### Outcome #### What did you expect? consistently grey shield #### What happened instead? ![image](https://user-images.githubusercontent.com/2803622/184224342-8cf93e49-98ae-4f0f-96c6-a78908ea8b5d.png) ### Operating system _No response_ ### Application version Element Nightly version: 2022081002 Olm version: 3.2.12 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
Verification modal shows red shield while info panel shows grey shield - ### Steps to reproduce Alice is using some device that does not support cross-signing. Alice has 3 sessions. Bob verifies Alice's first session. In the info panel, Alices 3 devices shields are now shown to Bob: Green, Grey, Grey. Bob clicks on one of Alice's grey sessions. ### Outcome #### What did you expect? consistently grey shield #### What happened instead? ![image](https://user-images.githubusercontent.com/2803622/184224342-8cf93e49-98ae-4f0f-96c6-a78908ea8b5d.png) ### Operating system _No response_ ### Application version Element Nightly version: 2022081002 Olm version: 3.2.12 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? No
non_process
verification modal shows red shield while info panel shows grey shield steps to reproduce alice is using some device that does not support cross signing alice has sessions bob verifies alice s first session in the info panel alices devices shields are now shown to bob green grey grey bob clicks on one of alice s grey sessions outcome what did you expect consistently grey shield what happened instead operating system no response application version element nightly version olm version how did you install the app no response homeserver no response will you send logs no
0
127,353
5,028,957,190
IssuesEvent
2016-12-15 19:44:40
USGCRP/gcis
https://api.github.com/repos/USGCRP/gcis
closed
Align the contributors in display table based on type of publication
an enhancement context Front End priority medium type technical
Ensure adding contributor (organization, role-publisher) to a journal and adding relationship (pub-article, role-publisher) to the organization should display in the same table and same group in the display. Currently the contents that are added in the same sequence only display under the same pub type. If we add in a different sequence, say, a journal, a book, a journal, all the journals won't align in the same group, instead splits into many as shown below. The screenshot below shows an example of how 'journal' pub type is split by the order of when it was added. Basically we should align these types in a group. ![screen shot 2016-11-14 at 4 22 42 pm](https://cloud.githubusercontent.com/assets/18075324/20283351/9f70c656-aa86-11e6-82aa-ee8d4b532325.png)
1.0
Align the contributors in display table based on type of publication - Ensure adding contributor (organization, role-publisher) to a journal and adding relationship (pub-article, role-publisher) to the organization should display in the same table and same group in the display. Currently the contents that are added in the same sequence only display under the same pub type. If we add in a different sequence, say, a journal, a book, a journal, all the journals won't align in the same group, instead splits into many as shown below. The screenshot below shows an example of how 'journal' pub type is split by the order of when it was added. Basically we should align these types in a group. ![screen shot 2016-11-14 at 4 22 42 pm](https://cloud.githubusercontent.com/assets/18075324/20283351/9f70c656-aa86-11e6-82aa-ee8d4b532325.png)
non_process
align the contributors in display table based on type of publication ensure adding contributor organization role publisher to a journal and adding relationship pub article role publisher to the organization should display in the same table and same group in the display currently the contents that are added in the same sequence only display under the same pub type if we add in a different sequence say a journal a book a journal all the journals won t align in the same group instead splits into many as shown below the screenshot below shows an example of how journal pub type is split by the order of when it was added basically we should align these types in a group
0
1,101
3,037,838,191
IssuesEvent
2015-08-06 19:10:26
mailpile/Mailpile
https://api.github.com/repos/mailpile/Mailpile
closed
Quoted Text in PGP Email Replies Should Be Off by Default (opt-in)
Back End Front End Privacy / Security
"Quoted Text" in email replies actually presents a security hazard, and should be off by default (opt-in) during PGP mode. This problem has not yet been mitigated in major PGP-supporting email clients, and it would be good to see mailpile rectify this. If Bob's private key is revealed, a passive attacker can use quoted text to read Bob's mails to Alice as well. This is unnecessary. Automated threading (gmail-like conversation view) should be used for keeping track of conversation flow instead of quoted text. This practice also reduces redundancy. For 1. the threat model and security hazard of quoted text @smari 2. the mitigation: using conversation view instead for UX @brennannovak 3. how to implement this using existing or novel headers @BjarniRunar a whitepaper is available at https://github.com/uktu/blog/blob/master/security-quotedtext
True
Quoted Text in PGP Email Replies Should Be Off by Default (opt-in) - "Quoted Text" in email replies actually presents a security hazard, and should be off by default (opt-in) during PGP mode. This problem has not yet been mitigated in major PGP-supporting email clients, and it would be good to see mailpile rectify this. If Bob's private key is revealed, a passive attacker can use quoted text to read Bob's mails to Alice as well. This is unnecessary. Automated threading (gmail-like conversation view) should be used for keeping track of conversation flow instead of quoted text. This practice also reduces redundancy. For 1. the threat model and security hazard of quoted text @smari 2. the mitigation: using conversation view instead for UX @brennannovak 3. how to implement this using existing or novel headers @BjarniRunar a whitepaper is available at https://github.com/uktu/blog/blob/master/security-quotedtext
non_process
quoted text in pgp email replies should be off by default opt in quoted text in email replies actually presents a security hazard and should be off by default opt in during pgp mode this problem has not yet been mitigated in major pgp supporting email clients and it would be good to see mailpile rectify this if bob s private key is revealed a passive attacker can use quoted text to read bob s mails to alice as well this is unnecessary automated threading gmail like conversation view should be used for keeping track of conversation flow instead of quoted text this practice also reduces redundancy for the threat model and security hazard of quoted text smari the mitigation using conversation view instead for ux brennannovak how to implement this using existing or novel headers bjarnirunar a whitepaper is available at
0
201,526
15,802,768,751
IssuesEvent
2021-04-03 11:20:17
bounswe/2021SpringGroup3
https://api.github.com/repos/bounswe/2021SpringGroup3
opened
Identify and document system requirements
Priority: High Status: Available Type: Documentation Type: Organization
Identify and document system requirements under the functional requirements section.
1.0
Identify and document system requirements - Identify and document system requirements under the functional requirements section.
non_process
identify and document system requirements identify and document system requirements under the functional requirements section
0
108,425
16,777,815,594
IssuesEvent
2021-06-15 01:03:49
rsoreq/zaproxy
https://api.github.com/repos/rsoreq/zaproxy
opened
CVE-2021-23369 (High) detected in handlebars-4.1.2.jar
security vulnerability
## CVE-2021-23369 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.2.jar</b></p></summary> <p>Logic-less and semantic templates with Java</p> <p>Library home page: <a href="https://github.com/jknack/handlebars.java">https://github.com/jknack/handlebars.java</a></p> <p>Path to dependency file: zaproxy</p> <p>Path to vulnerable library: /tmp/ws-ua_20210505055419_RKZHCN/downloadResource_XVLILT/20210505055836/handlebars-4.1.2.jar</p> <p> Dependency Hierarchy: - wiremock-jre8-2.25.1.jar (Root Library) - :x: **handlebars-4.1.2.jar** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source. <p>Publish Date: 2021-04-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369>CVE-2021-23369</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369</a></p> <p>Release Date: 2021-04-12</p> <p>Fix Resolution: handlebars - 4.7.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.github.jknack","packageName":"handlebars","packageVersion":"4.1.2","packageFilePaths":["zaproxy"],"isTransitiveDependency":true,"dependencyTree":"com.github.tomakehurst:wiremock-jre8:2.25.1;com.github.jknack:handlebars:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-23369","vulnerabilityDetails":"The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-23369 (High) detected in handlebars-4.1.2.jar - ## CVE-2021-23369 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.2.jar</b></p></summary> <p>Logic-less and semantic templates with Java</p> <p>Library home page: <a href="https://github.com/jknack/handlebars.java">https://github.com/jknack/handlebars.java</a></p> <p>Path to dependency file: zaproxy</p> <p>Path to vulnerable library: /tmp/ws-ua_20210505055419_RKZHCN/downloadResource_XVLILT/20210505055836/handlebars-4.1.2.jar</p> <p> Dependency Hierarchy: - wiremock-jre8-2.25.1.jar (Root Library) - :x: **handlebars-4.1.2.jar** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source. <p>Publish Date: 2021-04-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369>CVE-2021-23369</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369</a></p> <p>Release Date: 2021-04-12</p> <p>Fix Resolution: handlebars - 4.7.7</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.github.jknack","packageName":"handlebars","packageVersion":"4.1.2","packageFilePaths":["zaproxy"],"isTransitiveDependency":true,"dependencyTree":"com.github.tomakehurst:wiremock-jre8:2.25.1;com.github.jknack:handlebars:4.1.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.7.7"}],"baseBranches":["develop"],"vulnerabilityIdentifier":"CVE-2021-23369","vulnerabilityDetails":"The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in handlebars jar cve high severity vulnerability vulnerable library handlebars jar logic less and semantic templates with java library home page a href path to dependency file zaproxy path to vulnerable library tmp ws ua rkzhcn downloadresource xvlilt handlebars jar dependency hierarchy wiremock jar root library x handlebars jar vulnerable library found in base branch develop vulnerability details the package handlebars before are vulnerable to remote code execution rce when selecting certain compiling options to compile templates coming from an untrusted source publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com github tomakehurst wiremock com github jknack handlebars isminimumfixversionavailable true minimumfixversion handlebars basebranches vulnerabilityidentifier cve vulnerabilitydetails the package handlebars before are vulnerable to remote code execution rce when selecting certain compiling options to compile templates coming from an untrusted source vulnerabilityurl
0
178,237
21,509,348,058
IssuesEvent
2022-04-28 01:31:15
exadel-inc/ui-playground
https://api.github.com/repos/exadel-inc/ui-playground
opened
CVE-2022-29078 (High) detected in ejs-3.1.6.tgz
security vulnerability
## CVE-2022-29078 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ejs-3.1.6.tgz</b></p></summary> <p>Embedded JavaScript templates</p> <p>Library home page: <a href="https://registry.npmjs.org/ejs/-/ejs-3.1.6.tgz">https://registry.npmjs.org/ejs/-/ejs-3.1.6.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ejs/package.json</p> <p> Dependency Hierarchy: - eleventy-1.0.1.tgz (Root Library) - :x: **ejs-3.1.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/exadel-inc/ui-playground/commit/5a2292c232a69bab96f26542145fb5218c2edf44">5a2292c232a69bab96f26542145fb5218c2edf44</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ejs (aka Embedded JavaScript templates) package 3.1.6 for Node.js allows server-side template injection in settings[view options][outputFunctionName]. This is parsed as an internal option, and overwrites the outputFunctionName option with an arbitrary OS command (which is executed upon template compilation). <p>Publish Date: 2022-04-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29078>CVE-2022-29078</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29078~">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29078~</a></p> <p>Release Date: 2022-04-25</p> <p>Fix Resolution: ejs - v3.1.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-29078 (High) detected in ejs-3.1.6.tgz - ## CVE-2022-29078 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ejs-3.1.6.tgz</b></p></summary> <p>Embedded JavaScript templates</p> <p>Library home page: <a href="https://registry.npmjs.org/ejs/-/ejs-3.1.6.tgz">https://registry.npmjs.org/ejs/-/ejs-3.1.6.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ejs/package.json</p> <p> Dependency Hierarchy: - eleventy-1.0.1.tgz (Root Library) - :x: **ejs-3.1.6.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/exadel-inc/ui-playground/commit/5a2292c232a69bab96f26542145fb5218c2edf44">5a2292c232a69bab96f26542145fb5218c2edf44</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The ejs (aka Embedded JavaScript templates) package 3.1.6 for Node.js allows server-side template injection in settings[view options][outputFunctionName]. This is parsed as an internal option, and overwrites the outputFunctionName option with an arbitrary OS command (which is executed upon template compilation). <p>Publish Date: 2022-04-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-29078>CVE-2022-29078</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29078~">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-29078~</a></p> <p>Release Date: 2022-04-25</p> <p>Fix Resolution: ejs - v3.1.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in ejs tgz cve high severity vulnerability vulnerable library ejs tgz embedded javascript templates library home page a href path to dependency file package json path to vulnerable library node modules ejs package json dependency hierarchy eleventy tgz root library x ejs tgz vulnerable library found in head commit a href found in base branch main vulnerability details the ejs aka embedded javascript templates package for node js allows server side template injection in settings this is parsed as an internal option and overwrites the outputfunctionname option with an arbitrary os command which is executed upon template compilation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ejs step up your open source security game with whitesource
0
334,613
24,427,838,533
IssuesEvent
2022-10-06 05:25:40
Anon-Planet/thgtoa
https://api.github.com/repos/Anon-Planet/thgtoa
closed
New appendix: Spreading disinformation through Social Media: a how-to for protecting pseudonymous identities
documentation enhancement medium priority
## Misinformation and disinformation in social media to protect identities **We could have a section or appendix for user identities and spreading misinformation to protect them.** For example, when creating anonymous identities, it's much more than just making up a name and pretending to be a certain age. Having accounts for that person is good. But how do you remember to use these accounts and make it look believable? It can include making fake user accounts on social media that resemble your own. It can also include scheduling posts on Twitter, Reddit, etc. **I propose a direct guide section be made for putting out mis-/disinformation regarding identities online.** How to generate believable names, nicknames, and how to maintain social media presence in a way that's straightforward without being a burden on the person trying to do so. This might include: making reminders to post on your fake accounts to keep up appearances; how and when you should put out misinformation and when it's a good idea to misdirect. **Having a fake name isn't enough.** You have to keep up with the social media presence, including regularly posting but also maintaining the perspective of your fake identity, which can be much more difficult than just telling lies on the internet. We live in a society full of misinformation. Most of it, sadly, is intentional on behalf of governments to keep the peace or push a narrative. This is something that main-stream media has done since the dawn of World War I - it was the first time that highly organized state propaganda institutions were developed. You should be doing the same in the realm of keeping up anonymous identities. It's about knowing not only how to misdirect, like a magician, but also when and where - as well as making it a natural behavior. In other words, you won't be so obvious and it will be more believable and more effective as a result. **Give any additional relevant context.** _"...the most insidious and underrated problem in our information ecosystem is that we do not give the right kind of attention to the right things at the right time"_ - [Tim Harford, What magic teaches us about misinformation](https://timharford.com/2021/06/what-magic-teaches-us-about-misinformation/) _"We are living in a magical world where with the stroke of a wand the social media universe shifts either for the good or for the worst, and though misinformation on social media has been a problem for 15+ years it’s more prevalent now than ever."_ - [Soteria Intelligence, Social Media Magicians: The Age of Misinformation](https://www.soteriaintelligence.com/blog/social-media-magicians-the-age-of-misinformation/) _"Understanding how disinformation online has endangered democracy and democratic norms is incredibly nuanced."_ - [How disinformation can hinder democracy](https://news.stanford.edu/2022/04/13/know-disinformation-address/) _"...when a lie is repeated and seen multiple times, people unintentionally begin to accept it as truth..."_ - [UCSUSA, We all have the power to counter and interrupt disinformation](https://www.ucsusa.org/resources/how-stop-disinformation) _"Misdirection—manipulating the spectator away from the cause of a magic effect—is widely considered a central element of the practice of magic: “[m]isdirection is a principle element in the art of deception"_ (Randal, 1976, p. 380)," - [A psychologically-based taxonomy of misdirection](https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4260479/) The primary resource I use to maintain all of the above - an excellent book by C. Silverman, [Verification Handbook - For Disinformation And Media Manipulation](https://datajournalism.com/read/handbook/verification-3) - which is, for all intents and purposes, meant to *detect* media manipulation and misinformation, but we're going to use it to intentionally *misdirect*.
1.0
New appendix: Spreading disinformation through Social Media: a how-to for protecting pseudonymous identities - ## Misinformation and disinformation in social media to protect identities **We could have a section or appendix for user identities and spreading misinformation to protect them.** For example, when creating anonymous identities, it's much more than just making up a name and pretending to be a certain age. Having accounts for that person is good. But how do you remember to use these accounts and make it look believable? It can include making fake user accounts on social media that resemble your own. It can also include scheduling posts on Twitter, Reddit, etc. **I propose a direct guide section be made for putting out mis-/disinformation regarding identities online.** How to generate believable names, nicknames, and how to maintain social media presence in a way that's straightforward without being a burden on the person trying to do so. This might include: making reminders to post on your fake accounts to keep up appearances; how and when you should put out misinformation and when it's a good idea to misdirect. **Having a fake name isn't enough.** You have to keep up with the social media presence, including regularly posting but also maintaining the perspective of your fake identity, which can be much more difficult than just telling lies on the internet. We live in a society full of misinformation. Most of it, sadly, is intentional on behalf of governments to keep the peace or push a narrative. This is something that main-stream media has done since the dawn of World War I - it was the first time that highly organized state propaganda institutions were developed. You should be doing the same in the realm of keeping up anonymous identities. It's about knowing not only how to misdirect, like a magician, but also when and where - as well as making it a natural behavior. In other words, you won't be so obvious and it will be more believable and more effective as a result. **Give any additional relevant context.** _"...the most insidious and underrated problem in our information ecosystem is that we do not give the right kind of attention to the right things at the right time"_ - [Tim Harford, What magic teaches us about misinformation](https://timharford.com/2021/06/what-magic-teaches-us-about-misinformation/) _"We are living in a magical world where with the stroke of a wand the social media universe shifts either for the good or for the worst, and though misinformation on social media has been a problem for 15+ years it’s more prevalent now than ever."_ - [Soteria Intelligence, Social Media Magicians: The Age of Misinformation](https://www.soteriaintelligence.com/blog/social-media-magicians-the-age-of-misinformation/) _"Understanding how disinformation online has endangered democracy and democratic norms is incredibly nuanced."_ - [How disinformation can hinder democracy](https://news.stanford.edu/2022/04/13/know-disinformation-address/) _"...when a lie is repeated and seen multiple times, people unintentionally begin to accept it as truth..."_ - [UCSUSA, We all have the power to counter and interrupt disinformation](https://www.ucsusa.org/resources/how-stop-disinformation) _"Misdirection—manipulating the spectator away from the cause of a magic effect—is widely considered a central element of the practice of magic: “[m]isdirection is a principle element in the art of deception"_ (Randal, 1976, p. 380)," - [A psychologically-based taxonomy of misdirection](https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4260479/) The primary resource I use to maintain all of the above - an excellent book by C. Silverman, [Verification Handbook - For Disinformation And Media Manipulation](https://datajournalism.com/read/handbook/verification-3) - which is, for all intents and purposes, meant to *detect* media manipulation and misinformation, but we're going to use it to intentionally *misdirect*.
non_process
new appendix spreading disinformation through social media a how to for protecting pseudonymous identities misinformation and disinformation in social media to protect identities we could have a section or appendix for user identities and spreading misinformation to protect them for example when creating anonymous identities it s much more than just making up a name and pretending to be a certain age having accounts for that person is good but how do you remember to use these accounts and make it look believable it can include making fake user accounts on social media that resemble your own it can also include scheduling posts on twitter reddit etc i propose a direct guide section be made for putting out mis disinformation regarding identities online how to generate believable names nicknames and how to maintain social media presence in a way that s straightforward without being a burden on the person trying to do so this might include making reminders to post on your fake accounts to keep up appearances how and when you should put out misinformation and when it s a good idea to misdirect having a fake name isn t enough you have to keep up with the social media presence including regularly posting but also maintaining the perspective of your fake identity which can be much more difficult than just telling lies on the internet we live in a society full of misinformation most of it sadly is intentional on behalf of governments to keep the peace or push a narrative this is something that main stream media has done since the dawn of world war i it was the first time that highly organized state propaganda institutions were developed you should be doing the same in the realm of keeping up anonymous identities it s about knowing not only how to misdirect like a magician but also when and where as well as making it a natural behavior in other words you won t be so obvious and it will be more believable and more effective as a result give any additional relevant context the most insidious and underrated problem in our information ecosystem is that we do not give the right kind of attention to the right things at the right time we are living in a magical world where with the stroke of a wand the social media universe shifts either for the good or for the worst and though misinformation on social media has been a problem for years it’s more prevalent now than ever understanding how disinformation online has endangered democracy and democratic norms is incredibly nuanced when a lie is repeated and seen multiple times people unintentionally begin to accept it as truth misdirection—manipulating the spectator away from the cause of a magic effect—is widely considered a central element of the practice of magic “ isdirection is a principle element in the art of deception randal p the primary resource i use to maintain all of the above an excellent book by c silverman which is for all intents and purposes meant to detect media manipulation and misinformation but we re going to use it to intentionally misdirect
0
19,221
25,358,606,105
IssuesEvent
2022-11-20 16:33:04
streamnative/pulsar-spark
https://api.github.com/repos/streamnative/pulsar-spark
closed
Need a newer pulsar-spark version with pulsar-cient version 2.9+
type/feature compute/data-processing
Creating a new ticket as #88 was wrongfully closed. We are trying to use this adapter with pyspark. One of the challenges we are facing is setting custom subscription name for the individual readers that are created. The current version - 3.1.1.4 uses pulsar client 2.4 which doesn't have a way to set the full subscription name to something custom. It always appends a random UUID string. Latest pulsar client 2.9 fixes that. This repo's master branch is already updated to 2.9 pulsar client but there has been no release out of master since. When we tried locally building from master which has the pulsar client 2.9 dependency, (or taking the 3.1 branch and updating pulsar-client version to 2.9), we get the following error from pyspark when we eventually use the jar: ``` java.lang.ExceptionInInitializerError at org.apache.spark.sql.pulsar.PulsarProvider$.$anonfun$getClientParams$3(PulsarProvider.scala:259) at scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286) at scala.collection.immutable.Map$Map2.foreach(Map.scala:273) at scala.collection.TraversableLike.map(TraversableLike.scala:286) at scala.collection.TraversableLike.map$(TraversableLike.scala:279) at scala.collection.AbstractTraversable.map(Traversable.scala:108) at org.apache.spark.sql.pulsar.PulsarProvider$.getClientParams(PulsarProvider.scala:258) at org.apache.spark.sql.pulsar.PulsarProvider$.org$apache$spark$sql$pulsar$PulsarProvider$$prepareConfForReader(PulsarProvider.scala:513) at org.apache.spark.sql.pulsar.PulsarProvider.sourceSchema(PulsarProvider.scala:60) at org.apache.spark.sql.execution.datasources.DataSource.sourceSchema(DataSource.scala:236) at org.apache.spark.sql.execution.datasources.DataSource.sourceInfo$lzycompute(DataSource.scala:118) at org.apache.spark.sql.execution.datasources.DataSource.sourceInfo(DataSource.scala:118) at org.apache.spark.sql.execution.streaming.StreamingRelation$.apply(StreamingRelation.scala:34) at org.apache.spark.sql.streaming.DataStreamReader.loadInternal(DataStreamReader.scala:168) at org.apache.spark.sql.streaming.DataStreamReader.load(DataStreamReader.scala:144) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) at py4j.Gateway.invoke(Gateway.java:282) at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) at py4j.commands.CallCommand.execute(CallCommand.java:79) at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182) at py4j.ClientServerConnection.run(ClientServerConnection.java:106) at java.lang.Thread.run(Thread.java:748) Caused by: scala.reflect.internal.Symbols$CyclicReference: illegal cyclic reference involving class InterfaceAudience ```
1.0
Need a newer pulsar-spark version with pulsar-cient version 2.9+ - Creating a new ticket as #88 was wrongfully closed. We are trying to use this adapter with pyspark. One of the challenges we are facing is setting custom subscription name for the individual readers that are created. The current version - 3.1.1.4 uses pulsar client 2.4 which doesn't have a way to set the full subscription name to something custom. It always appends a random UUID string. Latest pulsar client 2.9 fixes that. This repo's master branch is already updated to 2.9 pulsar client but there has been no release out of master since. When we tried locally building from master which has the pulsar client 2.9 dependency, (or taking the 3.1 branch and updating pulsar-client version to 2.9), we get the following error from pyspark when we eventually use the jar: ``` java.lang.ExceptionInInitializerError at org.apache.spark.sql.pulsar.PulsarProvider$.$anonfun$getClientParams$3(PulsarProvider.scala:259) at scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286) at scala.collection.immutable.Map$Map2.foreach(Map.scala:273) at scala.collection.TraversableLike.map(TraversableLike.scala:286) at scala.collection.TraversableLike.map$(TraversableLike.scala:279) at scala.collection.AbstractTraversable.map(Traversable.scala:108) at org.apache.spark.sql.pulsar.PulsarProvider$.getClientParams(PulsarProvider.scala:258) at org.apache.spark.sql.pulsar.PulsarProvider$.org$apache$spark$sql$pulsar$PulsarProvider$$prepareConfForReader(PulsarProvider.scala:513) at org.apache.spark.sql.pulsar.PulsarProvider.sourceSchema(PulsarProvider.scala:60) at org.apache.spark.sql.execution.datasources.DataSource.sourceSchema(DataSource.scala:236) at org.apache.spark.sql.execution.datasources.DataSource.sourceInfo$lzycompute(DataSource.scala:118) at org.apache.spark.sql.execution.datasources.DataSource.sourceInfo(DataSource.scala:118) at org.apache.spark.sql.execution.streaming.StreamingRelation$.apply(StreamingRelation.scala:34) at org.apache.spark.sql.streaming.DataStreamReader.loadInternal(DataStreamReader.scala:168) at org.apache.spark.sql.streaming.DataStreamReader.load(DataStreamReader.scala:144) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244) at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357) at py4j.Gateway.invoke(Gateway.java:282) at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132) at py4j.commands.CallCommand.execute(CallCommand.java:79) at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182) at py4j.ClientServerConnection.run(ClientServerConnection.java:106) at java.lang.Thread.run(Thread.java:748) Caused by: scala.reflect.internal.Symbols$CyclicReference: illegal cyclic reference involving class InterfaceAudience ```
process
need a newer pulsar spark version with pulsar cient version creating a new ticket as was wrongfully closed we are trying to use this adapter with pyspark one of the challenges we are facing is setting custom subscription name for the individual readers that are created the current version uses pulsar client which doesn t have a way to set the full subscription name to something custom it always appends a random uuid string latest pulsar client fixes that this repo s master branch is already updated to pulsar client but there has been no release out of master since when we tried locally building from master which has the pulsar client dependency or taking the branch and updating pulsar client version to we get the following error from pyspark when we eventually use the jar java lang exceptionininitializererror at org apache spark sql pulsar pulsarprovider anonfun getclientparams pulsarprovider scala at scala collection traversablelike anonfun map traversablelike scala at scala collection immutable map foreach map scala at scala collection traversablelike map traversablelike scala at scala collection traversablelike map traversablelike scala at scala collection abstracttraversable map traversable scala at org apache spark sql pulsar pulsarprovider getclientparams pulsarprovider scala at org apache spark sql pulsar pulsarprovider org apache spark sql pulsar pulsarprovider prepareconfforreader pulsarprovider scala at org apache spark sql pulsar pulsarprovider sourceschema pulsarprovider scala at org apache spark sql execution datasources datasource sourceschema datasource scala at org apache spark sql execution datasources datasource sourceinfo lzycompute datasource scala at org apache spark sql execution datasources datasource sourceinfo datasource scala at org apache spark sql execution streaming streamingrelation apply streamingrelation scala at org apache spark sql streaming datastreamreader loadinternal datastreamreader scala at org apache spark sql streaming datastreamreader load datastreamreader scala at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at reflection methodinvoker invoke methodinvoker java at reflection reflectionengine invoke reflectionengine java at gateway invoke gateway java at commands abstractcommand invokemethod abstractcommand java at commands callcommand execute callcommand java at clientserverconnection waitforcommands clientserverconnection java at clientserverconnection run clientserverconnection java at java lang thread run thread java caused by scala reflect internal symbols cyclicreference illegal cyclic reference involving class interfaceaudience
1
242,493
18,666,789,773
IssuesEvent
2021-10-30 00:56:29
paulrobertlloyd/govuk-prototype-rig
https://api.github.com/repos/paulrobertlloyd/govuk-prototype-rig
opened
Document how to add form validation
documentation
There’s already an example, but some further explanation may be needed – i.e. what is the `errorList` helper for.
1.0
Document how to add form validation - There’s already an example, but some further explanation may be needed – i.e. what is the `errorList` helper for.
non_process
document how to add form validation there’s already an example but some further explanation may be needed – i e what is the errorlist helper for
0
4,866
7,749,673,444
IssuesEvent
2018-05-30 12:17:27
dzhw/zofar
https://api.github.com/repos/dzhw/zofar
opened
zofar function episodes (cal)
category: technical.processes prio: ? status: development type: backlog.task
translating episode data into usable variables - such as timestamps, timeframes, number of episodes, etc...
1.0
zofar function episodes (cal) - translating episode data into usable variables - such as timestamps, timeframes, number of episodes, etc...
process
zofar function episodes cal translating episode data into usable variables such as timestamps timeframes number of episodes etc
1
9,060
12,134,213,492
IssuesEvent
2020-04-23 10:19:33
MHRA/products
https://api.github.com/repos/MHRA/products
opened
Incoming requests should be logged before they are rejected
BUG :bug: EPIC - Auto Batch Process :oncoming_automobile:
**Describe the bug** A 4XX/5XX response is returned to the caller (due to e.g. unauthorized or incompatible headers) but there is no log of the error or the request that created it. **To Reproduce** 1. Make a request to the doc-index-updater with an invalid `Content-Type` (which returns a 415). 2. Observe no entry in the logs of the error or the incoming request **Expected behavior** The incoming request should be logged as info, a 4XX should be logged at an appropriate level (info/warn?) and a 5XX should be logged as an error. **Screenshots** N/A
1.0
Incoming requests should be logged before they are rejected - **Describe the bug** A 4XX/5XX response is returned to the caller (due to e.g. unauthorized or incompatible headers) but there is no log of the error or the request that created it. **To Reproduce** 1. Make a request to the doc-index-updater with an invalid `Content-Type` (which returns a 415). 2. Observe no entry in the logs of the error or the incoming request **Expected behavior** The incoming request should be logged as info, a 4XX should be logged at an appropriate level (info/warn?) and a 5XX should be logged as an error. **Screenshots** N/A
process
incoming requests should be logged before they are rejected describe the bug a response is returned to the caller due to e g unauthorized or incompatible headers but there is no log of the error or the request that created it to reproduce make a request to the doc index updater with an invalid content type which returns a observe no entry in the logs of the error or the incoming request expected behavior the incoming request should be logged as info a should be logged at an appropriate level info warn and a should be logged as an error screenshots n a
1
11,809
14,628,748,785
IssuesEvent
2020-12-23 14:41:54
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
resource trigger pipeline variables not present in devops server 2020 rc2
Pri2 devops-cicd-process/tech devops/prod doc-bug
In the documentation i read that when a resource triggers a pipeline should be available two variables: resources.triggeringAlias resources.triggeringCategory but printing env in the triggered pipeline, it doesn't print values, neither the variables name. Also pipeline doesn't substitute them in a task definition Is it a documentation bug so those variables do not exist in devops server 2020 rc2 or they should exist and there is a bug in the devops server 2020 rc2 ? Thank you --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2 * Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee * Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema&viewFallbackFrom=azure-devops-2020#resources-pipelines) * Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
resource trigger pipeline variables not present in devops server 2020 rc2 - In the documentation i read that when a resource triggers a pipeline should be available two variables: resources.triggeringAlias resources.triggeringCategory but printing env in the triggered pipeline, it doesn't print values, neither the variables name. Also pipeline doesn't substitute them in a task definition Is it a documentation bug so those variables do not exist in devops server 2020 rc2 or they should exist and there is a bug in the devops server 2020 rc2 ? Thank you --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ee4ec9d0-e0d5-4fb4-7c3e-b84abfa290c2 * Version Independent ID: 3e2b80d9-30e5-0c48-49f0-4fcdfedf5eee * Content: [Resources - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/resources?view=azure-devops&tabs=schema&viewFallbackFrom=azure-devops-2020#resources-pipelines) * Content Source: [docs/pipelines/process/resources.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/resources.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
resource trigger pipeline variables not present in devops server in the documentation i read that when a resource triggers a pipeline should be available two variables resources triggeringalias resources triggeringcategory but printing env in the triggered pipeline it doesn t print values neither the variables name also pipeline doesn t substitute them in a task definition is it a documentation bug so those variables do not exist in devops server or they should exist and there is a bug in the devops server thank you document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
102,429
16,566,211,613
IssuesEvent
2021-05-29 13:14:48
AlexRogalskiy/github-action-random-quote
https://api.github.com/repos/AlexRogalskiy/github-action-random-quote
opened
CVE-2015-9251 (Medium) detected in jquery-1.8.1.min.js
security vulnerability
## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: github-action-random-quote/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: github-action-random-quote/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-random-quote/commit/99c135aa691d6abe5673bace94251bc0966b49e0">99c135aa691d6abe5673bace94251bc0966b49e0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-9251 (Medium) detected in jquery-1.8.1.min.js - ## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: github-action-random-quote/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: github-action-random-quote/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-random-quote/commit/99c135aa691d6abe5673bace94251bc0966b49e0">99c135aa691d6abe5673bace94251bc0966b49e0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file github action random quote node modules redeyed examples browser index html path to vulnerable library github action random quote node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
21,034
27,979,065,674
IssuesEvent
2023-03-25 23:41:05
nsarrazin/serge
https://api.github.com/repos/nsarrazin/serge
closed
CI/CD Docker pipeline
processes enhancement
Love the project and I would love to contribute! My first thought its building a quick CI/CD pipeline with GitHub actions that allows any merge into a "release" branch would trigger a build of a docker image and upload it to the GitHub package repo. This way people can use their own docker-compose.yml (or the provided one in the repo) and pull the images without building them. Would I be able to get this setup? Edit: Grammar
1.0
CI/CD Docker pipeline - Love the project and I would love to contribute! My first thought its building a quick CI/CD pipeline with GitHub actions that allows any merge into a "release" branch would trigger a build of a docker image and upload it to the GitHub package repo. This way people can use their own docker-compose.yml (or the provided one in the repo) and pull the images without building them. Would I be able to get this setup? Edit: Grammar
process
ci cd docker pipeline love the project and i would love to contribute my first thought its building a quick ci cd pipeline with github actions that allows any merge into a release branch would trigger a build of a docker image and upload it to the github package repo this way people can use their own docker compose yml or the provided one in the repo and pull the images without building them would i be able to get this setup edit grammar
1
7,100
10,255,766,051
IssuesEvent
2019-08-21 16:04:46
prisma/specs
https://api.github.com/repos/prisma/specs
closed
Repo spec
area/process
Prisma uses multiple repos to store things, so it makes sense to write that down to be aware what our target is and to update this if we change that.
1.0
Repo spec - Prisma uses multiple repos to store things, so it makes sense to write that down to be aware what our target is and to update this if we change that.
process
repo spec prisma uses multiple repos to store things so it makes sense to write that down to be aware what our target is and to update this if we change that
1
103,152
4,164,939,035
IssuesEvent
2016-06-19 05:34:47
matan-sh/commitment-wall
https://api.github.com/repos/matan-sh/commitment-wall
closed
access control system
Done Page: Administration Priority: High
Only administrators can access to the administrator page with a full name and password
1.0
access control system - Only administrators can access to the administrator page with a full name and password
non_process
access control system only administrators can access to the administrator page with a full name and password
0
6,415
9,515,045,926
IssuesEvent
2019-04-26 03:30:25
arxiv-vanity/engrafo
https://api.github.com/repos/arxiv-vanity/engrafo
closed
Linkify old arXiv URLs
area/postprocessor good first issue help wanted type/enhancement
e.g. `gr-qc/9704013` in https://www.arxiv-vanity.com/papers/gr-qc/9806119/. They're sufficiently weird that a regex should work fine, but perhaps we can constrain to just the bibliography. - https://www.arxiv-vanity.com/papers/1601.00921/ - https://www.arxiv-vanity.com/papers/hep-th/9808085/ - https://www.arxiv-vanity.com/papers/1812.11173/
1.0
Linkify old arXiv URLs - e.g. `gr-qc/9704013` in https://www.arxiv-vanity.com/papers/gr-qc/9806119/. They're sufficiently weird that a regex should work fine, but perhaps we can constrain to just the bibliography. - https://www.arxiv-vanity.com/papers/1601.00921/ - https://www.arxiv-vanity.com/papers/hep-th/9808085/ - https://www.arxiv-vanity.com/papers/1812.11173/
process
linkify old arxiv urls e g gr qc in they re sufficiently weird that a regex should work fine but perhaps we can constrain to just the bibliography
1
210,040
16,331,833,400
IssuesEvent
2021-05-12 10:10:01
FiveIT/eseuri
https://api.github.com/repos/FiveIT/eseuri
closed
Write documentation (due by 30th April)
documentation
Documentation requirements: - [x] Team's name - [x] Team's members - [x] Team's mentor - [x] Project's title - [x] Project summary (5-10 lines) - [x] Technology stack (software, hardware) - [x] Project description (implemented modules/functionality) [3-4 pages] - [x] Each team members' role in the making of the project (max. 10 lines) - [x] Hiccups/difficulties faced (2-4 lines) - [x] Future development (2-4 lines) - [x] Reference/bibliography (links to the used technologies' documentations etc.) [Template document][1] (fill in with Calibri 11) [1]: https://www.facebook.com/download/275338504297948/Model_documentatie_4IT50.docx?av=100006250506848&eav=AfZO2GfIKvvQRGecDYackUZXsAt5jOYlFvjZXc7MWepNwmF49bAsmV7oQgdS2L3f4SE&hash=AcpfVryb7ZLrnnQhv3k&__cft__[0]=AZXL_bOAqBM5UOqDUxPlyK1vQcQTkAmmIuFXexsj7_oOvFQw2RbCK_nXziNbw8prh769UpQvZFgoyW_DpyF-Jsvlc4RhxG-5iB3XRL-kSFz2EvDpRNq0sLrOWW1UWjk2Dnt6Ks6MfJSN_IyKeS3pm1X3xJsWvDn2In65Zsmg3j_KdkNWhGHOGVzQBVoJY33nlCA&__tn__=H-R
1.0
Write documentation (due by 30th April) - Documentation requirements: - [x] Team's name - [x] Team's members - [x] Team's mentor - [x] Project's title - [x] Project summary (5-10 lines) - [x] Technology stack (software, hardware) - [x] Project description (implemented modules/functionality) [3-4 pages] - [x] Each team members' role in the making of the project (max. 10 lines) - [x] Hiccups/difficulties faced (2-4 lines) - [x] Future development (2-4 lines) - [x] Reference/bibliography (links to the used technologies' documentations etc.) [Template document][1] (fill in with Calibri 11) [1]: https://www.facebook.com/download/275338504297948/Model_documentatie_4IT50.docx?av=100006250506848&eav=AfZO2GfIKvvQRGecDYackUZXsAt5jOYlFvjZXc7MWepNwmF49bAsmV7oQgdS2L3f4SE&hash=AcpfVryb7ZLrnnQhv3k&__cft__[0]=AZXL_bOAqBM5UOqDUxPlyK1vQcQTkAmmIuFXexsj7_oOvFQw2RbCK_nXziNbw8prh769UpQvZFgoyW_DpyF-Jsvlc4RhxG-5iB3XRL-kSFz2EvDpRNq0sLrOWW1UWjk2Dnt6Ks6MfJSN_IyKeS3pm1X3xJsWvDn2In65Zsmg3j_KdkNWhGHOGVzQBVoJY33nlCA&__tn__=H-R
non_process
write documentation due by april documentation requirements team s name team s members team s mentor project s title project summary lines technology stack software hardware project description implemented modules functionality each team members role in the making of the project max lines hiccups difficulties faced lines future development lines reference bibliography links to the used technologies documentations etc fill in with calibri azxl dpyf tn h r
0
6,288
9,292,403,755
IssuesEvent
2019-03-22 02:57:51
cypress-io/cypress-example-recipes
https://api.github.com/repos/cypress-io/cypress-example-recipes
closed
minimize Renovate Bot noise
ci process
Group updates, make sure important dependencies (like cypress) are updated frequently, and other dependencies are updated as needed
1.0
minimize Renovate Bot noise - Group updates, make sure important dependencies (like cypress) are updated frequently, and other dependencies are updated as needed
process
minimize renovate bot noise group updates make sure important dependencies like cypress are updated frequently and other dependencies are updated as needed
1
226,403
7,518,890,489
IssuesEvent
2018-04-12 09:47:53
WordImpress/Give
https://api.github.com/repos/WordImpress/Give
closed
feat(form): allow forms to be queried based on close status
3-reported high-priority
## Issue Overview Occasionally, users want to have an archive of Give forms, but exclude those that are closed because they achieved their goal. We currently have `post_meta` called `_give_close_form_when_goal_achieved` which returns just `enabled|disabled`, but when the goal is actually achieved and the form is closed, there's no way to query and exclude those forms. Adding `_give_form_is_closed` with `true|false` would benefit these users greatly.
1.0
feat(form): allow forms to be queried based on close status - ## Issue Overview Occasionally, users want to have an archive of Give forms, but exclude those that are closed because they achieved their goal. We currently have `post_meta` called `_give_close_form_when_goal_achieved` which returns just `enabled|disabled`, but when the goal is actually achieved and the form is closed, there's no way to query and exclude those forms. Adding `_give_form_is_closed` with `true|false` would benefit these users greatly.
non_process
feat form allow forms to be queried based on close status issue overview occasionally users want to have an archive of give forms but exclude those that are closed because they achieved their goal we currently have post meta called give close form when goal achieved which returns just enabled disabled but when the goal is actually achieved and the form is closed there s no way to query and exclude those forms adding give form is closed with true false would benefit these users greatly
0
45,045
9,667,772,764
IssuesEvent
2019-05-21 13:54:33
IBAIT18/group5
https://api.github.com/repos/IBAIT18/group5
closed
GUI mit Datenbank verknüpfen
GUI Hintergrundcode
Eingaben aus der GUI sollen in der Datenbank gespeichert und aus der Datenbank gezogen werden können.
1.0
GUI mit Datenbank verknüpfen - Eingaben aus der GUI sollen in der Datenbank gespeichert und aus der Datenbank gezogen werden können.
non_process
gui mit datenbank verknüpfen eingaben aus der gui sollen in der datenbank gespeichert und aus der datenbank gezogen werden können
0
18,973
24,952,707,477
IssuesEvent
2022-11-01 08:56:07
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
[FALSE-POSITIVE?] lowendbox.com
whitelisting process
**Domains or links** <!-- Please list below any domains and links listed here which you believe are a false positive. --> 1. lowendbox.com **More Information** 1. It's just forums website that discusses vps hosting and more
1.0
[FALSE-POSITIVE?] lowendbox.com - **Domains or links** <!-- Please list below any domains and links listed here which you believe are a false positive. --> 1. lowendbox.com **More Information** 1. It's just forums website that discusses vps hosting and more
process
lowendbox com domains or links lowendbox com more information it s just forums website that discusses vps hosting and more
1
15,644
19,846,202,660
IssuesEvent
2022-01-21 06:45:10
ooi-data/RS01SBPD-DP01A-03-FLCDRA102-recovered_inst-dpc_flcdrtd_instrument_recovered
https://api.github.com/repos/ooi-data/RS01SBPD-DP01A-03-FLCDRA102-recovered_inst-dpc_flcdrtd_instrument_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T06:45:10.093142. ## Details Flow name: `RS01SBPD-DP01A-03-FLCDRA102-recovered_inst-dpc_flcdrtd_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T06:45:10.093142. ## Details Flow name: `RS01SBPD-DP01A-03-FLCDRA102-recovered_inst-dpc_flcdrtd_instrument_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered inst dpc flcdrtd instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
20,430
27,095,311,000
IssuesEvent
2023-02-15 02:00:08
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Wed, 15 Feb 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### Make Your Brief Stroke Real and Stereoscopic: 3D-Aware Simplified Sketch to Portrait Generation - **Authors:** Yasheng Sun, Qianyi Wu, Hang Zhou, Kaisiyuan Wang, Tianshu Hu, Chen-Chieh Liao, Dongliang He, Jingtuo Liu, Errui Ding, Jingdong Wang, Shio Miyafuji, Ziwei Liu, Hideki Koike - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.06857 - **Pdf link:** https://arxiv.org/pdf/2302.06857 - **Abstract** Creating the photo-realistic version of people sketched portraits is useful to various entertainment purposes. Existing studies only generate portraits in the 2D plane with fixed views, making the results less vivid. In this paper, we present Stereoscopic Simplified Sketch-to-Portrait (SSSP), which explores the possibility of creating Stereoscopic 3D-aware portraits from simple contour sketches by involving 3D generative models. Our key insight is to design sketch-aware constraints that can fully exploit the prior knowledge of a tri-plane-based 3D-aware generative model. Specifically, our designed region-aware volume rendering strategy and global consistency constraint further enhance detail correspondences during sketch encoding. Moreover, in order to facilitate the usage of layman users, we propose a Contour-to-Sketch module with vector quantized representations, so that easily drawn contours can directly guide the generation of 3D portraits. Extensive comparisons show that our method generates high-quality results that match the sketch. Our usability study verifies that our system is greatly preferred by user. ### Text-Guided Scene Sketch-to-Photo Synthesis - **Authors:** AprilPyone MaungMaung, Makoto Shing, Kentaro Mitsui, Kei Sawada, Fumio Okura - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.06883 - **Pdf link:** https://arxiv.org/pdf/2302.06883 - **Abstract** We propose a method for scene-level sketch-to-photo synthesis with text guidance. Although object-level sketch-to-photo synthesis has been widely studied, whole-scene synthesis is still challenging without reference photos that adequately reflect the target style. To this end, we leverage knowledge from recent large-scale pre-trained generative models, resulting in text-guided sketch-to-photo synthesis without the need for reference images. To train our model, we use self-supervised learning from a set of photographs. Specifically, we use a pre-trained edge detector that maps both color and sketch images into a standardized edge domain, which reduces the gap between photograph-based edge images (during training) and hand-drawn sketch images (during inference). We implement our method by fine-tuning a latent diffusion model (i.e., Stable Diffusion) with sketch and text conditions. Experiments show that the proposed method translates original sketch images that are not extracted from color images into photos with compelling visual quality. ### Camera Calibration without Camera Access -- A Robust Validation Technique for Extended PnP Methods - **Authors:** Emil Brissman, Per-Erik Forssén, Johan Edstedt - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.06949 - **Pdf link:** https://arxiv.org/pdf/2302.06949 - **Abstract** A challenge in image based metrology and forensics is intrinsic camera calibration when the used camera is unavailable. The unavailability raises two questions. The first question is how to find the projection model that describes the camera, and the second is to detect incorrect models. In this work, we use off-the-shelf extended PnP-methods to find the model from 2D-3D correspondences, and propose a method for model validation. The most common strategy for evaluating a projection model is comparing different models' residual variances - however, this naive strategy cannot distinguish whether the projection model is potentially underfitted or overfitted. To this end, we model the residual errors for each correspondence, individually scale all residuals using a predicted variance and test if the new residuals are drawn from a standard normal distribution. We demonstrate the effectiveness of our proposed validation in experiments on synthetic data, simulating 2D detection and Lidar measurements. Additionally, we provide experiments using data from an actual scene and compare non-camera access and camera access calibrations. Last, we use our method to validate annotations in MegaDepth. ## Keyword: raw image There is no result
2.0
New submissions for Wed, 15 Feb 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### Make Your Brief Stroke Real and Stereoscopic: 3D-Aware Simplified Sketch to Portrait Generation - **Authors:** Yasheng Sun, Qianyi Wu, Hang Zhou, Kaisiyuan Wang, Tianshu Hu, Chen-Chieh Liao, Dongliang He, Jingtuo Liu, Errui Ding, Jingdong Wang, Shio Miyafuji, Ziwei Liu, Hideki Koike - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.06857 - **Pdf link:** https://arxiv.org/pdf/2302.06857 - **Abstract** Creating the photo-realistic version of people sketched portraits is useful to various entertainment purposes. Existing studies only generate portraits in the 2D plane with fixed views, making the results less vivid. In this paper, we present Stereoscopic Simplified Sketch-to-Portrait (SSSP), which explores the possibility of creating Stereoscopic 3D-aware portraits from simple contour sketches by involving 3D generative models. Our key insight is to design sketch-aware constraints that can fully exploit the prior knowledge of a tri-plane-based 3D-aware generative model. Specifically, our designed region-aware volume rendering strategy and global consistency constraint further enhance detail correspondences during sketch encoding. Moreover, in order to facilitate the usage of layman users, we propose a Contour-to-Sketch module with vector quantized representations, so that easily drawn contours can directly guide the generation of 3D portraits. Extensive comparisons show that our method generates high-quality results that match the sketch. Our usability study verifies that our system is greatly preferred by user. ### Text-Guided Scene Sketch-to-Photo Synthesis - **Authors:** AprilPyone MaungMaung, Makoto Shing, Kentaro Mitsui, Kei Sawada, Fumio Okura - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.06883 - **Pdf link:** https://arxiv.org/pdf/2302.06883 - **Abstract** We propose a method for scene-level sketch-to-photo synthesis with text guidance. Although object-level sketch-to-photo synthesis has been widely studied, whole-scene synthesis is still challenging without reference photos that adequately reflect the target style. To this end, we leverage knowledge from recent large-scale pre-trained generative models, resulting in text-guided sketch-to-photo synthesis without the need for reference images. To train our model, we use self-supervised learning from a set of photographs. Specifically, we use a pre-trained edge detector that maps both color and sketch images into a standardized edge domain, which reduces the gap between photograph-based edge images (during training) and hand-drawn sketch images (during inference). We implement our method by fine-tuning a latent diffusion model (i.e., Stable Diffusion) with sketch and text conditions. Experiments show that the proposed method translates original sketch images that are not extracted from color images into photos with compelling visual quality. ### Camera Calibration without Camera Access -- A Robust Validation Technique for Extended PnP Methods - **Authors:** Emil Brissman, Per-Erik Forssén, Johan Edstedt - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2302.06949 - **Pdf link:** https://arxiv.org/pdf/2302.06949 - **Abstract** A challenge in image based metrology and forensics is intrinsic camera calibration when the used camera is unavailable. The unavailability raises two questions. The first question is how to find the projection model that describes the camera, and the second is to detect incorrect models. In this work, we use off-the-shelf extended PnP-methods to find the model from 2D-3D correspondences, and propose a method for model validation. The most common strategy for evaluating a projection model is comparing different models' residual variances - however, this naive strategy cannot distinguish whether the projection model is potentially underfitted or overfitted. To this end, we model the residual errors for each correspondence, individually scale all residuals using a predicted variance and test if the new residuals are drawn from a standard normal distribution. We demonstrate the effectiveness of our proposed validation in experiments on synthetic data, simulating 2D detection and Lidar measurements. Additionally, we provide experiments using data from an actual scene and compare non-camera access and camera access calibrations. Last, we use our method to validate annotations in MegaDepth. ## Keyword: raw image There is no result
process
new submissions for wed feb keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw make your brief stroke real and stereoscopic aware simplified sketch to portrait generation authors yasheng sun qianyi wu hang zhou kaisiyuan wang tianshu hu chen chieh liao dongliang he jingtuo liu errui ding jingdong wang shio miyafuji ziwei liu hideki koike subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract creating the photo realistic version of people sketched portraits is useful to various entertainment purposes existing studies only generate portraits in the plane with fixed views making the results less vivid in this paper we present stereoscopic simplified sketch to portrait sssp which explores the possibility of creating stereoscopic aware portraits from simple contour sketches by involving generative models our key insight is to design sketch aware constraints that can fully exploit the prior knowledge of a tri plane based aware generative model specifically our designed region aware volume rendering strategy and global consistency constraint further enhance detail correspondences during sketch encoding moreover in order to facilitate the usage of layman users we propose a contour to sketch module with vector quantized representations so that easily drawn contours can directly guide the generation of portraits extensive comparisons show that our method generates high quality results that match the sketch our usability study verifies that our system is greatly preferred by user text guided scene sketch to photo synthesis authors aprilpyone maungmaung makoto shing kentaro mitsui kei sawada fumio okura subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract we propose a method for scene level sketch to photo synthesis with text guidance although object level sketch to photo synthesis has been widely studied whole scene synthesis is still challenging without reference photos that adequately reflect the target style to this end we leverage knowledge from recent large scale pre trained generative models resulting in text guided sketch to photo synthesis without the need for reference images to train our model we use self supervised learning from a set of photographs specifically we use a pre trained edge detector that maps both color and sketch images into a standardized edge domain which reduces the gap between photograph based edge images during training and hand drawn sketch images during inference we implement our method by fine tuning a latent diffusion model i e stable diffusion with sketch and text conditions experiments show that the proposed method translates original sketch images that are not extracted from color images into photos with compelling visual quality camera calibration without camera access a robust validation technique for extended pnp methods authors emil brissman per erik forssén johan edstedt subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract a challenge in image based metrology and forensics is intrinsic camera calibration when the used camera is unavailable the unavailability raises two questions the first question is how to find the projection model that describes the camera and the second is to detect incorrect models in this work we use off the shelf extended pnp methods to find the model from correspondences and propose a method for model validation the most common strategy for evaluating a projection model is comparing different models residual variances however this naive strategy cannot distinguish whether the projection model is potentially underfitted or overfitted to this end we model the residual errors for each correspondence individually scale all residuals using a predicted variance and test if the new residuals are drawn from a standard normal distribution we demonstrate the effectiveness of our proposed validation in experiments on synthetic data simulating detection and lidar measurements additionally we provide experiments using data from an actual scene and compare non camera access and camera access calibrations last we use our method to validate annotations in megadepth keyword raw image there is no result
1
9,937
11,943,898,510
IssuesEvent
2020-04-03 00:42:32
AzureAD/microsoft-authentication-library-for-js
https://api.github.com/repos/AzureAD/microsoft-authentication-library-for-js
closed
Login Flow not working for MS Teams App with 1.2.0-Beta.3
bug compatibility p2
<!-- PLEASE HELP US TO HELP YOU BETTER AND FASTER BY PROVIDING THE FOLLOWING INFORMATION. --> ## I'm submitting a... <!-- Check one of the following options with "x" --> <pre><code> [ ] Regression (a behavior that used to work and stopped working in a new release) [X] Bug report <!-- Please search GitHub for a similar issue or PR before submitting --> [ ] Performance issue [ ] Feature request [ ] Documentation issue or request [ ] Other... Please describe: </code></pre> ## Browser: - [X] Chrome version 78.0 - [X] Firefox version 70.0 - [ ] IE version XX - [ ] Edge version XX - [ ] Safari version XX ## Library version <pre><code> Library version: 1.2.0-beta.3 <!-- Check whether this is still an issue in the most recent version --> </code></pre> ## Current behavior <!-- Describe how the issue manifests. --> I am trying to make the authentication work in a MS Teams Tab App using MSAL.JS MS Teams requires me to open a popup for authentication with it's own method (` microsoftTeams.authentication.authenticate`) as normal popups are blocked by the MS Teams client. So I can not use the `loginPopup()`-Method of MSAL but have to use `loginRedirect()` from within the MS Teams Popup. The authentication itself works and redirects back with a valid id_token in the url-hash. However, the hash is not handled and the redirect callback nevers gets called. The reason is [Line227](https://github.com/AzureAD/microsoft-authentication-library-for-js/blob/160efdc23963af376b242328de91e51ff9cb6eff/lib/msal-core/src/UserAgentApplication.ts#L227) in the `constructor` of the `UserAgentApplication`: ```JavaScript if (!this.config.framework.isAngular && urlContainsHash && !WindowUtils.isInIframe() && !WindowUtils.isInPopup()) { this.handleAuthenticationResponse(urlHash); } ``` The last condition `WindowUtils.isInPopup()` is `true` in this case, and therefore `handleAuthenticationResponse()` is never called. I don't know the reason for this check in this particular case - but is there a way to work around this issue? Falling back to an older version of MSAL is also no option as I also need the new IFrame-Capapbility to use `acquireTokenSilent()`. I also tried adal.js, but that is currently not working on mobile due to another IFrame issue. ## Expected behavior <!-- Describe what the desired behavior would be. --> The callback of `loginRedirect()` is also called within a popup-window. ## Minimal reproduction of the problem with instructions <!-- please provide the *STEPS TO REPRODUCE* --> Use `loginRedirect()` from within a popup window.
True
Login Flow not working for MS Teams App with 1.2.0-Beta.3 - <!-- PLEASE HELP US TO HELP YOU BETTER AND FASTER BY PROVIDING THE FOLLOWING INFORMATION. --> ## I'm submitting a... <!-- Check one of the following options with "x" --> <pre><code> [ ] Regression (a behavior that used to work and stopped working in a new release) [X] Bug report <!-- Please search GitHub for a similar issue or PR before submitting --> [ ] Performance issue [ ] Feature request [ ] Documentation issue or request [ ] Other... Please describe: </code></pre> ## Browser: - [X] Chrome version 78.0 - [X] Firefox version 70.0 - [ ] IE version XX - [ ] Edge version XX - [ ] Safari version XX ## Library version <pre><code> Library version: 1.2.0-beta.3 <!-- Check whether this is still an issue in the most recent version --> </code></pre> ## Current behavior <!-- Describe how the issue manifests. --> I am trying to make the authentication work in a MS Teams Tab App using MSAL.JS MS Teams requires me to open a popup for authentication with it's own method (` microsoftTeams.authentication.authenticate`) as normal popups are blocked by the MS Teams client. So I can not use the `loginPopup()`-Method of MSAL but have to use `loginRedirect()` from within the MS Teams Popup. The authentication itself works and redirects back with a valid id_token in the url-hash. However, the hash is not handled and the redirect callback nevers gets called. The reason is [Line227](https://github.com/AzureAD/microsoft-authentication-library-for-js/blob/160efdc23963af376b242328de91e51ff9cb6eff/lib/msal-core/src/UserAgentApplication.ts#L227) in the `constructor` of the `UserAgentApplication`: ```JavaScript if (!this.config.framework.isAngular && urlContainsHash && !WindowUtils.isInIframe() && !WindowUtils.isInPopup()) { this.handleAuthenticationResponse(urlHash); } ``` The last condition `WindowUtils.isInPopup()` is `true` in this case, and therefore `handleAuthenticationResponse()` is never called. I don't know the reason for this check in this particular case - but is there a way to work around this issue? Falling back to an older version of MSAL is also no option as I also need the new IFrame-Capapbility to use `acquireTokenSilent()`. I also tried adal.js, but that is currently not working on mobile due to another IFrame issue. ## Expected behavior <!-- Describe what the desired behavior would be. --> The callback of `loginRedirect()` is also called within a popup-window. ## Minimal reproduction of the problem with instructions <!-- please provide the *STEPS TO REPRODUCE* --> Use `loginRedirect()` from within a popup window.
non_process
login flow not working for ms teams app with beta please help us to help you better and faster by providing the following information i m submitting a regression a behavior that used to work and stopped working in a new release bug report performance issue feature request documentation issue or request other please describe browser chrome version firefox version ie version xx edge version xx safari version xx library version library version beta current behavior i am trying to make the authentication work in a ms teams tab app using msal js ms teams requires me to open a popup for authentication with it s own method microsoftteams authentication authenticate as normal popups are blocked by the ms teams client so i can not use the loginpopup method of msal but have to use loginredirect from within the ms teams popup the authentication itself works and redirects back with a valid id token in the url hash however the hash is not handled and the redirect callback nevers gets called the reason is in the constructor of the useragentapplication javascript if this config framework isangular urlcontainshash windowutils isiniframe windowutils isinpopup this handleauthenticationresponse urlhash the last condition windowutils isinpopup is true in this case and therefore handleauthenticationresponse is never called i don t know the reason for this check in this particular case but is there a way to work around this issue falling back to an older version of msal is also no option as i also need the new iframe capapbility to use acquiretokensilent i also tried adal js but that is currently not working on mobile due to another iframe issue expected behavior the callback of loginredirect is also called within a popup window minimal reproduction of the problem with instructions use loginredirect from within a popup window
0
5,427
8,289,066,175
IssuesEvent
2018-09-19 13:47:41
Rokid/ShadowNode
https://api.github.com/repos/Rokid/ShadowNode
opened
process: port signals to module process
process
<!-- Thank you for suggesting an idea to make ShadowNode better. Please fill in as much of the template below as you're able. --> **Describe the solution you'd like** Please describe the desired behavior. Signals like `SIGTERM` and `SIGINT` shall be passed through to JS engine and emitted on global process event emitter.
1.0
process: port signals to module process - <!-- Thank you for suggesting an idea to make ShadowNode better. Please fill in as much of the template below as you're able. --> **Describe the solution you'd like** Please describe the desired behavior. Signals like `SIGTERM` and `SIGINT` shall be passed through to JS engine and emitted on global process event emitter.
process
process port signals to module process thank you for suggesting an idea to make shadownode better please fill in as much of the template below as you re able describe the solution you d like please describe the desired behavior signals like sigterm and sigint shall be passed through to js engine and emitted on global process event emitter
1
150,057
5,735,731,651
IssuesEvent
2017-04-22 01:06:22
dmillerw/Impractical-Storage
https://api.github.com/repos/dmillerw/Impractical-Storage
closed
Changing Boundaries Has No Effect & BetterFoliage Crash
confirmed-bug high-priority
Love the mod! Been loving all the new updates! Anyway, for whatever reason, on my custom modpack I am not able to change any of the settings for the storage controller. The preview does not change as intended, and when I tried to change the sort mode whilst items were being piped in (to test if it was just a visual glitch), the game crashed (logs attached). Any time I exit the GUI, opening it again defaults to 8,8,8 ; 0,0,0 rows as default. I'd love to be able to get this to work, but I cant work out from the logs what is causing this to fail :/ [crash-2017-04-11_21.59.23-client.txt](https://github.com/dmillerw/Impractical-Storage/files/914684/crash-2017-04-11_21.59.23-client.txt) Thanks!
1.0
Changing Boundaries Has No Effect & BetterFoliage Crash - Love the mod! Been loving all the new updates! Anyway, for whatever reason, on my custom modpack I am not able to change any of the settings for the storage controller. The preview does not change as intended, and when I tried to change the sort mode whilst items were being piped in (to test if it was just a visual glitch), the game crashed (logs attached). Any time I exit the GUI, opening it again defaults to 8,8,8 ; 0,0,0 rows as default. I'd love to be able to get this to work, but I cant work out from the logs what is causing this to fail :/ [crash-2017-04-11_21.59.23-client.txt](https://github.com/dmillerw/Impractical-Storage/files/914684/crash-2017-04-11_21.59.23-client.txt) Thanks!
non_process
changing boundaries has no effect betterfoliage crash love the mod been loving all the new updates anyway for whatever reason on my custom modpack i am not able to change any of the settings for the storage controller the preview does not change as intended and when i tried to change the sort mode whilst items were being piped in to test if it was just a visual glitch the game crashed logs attached any time i exit the gui opening it again defaults to rows as default i d love to be able to get this to work but i cant work out from the logs what is causing this to fail thanks
0
101,037
30,855,868,000
IssuesEvent
2023-08-02 20:32:56
elastic/beats
https://api.github.com/repos/elastic/beats
closed
Build 38 for 8.9 with status FAILURE - Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/postgresql/statement
automation ci-reported Team:Elastic-Agent-Data-Plane build-failures
## :broken_heart: Tests Failed <!-- BUILD BADGES--> > _the below badges are clickable and redirect to their specific view in the CI or DOCS_ [![Pipeline View](https://img.shields.io/badge/pipeline-pipeline%20-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//pipeline) [![Test View](https://img.shields.io/badge/test-test-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//tests) [![Changes](https://img.shields.io/badge/changes-changes-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//changes) [![Artifacts](https://img.shields.io/badge/artifacts-artifacts-yellow)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//artifacts) [![preview](https://img.shields.io/badge/docs-preview-yellowgreen)](http://beats_null.docs-preview.app.elstc.co/diff) [![preview](https://img.shields.io/badge/elastic-observability-blue)](https://ci-stats.elastic.co/app/apm/services/beats-ci/transactions/view?rangeFrom=2023-07-25T04:04:41.298Z&rangeTo=2023-07-25T04:24:41.298Z&transactionName=BUILD+Beats%2Fbeats%2F8.9&transactionType=job&latencyAggregationType=avg&traceId=16f6a010777d86eed8228c11a9eb4d91&transactionId=314de9e9fb8c2a3d) <!-- BUILD SUMMARY--> <details><summary>Expand to view the summary</summary> <p> #### Build stats * Start Time: 2023-07-25T04:14:41.298+0000 * Duration: 95 min 57 sec #### Test stats :test_tube: | Test | Results | | ------------ | :-----------------------------: | | Failed | 1 | | Passed | 30141 | | Skipped | 2356 | | Total | 32498 | </p> </details> <!-- TEST RESULTS IF ANY--> ### Test errors [![1](https://img.shields.io/badge/1%20-red)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//tests) <details><summary>Expand to view the tests failures</summary><p> ##### `Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/postgresql/statement` <ul> <details><summary>Expand to view the error details</summary><p> ``` Failed ``` </p></details> <details><summary>Expand to view the stacktrace</summary><p> ``` === RUN TestFetch statement_integration_test.go:41: Error Trace: /go/src/github.com/elastic/beats/metricbeat/module/postgresql/statement/statement_integration_test.go:41 Error: Should NOT be empty, but was [] Test: TestFetch --- FAIL: TestFetch (7.79s) panic: runtime error: index out of range [0] with length 0 [recovered] panic: runtime error: index out of range [0] with length 0 goroutine 20 [running]: testing.tRunner.func1.2({0x10ba6a0, 0xc00003c480}) /usr/local/go/src/testing/testing.go:1396 +0x24e testing.tRunner.func1() /usr/local/go/src/testing/testing.go:1399 +0x39f panic({0x10ba6a0, 0xc00003c480}) /usr/local/go/src/runtime/panic.go:884 +0x212 github.com/elastic/beats/v7/metricbeat/module/postgresql/statement.TestFetch(0xc00032c9c0) /go/src/github.com/elastic/beats/metricbeat/module/postgresql/statement/statement_integration_test.go:42 +0xbb7 testing.tRunner(0xc00032c9c0, 0x117fae8) /usr/local/go/src/testing/testing.go:1446 +0x10b created by testing.(*T).Run /usr/local/go/src/testing/testing.go:1493 +0x35f ``` </p></details> </ul> </p></details> <!-- STEPS ERRORS IF ANY --> ### Steps errors [![5](https://img.shields.io/badge/5%20-red)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//pipeline) <details><summary>Expand to view the steps failures</summary> <p> ##### `libbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 6 min 49 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/12938/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 29 min 57 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/13948/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 18 min 7 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/21785/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 18 min 19 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/21987/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `Error signal` <ul> <li>Took 0 min 0 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/22003/log/?start=0">here</a></li> <li>Description: <code>Error "hudson.AbortException: script returned exit code 1"</code></l1> </ul> </p> </details>
1.0
Build 38 for 8.9 with status FAILURE - Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/postgresql/statement - ## :broken_heart: Tests Failed <!-- BUILD BADGES--> > _the below badges are clickable and redirect to their specific view in the CI or DOCS_ [![Pipeline View](https://img.shields.io/badge/pipeline-pipeline%20-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//pipeline) [![Test View](https://img.shields.io/badge/test-test-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//tests) [![Changes](https://img.shields.io/badge/changes-changes-green)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//changes) [![Artifacts](https://img.shields.io/badge/artifacts-artifacts-yellow)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//artifacts) [![preview](https://img.shields.io/badge/docs-preview-yellowgreen)](http://beats_null.docs-preview.app.elstc.co/diff) [![preview](https://img.shields.io/badge/elastic-observability-blue)](https://ci-stats.elastic.co/app/apm/services/beats-ci/transactions/view?rangeFrom=2023-07-25T04:04:41.298Z&rangeTo=2023-07-25T04:24:41.298Z&transactionName=BUILD+Beats%2Fbeats%2F8.9&transactionType=job&latencyAggregationType=avg&traceId=16f6a010777d86eed8228c11a9eb4d91&transactionId=314de9e9fb8c2a3d) <!-- BUILD SUMMARY--> <details><summary>Expand to view the summary</summary> <p> #### Build stats * Start Time: 2023-07-25T04:14:41.298+0000 * Duration: 95 min 57 sec #### Test stats :test_tube: | Test | Results | | ------------ | :-----------------------------: | | Failed | 1 | | Passed | 30141 | | Skipped | 2356 | | Total | 32498 | </p> </details> <!-- TEST RESULTS IF ANY--> ### Test errors [![1](https://img.shields.io/badge/1%20-red)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//tests) <details><summary>Expand to view the tests failures</summary><p> ##### `Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/postgresql/statement` <ul> <details><summary>Expand to view the error details</summary><p> ``` Failed ``` </p></details> <details><summary>Expand to view the stacktrace</summary><p> ``` === RUN TestFetch statement_integration_test.go:41: Error Trace: /go/src/github.com/elastic/beats/metricbeat/module/postgresql/statement/statement_integration_test.go:41 Error: Should NOT be empty, but was [] Test: TestFetch --- FAIL: TestFetch (7.79s) panic: runtime error: index out of range [0] with length 0 [recovered] panic: runtime error: index out of range [0] with length 0 goroutine 20 [running]: testing.tRunner.func1.2({0x10ba6a0, 0xc00003c480}) /usr/local/go/src/testing/testing.go:1396 +0x24e testing.tRunner.func1() /usr/local/go/src/testing/testing.go:1399 +0x39f panic({0x10ba6a0, 0xc00003c480}) /usr/local/go/src/runtime/panic.go:884 +0x212 github.com/elastic/beats/v7/metricbeat/module/postgresql/statement.TestFetch(0xc00032c9c0) /go/src/github.com/elastic/beats/metricbeat/module/postgresql/statement/statement_integration_test.go:42 +0xbb7 testing.tRunner(0xc00032c9c0, 0x117fae8) /usr/local/go/src/testing/testing.go:1446 +0x10b created by testing.(*T).Run /usr/local/go/src/testing/testing.go:1493 +0x35f ``` </p></details> </ul> </p></details> <!-- STEPS ERRORS IF ANY --> ### Steps errors [![5](https://img.shields.io/badge/5%20-red)](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.9/detail/8.9/38//pipeline) <details><summary>Expand to view the steps failures</summary> <p> ##### `libbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 6 min 49 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/12938/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 29 min 57 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/13948/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 18 min 7 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/21785/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `metricbeat-goIntegTest - mage goIntegTest` <ul> <li>Took 18 min 19 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/21987/log/?start=0">here</a></li> <li>Description: <code>mage goIntegTest</code></l1> </ul> ##### `Error signal` <ul> <li>Took 0 min 0 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.9/runs/38/steps/22003/log/?start=0">here</a></li> <li>Description: <code>Error "hudson.AbortException: script returned exit code 1"</code></l1> </ul> </p> </details>
non_process
build for with status failure build test metricbeat gointegtest testfetch – github com elastic beats metricbeat module postgresql statement broken heart tests failed the below badges are clickable and redirect to their specific view in the ci or docs expand to view the summary build stats start time duration min sec test stats test tube test results failed passed skipped total test errors expand to view the tests failures build test metricbeat gointegtest testfetch – github com elastic beats metricbeat module postgresql statement expand to view the error details failed expand to view the stacktrace run testfetch statement integration test go error trace go src github com elastic beats metricbeat module postgresql statement statement integration test go error should not be empty but was test testfetch fail testfetch panic runtime error index out of range with length panic runtime error index out of range with length goroutine testing trunner usr local go src testing testing go testing trunner usr local go src testing testing go panic usr local go src runtime panic go github com elastic beats metricbeat module postgresql statement testfetch go src github com elastic beats metricbeat module postgresql statement statement integration test go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go steps errors expand to view the steps failures libbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest error signal took min sec view more details a href description error hudson abortexception script returned exit code
0
83,904
10,342,256,387
IssuesEvent
2019-09-04 05:48:41
govau/design-system-site
https://api.github.com/repos/govau/design-system-site
closed
Add ACT Government logo to "Who's using it" section
documentation
Approval sought and received from relevant ACT Government contact. Brand guidance available via: https://www.act.gov.au/__data/assets/pdf_file/0010/1106110/ACTGov-WHOG-Brand-Guidelines.pdf Logos available via: https://www.act.gov.au/branding/logos Stacked variant recommended. Logo to link to https://www.act.gov.au
1.0
Add ACT Government logo to "Who's using it" section - Approval sought and received from relevant ACT Government contact. Brand guidance available via: https://www.act.gov.au/__data/assets/pdf_file/0010/1106110/ACTGov-WHOG-Brand-Guidelines.pdf Logos available via: https://www.act.gov.au/branding/logos Stacked variant recommended. Logo to link to https://www.act.gov.au
non_process
add act government logo to who s using it section approval sought and received from relevant act government contact brand guidance available via logos available via stacked variant recommended logo to link to
0
22,583
31,810,496,483
IssuesEvent
2023-09-13 16:29:03
nextflow-io/nextflow
https://api.github.com/repos/nextflow-io/nextflow
closed
Resume does not work with "exec:" (i.e. native groovy processes)
lang/processes
## Bug report ### Expected behavior and actual behavior `-resume` should work with `exec: ...` but it does not. ### Steps to reproduce the problem Minimal.nf: ``` process testExec { exec: println() } process testStd { """ """ } workflow { testExec() testStd() } ``` ### Program output ``` > ./nextflow run minimal.nf -resume N E X T F L O W ~ version 23.04.3 Launching `minimal.nf` [exotic_mercator] DSL2 - revision: 1be12e5277 executor > local (2) [ae/5ca330] process > testExec [100%] 1 of 1 ✔ [d9/387d71] process > testStd [100%] 1 of 1 ✔ > ./nextflow run minimal.nf -resume N E X T F L O W ~ version 23.04.3 Launching `minimal.nf` [jolly_easley] DSL2 - revision: 1be12e5277 executor > local (1) [41/bb36e7] process > testExec [100%] 1 of 1 ✔ [d9/387d71] process > testStd [100%] 1 of 1, cached: 1 ✔ ``` Note that the process based on `exec:` did not resume. ### Environment * Nextflow version: 23.04.3 * Java version: openjdk version "17.0.8" 2023-07-18 * Operating system: macOS * Bash version: (use the command zsh 5.9 (x86_64-apple-darwin22.0)) ### Additional comments This is not just a corner case, it seems native groovy blocks never get cached. Thank you!
1.0
Resume does not work with "exec:" (i.e. native groovy processes) - ## Bug report ### Expected behavior and actual behavior `-resume` should work with `exec: ...` but it does not. ### Steps to reproduce the problem Minimal.nf: ``` process testExec { exec: println() } process testStd { """ """ } workflow { testExec() testStd() } ``` ### Program output ``` > ./nextflow run minimal.nf -resume N E X T F L O W ~ version 23.04.3 Launching `minimal.nf` [exotic_mercator] DSL2 - revision: 1be12e5277 executor > local (2) [ae/5ca330] process > testExec [100%] 1 of 1 ✔ [d9/387d71] process > testStd [100%] 1 of 1 ✔ > ./nextflow run minimal.nf -resume N E X T F L O W ~ version 23.04.3 Launching `minimal.nf` [jolly_easley] DSL2 - revision: 1be12e5277 executor > local (1) [41/bb36e7] process > testExec [100%] 1 of 1 ✔ [d9/387d71] process > testStd [100%] 1 of 1, cached: 1 ✔ ``` Note that the process based on `exec:` did not resume. ### Environment * Nextflow version: 23.04.3 * Java version: openjdk version "17.0.8" 2023-07-18 * Operating system: macOS * Bash version: (use the command zsh 5.9 (x86_64-apple-darwin22.0)) ### Additional comments This is not just a corner case, it seems native groovy blocks never get cached. Thank you!
process
resume does not work with exec i e native groovy processes bug report expected behavior and actual behavior resume should work with exec but it does not steps to reproduce the problem minimal nf process testexec exec println process teststd workflow testexec teststd program output nextflow run minimal nf resume n e x t f l o w version launching minimal nf revision executor local process testexec of ✔ process teststd of ✔ nextflow run minimal nf resume n e x t f l o w version launching minimal nf revision executor local process testexec of ✔ process teststd of cached ✔ note that the process based on exec did not resume environment nextflow version java version openjdk version operating system macos bash version use the command zsh apple additional comments this is not just a corner case it seems native groovy blocks never get cached thank you
1
15,353
19,524,812,515
IssuesEvent
2021-12-30 04:46:18
beyondhb1079/s4us
https://api.github.com/repos/beyondhb1079/s4us
opened
Add full test coverage
process
Multiple tasks involved here: - [ ] Add test coverage analysis - [ ] Add tests for files missing them (multiple issues, this can be sharded/delegated) - [ ] #384 - [ ] #386 - [ ] #387 - [ ] #388 - [ ] #389 - [ ] #390 - [x] #391 - [ ] #392 - [ ] #393 - [ ] #394 - [ ] #395 - [ ] #665 - [ ] #705 - [ ] Add missing test cases for files - [ ] #661 - [ ] Add validation check to ensure test coverage remains high
1.0
Add full test coverage - Multiple tasks involved here: - [ ] Add test coverage analysis - [ ] Add tests for files missing them (multiple issues, this can be sharded/delegated) - [ ] #384 - [ ] #386 - [ ] #387 - [ ] #388 - [ ] #389 - [ ] #390 - [x] #391 - [ ] #392 - [ ] #393 - [ ] #394 - [ ] #395 - [ ] #665 - [ ] #705 - [ ] Add missing test cases for files - [ ] #661 - [ ] Add validation check to ensure test coverage remains high
process
add full test coverage multiple tasks involved here add test coverage analysis add tests for files missing them multiple issues this can be sharded delegated add missing test cases for files add validation check to ensure test coverage remains high
1
279,559
21,164,890,221
IssuesEvent
2022-04-07 12:48:16
tektoncd/resolution
https://api.github.com/repos/tektoncd/resolution
closed
Update docs for every resolver to indicate the `type` it responds to
kind/documentation
### Description Each resolver's docs should clearly indicate the `type` label string that it responds to. Right now we document the `params` that each resolver expects but one oversight is that we don't document the `type`. e.g. `git` for git resolution, `bundles` for bundle resolution, etc...
1.0
Update docs for every resolver to indicate the `type` it responds to - ### Description Each resolver's docs should clearly indicate the `type` label string that it responds to. Right now we document the `params` that each resolver expects but one oversight is that we don't document the `type`. e.g. `git` for git resolution, `bundles` for bundle resolution, etc...
non_process
update docs for every resolver to indicate the type it responds to description each resolver s docs should clearly indicate the type label string that it responds to right now we document the params that each resolver expects but one oversight is that we don t document the type e g git for git resolution bundles for bundle resolution etc
0
2,749
5,659,575,131
IssuesEvent
2017-04-10 13:18:55
nodejs/node
https://api.github.com/repos/nodejs/node
closed
doc: process.nextTick() example question
process question
[This section](https://nodejs.org/dist/latest-v7.x/docs/api/process.html#process_process_nexttick_callback_args) says: > ```js > // WARNING! DO NOT USE! BAD UNSAFE HAZARD! > function maybeSync(arg, cb) { > if (arg) { > cb(); > return; > } > > fs.stat('file', cb); > } > ``` > This API is hazardous because in the following case: > ```js > maybeSync(true, () => { > foo(); > }); > bar(); > ``` > It is not clear whether `foo()` or `bar()` will be called first. For me, if `arg` is `true`, the order is unambiguously `maybeSync() -> cb() -> foo() -> bar()`, i.e. no async at all. What do I miss?
1.0
doc: process.nextTick() example question - [This section](https://nodejs.org/dist/latest-v7.x/docs/api/process.html#process_process_nexttick_callback_args) says: > ```js > // WARNING! DO NOT USE! BAD UNSAFE HAZARD! > function maybeSync(arg, cb) { > if (arg) { > cb(); > return; > } > > fs.stat('file', cb); > } > ``` > This API is hazardous because in the following case: > ```js > maybeSync(true, () => { > foo(); > }); > bar(); > ``` > It is not clear whether `foo()` or `bar()` will be called first. For me, if `arg` is `true`, the order is unambiguously `maybeSync() -> cb() -> foo() -> bar()`, i.e. no async at all. What do I miss?
process
doc process nexttick example question says js warning do not use bad unsafe hazard function maybesync arg cb if arg cb return fs stat file cb this api is hazardous because in the following case js maybesync true foo bar it is not clear whether foo or bar will be called first for me if arg is true the order is unambiguously maybesync cb foo bar i e no async at all what do i miss
1
8,756
11,874,264,207
IssuesEvent
2020-03-26 18:41:39
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
Qemu builder : Post-processor failed: Unknown artifact type, can't build box
post-processor/vagrant question waiting-reply
How i can use post processor of type "vagrant" with qemu artifacts, they seem not supported by my packer configuration. During packer build after qemu builder i use some post-processors to create ovf from a vdmk. [with this tool](https://github.com/EmmanuelKasper/import2vbox/blob/master/import2vbox.pl) Option `keep artifact" : true` seem to doesn't work and during qemu builder the expected qcow2 file doesn't appear in my build directory (just a file without extension "buster" is created) - I'm on mac OS X Catalina 10.15.4 - I'm trying to build a debian buster 10.0.3. - vagrant post processor work with virtualbox-ovf - packer 1.5.4 Thanks for your help. Command: `packer build -parallel=false -force config.json` The error : ``` ==> qemu: Running post-processor: {{.Provider}}.box (type vagrant) Build 'qemu' errored: 1 error(s) occurred: * Post-processor failed: Unknown artifact type, can't build box: Some builds didn't complete successfully and had errors: Mar 25 17:01:45 --> qemu: 1 error(s) occurred: * Post-processor failed: Unknown artifact type, can't build box: ==> Builds finished but no artifacts were created. ``` My config.json ``` { "builders": [{ "boot_command": [ "<esc><wait>", "install <wait>", "preseed/url=http://{{ .HTTPIP }}:{{ .HTTPPort }}/buster-preseed.cfg <wait>", "auto <wait>", "locale=en_US.UTF-8 <wait>", "netcfg/get_hostname={{ .Name }} <wait>", "netcfg/get_domain=vagrantup.com <wait>", "debconf/frontend=noninteractive <wait>", "console-setup/ask_detect=false <wait>", "kbd-chooser/method=us <wait>", "console-keymaps-at/keymap=us <wait>", "keyboard-configuration/xkb-keymap=us <wait>", "net.ifnames=0 <wait>", "<enter><wait>" ], "disk_size": "20280", "disk_interface": "virtio", "headless": true, "http_directory": "http", "iso_checksum": "{{user `iso_checksum`}}", "iso_checksum_type": "sha256", "iso_url": "https://cdimage.debian.org/debian-cd/current/amd64/iso-cd/debian-10.3.0-amd64-netinst.iso", "type": "qemu", "format": "qcow2", "accelerator": "hvf", "qemuargs": [ ["-m", "1024M"], ["-smp", "2"] ], "ssh_password": "vagrant", "ssh_username": "vagrant", "ssh_port": 22, "ssh_wait_timeout": "10000s", "shutdown_command": "echo 'vagrant' | sudo /sbin/halt -p", "output_directory": "{{user `output_dir`}}", "vm_name": "{{user `vm_name`}}" }, { "boot_command": [ "<up><wait><tab> text ks=http://{{ .HTTPIP }}:{{ .HTTPPort }}/ks.cfg PACKER_USER=vagrant PACKER_AUTHORIZED_KEY={{ .SSHPublicKey | urlquery }}<enter>" ], "headless": true, "checksum": "{{user `iso_checksum`}}", "checksum_type": "sha256", "type": "virtualbox-ovf", "format": "ovf", "source_path": "{{user `vm_name`}}.ovf", "guest_additions_mode": "disable", "virtualbox_version_file": "", "communicator": "none", "output_directory": "{{user `output_dir`}}", "shutdown_command": "", "shutdown_timeout": "1s", "vm_name": "{{user `vm_name`}}" } ], "provisioners": [{ "type": "file", "source": "res", "destination": "/tmp/res", "only": ["qemu"] }, { "type": "shell", "pause_before": "5s", "execute_command": "echo 'vagrant'| {{.Vars}} sudo --preserve-env --stdin sh '{{.Path}}'", "environment_vars": [ "VAGRANT_BUILDER_FS=/" ], "scripts": [ "../helpers/vagrant-setup", "scripts/vagrant.sh", "scripts/temp_install.sh", "scripts/networking.sh", "scripts/cleanup.sh", "scripts/minimize.sh" ], "only": ["qemu"] } ], "post-processors": [ [{ "type": "shell-local", "environment_vars": [ "VM_NAME={{user `vm_name`}}", "OUTPUT={{user `output_dir`}}" ], "script": "scripts/to-ovf-init.sh", "keep_input_artifact": true, "only": ["qemu"] }, { "type": "shell-local", "environment_vars": [ "VM_NAME={{user `vm_name`}}", "OUTPUT={{user `output_dir`}}", "WORK_PATH=./" ], "script": "scripts/to-ovf-linux.sh", "only_on": "linux", "keep_input_artifact": true, "only": ["qemu"] }, { "type": "shell-local", "environment_vars": [ "VM_NAME={{user `vm_name`}}", "OUTPUT={{user `output_dir`}}" ], "script": "scripts/to-ovf-darwin.sh", "only_on": "darwin", "keep_input_artifact": true, "only": ["qemu"] }, { "name": "{{.Provider}}.box", "output": "{{user `vm_name`}}.exports/{{.Provider}}.{{user `vm_name`}}.box", "type": "vagrant", "keep_input_artifact": true }, { "type": "vagrant-cloud", "box_tag": "loic-roux-404/deb64-buster", "access_token": "{{user `cloud_token`}}", "version": "{{user `box_version`}}", "keep_input_artifact": true } ] ], {variables...}} ```
1.0
Qemu builder : Post-processor failed: Unknown artifact type, can't build box - How i can use post processor of type "vagrant" with qemu artifacts, they seem not supported by my packer configuration. During packer build after qemu builder i use some post-processors to create ovf from a vdmk. [with this tool](https://github.com/EmmanuelKasper/import2vbox/blob/master/import2vbox.pl) Option `keep artifact" : true` seem to doesn't work and during qemu builder the expected qcow2 file doesn't appear in my build directory (just a file without extension "buster" is created) - I'm on mac OS X Catalina 10.15.4 - I'm trying to build a debian buster 10.0.3. - vagrant post processor work with virtualbox-ovf - packer 1.5.4 Thanks for your help. Command: `packer build -parallel=false -force config.json` The error : ``` ==> qemu: Running post-processor: {{.Provider}}.box (type vagrant) Build 'qemu' errored: 1 error(s) occurred: * Post-processor failed: Unknown artifact type, can't build box: Some builds didn't complete successfully and had errors: Mar 25 17:01:45 --> qemu: 1 error(s) occurred: * Post-processor failed: Unknown artifact type, can't build box: ==> Builds finished but no artifacts were created. ``` My config.json ``` { "builders": [{ "boot_command": [ "<esc><wait>", "install <wait>", "preseed/url=http://{{ .HTTPIP }}:{{ .HTTPPort }}/buster-preseed.cfg <wait>", "auto <wait>", "locale=en_US.UTF-8 <wait>", "netcfg/get_hostname={{ .Name }} <wait>", "netcfg/get_domain=vagrantup.com <wait>", "debconf/frontend=noninteractive <wait>", "console-setup/ask_detect=false <wait>", "kbd-chooser/method=us <wait>", "console-keymaps-at/keymap=us <wait>", "keyboard-configuration/xkb-keymap=us <wait>", "net.ifnames=0 <wait>", "<enter><wait>" ], "disk_size": "20280", "disk_interface": "virtio", "headless": true, "http_directory": "http", "iso_checksum": "{{user `iso_checksum`}}", "iso_checksum_type": "sha256", "iso_url": "https://cdimage.debian.org/debian-cd/current/amd64/iso-cd/debian-10.3.0-amd64-netinst.iso", "type": "qemu", "format": "qcow2", "accelerator": "hvf", "qemuargs": [ ["-m", "1024M"], ["-smp", "2"] ], "ssh_password": "vagrant", "ssh_username": "vagrant", "ssh_port": 22, "ssh_wait_timeout": "10000s", "shutdown_command": "echo 'vagrant' | sudo /sbin/halt -p", "output_directory": "{{user `output_dir`}}", "vm_name": "{{user `vm_name`}}" }, { "boot_command": [ "<up><wait><tab> text ks=http://{{ .HTTPIP }}:{{ .HTTPPort }}/ks.cfg PACKER_USER=vagrant PACKER_AUTHORIZED_KEY={{ .SSHPublicKey | urlquery }}<enter>" ], "headless": true, "checksum": "{{user `iso_checksum`}}", "checksum_type": "sha256", "type": "virtualbox-ovf", "format": "ovf", "source_path": "{{user `vm_name`}}.ovf", "guest_additions_mode": "disable", "virtualbox_version_file": "", "communicator": "none", "output_directory": "{{user `output_dir`}}", "shutdown_command": "", "shutdown_timeout": "1s", "vm_name": "{{user `vm_name`}}" } ], "provisioners": [{ "type": "file", "source": "res", "destination": "/tmp/res", "only": ["qemu"] }, { "type": "shell", "pause_before": "5s", "execute_command": "echo 'vagrant'| {{.Vars}} sudo --preserve-env --stdin sh '{{.Path}}'", "environment_vars": [ "VAGRANT_BUILDER_FS=/" ], "scripts": [ "../helpers/vagrant-setup", "scripts/vagrant.sh", "scripts/temp_install.sh", "scripts/networking.sh", "scripts/cleanup.sh", "scripts/minimize.sh" ], "only": ["qemu"] } ], "post-processors": [ [{ "type": "shell-local", "environment_vars": [ "VM_NAME={{user `vm_name`}}", "OUTPUT={{user `output_dir`}}" ], "script": "scripts/to-ovf-init.sh", "keep_input_artifact": true, "only": ["qemu"] }, { "type": "shell-local", "environment_vars": [ "VM_NAME={{user `vm_name`}}", "OUTPUT={{user `output_dir`}}", "WORK_PATH=./" ], "script": "scripts/to-ovf-linux.sh", "only_on": "linux", "keep_input_artifact": true, "only": ["qemu"] }, { "type": "shell-local", "environment_vars": [ "VM_NAME={{user `vm_name`}}", "OUTPUT={{user `output_dir`}}" ], "script": "scripts/to-ovf-darwin.sh", "only_on": "darwin", "keep_input_artifact": true, "only": ["qemu"] }, { "name": "{{.Provider}}.box", "output": "{{user `vm_name`}}.exports/{{.Provider}}.{{user `vm_name`}}.box", "type": "vagrant", "keep_input_artifact": true }, { "type": "vagrant-cloud", "box_tag": "loic-roux-404/deb64-buster", "access_token": "{{user `cloud_token`}}", "version": "{{user `box_version`}}", "keep_input_artifact": true } ] ], {variables...}} ```
process
qemu builder post processor failed unknown artifact type can t build box how i can use post processor of type vagrant with qemu artifacts they seem not supported by my packer configuration during packer build after qemu builder i use some post processors to create ovf from a vdmk option keep artifact true seem to doesn t work and during qemu builder the expected file doesn t appear in my build directory just a file without extension buster is created i m on mac os x catalina i m trying to build a debian buster vagrant post processor work with virtualbox ovf packer thanks for your help command packer build parallel false force config json the error qemu running post processor provider box type vagrant build qemu errored error s occurred post processor failed unknown artifact type can t build box some builds didn t complete successfully and had errors mar qemu error s occurred post processor failed unknown artifact type can t build box builds finished but no artifacts were created my config json builders boot command install preseed url httpip httpport buster preseed cfg auto locale en us utf netcfg get hostname name netcfg get domain vagrantup com debconf frontend noninteractive console setup ask detect false kbd chooser method us console keymaps at keymap us keyboard configuration xkb keymap us net ifnames disk size disk interface virtio headless true http directory http iso checksum user iso checksum iso checksum type iso url type qemu format accelerator hvf qemuargs ssh password vagrant ssh username vagrant ssh port ssh wait timeout shutdown command echo vagrant sudo sbin halt p output directory user output dir vm name user vm name boot command text ks httpip httpport ks cfg packer user vagrant packer authorized key sshpublickey urlquery headless true checksum user iso checksum checksum type type virtualbox ovf format ovf source path user vm name ovf guest additions mode disable virtualbox version file communicator none output directory user output dir shutdown command shutdown timeout vm name user vm name provisioners type file source res destination tmp res only type shell pause before execute command echo vagrant vars sudo preserve env stdin sh path environment vars vagrant builder fs scripts helpers vagrant setup scripts vagrant sh scripts temp install sh scripts networking sh scripts cleanup sh scripts minimize sh only post processors type shell local environment vars vm name user vm name output user output dir script scripts to ovf init sh keep input artifact true only type shell local environment vars vm name user vm name output user output dir work path script scripts to ovf linux sh only on linux keep input artifact true only type shell local environment vars vm name user vm name output user output dir script scripts to ovf darwin sh only on darwin keep input artifact true only name provider box output user vm name exports provider user vm name box type vagrant keep input artifact true type vagrant cloud box tag loic roux buster access token user cloud token version user box version keep input artifact true variables
1
1,305
3,857,511,881
IssuesEvent
2016-04-07 06:33:00
PlagueHO/LabBuilder
https://api.github.com/repos/PlagueHO/LabBuilder
closed
Always add DSC Package to any Nano Server VMs
bug In Process
To support DSC Nano Servers must have the DSC package added by default.
1.0
Always add DSC Package to any Nano Server VMs - To support DSC Nano Servers must have the DSC package added by default.
process
always add dsc package to any nano server vms to support dsc nano servers must have the dsc package added by default
1
16,394
21,163,487,463
IssuesEvent
2022-04-07 11:34:22
paul-buerkner/brms
https://api.github.com/repos/paul-buerkner/brms
closed
support for the new ppd plots from Bayesplot
feature post-processing
`bayesplot` has [prior predictive density plots](https://mc-stan.org/bayesplot/reference/PPD-distributions.html), it's basically the same as a post pred check but without plotting the original data. It would be nice to have support for this as well.
1.0
support for the new ppd plots from Bayesplot - `bayesplot` has [prior predictive density plots](https://mc-stan.org/bayesplot/reference/PPD-distributions.html), it's basically the same as a post pred check but without plotting the original data. It would be nice to have support for this as well.
process
support for the new ppd plots from bayesplot bayesplot has it s basically the same as a post pred check but without plotting the original data it would be nice to have support for this as well
1
373,802
26,084,918,922
IssuesEvent
2022-12-26 00:34:51
codeigniter4/shield
https://api.github.com/repos/codeigniter4/shield
closed
preg_match(): Passing null to parameter #2 ($subject) of type string is deprecated
documentation
Does this application not support using PHP 8.1? I just installed Shield and tried to use Email2Fa, and I got this error message. The description says that the error is in SYSTEMPATH\Email\Email.php at line 466. ![ufY3l6](https://user-images.githubusercontent.com/9530214/209454604-6a67ae0a-9485-452b-add4-0571604a62fe.png) This a screen shoot : https://im.ge/i/ufY3l6
1.0
preg_match(): Passing null to parameter #2 ($subject) of type string is deprecated - Does this application not support using PHP 8.1? I just installed Shield and tried to use Email2Fa, and I got this error message. The description says that the error is in SYSTEMPATH\Email\Email.php at line 466. ![ufY3l6](https://user-images.githubusercontent.com/9530214/209454604-6a67ae0a-9485-452b-add4-0571604a62fe.png) This a screen shoot : https://im.ge/i/ufY3l6
non_process
preg match passing null to parameter subject of type string is deprecated does this application not support using php i just installed shield and tried to use and i got this error message the description says that the error is in systempath email email php at line this a screen shoot
0
19,456
25,737,162,077
IssuesEvent
2022-12-08 02:00:07
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 8 Dec 22
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events ### Unsupervised Flood Detection on SAR Time Series - **Authors:** Ritu Yadav, Andrea Nascetti, Hossein Azizpour, Yifang Ban - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.03675 - **Pdf link:** https://arxiv.org/pdf/2212.03675 - **Abstract** Human civilization has an increasingly powerful influence on the earth system. Affected by climate change and land-use change, natural disasters such as flooding have been increasing in recent years. Earth observations are an invaluable source for assessing and mitigating negative impacts. Detecting changes from Earth observation data is one way to monitor the possible impact. Effective and reliable Change Detection (CD) methods can help in identifying the risk of disaster events at an early stage. In this work, we propose a novel unsupervised CD method on time series Synthetic Aperture Radar~(SAR) data. Our proposed method is a probabilistic model trained with unsupervised learning techniques, reconstruction, and contrastive learning. The change map is generated with the help of the distribution difference between pre-incident and post-incident data. Our proposed CD model is evaluated on flood detection data. We verified the efficacy of our model on 8 different flood sites, including three recent flood events from Copernicus Emergency Management Services and six from the Sen1Floods11 dataset. Our proposed model achieved an average of 64.53\% Intersection Over Union(IoU) value and 75.43\% F1 score. Our achieved IoU score is approximately 6-27\% and F1 score is approximately 7-22\% better than the compared unsupervised and supervised existing CD methods. The results and extensive discussion presented in the study show the effectiveness of the proposed unsupervised CD method. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Face Forgery Detection Based on Facial Region Displacement Trajectory Series - **Authors:** YuYang Sun, ZhiYong Zhang, Isao Echizen, Huy H.Nguyen, ChangZhen Qiu, Lu Sun - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2212.03678 - **Pdf link:** https://arxiv.org/pdf/2212.03678 - **Abstract** Deep-learning-based technologies such as deepfakes ones have been attracting widespread attention in both society and academia, particularly ones used to synthesize forged face images. These automatic and professional-skill-free face manipulation technologies can be used to replace the face in an original image or video with any target object while maintaining the expression and demeanor. Since human faces are closely related to identity characteristics, maliciously disseminated identity manipulated videos could trigger a crisis of public trust in the media and could even have serious political, social, and legal implications. To effectively detect manipulated videos, we focus on the position offset in the face blending process, resulting from the forced affine transformation of the normalized forged face. We introduce a method for detecting manipulated videos that is based on the trajectory of the facial region displacement. Specifically, we develop a virtual-anchor-based method for extracting the facial trajectory, which can robustly represent displacement information. This information was used to construct a network for exposing multidimensional artifacts in the trajectory sequences of manipulated videos that is based on dual-stream spatial-temporal graph attention and a gated recurrent unit backbone. Testing of our method on various manipulation datasets demonstrated that its accuracy and generalization ability is competitive with that of the leading detection methods. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### ABN: Anti-Blur Neural Networks for Multi-Stage Deformable Image Registration - **Authors:** Yao Su, Xin Dai, Lifang He, Xiangnan Kong - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.03277 - **Pdf link:** https://arxiv.org/pdf/2212.03277 - **Abstract** Deformable image registration, i.e., the task of aligning multiple images into one coordinate system by non-linear transformation, serves as an essential preprocessing step for neuroimaging data. Recent research on deformable image registration is mainly focused on improving the registration accuracy using multi-stage alignment methods, where the source image is repeatedly deformed in stages by a same neural network until it is well-aligned with the target image. Conventional methods for multi-stage registration can often blur the source image as the pixel/voxel values are repeatedly interpolated from the image generated by the previous stage. However, maintaining image quality such as sharpness during image registration is crucial to medical data analysis. In this paper, we study the problem of anti-blur deformable image registration and propose a novel solution, called Anti-Blur Network (ABN), for multi-stage image registration. Specifically, we use a pair of short-term registration and long-term memory networks to learn the nonlinear deformations at each stage, where the short-term registration network learns how to improve the registration accuracy incrementally and the long-term memory network combines all the previous deformations to allow an interpolation to perform on the raw image directly and preserve image sharpness. Extensive experiments on both natural and medical image datasets demonstrated that ABN can accurately register images while preserving their sharpness. Our code and data can be found at https://github.com/anonymous3214/ABN ### DroneAttention: Sparse Weighted Temporal Attention for Drone-Camera Based Activity Recognition - **Authors:** Santosh Kumar Yadav, Achleshwar Luthra, Esha Pahwa, Kamlesh Tiwari, Heena Rathore, Hari Mohan Pandey, Peter Corcoran - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.03384 - **Pdf link:** https://arxiv.org/pdf/2212.03384 - **Abstract** Human activity recognition (HAR) using drone-mounted cameras has attracted considerable interest from the computer vision research community in recent years. A robust and efficient HAR system has a pivotal role in fields like video surveillance, crowd behavior analysis, sports analysis, and human-computer interaction. What makes it challenging are the complex poses, understanding different viewpoints, and the environmental scenarios where the action is taking place. To address such complexities, in this paper, we propose a novel Sparse Weighted Temporal Attention (SWTA) module to utilize sparsely sampled video frames for obtaining global weighted temporal attention. The proposed SWTA is comprised of two parts. First, temporal segment network that sparsely samples a given set of frames. Second, weighted temporal attention, which incorporates a fusion of attention maps derived from optical flow, with raw RGB images. This is followed by a basenet network, which comprises a convolutional neural network (CNN) module along with fully connected layers that provide us with activity recognition. The SWTA network can be used as a plug-in module to the existing deep CNN architectures, for optimizing them to learn temporal information by eliminating the need for a separate temporal stream. It has been evaluated on three publicly available benchmark datasets, namely Okutama, MOD20, and Drone-Action. The proposed model has received an accuracy of 72.76%, 92.56%, and 78.86% on the respective datasets thereby surpassing the previous state-of-the-art performances by a margin of 25.26%, 18.56%, and 2.94%, respectively. ### iQuery: Instruments as Queries for Audio-Visual Sound Separation - **Authors:** Jiaben Chen, Renrui Zhang, Dongze Lian, Jiaqi Yang, Ziyao Zeng, Jianbo Shi - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2212.03814 - **Pdf link:** https://arxiv.org/pdf/2212.03814 - **Abstract** Current audio-visual separation methods share a standard architecture design where an audio encoder-decoder network is fused with visual encoding features at the encoder bottleneck. This design confounds the learning of multi-modal feature encoding with robust sound decoding for audio separation. To generalize to a new instrument: one must finetune the entire visual and audio network for all musical instruments. We re-formulate visual-sound separation task and propose Instrument as Query (iQuery) with a flexible query expansion mechanism. Our approach ensures cross-modal consistency and cross-instrument disentanglement. We utilize "visually named" queries to initiate the learning of audio queries and use cross-modal attention to remove potential sound source interference at the estimated waveforms. To generalize to a new instrument or event class, drawing inspiration from the text-prompt design, we insert an additional query as an audio prompt while freezing the attention mechanism. Experimental results on three benchmarks demonstrate that our iQuery improves audio-visual sound source separation performance. ### X-Paste: Revisit Copy-Paste at Scale with CLIP and StableDiffusion - **Authors:** Hanqing Zhao, Dianmo Sheng, Jianmin Bao, Dongdong Chen, Dong Chen, Fang Wen, Lu Yuan, Ce Liu, Wenbo Zhou, Qi Chu, Weiming Zhang, Nenghai Yu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2212.03863 - **Pdf link:** https://arxiv.org/pdf/2212.03863 - **Abstract** Copy-Paste is a simple and effective data augmentation strategy for instance segmentation. By randomly pasting object instances onto new background images, it creates new training data for free and significantly boosts the segmentation performance, especially for rare object categories. Although diverse, high-quality object instances used in Copy-Paste result in more performance gain, previous works utilize object instances either from human-annotated instance segmentation datasets or rendered from 3D object models, and both approaches are too expensive to scale up to obtain good diversity. In this paper, we revisit Copy-Paste at scale with the power of newly emerged zero-shot recognition models (e.g., CLIP) and text2image models (e.g., StableDiffusion). We demonstrate for the first time that using a text2image model to generate images or zero-shot recognition model to filter noisily crawled images for different object categories is a feasible way to make Copy-Paste truly scalable. To make such success happen, we design a data acquisition and processing framework, dubbed "X-Paste", upon which a systematic study is conducted. On the LVIS dataset, X-Paste provides impressive improvements over the strong baseline CenterNet2 with Swin-L as the backbone. Specifically, it archives +2.6 box AP and +2.1 mask AP gains on all classes and even more significant gains with +6.8 box AP +6.5 mask AP on long-tail classes. ## Keyword: raw image ### ABN: Anti-Blur Neural Networks for Multi-Stage Deformable Image Registration - **Authors:** Yao Su, Xin Dai, Lifang He, Xiangnan Kong - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.03277 - **Pdf link:** https://arxiv.org/pdf/2212.03277 - **Abstract** Deformable image registration, i.e., the task of aligning multiple images into one coordinate system by non-linear transformation, serves as an essential preprocessing step for neuroimaging data. Recent research on deformable image registration is mainly focused on improving the registration accuracy using multi-stage alignment methods, where the source image is repeatedly deformed in stages by a same neural network until it is well-aligned with the target image. Conventional methods for multi-stage registration can often blur the source image as the pixel/voxel values are repeatedly interpolated from the image generated by the previous stage. However, maintaining image quality such as sharpness during image registration is crucial to medical data analysis. In this paper, we study the problem of anti-blur deformable image registration and propose a novel solution, called Anti-Blur Network (ABN), for multi-stage image registration. Specifically, we use a pair of short-term registration and long-term memory networks to learn the nonlinear deformations at each stage, where the short-term registration network learns how to improve the registration accuracy incrementally and the long-term memory network combines all the previous deformations to allow an interpolation to perform on the raw image directly and preserve image sharpness. Extensive experiments on both natural and medical image datasets demonstrated that ABN can accurately register images while preserving their sharpness. Our code and data can be found at https://github.com/anonymous3214/ABN
2.0
New submissions for Thu, 8 Dec 22 - ## Keyword: events ### Unsupervised Flood Detection on SAR Time Series - **Authors:** Ritu Yadav, Andrea Nascetti, Hossein Azizpour, Yifang Ban - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.03675 - **Pdf link:** https://arxiv.org/pdf/2212.03675 - **Abstract** Human civilization has an increasingly powerful influence on the earth system. Affected by climate change and land-use change, natural disasters such as flooding have been increasing in recent years. Earth observations are an invaluable source for assessing and mitigating negative impacts. Detecting changes from Earth observation data is one way to monitor the possible impact. Effective and reliable Change Detection (CD) methods can help in identifying the risk of disaster events at an early stage. In this work, we propose a novel unsupervised CD method on time series Synthetic Aperture Radar~(SAR) data. Our proposed method is a probabilistic model trained with unsupervised learning techniques, reconstruction, and contrastive learning. The change map is generated with the help of the distribution difference between pre-incident and post-incident data. Our proposed CD model is evaluated on flood detection data. We verified the efficacy of our model on 8 different flood sites, including three recent flood events from Copernicus Emergency Management Services and six from the Sen1Floods11 dataset. Our proposed model achieved an average of 64.53\% Intersection Over Union(IoU) value and 75.43\% F1 score. Our achieved IoU score is approximately 6-27\% and F1 score is approximately 7-22\% better than the compared unsupervised and supervised existing CD methods. The results and extensive discussion presented in the study show the effectiveness of the proposed unsupervised CD method. ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB There is no result ## Keyword: ISP ### Face Forgery Detection Based on Facial Region Displacement Trajectory Series - **Authors:** YuYang Sun, ZhiYong Zhang, Isao Echizen, Huy H.Nguyen, ChangZhen Qiu, Lu Sun - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI) - **Arxiv link:** https://arxiv.org/abs/2212.03678 - **Pdf link:** https://arxiv.org/pdf/2212.03678 - **Abstract** Deep-learning-based technologies such as deepfakes ones have been attracting widespread attention in both society and academia, particularly ones used to synthesize forged face images. These automatic and professional-skill-free face manipulation technologies can be used to replace the face in an original image or video with any target object while maintaining the expression and demeanor. Since human faces are closely related to identity characteristics, maliciously disseminated identity manipulated videos could trigger a crisis of public trust in the media and could even have serious political, social, and legal implications. To effectively detect manipulated videos, we focus on the position offset in the face blending process, resulting from the forced affine transformation of the normalized forged face. We introduce a method for detecting manipulated videos that is based on the trajectory of the facial region displacement. Specifically, we develop a virtual-anchor-based method for extracting the facial trajectory, which can robustly represent displacement information. This information was used to construct a network for exposing multidimensional artifacts in the trajectory sequences of manipulated videos that is based on dual-stream spatial-temporal graph attention and a gated recurrent unit backbone. Testing of our method on various manipulation datasets demonstrated that its accuracy and generalization ability is competitive with that of the leading detection methods. ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression There is no result ## Keyword: RAW ### ABN: Anti-Blur Neural Networks for Multi-Stage Deformable Image Registration - **Authors:** Yao Su, Xin Dai, Lifang He, Xiangnan Kong - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.03277 - **Pdf link:** https://arxiv.org/pdf/2212.03277 - **Abstract** Deformable image registration, i.e., the task of aligning multiple images into one coordinate system by non-linear transformation, serves as an essential preprocessing step for neuroimaging data. Recent research on deformable image registration is mainly focused on improving the registration accuracy using multi-stage alignment methods, where the source image is repeatedly deformed in stages by a same neural network until it is well-aligned with the target image. Conventional methods for multi-stage registration can often blur the source image as the pixel/voxel values are repeatedly interpolated from the image generated by the previous stage. However, maintaining image quality such as sharpness during image registration is crucial to medical data analysis. In this paper, we study the problem of anti-blur deformable image registration and propose a novel solution, called Anti-Blur Network (ABN), for multi-stage image registration. Specifically, we use a pair of short-term registration and long-term memory networks to learn the nonlinear deformations at each stage, where the short-term registration network learns how to improve the registration accuracy incrementally and the long-term memory network combines all the previous deformations to allow an interpolation to perform on the raw image directly and preserve image sharpness. Extensive experiments on both natural and medical image datasets demonstrated that ABN can accurately register images while preserving their sharpness. Our code and data can be found at https://github.com/anonymous3214/ABN ### DroneAttention: Sparse Weighted Temporal Attention for Drone-Camera Based Activity Recognition - **Authors:** Santosh Kumar Yadav, Achleshwar Luthra, Esha Pahwa, Kamlesh Tiwari, Heena Rathore, Hari Mohan Pandey, Peter Corcoran - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.03384 - **Pdf link:** https://arxiv.org/pdf/2212.03384 - **Abstract** Human activity recognition (HAR) using drone-mounted cameras has attracted considerable interest from the computer vision research community in recent years. A robust and efficient HAR system has a pivotal role in fields like video surveillance, crowd behavior analysis, sports analysis, and human-computer interaction. What makes it challenging are the complex poses, understanding different viewpoints, and the environmental scenarios where the action is taking place. To address such complexities, in this paper, we propose a novel Sparse Weighted Temporal Attention (SWTA) module to utilize sparsely sampled video frames for obtaining global weighted temporal attention. The proposed SWTA is comprised of two parts. First, temporal segment network that sparsely samples a given set of frames. Second, weighted temporal attention, which incorporates a fusion of attention maps derived from optical flow, with raw RGB images. This is followed by a basenet network, which comprises a convolutional neural network (CNN) module along with fully connected layers that provide us with activity recognition. The SWTA network can be used as a plug-in module to the existing deep CNN architectures, for optimizing them to learn temporal information by eliminating the need for a separate temporal stream. It has been evaluated on three publicly available benchmark datasets, namely Okutama, MOD20, and Drone-Action. The proposed model has received an accuracy of 72.76%, 92.56%, and 78.86% on the respective datasets thereby surpassing the previous state-of-the-art performances by a margin of 25.26%, 18.56%, and 2.94%, respectively. ### iQuery: Instruments as Queries for Audio-Visual Sound Separation - **Authors:** Jiaben Chen, Renrui Zhang, Dongze Lian, Jiaqi Yang, Ziyao Zeng, Jianbo Shi - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Multimedia (cs.MM); Sound (cs.SD); Audio and Speech Processing (eess.AS) - **Arxiv link:** https://arxiv.org/abs/2212.03814 - **Pdf link:** https://arxiv.org/pdf/2212.03814 - **Abstract** Current audio-visual separation methods share a standard architecture design where an audio encoder-decoder network is fused with visual encoding features at the encoder bottleneck. This design confounds the learning of multi-modal feature encoding with robust sound decoding for audio separation. To generalize to a new instrument: one must finetune the entire visual and audio network for all musical instruments. We re-formulate visual-sound separation task and propose Instrument as Query (iQuery) with a flexible query expansion mechanism. Our approach ensures cross-modal consistency and cross-instrument disentanglement. We utilize "visually named" queries to initiate the learning of audio queries and use cross-modal attention to remove potential sound source interference at the estimated waveforms. To generalize to a new instrument or event class, drawing inspiration from the text-prompt design, we insert an additional query as an audio prompt while freezing the attention mechanism. Experimental results on three benchmarks demonstrate that our iQuery improves audio-visual sound source separation performance. ### X-Paste: Revisit Copy-Paste at Scale with CLIP and StableDiffusion - **Authors:** Hanqing Zhao, Dianmo Sheng, Jianmin Bao, Dongdong Chen, Dong Chen, Fang Wen, Lu Yuan, Ce Liu, Wenbo Zhou, Qi Chu, Weiming Zhang, Nenghai Yu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2212.03863 - **Pdf link:** https://arxiv.org/pdf/2212.03863 - **Abstract** Copy-Paste is a simple and effective data augmentation strategy for instance segmentation. By randomly pasting object instances onto new background images, it creates new training data for free and significantly boosts the segmentation performance, especially for rare object categories. Although diverse, high-quality object instances used in Copy-Paste result in more performance gain, previous works utilize object instances either from human-annotated instance segmentation datasets or rendered from 3D object models, and both approaches are too expensive to scale up to obtain good diversity. In this paper, we revisit Copy-Paste at scale with the power of newly emerged zero-shot recognition models (e.g., CLIP) and text2image models (e.g., StableDiffusion). We demonstrate for the first time that using a text2image model to generate images or zero-shot recognition model to filter noisily crawled images for different object categories is a feasible way to make Copy-Paste truly scalable. To make such success happen, we design a data acquisition and processing framework, dubbed "X-Paste", upon which a systematic study is conducted. On the LVIS dataset, X-Paste provides impressive improvements over the strong baseline CenterNet2 with Swin-L as the backbone. Specifically, it archives +2.6 box AP and +2.1 mask AP gains on all classes and even more significant gains with +6.8 box AP +6.5 mask AP on long-tail classes. ## Keyword: raw image ### ABN: Anti-Blur Neural Networks for Multi-Stage Deformable Image Registration - **Authors:** Yao Su, Xin Dai, Lifang He, Xiangnan Kong - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2212.03277 - **Pdf link:** https://arxiv.org/pdf/2212.03277 - **Abstract** Deformable image registration, i.e., the task of aligning multiple images into one coordinate system by non-linear transformation, serves as an essential preprocessing step for neuroimaging data. Recent research on deformable image registration is mainly focused on improving the registration accuracy using multi-stage alignment methods, where the source image is repeatedly deformed in stages by a same neural network until it is well-aligned with the target image. Conventional methods for multi-stage registration can often blur the source image as the pixel/voxel values are repeatedly interpolated from the image generated by the previous stage. However, maintaining image quality such as sharpness during image registration is crucial to medical data analysis. In this paper, we study the problem of anti-blur deformable image registration and propose a novel solution, called Anti-Blur Network (ABN), for multi-stage image registration. Specifically, we use a pair of short-term registration and long-term memory networks to learn the nonlinear deformations at each stage, where the short-term registration network learns how to improve the registration accuracy incrementally and the long-term memory network combines all the previous deformations to allow an interpolation to perform on the raw image directly and preserve image sharpness. Extensive experiments on both natural and medical image datasets demonstrated that ABN can accurately register images while preserving their sharpness. Our code and data can be found at https://github.com/anonymous3214/ABN
process
new submissions for thu dec keyword events unsupervised flood detection on sar time series authors ritu yadav andrea nascetti hossein azizpour yifang ban subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract human civilization has an increasingly powerful influence on the earth system affected by climate change and land use change natural disasters such as flooding have been increasing in recent years earth observations are an invaluable source for assessing and mitigating negative impacts detecting changes from earth observation data is one way to monitor the possible impact effective and reliable change detection cd methods can help in identifying the risk of disaster events at an early stage in this work we propose a novel unsupervised cd method on time series synthetic aperture radar sar data our proposed method is a probabilistic model trained with unsupervised learning techniques reconstruction and contrastive learning the change map is generated with the help of the distribution difference between pre incident and post incident data our proposed cd model is evaluated on flood detection data we verified the efficacy of our model on different flood sites including three recent flood events from copernicus emergency management services and six from the dataset our proposed model achieved an average of intersection over union iou value and score our achieved iou score is approximately and score is approximately better than the compared unsupervised and supervised existing cd methods the results and extensive discussion presented in the study show the effectiveness of the proposed unsupervised cd method keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp face forgery detection based on facial region displacement trajectory series authors yuyang sun zhiyong zhang isao echizen huy h nguyen changzhen qiu lu sun subjects computer vision and pattern recognition cs cv artificial intelligence cs ai arxiv link pdf link abstract deep learning based technologies such as deepfakes ones have been attracting widespread attention in both society and academia particularly ones used to synthesize forged face images these automatic and professional skill free face manipulation technologies can be used to replace the face in an original image or video with any target object while maintaining the expression and demeanor since human faces are closely related to identity characteristics maliciously disseminated identity manipulated videos could trigger a crisis of public trust in the media and could even have serious political social and legal implications to effectively detect manipulated videos we focus on the position offset in the face blending process resulting from the forced affine transformation of the normalized forged face we introduce a method for detecting manipulated videos that is based on the trajectory of the facial region displacement specifically we develop a virtual anchor based method for extracting the facial trajectory which can robustly represent displacement information this information was used to construct a network for exposing multidimensional artifacts in the trajectory sequences of manipulated videos that is based on dual stream spatial temporal graph attention and a gated recurrent unit backbone testing of our method on various manipulation datasets demonstrated that its accuracy and generalization ability is competitive with that of the leading detection methods keyword image signal processing there is no result keyword image signal process there is no result keyword compression there is no result keyword raw abn anti blur neural networks for multi stage deformable image registration authors yao su xin dai lifang he xiangnan kong subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deformable image registration i e the task of aligning multiple images into one coordinate system by non linear transformation serves as an essential preprocessing step for neuroimaging data recent research on deformable image registration is mainly focused on improving the registration accuracy using multi stage alignment methods where the source image is repeatedly deformed in stages by a same neural network until it is well aligned with the target image conventional methods for multi stage registration can often blur the source image as the pixel voxel values are repeatedly interpolated from the image generated by the previous stage however maintaining image quality such as sharpness during image registration is crucial to medical data analysis in this paper we study the problem of anti blur deformable image registration and propose a novel solution called anti blur network abn for multi stage image registration specifically we use a pair of short term registration and long term memory networks to learn the nonlinear deformations at each stage where the short term registration network learns how to improve the registration accuracy incrementally and the long term memory network combines all the previous deformations to allow an interpolation to perform on the raw image directly and preserve image sharpness extensive experiments on both natural and medical image datasets demonstrated that abn can accurately register images while preserving their sharpness our code and data can be found at droneattention sparse weighted temporal attention for drone camera based activity recognition authors santosh kumar yadav achleshwar luthra esha pahwa kamlesh tiwari heena rathore hari mohan pandey peter corcoran subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract human activity recognition har using drone mounted cameras has attracted considerable interest from the computer vision research community in recent years a robust and efficient har system has a pivotal role in fields like video surveillance crowd behavior analysis sports analysis and human computer interaction what makes it challenging are the complex poses understanding different viewpoints and the environmental scenarios where the action is taking place to address such complexities in this paper we propose a novel sparse weighted temporal attention swta module to utilize sparsely sampled video frames for obtaining global weighted temporal attention the proposed swta is comprised of two parts first temporal segment network that sparsely samples a given set of frames second weighted temporal attention which incorporates a fusion of attention maps derived from optical flow with raw rgb images this is followed by a basenet network which comprises a convolutional neural network cnn module along with fully connected layers that provide us with activity recognition the swta network can be used as a plug in module to the existing deep cnn architectures for optimizing them to learn temporal information by eliminating the need for a separate temporal stream it has been evaluated on three publicly available benchmark datasets namely okutama and drone action the proposed model has received an accuracy of and on the respective datasets thereby surpassing the previous state of the art performances by a margin of and respectively iquery instruments as queries for audio visual sound separation authors jiaben chen renrui zhang dongze lian jiaqi yang ziyao zeng jianbo shi subjects computer vision and pattern recognition cs cv multimedia cs mm sound cs sd audio and speech processing eess as arxiv link pdf link abstract current audio visual separation methods share a standard architecture design where an audio encoder decoder network is fused with visual encoding features at the encoder bottleneck this design confounds the learning of multi modal feature encoding with robust sound decoding for audio separation to generalize to a new instrument one must finetune the entire visual and audio network for all musical instruments we re formulate visual sound separation task and propose instrument as query iquery with a flexible query expansion mechanism our approach ensures cross modal consistency and cross instrument disentanglement we utilize visually named queries to initiate the learning of audio queries and use cross modal attention to remove potential sound source interference at the estimated waveforms to generalize to a new instrument or event class drawing inspiration from the text prompt design we insert an additional query as an audio prompt while freezing the attention mechanism experimental results on three benchmarks demonstrate that our iquery improves audio visual sound source separation performance x paste revisit copy paste at scale with clip and stablediffusion authors hanqing zhao dianmo sheng jianmin bao dongdong chen dong chen fang wen lu yuan ce liu wenbo zhou qi chu weiming zhang nenghai yu subjects computer vision and pattern recognition cs cv machine learning cs lg arxiv link pdf link abstract copy paste is a simple and effective data augmentation strategy for instance segmentation by randomly pasting object instances onto new background images it creates new training data for free and significantly boosts the segmentation performance especially for rare object categories although diverse high quality object instances used in copy paste result in more performance gain previous works utilize object instances either from human annotated instance segmentation datasets or rendered from object models and both approaches are too expensive to scale up to obtain good diversity in this paper we revisit copy paste at scale with the power of newly emerged zero shot recognition models e g clip and models e g stablediffusion we demonstrate for the first time that using a model to generate images or zero shot recognition model to filter noisily crawled images for different object categories is a feasible way to make copy paste truly scalable to make such success happen we design a data acquisition and processing framework dubbed x paste upon which a systematic study is conducted on the lvis dataset x paste provides impressive improvements over the strong baseline with swin l as the backbone specifically it archives box ap and mask ap gains on all classes and even more significant gains with box ap mask ap on long tail classes keyword raw image abn anti blur neural networks for multi stage deformable image registration authors yao su xin dai lifang he xiangnan kong subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deformable image registration i e the task of aligning multiple images into one coordinate system by non linear transformation serves as an essential preprocessing step for neuroimaging data recent research on deformable image registration is mainly focused on improving the registration accuracy using multi stage alignment methods where the source image is repeatedly deformed in stages by a same neural network until it is well aligned with the target image conventional methods for multi stage registration can often blur the source image as the pixel voxel values are repeatedly interpolated from the image generated by the previous stage however maintaining image quality such as sharpness during image registration is crucial to medical data analysis in this paper we study the problem of anti blur deformable image registration and propose a novel solution called anti blur network abn for multi stage image registration specifically we use a pair of short term registration and long term memory networks to learn the nonlinear deformations at each stage where the short term registration network learns how to improve the registration accuracy incrementally and the long term memory network combines all the previous deformations to allow an interpolation to perform on the raw image directly and preserve image sharpness extensive experiments on both natural and medical image datasets demonstrated that abn can accurately register images while preserving their sharpness our code and data can be found at
1
19,768
26,143,761,107
IssuesEvent
2022-12-29 23:21:31
vectordotdev/vector
https://api.github.com/repos/vectordotdev/vector
closed
New `collaborate` transform (pipelines)
domain: transforms type: feature domain: processing
See the [Pipelines 2.0 brief](https://docs.google.com/document/d/19L5p-kqvROkygDy9t21nV9EOmxgb_DDbsvoV65ixrk0/edit?usp=sharing). The naming for this transform is still TBD. - [x] #9273
1.0
New `collaborate` transform (pipelines) - See the [Pipelines 2.0 brief](https://docs.google.com/document/d/19L5p-kqvROkygDy9t21nV9EOmxgb_DDbsvoV65ixrk0/edit?usp=sharing). The naming for this transform is still TBD. - [x] #9273
process
new collaborate transform pipelines see the the naming for this transform is still tbd
1
51,814
13,211,318,393
IssuesEvent
2020-08-15 22:16:34
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[cmake] policy CMP0054 (Trac #1109)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1109">https://code.icecube.wisc.edu/projects/icecube/ticket/1109</a>, reported by david.schultzand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:10", "_ts": "1458335650323600", "description": "{{{\nCMake Warning (dev) at cmake/config.cmake:245 (if):\n Policy CMP0054 is not set: Only interpret if() arguments as variables or\n keywords when unquoted. Run \"cmake --help-policy CMP0054\" for policy\n details. Use the cmake_policy command to set the policy and suppress this\n warning.\n\n Quoted variables like \"offline-software.trunk\" will no longer be\n dereferenced when the policy is set to NEW. Since the policy is not set\n the OLD behavior will be used.\nCall Stack (most recent call first):\n cmake/toplevel.cmake:63 (include)\n CMakeLists.txt:22 (include)\nThis warning is for project developers. Use -Wno-dev to suppress it.\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "time": "2015-08-11T20:31:09", "component": "cmake", "summary": "[cmake] policy CMP0054", "priority": "major", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
[cmake] policy CMP0054 (Trac #1109) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1109">https://code.icecube.wisc.edu/projects/icecube/ticket/1109</a>, reported by david.schultzand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:10", "_ts": "1458335650323600", "description": "{{{\nCMake Warning (dev) at cmake/config.cmake:245 (if):\n Policy CMP0054 is not set: Only interpret if() arguments as variables or\n keywords when unquoted. Run \"cmake --help-policy CMP0054\" for policy\n details. Use the cmake_policy command to set the policy and suppress this\n warning.\n\n Quoted variables like \"offline-software.trunk\" will no longer be\n dereferenced when the policy is set to NEW. Since the policy is not set\n the OLD behavior will be used.\nCall Stack (most recent call first):\n cmake/toplevel.cmake:63 (include)\n CMakeLists.txt:22 (include)\nThis warning is for project developers. Use -Wno-dev to suppress it.\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "time": "2015-08-11T20:31:09", "component": "cmake", "summary": "[cmake] policy CMP0054", "priority": "major", "keywords": "", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
non_process
policy trac migrated from json status closed changetime ts description ncmake warning dev at cmake config cmake if n policy is not set only interpret if arguments as variables or n keywords when unquoted run cmake help policy for policy n details use the cmake policy command to set the policy and suppress this n warning n n quoted variables like offline software trunk will no longer be n dereferenced when the policy is set to new since the policy is not set n the old behavior will be used ncall stack most recent call first n cmake toplevel cmake include n cmakelists txt include nthis warning is for project developers use wno dev to suppress it n reporter david schultz cc resolution fixed time component cmake summary policy priority major keywords milestone owner nega type defect
0
9,348
12,359,948,417
IssuesEvent
2020-05-17 13:16:36
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
Symfony Process - "mv" command
Bug Process Status: Reviewed
**Symfony version(s) affected**: 5.0.7 **Description** Symfony Process doesn't execute `mv` command if there is `/*` at the end of source directory: ``` $process = new Process(['mv', '/var/tmp/test/*', '/var/shared']); $process->start(); foreach ($process as $type => $data) { if ($process::OUT === $type) { dump("Read from stdout: ".$data); } else { // $process::ERR === $type dump("Read from stderr: ".$data); } } ``` **Output:** ``` Read from stderr: mv: can't rename '/var/tmp/test/*': No such file or directory ``` Both directories (source `/var/tmp/test/` and target `/var/shared`) exist. Directory `/var/tmp/test/` contains files and sub-directories. If I try to move the whole `test` directory, it works: ``` $process = new Process(['mv', '/var/tmp/test', '/var/shared']); ``` But it also moves the `test` directory that I don't need. I need to move only the content (files and directories) of `/var/tmp/test` directory. If I call `exec('mv /var/tmp/test/* /var/shared')` instead, it works as expected, but I would like to use the Symfony Process.
1.0
Symfony Process - "mv" command - **Symfony version(s) affected**: 5.0.7 **Description** Symfony Process doesn't execute `mv` command if there is `/*` at the end of source directory: ``` $process = new Process(['mv', '/var/tmp/test/*', '/var/shared']); $process->start(); foreach ($process as $type => $data) { if ($process::OUT === $type) { dump("Read from stdout: ".$data); } else { // $process::ERR === $type dump("Read from stderr: ".$data); } } ``` **Output:** ``` Read from stderr: mv: can't rename '/var/tmp/test/*': No such file or directory ``` Both directories (source `/var/tmp/test/` and target `/var/shared`) exist. Directory `/var/tmp/test/` contains files and sub-directories. If I try to move the whole `test` directory, it works: ``` $process = new Process(['mv', '/var/tmp/test', '/var/shared']); ``` But it also moves the `test` directory that I don't need. I need to move only the content (files and directories) of `/var/tmp/test` directory. If I call `exec('mv /var/tmp/test/* /var/shared')` instead, it works as expected, but I would like to use the Symfony Process.
process
symfony process mv command symfony version s affected description symfony process doesn t execute mv command if there is at the end of source directory process new process process start foreach process as type data if process out type dump read from stdout data else process err type dump read from stderr data output read from stderr mv can t rename var tmp test no such file or directory both directories source var tmp test and target var shared exist directory var tmp test contains files and sub directories if i try to move the whole test directory it works process new process but it also moves the test directory that i don t need i need to move only the content files and directories of var tmp test directory if i call exec mv var tmp test var shared instead it works as expected but i would like to use the symfony process
1
24,830
5,105,730,567
IssuesEvent
2017-01-05 08:53:37
gatling/gatling
https://api.github.com/repos/gatling/gatling
closed
Doc: Re-add snapshots to download page
Documentation Type - Feature
Switched highcharts modules to 3.0.0-SNAPSHOT, available [on Sonatype repo](https://oss.sonatype.org/content/repositories/snapshots/io/gatling/highcharts/gatling-charts-highcharts-bundle/3.0.0-SNAPSHOT/). Add snapshots download section back on [download page](http://gatling.io/#/resources/download).
1.0
Doc: Re-add snapshots to download page - Switched highcharts modules to 3.0.0-SNAPSHOT, available [on Sonatype repo](https://oss.sonatype.org/content/repositories/snapshots/io/gatling/highcharts/gatling-charts-highcharts-bundle/3.0.0-SNAPSHOT/). Add snapshots download section back on [download page](http://gatling.io/#/resources/download).
non_process
doc re add snapshots to download page switched highcharts modules to snapshot available add snapshots download section back on
0
19,143
25,204,739,386
IssuesEvent
2022-11-13 14:46:52
eobermuhlner/big-math
https://api.github.com/repos/eobermuhlner/big-math
closed
Prepare release 2.3.2
development process
- [x] add release number header to release note - [x] rename release note - [x] create empty release note for next release from `template_release_note.md` - [x] change version in `build.gradle` of `big-math` subproject - [x] change version in pom.xml - [x] upload artifacts to maven central - [x] run `./gradlew clean` - [x] run `./gradlew :ch.obermuhlner.math.big:uploadArchives` - [x] go to https://oss.sonatype.org/ - [x] in tab 'Staging Repositories' locate own Repository (typically at the end of the list) - [x] verify content of own Repository (version number!) - [x] `Close` own Repository - [x] `Refresh` until `Release` becomes enabled - [x] `Release` own Repository - [ ] create github release from same artifacts - [x] Create new draft release - [x] Create new tag "v1.2.3" (will be created when published) - [x] Create new title for release e.g. "Release 1.2.3" - [x] Copy content of release note into draft release (without the first header "# Release 1.2.3") - [x] Add artefacts from gradle build to draft release - [x] big-math-*.jar - [x] big-math-*-javadoc.jar - [x] big-math-*-sources.jar - [x] Verify all changes are committed and pushed - [x] Publish release - [x] update readme - [x] add generated javadoc to `docs/javadoc` (copy folder `javadoc` from `build/docs` and renamed to e.g. `v1.2.3`) - [x] update `docs/index.md` - [ ] update dependent projects - [x] create regression project for new release - [x] edit new `build.gradle` to use new release library - [x] edit `settings.gradle` to add new regression project - [x] run `./run_regression_analysis.sh` - [x] commit and push chages
1.0
Prepare release 2.3.2 - - [x] add release number header to release note - [x] rename release note - [x] create empty release note for next release from `template_release_note.md` - [x] change version in `build.gradle` of `big-math` subproject - [x] change version in pom.xml - [x] upload artifacts to maven central - [x] run `./gradlew clean` - [x] run `./gradlew :ch.obermuhlner.math.big:uploadArchives` - [x] go to https://oss.sonatype.org/ - [x] in tab 'Staging Repositories' locate own Repository (typically at the end of the list) - [x] verify content of own Repository (version number!) - [x] `Close` own Repository - [x] `Refresh` until `Release` becomes enabled - [x] `Release` own Repository - [ ] create github release from same artifacts - [x] Create new draft release - [x] Create new tag "v1.2.3" (will be created when published) - [x] Create new title for release e.g. "Release 1.2.3" - [x] Copy content of release note into draft release (without the first header "# Release 1.2.3") - [x] Add artefacts from gradle build to draft release - [x] big-math-*.jar - [x] big-math-*-javadoc.jar - [x] big-math-*-sources.jar - [x] Verify all changes are committed and pushed - [x] Publish release - [x] update readme - [x] add generated javadoc to `docs/javadoc` (copy folder `javadoc` from `build/docs` and renamed to e.g. `v1.2.3`) - [x] update `docs/index.md` - [ ] update dependent projects - [x] create regression project for new release - [x] edit new `build.gradle` to use new release library - [x] edit `settings.gradle` to add new regression project - [x] run `./run_regression_analysis.sh` - [x] commit and push chages
process
prepare release add release number header to release note rename release note create empty release note for next release from template release note md change version in build gradle of big math subproject change version in pom xml upload artifacts to maven central run gradlew clean run gradlew ch obermuhlner math big uploadarchives go to in tab staging repositories locate own repository typically at the end of the list verify content of own repository version number close own repository refresh until release becomes enabled release own repository create github release from same artifacts create new draft release create new tag will be created when published create new title for release e g release copy content of release note into draft release without the first header release add artefacts from gradle build to draft release big math jar big math javadoc jar big math sources jar verify all changes are committed and pushed publish release update readme add generated javadoc to docs javadoc copy folder javadoc from build docs and renamed to e g update docs index md update dependent projects create regression project for new release edit new build gradle to use new release library edit settings gradle to add new regression project run run regression analysis sh commit and push chages
1
71,400
7,244,074,426
IssuesEvent
2018-02-14 14:03:18
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
TestHarness Scheduler exception while running tests
C: TestHarness
## Rationale <!--What is the reason for this enhancement or what error are you reporting?--> I have seen this a few times now. An exception happens while running tests causing a non zero exit status. The summary still gets written out and no jobs are set to failed. ## Description <!--Provide details of the enhancement or instructions for reproducing the error.--> ``` runWorker Exception: Traceback (most recent call last): File "/opt/civet/build_0/moose/python/TestHarness/schedulers/Scheduler.py", line 605, in runWorker self.queueJobs(run_jobs=next_job_group) File "/opt/civet/build_0/moose/python/TestHarness/schedulers/Scheduler.py", line 503, in queueJobs self.run_pool.apply_async(self.runWorker, (job_container,)) File "/opt/moose/miniconda/lib/python2.7/multiprocessing/pool.py", line 295, in apply_async assert self._state == RUN AssertionError ``` ## Impact <!--How will the changes impact the code, developers, and users? Discuss changes to the internal interfaces and public API.--> More robust TestHarness
1.0
TestHarness Scheduler exception while running tests - ## Rationale <!--What is the reason for this enhancement or what error are you reporting?--> I have seen this a few times now. An exception happens while running tests causing a non zero exit status. The summary still gets written out and no jobs are set to failed. ## Description <!--Provide details of the enhancement or instructions for reproducing the error.--> ``` runWorker Exception: Traceback (most recent call last): File "/opt/civet/build_0/moose/python/TestHarness/schedulers/Scheduler.py", line 605, in runWorker self.queueJobs(run_jobs=next_job_group) File "/opt/civet/build_0/moose/python/TestHarness/schedulers/Scheduler.py", line 503, in queueJobs self.run_pool.apply_async(self.runWorker, (job_container,)) File "/opt/moose/miniconda/lib/python2.7/multiprocessing/pool.py", line 295, in apply_async assert self._state == RUN AssertionError ``` ## Impact <!--How will the changes impact the code, developers, and users? Discuss changes to the internal interfaces and public API.--> More robust TestHarness
non_process
testharness scheduler exception while running tests rationale i have seen this a few times now an exception happens while running tests causing a non zero exit status the summary still gets written out and no jobs are set to failed description runworker exception traceback most recent call last file opt civet build moose python testharness schedulers scheduler py line in runworker self queuejobs run jobs next job group file opt civet build moose python testharness schedulers scheduler py line in queuejobs self run pool apply async self runworker job container file opt moose miniconda lib multiprocessing pool py line in apply async assert self state run assertionerror impact how will the changes impact the code developers and users discuss changes to the internal interfaces and public api more robust testharness
0
9,332
12,340,660,953
IssuesEvent
2020-05-14 20:21:09
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
"Permissions required to create an Automation account" / shouldn't "Global Admin" be replaced by "Application developer"
Pri2 assigned-to-author automation/svc doc-enhancement process-automation/subsvc triaged
replace "If App registrations is set to No, the user who performs this action must be a global administrator in Azure AD" by "If App registrations is set to No, the user who performs this action must have at least Application Developer role in Azure AD." --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9b4440e0-1ff5-0fd3-6983-d5f6ed86e818 * Version Independent ID: 8d6aecae-1a58-83aa-45f7-306fb6c92d38 * Content: [Create a standalone Azure Automation account](https://docs.microsoft.com/en-us/azure/automation/automation-create-standalone-account#feedback) * Content Source: [articles/automation/automation-create-standalone-account.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-create-standalone-account.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
"Permissions required to create an Automation account" / shouldn't "Global Admin" be replaced by "Application developer" - replace "If App registrations is set to No, the user who performs this action must be a global administrator in Azure AD" by "If App registrations is set to No, the user who performs this action must have at least Application Developer role in Azure AD." --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9b4440e0-1ff5-0fd3-6983-d5f6ed86e818 * Version Independent ID: 8d6aecae-1a58-83aa-45f7-306fb6c92d38 * Content: [Create a standalone Azure Automation account](https://docs.microsoft.com/en-us/azure/automation/automation-create-standalone-account#feedback) * Content Source: [articles/automation/automation-create-standalone-account.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-create-standalone-account.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
permissions required to create an automation account shouldn t global admin be replaced by application developer replace if app registrations is set to no the user who performs this action must be a global administrator in azure ad by if app registrations is set to no the user who performs this action must have at least application developer role in azure ad document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
193,707
22,216,283,750
IssuesEvent
2022-06-08 02:14:27
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2021-3348 (High) detected in linux-stable-rtv4.1.33
security vulnerability
## CVE-2021-3348 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/nbd.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/nbd.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> nbd_add_socket in drivers/block/nbd.c in the Linux kernel through 5.10.12 has an ndb_queue_rq use-after-free that could be triggered by local attackers (with access to the nbd device) via an I/O request at a certain point during device setup, aka CID-b98e762e3d71. <p>Publish Date: 2021-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3348>CVE-2021-3348</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3348">https://www.linuxkernelcves.com/cves/CVE-2021-3348</a></p> <p>Release Date: 2021-02-01</p> <p>Fix Resolution: v4.14.219,4.19.173,v5.4.95,v5.10.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3348 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2021-3348 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/nbd.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/block/nbd.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> nbd_add_socket in drivers/block/nbd.c in the Linux kernel through 5.10.12 has an ndb_queue_rq use-after-free that could be triggered by local attackers (with access to the nbd device) via an I/O request at a certain point during device setup, aka CID-b98e762e3d71. <p>Publish Date: 2021-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3348>CVE-2021-3348</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-3348">https://www.linuxkernelcves.com/cves/CVE-2021-3348</a></p> <p>Release Date: 2021-02-01</p> <p>Fix Resolution: v4.14.219,4.19.173,v5.4.95,v5.10.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers block nbd c drivers block nbd c vulnerability details nbd add socket in drivers block nbd c in the linux kernel through has an ndb queue rq use after free that could be triggered by local attackers with access to the nbd device via an i o request at a certain point during device setup aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
45,156
7,164,139,173
IssuesEvent
2018-01-29 10:10:55
doitintl/shamash
https://api.github.com/repos/doitintl/shamash
closed
Add Enabling APIs to deployment instructions
Documentation
Shamash requires both `Google Compute Engine`, `Google Cloud Pub/Sub`, `Dataproc API` and `Stackdriver` APIs to be enabled, otherwise the app is crashing. We need to add this to the deployment instructions or enable them in code (which I don't think is possible)
1.0
Add Enabling APIs to deployment instructions - Shamash requires both `Google Compute Engine`, `Google Cloud Pub/Sub`, `Dataproc API` and `Stackdriver` APIs to be enabled, otherwise the app is crashing. We need to add this to the deployment instructions or enable them in code (which I don't think is possible)
non_process
add enabling apis to deployment instructions shamash requires both google compute engine google cloud pub sub dataproc api and stackdriver apis to be enabled otherwise the app is crashing we need to add this to the deployment instructions or enable them in code which i don t think is possible
0
31,339
4,257,789,639
IssuesEvent
2016-07-11 01:25:44
geetsisbac/YE7GLKCP77LP3XTRJGXDSBHN
https://api.github.com/repos/geetsisbac/YE7GLKCP77LP3XTRJGXDSBHN
closed
ARX3HOThr1zUcfL6ZDgE+jIm/Heudzl3PFy4FeZTw9qbP222s5phYcWPwtP+kDdDucIgrkDSf/O0nAR0igQDuuaswy58i9qZh7SiP4ie1I0TXZWhkHmNVCk+BaRfbQ69iPwWC9cL7cG2Uw2+v8PN9Bbjd/fOkJnHGoERBeSkMnQ=
design
s6K3Rrn2LU5cJJoR2VKnH/xVIxf+I3Fq756rFR/rdKDCLwj5BhRTP9bHcV+KFid/XlNPpbENyQ55ovacq7fcuHyr2Up0fTxQH4RZpwQ76NFi85NLIDzB2QISLQ0bTcKEqcI0yRLU5wEu8ZdWUuWoEY21s86J5FCw8pWChV/kW/8JDF+Ab7qj7XrMsGZ2rGGA12d2JnQq4EqIUjpM6Xr38mRuMP6aekRr1JgHma0TIOEe22mLrMf4Pq3KAwY0VX5ghM/MmxCWfGQ9iLs+yJxMBKHcBP7+7S5bI5TfbBZ3bSTD2LWae6io/xVXOiDxYpsWKslRNvtcMLPa+DSml70bR3eOgV+Z2/pa6ZEkaJpkasexRKOBvpgA3Hhp3n93bEwVGiHfUkUquP5cd2dwJHjKN8PYtZp7qKj/FVc6IPFimxa5U11hWoIdBLmVG1ZzVO1DEPTEy7XRaRuK8U33mMWmSrFEo4G+mADceGnef3dsTBVdeE1sFSdNDBXckb8p6SKVw9i1mnuoqP8VVzog8WKbFkoKYg6UC7FkcVAU5rMZ3t6Sbm2niDuUqgUlEhz0ckWFsUSjgb6YANx4ad5/d2xMFawie/2tJ2PYdoABRwEuVB/D2LWae6io/xVXOiDxYpsWdbv1y1NGJjfkcdxk9ohvHvAQAPwFiyzW6cdDoHRdw4/jzFOwhLrHjppEsUt9ol/lYz9h5K2N3xRVeApxJqp4JMPYtZp7qKj/FVc6IPFimxZSJwO+DzD8QIfN6VE475g7HKXvjBOSWHVNMs7tCNs1gs9g6LwqCRZmKtejbA8UUfaxlx7Of6+X9wtAi+tTmEcaw9i1mnuoqP8VVzog8WKbFmzl9po+n5lQ5QfpDgBRj3uPNPSJYpstn+E4xSDl8XqrsUSjgb6YANx4ad5/d2xMFapzvx4YQfhms3yKp6O0KoHD2LWae6io/xVXOiDxYpsWzhT+cUUp6Dh+lDi1dSyPw8YFodtwvL2S75xNVtO4eS+xRKOBvpgA3Hhp3n93bEwVGmp2ZZVCFiENkoYaLF+X68PYtZp7qKj/FVc6IPFimxa1EfmCrqYK7u6U50PM4LQ1vPYycdU7ml/A2C501Dz7X+PMU7CEuseOmkSxS32iX+U5mPxIv4SIEI9JmsmoKEl7w9i1mnuoqP8VVzog8WKbFlK+31V6cCtk3ZcD8m9YEzq89jJx1TuaX8DYLnTUPPtf48xTsIS6x46aRLFLfaJf5bVjVERpfGXgn4OB0LiH1jnD2LWae6io/xVXOiDxYpsW6fYnXDHpcX0unk5OBQE2arz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/ljIXpBRLFhxav6cuj/H1OfcPYtZp7qKj/FVc6IPFimxZfU5HRz+L0zvVl6Lo0Xrf+vPYycdU7ml/A2C501Dz7X+PMU7CEuseOmkSxS32iX+UgcqC+p1aMZE0i/aBwP4LOw9i1mnuoqP8VVzog8WKbFmPmqcnVkQhDIyKT/eExe7C89jJx1TuaX8DYLnTUPPtf48xTsIS6x46aRLFLfaJf5RzvaFTeD4e20/keKaaDEHvD2LWae6io/xVXOiDxYpsWjWgoPXwhsu6WrgOoAFwaALz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/lXBoOwtQ0oqXHuJp3cWbz3cPYtZp7qKj/FVc6IPFimxZvT8etorln6qiZX22rqMDdURzrFAg+zqS7Vu6DF9kGF89g6LwqCRZmKtejbA8UUfYaLFcRyCsJimD3o7l2deBtw9i1mnuoqP8VVzog8WKbFm9D+fkpvWQkElEO9nJSBye89jJx1TuaX8DYLnTUPPtf48xTsIS6x46aRLFLfaJf5RxUrYC2vi0PEXIorKnOZkXD2LWae6io/xVXOiDxYpsWm8+vGr12PtbXHXGqiYSFOsjquSFT1swc0FYW0+29SJfPw5BQZ710XScJMJ8iAcg5kOeyWIus+IHCvPZp3VLMX8PYtZp7qKj/FVc6IPFimxYZeggrhLxEeQCS+ZHJzmBl7rvxFEkPm73XLIhqmqfuakQEUQVa2GyXwdsGsc6V38YOyrt5QQbPDz21eZeNgej5w9i1mnuoqP8VVzog8WKbFkIvWKehyGhpoY9iZRKv3x/6aLVuNRtIEEZ5UDIBq2t4z2DovCoJFmYq16NsDxRR9kW6gZMTX/OpjVEyvxtl/avD2LWae6io/xVXOiDxYpsWsu8BTLoMHflbwAjR/30h3fpotW41G0gQRnlQMgGra3jPYOi8KgkWZirXo2wPFFH2yM8dW1XRXewugQEkNKw568PYtZp7qKj/FVc6IPFimxYqmLod4PNaF3Mdg5l+34t/+mi1bjUbSBBGeVAyAatreM9g6LwqCRZmKtejbA8UUfYV2tusIK4QamB4nnWbXIdNw9i1mnuoqP8VVzog8WKbFqqocPWptdQJDGD0V+svppa89jJx1TuaX8DYLnTUPPtf48xTsIS6x46aRLFLfaJf5Sp2ytk0bG8Vm9L4cil8b+HD2LWae6io/xVXOiDxYpsW13juFWQlBJLAqnYL8C43rbz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/l87d2vicBL7tXo1BaFDpDysPYtZp7qKj/FVc6IPFimxavq9ymfoXOHjjrxMxqvtjs1pbMpQhYMfwf3N0hRzbOvePMU7CEuseOmkSxS32iX+UxB+9RVb1jX6L1+z8ZQNqmw9i1mnuoqP8VVzog8WKbFjVVM1ZQLGLzxIalPcWeeCe+DfdrTRKJhVUgtGb8mujWXxYUBhQEpwY1d3Cim+ZKYMeZALz0LtGd3fuGSn1R9NjD2LWae6io/xVXOiDxYpsWkfmuUUORDpG0+9bR3XNmP4Tor0w1ntGRFBW03CpMSh5j3RzsDKIQl1sRus21qeEC17dGdbW0N+mnd6Ih5MjCdsPYtZp7qKj/FVc6IPFimxYocgegQw9dSXX3VGAgC7BpvPYycdU7ml/A2C501Dz7X+PMU7CEuseOmkSxS32iX+UFLzcjoQUTjBSq8LJnw9SDw9i1mnuoqP8VVzog8WKbFnyWO0NISPEJdsqVVpxNcbuy8pjlaw9cFztfz3ozpbFJsUSjgb6YANx4ad5/d2xMFTiy5gAQhxSAGmnL8IjUe7jD2LWae6io/xVXOiDxYpsWSgpiDpQLsWRxUBTmsxne3oItPo+r1LI7lma/rWO6c3nqJJt5Nlncw4K6QYDwmBiHrQbrzdrSuGWZ5vBkkAQ9csPYtZp7qKj/FVc6IPFimxZuhNe7OQi3ISAzhinzrieuclDyir1DL1kY2YpOKsZ9Y8dAhfrVI428u0u54wQjAOzImMJmP76x9nOMZbgJGHlXw9i1mnuoqP8VVzog8WKbFm0WTfzyLssvZUpeNceQPf8OFV/GhfaA1EwzjgVa0mQNe56sn7eFNF1sXW/NQolQDlrLW9ZSLfVM92JbR+wRw1TD2LWae6io/xVXOiDxYpsWgz28C9XOfyvWGuAnhvDE1DWWoxocNesPUVgfJj/nqWhPJXCbO4bdoCd8PPNYr5yOOjoAI1XzkxWXM6wWAdTqk8PYtZp7qKj/FVc6IPFimxYaewruaDKqEDbyFUDs7oakhsCaB9HIpOScSoPZJk73YkCSdewck0gaE7vc5EvIrkkqtkcCV1IHOQwKB29rMlmxw9i1mnuoqP8VVzog8WKbFlXg4EhQUVDWmAviVnMPxx8RMnhm6QdihjZ4I51fZVzK48xTsIS6x46aRLFLfaJf5WDOTxbmN1nI4gJnx3FmFvLD2LWae6io/xVXOiDxYpsWnM17yFkZKOtBK7jHVXT48Lz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/lS3NtoekdrI0H58Ej045w1sPYtZp7qKj/FVc6IPFimxZQNqHrzwdXPCmijjRD7R1RM1xxQo83U/RcL/4LBOz9y7FEo4G+mADceGnef3dsTBVqAQY5UVkwd/CpSqgHnknJw9i1mnuoqP8VVzog8WKbFhRtaQp50gqTBYjn3TFtOJGrpRLlJOuIPFzr69y8mTkjJ3mIsZmjhCxc+lViCtu52JMJ19tpaS70LKPi/p4R6kvD2LWae6io/xVXOiDxYpsWPVx5sWLJW9Ew1wzNm5oG+o6p2l2vgMn1nwcqZmNdJtSxRKOBvpgA3Hhp3n93bEwVk/GDJYLE5TiEjTgwb5VnZcPYtZp7qKj/FVc6IPFimxaUpOg+A1wHGhjZycG+18mmWNcXlWaT5PA28G5J84PQz89g6LwqCRZmKtejbA8UUfbM5JJRGY+yzSImdg04B6Kdw9i1mnuoqP8VVzog8WKbFle26sR4vHbTwOL5Q3NSKtB9pPp5z04lmWFtUR4g4+wg48xTsIS6x46aRLFLfaJf5Z8x9NHI5kGTrdJ8081RDGDD2LWae6io/xVXOiDxYpsWveUwzL908Z9UfawtoaDoUBgBNAPRpIk8RtVmfVuMhZ/P0Q7kLGLy0ixe3gt7c+dEWuQSxErlue7fzVip+yja88PYtZp7qKj/FVc6IPFimxY4p3PdruRdh/UgumF8Mnhfo/Zo1xO1EW4AO9I0qsXidf8ZJjjBTjSdSLzS+Myvtoe7MZWcKVYYbMi8wPku2Lztw9i1mnuoqP8VVzog8WKbFgmYbNuEVA1e9T/lcxIDsrRG7h2qu679GTKNclRd4CyKGHqC2XbUtqks+HhricJ7m1PehOyxhojlecrm6aPH0NbD2LWae6io/xVXOiDxYpsWtrvpvSGNmWSV/Y52x+o8yKrUeAoa5qtehS9QfOmUzyyxRKOBvpgA3Hhp3n93bEwVXLmIdySU4siANiwG0xHS3cPYtZp7qKj/FVc6IPFimxZXEn08SBQI932p1ve+Z5JEdM/EK7jOxqnMdlBT+IGuOCnbB/va8UktX3N7USW2Q9gmAMdfQaAnkew5cGpHUmFRw9i1mnuoqP8VVzog8WKbFq6uQ3vJ+qgTf7vbKUcWd0J3oFcyvsHDo2v9vYZpvwsz/xkmOMFONJ1IvNL4zK+2h4bxHgRjk5ZVUAFUmnYsmxLD2LWae6io/xVXOiDxYpsWD/lB/aSqm7o1LE/yQNXU9a8WuNTfFGtiKhmeaAsx4ajPYOi8KgkWZirXo2wPFFH2RyXnT0iL5EdsQRp7+vlTiMPYtZp7qKj/FVc6IPFimxYWtDAG846hDCDPobb7F/1E3hvXB2UVqvhuSDiVjR5hqs9g6LwqCRZmKtejbA8UUfatuSRCKdK4uSvrBZtXSyZxw9i1mnuoqP8VVzog8WKbFjFLvIexI8KuN6dQZIMTgqlCGA6uXLit3GnT9ZWnRSiFz2DovCoJFmYq16NsDxRR9i3w8i2bixuEIDLPadjF6tnD2LWae6io/xVXOiDxYpsWU9G3RQVDFcOeTX/BQMISiYZ+oH/UtIOZCE8GxbxvY78YQBC0ZM8TNWY88IjsPhldBXVum53optfxsGvSabD3y8PYtZp7qKj/FVc6IPFimxZitV0fJANlbIWUy/I9APM1vPYycdU7ml/A2C501Dz7X+PMU7CEuseOmkSxS32iX+WWFi00Wg7B+TGc9sJIMH7Xw9i1mnuoqP8VVzog8WKbFh2pc0JQHiqHhOZMQqSNjjPDXzS3zEROI1f1Vq1u42mFsUSjgb6YANx4ad5/d2xMFZpoAH3R/t/GpUt1ZNPpbiDD2LWae6io/xVXOiDxYpsWfqtL2Pvdv2TOfyE6+uOAW7z2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/lh7DTcmvzq9mxmxWH3O6AT8PYtZp7qKj/FVc6IPFimxbfBzpttv2n9xdg+CB7N7OlnfN9dQ5fzNOMrmrhKCm+ySd5iLGZo4QsXPpVYgrbudguileYPPqh5IfHqzcumPGCw9i1mnuoqP8VVzog8WKbFlLuxyDOFz6/q5H9lI2z1bT5yV2tkluU21DUAhW1IHQW48xTsIS6x46aRLFLfaJf5e8ZkMgmu4mYlyXxpJRZLG/D2LWae6io/xVXOiDxYpsWYTikBMgWmyTas/NVrl2fTLz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/lBVzeNyI7Vbq4BHMGXKHIv8PYtZp7qKj/FVc6IPFimxaThaGlTtz6VnCRLGrazbF1IKpuNC5cuTfcsoYzENmcg7FEo4G+mADceGnef3dsTBVTnnSlZKheBzeH7M/IH3uBw9i1mnuoqP8VVzog8WKbFoGDluJ7zeGCpfrGTk38EkqArf0GX8aYudl706YiYWEt48xTsIS6x46aRLFLfaJf5aRLOiIyMdnot5dtcKXt74LD2LWae6io/xVXOiDxYpsWGYCZlWh4Vc0MaVlUIrOWLGkYt0Bzfs8StJVw/9G7hP0p2wf72vFJLV9ze1EltkPYV0s/T4eMCUKWlme/fVvuAcPYtZp7qKj/FVc6IPFimxYd6+szZRqHG3PE1EeL2LeaVIItI+yDfnR28ZrTB2fdXfq3uXHtCNXEBqYZQfVa0o9/dXrsuIr4ooe2oP+Y1jIDw9i1mnuoqP8VVzog8WKbFiXEPX7jjK/aiKSs+q09Uf6QIxp4B7VQtlbn/EvRf4pE7SHTx+z/5nNkM9pobJbesozHINuc8jDshVTr2U9TM8DD2LWae6io/xVXOiDxYpsWCk9DtF0eZ9LkXhJbNTwrQJAjGngHtVC2Vuf8S9F/ikS7Bwd3jahAhFJYCmw5tsMnZnkCpTgeJdAeFHv0sSdsOMPYtZp7qKj/FVc6IPFimxYjQmKvH9im0nw2XQUIF2H3NaW5fx7gt22kIaQtBYYG3hmevzOeskPr1bQMksXScXvsKIPLOoagdn+bpDO9rCz9w9i1mnuoqP8VVzog8WKbFoMeHhVW67lDKsVc64TM0PC6kXurc+irzp/nv/7yFhtySYjQOqFRUyVpzp2wUj8oJeFM+yMhMMAbWj3ErmHdpTXD2LWae6io/xVXOiDxYpsWjAAmgNGw0NjYKOlyzjRNJ5JubaeIO5SqBSUSHPRyRYWxRKOBvpgA3Hhp3n93bEwVG0lfZATSc6DtcBkMZX1vdrFEo4G+mADceGnef3dsTBXcyO9YwcBYYSvABe8uhJGnWuMAnjkuJyWieR50mt/ZLM9g6LwqCRZmKtejbA8UUfaGpClg0110nDnQxClLXEG2sUSjgb6YANx4ad5/d2xMFRfl2t1AbrBEaYah2wlsfs1rIUWHapM5jmbNy12OLTCNz2DovCoJFmYq16NsDxRR9tprpBgf+j5nVpXO0p9/QaOxRKOBvpgA3Hhp3n93bEwVdaxmla10uu7sSC+y1X2EqMuKlSiQxyk2IVQoqMKNIhuaFhoqMSgmj7wCoXZfS4FMBDU3+lV6tvhYpvyf++UNeMPYtZp7qKj/FVc6IPFimxb3R7sHh3FVrstfmz3Ss+KpBlAIDSnLfWxWFU2fyDdV97FEo4G+mADceGnef3dsTBW6mIgmOP7G4NekzBN1zXuesUSjgb6YANx4ad5/d2xMFS902QTSdEmPMdQI27yTd5NI3kW+kjd/6no8tcQWn+N0J3mIsZmjhCxc+lViCtu52D0bRbEVH1jGgE+AXYTmtoqxRKOBvpgA3Hhp3n93bEwVp0tEhG6UnhcL0tAVI7ufI5Wx3yYETH8hgQTthGLucUbjzFOwhLrHjppEsUt9ol/lIE89jXj/LW9RbsZapNi9xrFEo4G+mADceGnef3dsTBVjkXEogqYnHcqCdZde33eI/F+AqfMaespootIJPc+2nf8ZJjjBTjSdSLzS+MyvtoeYEvs1JO2kFY08rw0kWmxFsUSjgb6YANx4ad5/d2xMFdL+xrBG/sOjhbPQbsN9KoiVsd8mBEx/IYEE7YRi7nFG48xTsIS6x46aRLFLfaJf5XD3ZXGqOpJ46ekb0OnhqC+xRKOBvpgA3Hhp3n93bEwVoGMlFykX9GFV/OYphvZQqDEWMI1Gs4KEUm+E64z1BTPnoTK/RpkZFYb8B0bgNnUADxxGg8X8WWXmEHVFZuG8JbFEo4G+mADceGnef3dsTBXxVhkYN+HQiP2J5DyJ68rKUNVeMv2vW7TILk7R8tkSAb6n/HY5CeGwFfQtPHjO20jPxPNbDK20KoZlc7kbiA+isUSjgb6YANx4ad5/d2xMFRO6JxGGX/+eO/+LwDVK6PWmR41NBBWIdbX4tKzo45WVdhKTmVMLV0hyuWri0ApWjGrqPxEpmFN2kmXegn18LjexRKOBvpgA3Hhp3n93bEwVv4LaJDdD0YCWnfMGuMwLCN0z2cMuFOhcKUbScfTtKqCxRKOBvpgA3Hhp3n93bEwVqOA59WD+1bEIUMhxg0xfpLFEo4G+mADceGnef3dsTBVLXqIgHvkJJu32f885MiFjHc90/1/lgTAe8K+obAES1P8ZJjjBTjSdSLzS+MyvtoeplftJsI3bHGbrIVFgv0CZsUSjgb6YANx4ad5/d2xMFR6mKulUnUXtyaA+XlL7BjKe64AJiOXSq8iFU7MoQgVKz2DovCoJFmYq16NsDxRR9mv2vgOK9UebYY5eYS6KN5KxRKOBvpgA3Hhp3n93bEwVbka3Y/mMfLDzaDYTESJqnhrfY4/iH4ENxoTWUEXDp3hCD9bVkTIagOGSCz61ZsbTThf5MUpF1FnryPKxC/LgTrFEo4G+mADceGnef3dsTBVH8skSqZ8PNBcgKSQzsp9paLtulUdrYaZme9RZ15GKVbFEo4G+mADceGnef3dsTBWI6bS172w+TTUTAIGoscE5sUSjgb6YANx4ad5/d2xMFSbN32sSfim+AE7G87UG62bWANEKXvAtYKGR0WDs/p7lJ3mIsZmjhCxc+lViCtu52Lbr01UMOKE4vMpw5zftX2ixRKOBvpgA3Hhp3n93bEwVpwBhLoXbqhYHjSNdmfnWCZRYPJpIRyV6UF3PR1Rn05rPYOi8KgkWZirXo2wPFFH2kO0IfAxORxX2RCkt3Q8n1LFEo4G+mADceGnef3dsTBUSxX18pVUzmU6igoxALJYUk7VgvNWXaHki8ZjFCnp1Ff8ZJjjBTjSdSLzS+MyvtofwDVbw/WfhfMd13mJo9VdvsUSjgb6YANx4ad5/d2xMFWXBSeBgRWBlkKz6mreO6sgB6cwk1v7hS9iYrJ9Hd9IEz2DovCoJFmYq16NsDxRR9ghuzr/dt9LF1qgiB3+AqFCxRKOBvpgA3Hhp3n93bEwVUsZlbtPF1DmWuYdqo1HNbBCDMXxzd7px2y2oFqO4HuTPYOi8KgkWZirXo2wPFFH25d+5Fpy7Sgak/ZRmiepoobFEo4G+mADceGnef3dsTBXAlRX7dVKOUZb9jamgxvVtgJ0NWMBFAOIqAccuISjxDomMPC7EMamFagDtwu1KIfimlTXiqD/Wo09dSxvhVJ3nw9i1mnuoqP8VVzog8WKbFt4YwiZ0AE6RqTSHT8PK/cReyjecdbtdbOvUqE6Y3UwlJ3mIsZmjhCxc+lViCtu52GmIsB5J5S5qfWoctbvjNwixRKOBvpgA3Hhp3n93bEwVTXJU4TU5ICbK7MIZ8PQhsaZHjU0EFYh1tfi0rOjjlZV2EpOZUwtXSHK5auLQClaMBA5fWBM0yXiIXDkoDdDkybFEo4G+mADceGnef3dsTBXTdL9XIE6ZaXJXvBFnevLQ76JmRP9OH+vgVTIWiYY8h/bLzW5rBSv238IuOYt+1r30y6wGmm3TLiRxQC0lkBxnsUSjgb6YANx4ad5/d2xMFcP78Bqu/OliPIATVr4sFt9swBRTljv8IWtWa5k+paehsUSjgb6YANx4ad5/d2xMFcOk9UhA4JBvkWc67YGcI3ixRKOBvpgA3Hhp3n93bEwVOY6mZe5mMdl29kHUIUJgs++iZkT/Th/r4FUyFomGPIe7Uqu4fhYwzdcx2EzIjcOCd2yLW3L5zEBurnq59k5PMrFEo4G+mADceGnef3dsTBWyqYsSs382/vyKk0DjRRQU+cldrZJblNtQ1AIVtSB0FuPMU7CEuseOmkSxS32iX+Ws0Tp7Clk5mQ+C3PAwm0DOsUSjgb6YANx4ad5/d2xMFWX4ch/a362r1PokL3rsIkrvomZE/04f6+BVMhaJhjyHBQG6MS2enqq8kXx4MFaxoO3N/EwRR6ni+rigfX/r0jOxRKOBvpgA3Hhp3n93bEwVQoRovA8nuK9uUzCLUvp4yPnJXa2SW5TbUNQCFbUgdBbjzFOwhLrHjppEsUt9ol/luFwwPRGbU3qpUz6yEB3IXrFEo4G+mADceGnef3dsTBUlUxt3bt5r/hU1FuMv29wpbMAUU5Y7/CFrVmuZPqWnobFEo4G+mADceGnef3dsTBXN4lbU9Bf0VaaS7gFyC3CFsUSjgb6YANx4ad5/d2xMFX636q8GjjKioLdGS5YUFddswBRTljv8IWtWa5k+paehsUSjgb6YANx4ad5/d2xMFbceScC8QmzgQrPYbo0Nk0qxRKOBvpgA3Hhp3n93bEwVUkRTpbrizVmHsEitUD0JnmzAFFOWO/wha1ZrmT6lp6GxRKOBvpgA3Hhp3n93bEwVe8QPMdqvasuYvYJVAkLkTLFEo4G+mADceGnef3dsTBXGd+XW6SiOLYNEACZtLQVebMAUU5Y7/CFrVmuZPqWnobFEo4G+mADceGnef3dsTBUTaJ9QcKuvma4RR0/VHyXDsUSjgb6YANx4ad5/d2xMFXMAQ1jy/89BQTioLPxGJPfFn5HF0Y8GhtZ/g5nARQbtmYqHqn5bLgft+Y6T6iISuwrZ2rO7ldLshUzmnuby+PaxRKOBvpgA3Hhp3n93bEwV8ZNcDFPlTiLQV+HzlW7rtfnJXa2SW5TbUNQCFbUgdBbjzFOwhLrHjppEsUt9ol/lLNNr/+QDC2YfXHdskV7rbrFEo4G+mADceGnef3dsTBVhGn3FOuPAknt3rCSztvAwkNdkRTSmRDq97cmRF5wdI+PMU7CEuseOmkSxS32iX+WaT2JkBarcKboRFehtpH4tsUSjgb6YANx4ad5/d2xMFYVfWb0JX0lUnqPo4v71ylqQ12RFNKZEOr3tyZEXnB0j48xTsIS6x46aRLFLfaJf5TtNmZM0x7gWzcNORtx+xZaxRKOBvpgA3Hhp3n93bEwVhdK3AuZUhtz2CLwkovvB68datIZdKG4YF7fTlxI7P2XPYOi8KgkWZirXo2wPFFH2/q4kM/ij1QKnunlJz/jbMrFEo4G+mADceGnef3dsTBXGXJ4Som52zu1RxQbFBUfLx1q0hl0obhgXt9OXEjs/Zc9g6LwqCRZmKtejbA8UUfZCbL8mAOhdBme2KTLAEYFmsUSjgb6YANx4ad5/d2xMFaXbq4PjOoGzXA+h9R7YDTDKjzuCKsHbLv/fUj8f0mGD48xTsIS6x46aRLFLfaJf5c2HChRj0j/hXi9QbwhY3POxRKOBvpgA3Hhp3n93bEwVJYPNN3nTTe4uRPA8ToNQf8qPO4Iqwdsu/99SPx/SYYPjzFOwhLrHjppEsUt9ol/ltyZARE8naxD++yyxxuVfabFEo4G+mADceGnef3dsTBVac5Cj+QUU+H2+TDzYS2iAr4e2yG3NhCOsmQR3r5WY9Sd6jbuhsY5PCy+55G/xFVesTVeDuNhC6vUDfSg0AlWYw9i1mnuoqP8VVzog8WKbFtbGyfVERyVtp4ECf4WBuMKz2BlnL0amqOo9tgHU+orVJ3mIsZmjhCxc+lViCtu52I34IogOxtChl+e89MCdYyexRKOBvpgA3Hhp3n93bEwVpWnF6Tz4uHZEZ4lkaNZBK9KzwJI3IpyoKvcBikbk2PkneYixmaOELFz6VWIK27nYeJekmV3sb3akavEbZq7xY7FEo4G+mADceGnef3dsTBXIJ/mYJWFsUFBAdZhlQgOoQhgOrly4rdxp0/WVp0Uohc9g6LwqCRZmKtejbA8UUfaYqetV5GZK2OL2ZEEc/9z8w9i1mnuoqP8VVzog8WKbFl+Yh4ZQspspy/FrNZv5G6wMykH9lIZSmoV8SVhfYB3l48xTsIS6x46aRLFLfaJf5UUZaWQYwdvAxRMVa8+qEFixRKOBvpgA3Hhp3n93bEwVggWZYdHg7Rs+FmblP8x+zQzKQf2UhlKahXxJWF9gHeXjzFOwhLrHjppEsUt9ol/l4nA5LNMXxrn5GNuHFfKuabFEo4G+mADceGnef3dsTBUUrw31YCieapvVRk0QYlFiDMpB/ZSGUpqFfElYX2Ad5ePMU7CEuseOmkSxS32iX+U0yW/xoqAAd80e5sDJhVFvsUSjgb6YANx4ad5/d2xMFYWxmBu7LVhcLTEC6yz6pVsMykH9lIZSmoV8SVhfYB3l48xTsIS6x46aRLFLfaJf5X14GrkwJycN8Nj6Bk2gpOSxRKOBvpgA3Hhp3n93bEwVhcEUJsk+d4wu0mqvop/ZJsdatIZdKG4YF7fTlxI7P2XPYOi8KgkWZirXo2wPFFH2H3d/i+0Cof8mu41vsSBO7LFEo4G+mADceGnef3dsTBUVxhSTXgdiL7O6JdCnE1H3N+P6SYoJRuHQ77rxRYIxaWDyOOFaCzqTtHEF2gZAJJwfie1mfH76vnlSAr2C4AjNsUSjgb6YANx4ad5/d2xMFSD2pKzk1z2fNH545fDfeHOHtIU8WCck/1T7daefACTQz2DovCoJFmYq16NsDxRR9jsTghj4yDhOMLFdG/45E/WxRKOBvpgA3Hhp3n93bEwVvyKTfvghlfSLMiMWG7CPJEz7orJAqIDjtrlerAi+I9TjzFOwhLrHjppEsUt9ol/lftFhFYz1KtHZSHfHhp5mbcPYtZp7qKj/FVc6IPFimxbsH2FkuXsgAlehcGbSuk6o+cldrZJblNtQ1AIVtSB0FuPMU7CEuseOmkSxS32iX+Xr9sXBx9DxcrhH/Ibga2hBw9i1mnuoqP8VVzog8WKbFn/6UaKqc6mh/vq4v6n2BlP6aLVuNRtIEEZ5UDIBq2t4z2DovCoJFmYq16NsDxRR9v/jrpztTRqF1XC9PRLpFNXD2LWae6io/xVXOiDxYpsWutMZn5E8l/8g+IxbvFJqLQzKQf2UhlKahXxJWF9gHeXjzFOwhLrHjppEsUt9ol/l11YJxN6Fn2tt3iyDLqXlW7FEo4G+mADceGnef3dsTBWcQUZF1J6NEHGdA66Tdlaukm5tp4g7lKoFJRIc9HJFhbFEo4G+mADceGnef3dsTBV8SQEt7RrARa13dxrT5wqLsUSjgb6YANx4ad5/d2xMFXChGRnlcSyF+Lu3lFbNl/9a4wCeOS4nJaJ5HnSa39ksz2DovCoJFmYq16NsDxRR9kn2Lu4W8aCAGoiqrqpP2eWxRKOBvpgA3Hhp3n93bEwVOSEdzhZc2CS6tOBGemrCDkIaH9HSWeHDXq1BkOEHEQ/jzFOwhLrHjppEsUt9ol/lRA5uyC4bAp4bMjkK51d9ULFEo4G+mADceGnef3dsTBXucNluIONQjXSCsoYzQV4DkNdkRTSmRDq97cmRF5wdI+PMU7CEuseOmkSxS32iX+XHl68WurTvGTeajeAsBeXisUSjgb6YANx4ad5/d2xMFZUgcpU2sRHq0uuh77idRX9WpxfQalJMrIUh7ncHV8JK48xTsIS6x46aRLFLfaJf5ZLIO9jmbV4R7zaYSvFA3BaxRKOBvpgA3Hhp3n93bEwVM/cp4BNBgC0S5PWvPJFLmKrUeAoa5qtehS9QfOmUzyyxRKOBvpgA3Hhp3n93bEwVaEaoMWzUt3GuVPnZOvSrF7FEo4G+mADceGnef3dsTBWjlTeCG3IccYTM7NBKyoV6WAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBXycHfUVdCu3PvY9jRsDK1dsUSjgb6YANx4ad5/d2xMFWftmHhaRtM3DnQfLQsPAPtYCHifmbloZxNCP6uSi0uisUSjgb6YANx4ad5/d2xMFYB3Enk8rkyao+Tpk99xDQGxRKOBvpgA3Hhp3n93bEwVSc2OSZeWX5bjndavrdhFKFgIeJ+ZuWhnE0I/q5KLS6KxRKOBvpgA3Hhp3n93bEwVTRTknj82nAxzfgP+BaDWCrFEo4G+mADceGnef3dsTBWkOmsqKx/z4SVMOIu6YebbWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBXCw5EW+Vq2a60ycq3Zo2l4sUSjgb6YANx4ad5/d2xMFVPftg3xJUI0bi6wvafEW+ZYCHifmbloZxNCP6uSi0uisUSjgb6YANx4ad5/d2xMFepzFk2diJFRt8BEn026mS+xRKOBvpgA3Hhp3n93bEwV8bmhEehwvyGPnn2Wo1HnR1gIeJ+ZuWhnE0I/q5KLS6KxRKOBvpgA3Hhp3n93bEwVjg640klWJSyebdtD2cNkUbFEo4G+mADceGnef3dsTBW8rPwQaTR0qQWVpUgynK4lWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBWNfHEE4Te+m0AgsnrGFnj7sUSjgb6YANx4ad5/d2xMFfJekrn+c/k4zDBfVobvbeZYCHifmbloZxNCP6uSi0uisUSjgb6YANx4ad5/d2xMFe37x0JtQ6VS2zs5kIebXxmxRKOBvpgA3Hhp3n93bEwVAI/aaDjqu+uRgsCGgzpRq1gIeJ+ZuWhnE0I/q5KLS6KxRKOBvpgA3Hhp3n93bEwV5coaeQynnX+qbg2mQsSjBLFEo4G+mADceGnef3dsTBVmkm2h4tg7G+uRNgg3N/TuWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBVb4OrPHGZvhfojwXDFMe0usUSjgb6YANx4ad5/d2xMFY96s4Hqz4fz90hxq9cPil5YCHifmbloZxNCP6uSi0uisUSjgb6YANx4ad5/d2xMFbc7ZSZiyK9naOk7r1aIwpqxRKOBvpgA3Hhp3n93bEwVMbIDjcV+jsNfaEjIXiAaMVgIeJ+ZuWhnE0I/q5KLS6KxRKOBvpgA3Hhp3n93bEwVHM5E5fpwwYmUXIioswHpsbFEo4G+mADceGnef3dsTBU3Kd8GK8gz0uOSTB7D07onWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBU4Grnw1tpyLUTQ6CWI4x9KsUSjgb6YANx4ad5/d2xMFZhj/HJbVpW6G7k6iIxqk57t07RJTJXuYPoFvLdxllXM48xTsIS6x46aRLFLfaJf5fT76k7+Usa4IwmwCEDmamuxRKOBvpgA3Hhp3n93bEwV1M7FWSujPM17Dq0Cm4dCBKKpBbyGaDPBpXw+6WjASN/jzFOwhLrHjppEsUt9ol/lKw6MQpUQbuZB406YiUjZa7FEo4G+mADceGnef3dsTBXVmeAJwFHQak8uNrW/nvBRWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBWI50g5L4WEQeaYTkp1LnOJsUSjgb6YANx4ad5/d2xMFVrzIaxOk6aAqbZmWkonJT24rfhyB3hEAAa7zVQTv/XCsUSjgb6YANx4ad5/d2xMFYTjmVbWE7sOyKnjA5wTL06xRKOBvpgA3Hhp3n93bEwVxYIfmueBT/YmrnJnnnA5S/nJXa2SW5TbUNQCFbUgdBbjzFOwhLrHjppEsUt9ol/la1HdX7UiRM3QLIMjoYYT9bFEo4G+mADceGnef3dsTBWCTPqGV8fOyFxZX4w64IGVBx0I2zSA0mp7S/41Rp/ULc9g6LwqCRZmKtejbA8UUfZTxnVQUnq1wiujJhaOPky2sUSjgb6YANx4ad5/d2xMFQQyAN2y7zldavzxQdC/AEVCGA6uXLit3GnT9ZWnRSiFz2DovCoJFmYq16NsDxRR9mML5IgftG/yqOZYV0WmyabD2LWae6io/xVXOiDxYpsWIXEyR07yinln6FlRTgi1LcJH6T7yZcE89SJ1EQez5IR6JHrD/Nuiufkesf7NViHYBDpSYeHXi77i89b6dK1i3+UKRZxsDf8CZaHHUfJGDTM=
1.0
ARX3HOThr1zUcfL6ZDgE+jIm/Heudzl3PFy4FeZTw9qbP222s5phYcWPwtP+kDdDucIgrkDSf/O0nAR0igQDuuaswy58i9qZh7SiP4ie1I0TXZWhkHmNVCk+BaRfbQ69iPwWC9cL7cG2Uw2+v8PN9Bbjd/fOkJnHGoERBeSkMnQ= - s6K3Rrn2LU5cJJoR2VKnH/xVIxf+I3Fq756rFR/rdKDCLwj5BhRTP9bHcV+KFid/XlNPpbENyQ55ovacq7fcuHyr2Up0fTxQH4RZpwQ76NFi85NLIDzB2QISLQ0bTcKEqcI0yRLU5wEu8ZdWUuWoEY21s86J5FCw8pWChV/kW/8JDF+Ab7qj7XrMsGZ2rGGA12d2JnQq4EqIUjpM6Xr38mRuMP6aekRr1JgHma0TIOEe22mLrMf4Pq3KAwY0VX5ghM/MmxCWfGQ9iLs+yJxMBKHcBP7+7S5bI5TfbBZ3bSTD2LWae6io/xVXOiDxYpsWKslRNvtcMLPa+DSml70bR3eOgV+Z2/pa6ZEkaJpkasexRKOBvpgA3Hhp3n93bEwVGiHfUkUquP5cd2dwJHjKN8PYtZp7qKj/FVc6IPFimxa5U11hWoIdBLmVG1ZzVO1DEPTEy7XRaRuK8U33mMWmSrFEo4G+mADceGnef3dsTBVdeE1sFSdNDBXckb8p6SKVw9i1mnuoqP8VVzog8WKbFkoKYg6UC7FkcVAU5rMZ3t6Sbm2niDuUqgUlEhz0ckWFsUSjgb6YANx4ad5/d2xMFawie/2tJ2PYdoABRwEuVB/D2LWae6io/xVXOiDxYpsWdbv1y1NGJjfkcdxk9ohvHvAQAPwFiyzW6cdDoHRdw4/jzFOwhLrHjppEsUt9ol/lYz9h5K2N3xRVeApxJqp4JMPYtZp7qKj/FVc6IPFimxZSJwO+DzD8QIfN6VE475g7HKXvjBOSWHVNMs7tCNs1gs9g6LwqCRZmKtejbA8UUfaxlx7Of6+X9wtAi+tTmEcaw9i1mnuoqP8VVzog8WKbFmzl9po+n5lQ5QfpDgBRj3uPNPSJYpstn+E4xSDl8XqrsUSjgb6YANx4ad5/d2xMFapzvx4YQfhms3yKp6O0KoHD2LWae6io/xVXOiDxYpsWzhT+cUUp6Dh+lDi1dSyPw8YFodtwvL2S75xNVtO4eS+xRKOBvpgA3Hhp3n93bEwVGmp2ZZVCFiENkoYaLF+X68PYtZp7qKj/FVc6IPFimxa1EfmCrqYK7u6U50PM4LQ1vPYycdU7ml/A2C501Dz7X+PMU7CEuseOmkSxS32iX+U5mPxIv4SIEI9JmsmoKEl7w9i1mnuoqP8VVzog8WKbFlK+31V6cCtk3ZcD8m9YEzq89jJx1TuaX8DYLnTUPPtf48xTsIS6x46aRLFLfaJf5bVjVERpfGXgn4OB0LiH1jnD2LWae6io/xVXOiDxYpsW6fYnXDHpcX0unk5OBQE2arz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/ljIXpBRLFhxav6cuj/H1OfcPYtZp7qKj/FVc6IPFimxZfU5HRz+L0zvVl6Lo0Xrf+vPYycdU7ml/A2C501Dz7X+PMU7CEuseOmkSxS32iX+UgcqC+p1aMZE0i/aBwP4LOw9i1mnuoqP8VVzog8WKbFmPmqcnVkQhDIyKT/eExe7C89jJx1TuaX8DYLnTUPPtf48xTsIS6x46aRLFLfaJf5RzvaFTeD4e20/keKaaDEHvD2LWae6io/xVXOiDxYpsWjWgoPXwhsu6WrgOoAFwaALz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/lXBoOwtQ0oqXHuJp3cWbz3cPYtZp7qKj/FVc6IPFimxZvT8etorln6qiZX22rqMDdURzrFAg+zqS7Vu6DF9kGF89g6LwqCRZmKtejbA8UUfYaLFcRyCsJimD3o7l2deBtw9i1mnuoqP8VVzog8WKbFm9D+fkpvWQkElEO9nJSBye89jJx1TuaX8DYLnTUPPtf48xTsIS6x46aRLFLfaJf5RxUrYC2vi0PEXIorKnOZkXD2LWae6io/xVXOiDxYpsWm8+vGr12PtbXHXGqiYSFOsjquSFT1swc0FYW0+29SJfPw5BQZ710XScJMJ8iAcg5kOeyWIus+IHCvPZp3VLMX8PYtZp7qKj/FVc6IPFimxYZeggrhLxEeQCS+ZHJzmBl7rvxFEkPm73XLIhqmqfuakQEUQVa2GyXwdsGsc6V38YOyrt5QQbPDz21eZeNgej5w9i1mnuoqP8VVzog8WKbFkIvWKehyGhpoY9iZRKv3x/6aLVuNRtIEEZ5UDIBq2t4z2DovCoJFmYq16NsDxRR9kW6gZMTX/OpjVEyvxtl/avD2LWae6io/xVXOiDxYpsWsu8BTLoMHflbwAjR/30h3fpotW41G0gQRnlQMgGra3jPYOi8KgkWZirXo2wPFFH2yM8dW1XRXewugQEkNKw568PYtZp7qKj/FVc6IPFimxYqmLod4PNaF3Mdg5l+34t/+mi1bjUbSBBGeVAyAatreM9g6LwqCRZmKtejbA8UUfYV2tusIK4QamB4nnWbXIdNw9i1mnuoqP8VVzog8WKbFqqocPWptdQJDGD0V+svppa89jJx1TuaX8DYLnTUPPtf48xTsIS6x46aRLFLfaJf5Sp2ytk0bG8Vm9L4cil8b+HD2LWae6io/xVXOiDxYpsW13juFWQlBJLAqnYL8C43rbz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/l87d2vicBL7tXo1BaFDpDysPYtZp7qKj/FVc6IPFimxavq9ymfoXOHjjrxMxqvtjs1pbMpQhYMfwf3N0hRzbOvePMU7CEuseOmkSxS32iX+UxB+9RVb1jX6L1+z8ZQNqmw9i1mnuoqP8VVzog8WKbFjVVM1ZQLGLzxIalPcWeeCe+DfdrTRKJhVUgtGb8mujWXxYUBhQEpwY1d3Cim+ZKYMeZALz0LtGd3fuGSn1R9NjD2LWae6io/xVXOiDxYpsWkfmuUUORDpG0+9bR3XNmP4Tor0w1ntGRFBW03CpMSh5j3RzsDKIQl1sRus21qeEC17dGdbW0N+mnd6Ih5MjCdsPYtZp7qKj/FVc6IPFimxYocgegQw9dSXX3VGAgC7BpvPYycdU7ml/A2C501Dz7X+PMU7CEuseOmkSxS32iX+UFLzcjoQUTjBSq8LJnw9SDw9i1mnuoqP8VVzog8WKbFnyWO0NISPEJdsqVVpxNcbuy8pjlaw9cFztfz3ozpbFJsUSjgb6YANx4ad5/d2xMFTiy5gAQhxSAGmnL8IjUe7jD2LWae6io/xVXOiDxYpsWSgpiDpQLsWRxUBTmsxne3oItPo+r1LI7lma/rWO6c3nqJJt5Nlncw4K6QYDwmBiHrQbrzdrSuGWZ5vBkkAQ9csPYtZp7qKj/FVc6IPFimxZuhNe7OQi3ISAzhinzrieuclDyir1DL1kY2YpOKsZ9Y8dAhfrVI428u0u54wQjAOzImMJmP76x9nOMZbgJGHlXw9i1mnuoqP8VVzog8WKbFm0WTfzyLssvZUpeNceQPf8OFV/GhfaA1EwzjgVa0mQNe56sn7eFNF1sXW/NQolQDlrLW9ZSLfVM92JbR+wRw1TD2LWae6io/xVXOiDxYpsWgz28C9XOfyvWGuAnhvDE1DWWoxocNesPUVgfJj/nqWhPJXCbO4bdoCd8PPNYr5yOOjoAI1XzkxWXM6wWAdTqk8PYtZp7qKj/FVc6IPFimxYaewruaDKqEDbyFUDs7oakhsCaB9HIpOScSoPZJk73YkCSdewck0gaE7vc5EvIrkkqtkcCV1IHOQwKB29rMlmxw9i1mnuoqP8VVzog8WKbFlXg4EhQUVDWmAviVnMPxx8RMnhm6QdihjZ4I51fZVzK48xTsIS6x46aRLFLfaJf5WDOTxbmN1nI4gJnx3FmFvLD2LWae6io/xVXOiDxYpsWnM17yFkZKOtBK7jHVXT48Lz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/lS3NtoekdrI0H58Ej045w1sPYtZp7qKj/FVc6IPFimxZQNqHrzwdXPCmijjRD7R1RM1xxQo83U/RcL/4LBOz9y7FEo4G+mADceGnef3dsTBVqAQY5UVkwd/CpSqgHnknJw9i1mnuoqP8VVzog8WKbFhRtaQp50gqTBYjn3TFtOJGrpRLlJOuIPFzr69y8mTkjJ3mIsZmjhCxc+lViCtu52JMJ19tpaS70LKPi/p4R6kvD2LWae6io/xVXOiDxYpsWPVx5sWLJW9Ew1wzNm5oG+o6p2l2vgMn1nwcqZmNdJtSxRKOBvpgA3Hhp3n93bEwVk/GDJYLE5TiEjTgwb5VnZcPYtZp7qKj/FVc6IPFimxaUpOg+A1wHGhjZycG+18mmWNcXlWaT5PA28G5J84PQz89g6LwqCRZmKtejbA8UUfbM5JJRGY+yzSImdg04B6Kdw9i1mnuoqP8VVzog8WKbFle26sR4vHbTwOL5Q3NSKtB9pPp5z04lmWFtUR4g4+wg48xTsIS6x46aRLFLfaJf5Z8x9NHI5kGTrdJ8081RDGDD2LWae6io/xVXOiDxYpsWveUwzL908Z9UfawtoaDoUBgBNAPRpIk8RtVmfVuMhZ/P0Q7kLGLy0ixe3gt7c+dEWuQSxErlue7fzVip+yja88PYtZp7qKj/FVc6IPFimxY4p3PdruRdh/UgumF8Mnhfo/Zo1xO1EW4AO9I0qsXidf8ZJjjBTjSdSLzS+Myvtoe7MZWcKVYYbMi8wPku2Lztw9i1mnuoqP8VVzog8WKbFgmYbNuEVA1e9T/lcxIDsrRG7h2qu679GTKNclRd4CyKGHqC2XbUtqks+HhricJ7m1PehOyxhojlecrm6aPH0NbD2LWae6io/xVXOiDxYpsWtrvpvSGNmWSV/Y52x+o8yKrUeAoa5qtehS9QfOmUzyyxRKOBvpgA3Hhp3n93bEwVXLmIdySU4siANiwG0xHS3cPYtZp7qKj/FVc6IPFimxZXEn08SBQI932p1ve+Z5JEdM/EK7jOxqnMdlBT+IGuOCnbB/va8UktX3N7USW2Q9gmAMdfQaAnkew5cGpHUmFRw9i1mnuoqP8VVzog8WKbFq6uQ3vJ+qgTf7vbKUcWd0J3oFcyvsHDo2v9vYZpvwsz/xkmOMFONJ1IvNL4zK+2h4bxHgRjk5ZVUAFUmnYsmxLD2LWae6io/xVXOiDxYpsWD/lB/aSqm7o1LE/yQNXU9a8WuNTfFGtiKhmeaAsx4ajPYOi8KgkWZirXo2wPFFH2RyXnT0iL5EdsQRp7+vlTiMPYtZp7qKj/FVc6IPFimxYWtDAG846hDCDPobb7F/1E3hvXB2UVqvhuSDiVjR5hqs9g6LwqCRZmKtejbA8UUfatuSRCKdK4uSvrBZtXSyZxw9i1mnuoqP8VVzog8WKbFjFLvIexI8KuN6dQZIMTgqlCGA6uXLit3GnT9ZWnRSiFz2DovCoJFmYq16NsDxRR9i3w8i2bixuEIDLPadjF6tnD2LWae6io/xVXOiDxYpsWU9G3RQVDFcOeTX/BQMISiYZ+oH/UtIOZCE8GxbxvY78YQBC0ZM8TNWY88IjsPhldBXVum53optfxsGvSabD3y8PYtZp7qKj/FVc6IPFimxZitV0fJANlbIWUy/I9APM1vPYycdU7ml/A2C501Dz7X+PMU7CEuseOmkSxS32iX+WWFi00Wg7B+TGc9sJIMH7Xw9i1mnuoqP8VVzog8WKbFh2pc0JQHiqHhOZMQqSNjjPDXzS3zEROI1f1Vq1u42mFsUSjgb6YANx4ad5/d2xMFZpoAH3R/t/GpUt1ZNPpbiDD2LWae6io/xVXOiDxYpsWfqtL2Pvdv2TOfyE6+uOAW7z2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/lh7DTcmvzq9mxmxWH3O6AT8PYtZp7qKj/FVc6IPFimxbfBzpttv2n9xdg+CB7N7OlnfN9dQ5fzNOMrmrhKCm+ySd5iLGZo4QsXPpVYgrbudguileYPPqh5IfHqzcumPGCw9i1mnuoqP8VVzog8WKbFlLuxyDOFz6/q5H9lI2z1bT5yV2tkluU21DUAhW1IHQW48xTsIS6x46aRLFLfaJf5e8ZkMgmu4mYlyXxpJRZLG/D2LWae6io/xVXOiDxYpsWYTikBMgWmyTas/NVrl2fTLz2MnHVO5pfwNgudNQ8+1/jzFOwhLrHjppEsUt9ol/lBVzeNyI7Vbq4BHMGXKHIv8PYtZp7qKj/FVc6IPFimxaThaGlTtz6VnCRLGrazbF1IKpuNC5cuTfcsoYzENmcg7FEo4G+mADceGnef3dsTBVTnnSlZKheBzeH7M/IH3uBw9i1mnuoqP8VVzog8WKbFoGDluJ7zeGCpfrGTk38EkqArf0GX8aYudl706YiYWEt48xTsIS6x46aRLFLfaJf5aRLOiIyMdnot5dtcKXt74LD2LWae6io/xVXOiDxYpsWGYCZlWh4Vc0MaVlUIrOWLGkYt0Bzfs8StJVw/9G7hP0p2wf72vFJLV9ze1EltkPYV0s/T4eMCUKWlme/fVvuAcPYtZp7qKj/FVc6IPFimxYd6+szZRqHG3PE1EeL2LeaVIItI+yDfnR28ZrTB2fdXfq3uXHtCNXEBqYZQfVa0o9/dXrsuIr4ooe2oP+Y1jIDw9i1mnuoqP8VVzog8WKbFiXEPX7jjK/aiKSs+q09Uf6QIxp4B7VQtlbn/EvRf4pE7SHTx+z/5nNkM9pobJbesozHINuc8jDshVTr2U9TM8DD2LWae6io/xVXOiDxYpsWCk9DtF0eZ9LkXhJbNTwrQJAjGngHtVC2Vuf8S9F/ikS7Bwd3jahAhFJYCmw5tsMnZnkCpTgeJdAeFHv0sSdsOMPYtZp7qKj/FVc6IPFimxYjQmKvH9im0nw2XQUIF2H3NaW5fx7gt22kIaQtBYYG3hmevzOeskPr1bQMksXScXvsKIPLOoagdn+bpDO9rCz9w9i1mnuoqP8VVzog8WKbFoMeHhVW67lDKsVc64TM0PC6kXurc+irzp/nv/7yFhtySYjQOqFRUyVpzp2wUj8oJeFM+yMhMMAbWj3ErmHdpTXD2LWae6io/xVXOiDxYpsWjAAmgNGw0NjYKOlyzjRNJ5JubaeIO5SqBSUSHPRyRYWxRKOBvpgA3Hhp3n93bEwVG0lfZATSc6DtcBkMZX1vdrFEo4G+mADceGnef3dsTBXcyO9YwcBYYSvABe8uhJGnWuMAnjkuJyWieR50mt/ZLM9g6LwqCRZmKtejbA8UUfaGpClg0110nDnQxClLXEG2sUSjgb6YANx4ad5/d2xMFRfl2t1AbrBEaYah2wlsfs1rIUWHapM5jmbNy12OLTCNz2DovCoJFmYq16NsDxRR9tprpBgf+j5nVpXO0p9/QaOxRKOBvpgA3Hhp3n93bEwVdaxmla10uu7sSC+y1X2EqMuKlSiQxyk2IVQoqMKNIhuaFhoqMSgmj7wCoXZfS4FMBDU3+lV6tvhYpvyf++UNeMPYtZp7qKj/FVc6IPFimxb3R7sHh3FVrstfmz3Ss+KpBlAIDSnLfWxWFU2fyDdV97FEo4G+mADceGnef3dsTBW6mIgmOP7G4NekzBN1zXuesUSjgb6YANx4ad5/d2xMFS902QTSdEmPMdQI27yTd5NI3kW+kjd/6no8tcQWn+N0J3mIsZmjhCxc+lViCtu52D0bRbEVH1jGgE+AXYTmtoqxRKOBvpgA3Hhp3n93bEwVp0tEhG6UnhcL0tAVI7ufI5Wx3yYETH8hgQTthGLucUbjzFOwhLrHjppEsUt9ol/lIE89jXj/LW9RbsZapNi9xrFEo4G+mADceGnef3dsTBVjkXEogqYnHcqCdZde33eI/F+AqfMaespootIJPc+2nf8ZJjjBTjSdSLzS+MyvtoeYEvs1JO2kFY08rw0kWmxFsUSjgb6YANx4ad5/d2xMFdL+xrBG/sOjhbPQbsN9KoiVsd8mBEx/IYEE7YRi7nFG48xTsIS6x46aRLFLfaJf5XD3ZXGqOpJ46ekb0OnhqC+xRKOBvpgA3Hhp3n93bEwVoGMlFykX9GFV/OYphvZQqDEWMI1Gs4KEUm+E64z1BTPnoTK/RpkZFYb8B0bgNnUADxxGg8X8WWXmEHVFZuG8JbFEo4G+mADceGnef3dsTBXxVhkYN+HQiP2J5DyJ68rKUNVeMv2vW7TILk7R8tkSAb6n/HY5CeGwFfQtPHjO20jPxPNbDK20KoZlc7kbiA+isUSjgb6YANx4ad5/d2xMFRO6JxGGX/+eO/+LwDVK6PWmR41NBBWIdbX4tKzo45WVdhKTmVMLV0hyuWri0ApWjGrqPxEpmFN2kmXegn18LjexRKOBvpgA3Hhp3n93bEwVv4LaJDdD0YCWnfMGuMwLCN0z2cMuFOhcKUbScfTtKqCxRKOBvpgA3Hhp3n93bEwVqOA59WD+1bEIUMhxg0xfpLFEo4G+mADceGnef3dsTBVLXqIgHvkJJu32f885MiFjHc90/1/lgTAe8K+obAES1P8ZJjjBTjSdSLzS+MyvtoeplftJsI3bHGbrIVFgv0CZsUSjgb6YANx4ad5/d2xMFR6mKulUnUXtyaA+XlL7BjKe64AJiOXSq8iFU7MoQgVKz2DovCoJFmYq16NsDxRR9mv2vgOK9UebYY5eYS6KN5KxRKOBvpgA3Hhp3n93bEwVbka3Y/mMfLDzaDYTESJqnhrfY4/iH4ENxoTWUEXDp3hCD9bVkTIagOGSCz61ZsbTThf5MUpF1FnryPKxC/LgTrFEo4G+mADceGnef3dsTBVH8skSqZ8PNBcgKSQzsp9paLtulUdrYaZme9RZ15GKVbFEo4G+mADceGnef3dsTBWI6bS172w+TTUTAIGoscE5sUSjgb6YANx4ad5/d2xMFSbN32sSfim+AE7G87UG62bWANEKXvAtYKGR0WDs/p7lJ3mIsZmjhCxc+lViCtu52Lbr01UMOKE4vMpw5zftX2ixRKOBvpgA3Hhp3n93bEwVpwBhLoXbqhYHjSNdmfnWCZRYPJpIRyV6UF3PR1Rn05rPYOi8KgkWZirXo2wPFFH2kO0IfAxORxX2RCkt3Q8n1LFEo4G+mADceGnef3dsTBUSxX18pVUzmU6igoxALJYUk7VgvNWXaHki8ZjFCnp1Ff8ZJjjBTjSdSLzS+MyvtofwDVbw/WfhfMd13mJo9VdvsUSjgb6YANx4ad5/d2xMFWXBSeBgRWBlkKz6mreO6sgB6cwk1v7hS9iYrJ9Hd9IEz2DovCoJFmYq16NsDxRR9ghuzr/dt9LF1qgiB3+AqFCxRKOBvpgA3Hhp3n93bEwVUsZlbtPF1DmWuYdqo1HNbBCDMXxzd7px2y2oFqO4HuTPYOi8KgkWZirXo2wPFFH25d+5Fpy7Sgak/ZRmiepoobFEo4G+mADceGnef3dsTBXAlRX7dVKOUZb9jamgxvVtgJ0NWMBFAOIqAccuISjxDomMPC7EMamFagDtwu1KIfimlTXiqD/Wo09dSxvhVJ3nw9i1mnuoqP8VVzog8WKbFt4YwiZ0AE6RqTSHT8PK/cReyjecdbtdbOvUqE6Y3UwlJ3mIsZmjhCxc+lViCtu52GmIsB5J5S5qfWoctbvjNwixRKOBvpgA3Hhp3n93bEwVTXJU4TU5ICbK7MIZ8PQhsaZHjU0EFYh1tfi0rOjjlZV2EpOZUwtXSHK5auLQClaMBA5fWBM0yXiIXDkoDdDkybFEo4G+mADceGnef3dsTBXTdL9XIE6ZaXJXvBFnevLQ76JmRP9OH+vgVTIWiYY8h/bLzW5rBSv238IuOYt+1r30y6wGmm3TLiRxQC0lkBxnsUSjgb6YANx4ad5/d2xMFcP78Bqu/OliPIATVr4sFt9swBRTljv8IWtWa5k+paehsUSjgb6YANx4ad5/d2xMFcOk9UhA4JBvkWc67YGcI3ixRKOBvpgA3Hhp3n93bEwVOY6mZe5mMdl29kHUIUJgs++iZkT/Th/r4FUyFomGPIe7Uqu4fhYwzdcx2EzIjcOCd2yLW3L5zEBurnq59k5PMrFEo4G+mADceGnef3dsTBWyqYsSs382/vyKk0DjRRQU+cldrZJblNtQ1AIVtSB0FuPMU7CEuseOmkSxS32iX+Ws0Tp7Clk5mQ+C3PAwm0DOsUSjgb6YANx4ad5/d2xMFWX4ch/a362r1PokL3rsIkrvomZE/04f6+BVMhaJhjyHBQG6MS2enqq8kXx4MFaxoO3N/EwRR6ni+rigfX/r0jOxRKOBvpgA3Hhp3n93bEwVQoRovA8nuK9uUzCLUvp4yPnJXa2SW5TbUNQCFbUgdBbjzFOwhLrHjppEsUt9ol/luFwwPRGbU3qpUz6yEB3IXrFEo4G+mADceGnef3dsTBUlUxt3bt5r/hU1FuMv29wpbMAUU5Y7/CFrVmuZPqWnobFEo4G+mADceGnef3dsTBXN4lbU9Bf0VaaS7gFyC3CFsUSjgb6YANx4ad5/d2xMFX636q8GjjKioLdGS5YUFddswBRTljv8IWtWa5k+paehsUSjgb6YANx4ad5/d2xMFbceScC8QmzgQrPYbo0Nk0qxRKOBvpgA3Hhp3n93bEwVUkRTpbrizVmHsEitUD0JnmzAFFOWO/wha1ZrmT6lp6GxRKOBvpgA3Hhp3n93bEwVe8QPMdqvasuYvYJVAkLkTLFEo4G+mADceGnef3dsTBXGd+XW6SiOLYNEACZtLQVebMAUU5Y7/CFrVmuZPqWnobFEo4G+mADceGnef3dsTBUTaJ9QcKuvma4RR0/VHyXDsUSjgb6YANx4ad5/d2xMFXMAQ1jy/89BQTioLPxGJPfFn5HF0Y8GhtZ/g5nARQbtmYqHqn5bLgft+Y6T6iISuwrZ2rO7ldLshUzmnuby+PaxRKOBvpgA3Hhp3n93bEwV8ZNcDFPlTiLQV+HzlW7rtfnJXa2SW5TbUNQCFbUgdBbjzFOwhLrHjppEsUt9ol/lLNNr/+QDC2YfXHdskV7rbrFEo4G+mADceGnef3dsTBVhGn3FOuPAknt3rCSztvAwkNdkRTSmRDq97cmRF5wdI+PMU7CEuseOmkSxS32iX+WaT2JkBarcKboRFehtpH4tsUSjgb6YANx4ad5/d2xMFYVfWb0JX0lUnqPo4v71ylqQ12RFNKZEOr3tyZEXnB0j48xTsIS6x46aRLFLfaJf5TtNmZM0x7gWzcNORtx+xZaxRKOBvpgA3Hhp3n93bEwVhdK3AuZUhtz2CLwkovvB68datIZdKG4YF7fTlxI7P2XPYOi8KgkWZirXo2wPFFH2/q4kM/ij1QKnunlJz/jbMrFEo4G+mADceGnef3dsTBXGXJ4Som52zu1RxQbFBUfLx1q0hl0obhgXt9OXEjs/Zc9g6LwqCRZmKtejbA8UUfZCbL8mAOhdBme2KTLAEYFmsUSjgb6YANx4ad5/d2xMFaXbq4PjOoGzXA+h9R7YDTDKjzuCKsHbLv/fUj8f0mGD48xTsIS6x46aRLFLfaJf5c2HChRj0j/hXi9QbwhY3POxRKOBvpgA3Hhp3n93bEwVJYPNN3nTTe4uRPA8ToNQf8qPO4Iqwdsu/99SPx/SYYPjzFOwhLrHjppEsUt9ol/ltyZARE8naxD++yyxxuVfabFEo4G+mADceGnef3dsTBVac5Cj+QUU+H2+TDzYS2iAr4e2yG3NhCOsmQR3r5WY9Sd6jbuhsY5PCy+55G/xFVesTVeDuNhC6vUDfSg0AlWYw9i1mnuoqP8VVzog8WKbFtbGyfVERyVtp4ECf4WBuMKz2BlnL0amqOo9tgHU+orVJ3mIsZmjhCxc+lViCtu52I34IogOxtChl+e89MCdYyexRKOBvpgA3Hhp3n93bEwVpWnF6Tz4uHZEZ4lkaNZBK9KzwJI3IpyoKvcBikbk2PkneYixmaOELFz6VWIK27nYeJekmV3sb3akavEbZq7xY7FEo4G+mADceGnef3dsTBXIJ/mYJWFsUFBAdZhlQgOoQhgOrly4rdxp0/WVp0Uohc9g6LwqCRZmKtejbA8UUfaYqetV5GZK2OL2ZEEc/9z8w9i1mnuoqP8VVzog8WKbFl+Yh4ZQspspy/FrNZv5G6wMykH9lIZSmoV8SVhfYB3l48xTsIS6x46aRLFLfaJf5UUZaWQYwdvAxRMVa8+qEFixRKOBvpgA3Hhp3n93bEwVggWZYdHg7Rs+FmblP8x+zQzKQf2UhlKahXxJWF9gHeXjzFOwhLrHjppEsUt9ol/l4nA5LNMXxrn5GNuHFfKuabFEo4G+mADceGnef3dsTBUUrw31YCieapvVRk0QYlFiDMpB/ZSGUpqFfElYX2Ad5ePMU7CEuseOmkSxS32iX+U0yW/xoqAAd80e5sDJhVFvsUSjgb6YANx4ad5/d2xMFYWxmBu7LVhcLTEC6yz6pVsMykH9lIZSmoV8SVhfYB3l48xTsIS6x46aRLFLfaJf5X14GrkwJycN8Nj6Bk2gpOSxRKOBvpgA3Hhp3n93bEwVhcEUJsk+d4wu0mqvop/ZJsdatIZdKG4YF7fTlxI7P2XPYOi8KgkWZirXo2wPFFH2H3d/i+0Cof8mu41vsSBO7LFEo4G+mADceGnef3dsTBUVxhSTXgdiL7O6JdCnE1H3N+P6SYoJRuHQ77rxRYIxaWDyOOFaCzqTtHEF2gZAJJwfie1mfH76vnlSAr2C4AjNsUSjgb6YANx4ad5/d2xMFSD2pKzk1z2fNH545fDfeHOHtIU8WCck/1T7daefACTQz2DovCoJFmYq16NsDxRR9jsTghj4yDhOMLFdG/45E/WxRKOBvpgA3Hhp3n93bEwVvyKTfvghlfSLMiMWG7CPJEz7orJAqIDjtrlerAi+I9TjzFOwhLrHjppEsUt9ol/lftFhFYz1KtHZSHfHhp5mbcPYtZp7qKj/FVc6IPFimxbsH2FkuXsgAlehcGbSuk6o+cldrZJblNtQ1AIVtSB0FuPMU7CEuseOmkSxS32iX+Xr9sXBx9DxcrhH/Ibga2hBw9i1mnuoqP8VVzog8WKbFn/6UaKqc6mh/vq4v6n2BlP6aLVuNRtIEEZ5UDIBq2t4z2DovCoJFmYq16NsDxRR9v/jrpztTRqF1XC9PRLpFNXD2LWae6io/xVXOiDxYpsWutMZn5E8l/8g+IxbvFJqLQzKQf2UhlKahXxJWF9gHeXjzFOwhLrHjppEsUt9ol/l11YJxN6Fn2tt3iyDLqXlW7FEo4G+mADceGnef3dsTBWcQUZF1J6NEHGdA66Tdlaukm5tp4g7lKoFJRIc9HJFhbFEo4G+mADceGnef3dsTBV8SQEt7RrARa13dxrT5wqLsUSjgb6YANx4ad5/d2xMFXChGRnlcSyF+Lu3lFbNl/9a4wCeOS4nJaJ5HnSa39ksz2DovCoJFmYq16NsDxRR9kn2Lu4W8aCAGoiqrqpP2eWxRKOBvpgA3Hhp3n93bEwVOSEdzhZc2CS6tOBGemrCDkIaH9HSWeHDXq1BkOEHEQ/jzFOwhLrHjppEsUt9ol/lRA5uyC4bAp4bMjkK51d9ULFEo4G+mADceGnef3dsTBXucNluIONQjXSCsoYzQV4DkNdkRTSmRDq97cmRF5wdI+PMU7CEuseOmkSxS32iX+XHl68WurTvGTeajeAsBeXisUSjgb6YANx4ad5/d2xMFZUgcpU2sRHq0uuh77idRX9WpxfQalJMrIUh7ncHV8JK48xTsIS6x46aRLFLfaJf5ZLIO9jmbV4R7zaYSvFA3BaxRKOBvpgA3Hhp3n93bEwVM/cp4BNBgC0S5PWvPJFLmKrUeAoa5qtehS9QfOmUzyyxRKOBvpgA3Hhp3n93bEwVaEaoMWzUt3GuVPnZOvSrF7FEo4G+mADceGnef3dsTBWjlTeCG3IccYTM7NBKyoV6WAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBXycHfUVdCu3PvY9jRsDK1dsUSjgb6YANx4ad5/d2xMFWftmHhaRtM3DnQfLQsPAPtYCHifmbloZxNCP6uSi0uisUSjgb6YANx4ad5/d2xMFYB3Enk8rkyao+Tpk99xDQGxRKOBvpgA3Hhp3n93bEwVSc2OSZeWX5bjndavrdhFKFgIeJ+ZuWhnE0I/q5KLS6KxRKOBvpgA3Hhp3n93bEwVTRTknj82nAxzfgP+BaDWCrFEo4G+mADceGnef3dsTBWkOmsqKx/z4SVMOIu6YebbWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBXCw5EW+Vq2a60ycq3Zo2l4sUSjgb6YANx4ad5/d2xMFVPftg3xJUI0bi6wvafEW+ZYCHifmbloZxNCP6uSi0uisUSjgb6YANx4ad5/d2xMFepzFk2diJFRt8BEn026mS+xRKOBvpgA3Hhp3n93bEwV8bmhEehwvyGPnn2Wo1HnR1gIeJ+ZuWhnE0I/q5KLS6KxRKOBvpgA3Hhp3n93bEwVjg640klWJSyebdtD2cNkUbFEo4G+mADceGnef3dsTBW8rPwQaTR0qQWVpUgynK4lWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBWNfHEE4Te+m0AgsnrGFnj7sUSjgb6YANx4ad5/d2xMFfJekrn+c/k4zDBfVobvbeZYCHifmbloZxNCP6uSi0uisUSjgb6YANx4ad5/d2xMFe37x0JtQ6VS2zs5kIebXxmxRKOBvpgA3Hhp3n93bEwVAI/aaDjqu+uRgsCGgzpRq1gIeJ+ZuWhnE0I/q5KLS6KxRKOBvpgA3Hhp3n93bEwV5coaeQynnX+qbg2mQsSjBLFEo4G+mADceGnef3dsTBVmkm2h4tg7G+uRNgg3N/TuWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBVb4OrPHGZvhfojwXDFMe0usUSjgb6YANx4ad5/d2xMFY96s4Hqz4fz90hxq9cPil5YCHifmbloZxNCP6uSi0uisUSjgb6YANx4ad5/d2xMFbc7ZSZiyK9naOk7r1aIwpqxRKOBvpgA3Hhp3n93bEwVMbIDjcV+jsNfaEjIXiAaMVgIeJ+ZuWhnE0I/q5KLS6KxRKOBvpgA3Hhp3n93bEwVHM5E5fpwwYmUXIioswHpsbFEo4G+mADceGnef3dsTBU3Kd8GK8gz0uOSTB7D07onWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBU4Grnw1tpyLUTQ6CWI4x9KsUSjgb6YANx4ad5/d2xMFZhj/HJbVpW6G7k6iIxqk57t07RJTJXuYPoFvLdxllXM48xTsIS6x46aRLFLfaJf5fT76k7+Usa4IwmwCEDmamuxRKOBvpgA3Hhp3n93bEwV1M7FWSujPM17Dq0Cm4dCBKKpBbyGaDPBpXw+6WjASN/jzFOwhLrHjppEsUt9ol/lKw6MQpUQbuZB406YiUjZa7FEo4G+mADceGnef3dsTBXVmeAJwFHQak8uNrW/nvBRWAh4n5m5aGcTQj+rkotLorFEo4G+mADceGnef3dsTBWI50g5L4WEQeaYTkp1LnOJsUSjgb6YANx4ad5/d2xMFVrzIaxOk6aAqbZmWkonJT24rfhyB3hEAAa7zVQTv/XCsUSjgb6YANx4ad5/d2xMFYTjmVbWE7sOyKnjA5wTL06xRKOBvpgA3Hhp3n93bEwVxYIfmueBT/YmrnJnnnA5S/nJXa2SW5TbUNQCFbUgdBbjzFOwhLrHjppEsUt9ol/la1HdX7UiRM3QLIMjoYYT9bFEo4G+mADceGnef3dsTBWCTPqGV8fOyFxZX4w64IGVBx0I2zSA0mp7S/41Rp/ULc9g6LwqCRZmKtejbA8UUfZTxnVQUnq1wiujJhaOPky2sUSjgb6YANx4ad5/d2xMFQQyAN2y7zldavzxQdC/AEVCGA6uXLit3GnT9ZWnRSiFz2DovCoJFmYq16NsDxRR9mML5IgftG/yqOZYV0WmyabD2LWae6io/xVXOiDxYpsWIXEyR07yinln6FlRTgi1LcJH6T7yZcE89SJ1EQez5IR6JHrD/Nuiufkesf7NViHYBDpSYeHXi77i89b6dK1i3+UKRZxsDf8CZaHHUfJGDTM=
non_process
jim kddducigrkdsf fokjnhgoerbeskmnq xvixf kfid kw xvxoidxypswkslrnvtcmlpa xvxoidxypswzht ugcqc opjveyvxtl uxb rcl xvxoidxypswtrvpvsgnmwsv iguocnbb xvxoidxypswd lb bqmisiyz oh t xvxoidxypswytikbmgwmytas aikss z irzp nv kjd f aqfmaespootijpc xrbg eo myvtofwdvbw izkt th rigfx llnnr quu i c aadjqu jsnfaejixiaamvgiej
0
21,676
30,121,075,256
IssuesEvent
2023-06-30 15:12:28
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Pty host last pty id is not restored when pty host is restarted
bug terminal-process
The `lastPtyId` on the main process was being updated here: https://github.com/microsoft/vscode/blob/cf30025bd71c48b0ede8943f10cb96b7f6fd82fe/src/vs/platform/terminal/node/ptyHostService.ts#L215 But since we now use a message port for `IPtyService.createProcess`, this isn't updated when creating a process anymore. So restarting the pty host will reset the lastPtyId to 0. I'm not making this a candidate as this is only an issue when the pty host is restarted which rarely happens. Found when looking into https://github.com/microsoft/vscode/issues/133542
1.0
Pty host last pty id is not restored when pty host is restarted - The `lastPtyId` on the main process was being updated here: https://github.com/microsoft/vscode/blob/cf30025bd71c48b0ede8943f10cb96b7f6fd82fe/src/vs/platform/terminal/node/ptyHostService.ts#L215 But since we now use a message port for `IPtyService.createProcess`, this isn't updated when creating a process anymore. So restarting the pty host will reset the lastPtyId to 0. I'm not making this a candidate as this is only an issue when the pty host is restarted which rarely happens. Found when looking into https://github.com/microsoft/vscode/issues/133542
process
pty host last pty id is not restored when pty host is restarted the lastptyid on the main process was being updated here but since we now use a message port for iptyservice createprocess this isn t updated when creating a process anymore so restarting the pty host will reset the lastptyid to i m not making this a candidate as this is only an issue when the pty host is restarted which rarely happens found when looking into
1
719,942
24,774,188,846
IssuesEvent
2022-10-23 14:24:37
NUS-Fintech-Society/SD_DAO
https://api.github.com/repos/NUS-Fintech-Society/SD_DAO
closed
Create User model
back end priority medium
Create User model with getter and setter functions for async display of info.
1.0
Create User model - Create User model with getter and setter functions for async display of info.
non_process
create user model create user model with getter and setter functions for async display of info
0
12,923
15,295,167,260
IssuesEvent
2021-02-24 04:10:37
topcoder-platform/community-app
https://api.github.com/repos/topcoder-platform/community-app
opened
CORS issues on beta environment
ShapeupProcess challenge- recommender-tool
Can you allowed beta-community-app.topcoder.com to received access from recommender-api url: beta-community-app.topcoder.com/challenges?bucket=openForRegistration&tracks[DS]=true&tracks[Des]=true&tracks[Dev]=true&tracks[QA]=true&types[]=CH&types[]=F2F&types[]=TSK ![Screen Shot 2021-02-24 at 10 54 03](https://user-images.githubusercontent.com/4476442/108946378-e1c3b600-7690-11eb-84ba-0d991b290d99.png)
1.0
CORS issues on beta environment - Can you allowed beta-community-app.topcoder.com to received access from recommender-api url: beta-community-app.topcoder.com/challenges?bucket=openForRegistration&tracks[DS]=true&tracks[Des]=true&tracks[Dev]=true&tracks[QA]=true&types[]=CH&types[]=F2F&types[]=TSK ![Screen Shot 2021-02-24 at 10 54 03](https://user-images.githubusercontent.com/4476442/108946378-e1c3b600-7690-11eb-84ba-0d991b290d99.png)
process
cors issues on beta environment can you allowed beta community app topcoder com to received access from recommender api url beta community app topcoder com challenges bucket openforregistration tracks true tracks true tracks true tracks true types ch types types tsk
1
4,670
7,504,522,213
IssuesEvent
2018-04-10 04:09:49
UnbFeelings/unb-feelings-docs
https://api.github.com/repos/UnbFeelings/unb-feelings-docs
closed
Descrever Processo de Garantia da Qualidade
Processo
O processo de garantia da qualidade não está representado visualmente nem possui as atividades descritas.
1.0
Descrever Processo de Garantia da Qualidade - O processo de garantia da qualidade não está representado visualmente nem possui as atividades descritas.
process
descrever processo de garantia da qualidade o processo de garantia da qualidade não está representado visualmente nem possui as atividades descritas
1
232,655
25,596,789,606
IssuesEvent
2022-12-01 16:47:01
Tim2023-workshop/NodeGoat
https://api.github.com/repos/Tim2023-workshop/NodeGoat
opened
mongodb-2.2.36.tgz: 1 vulnerabilities (highest severity is: 9.8)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mongodb-2.2.36.tgz</b></p></summary> <p>The official MongoDB driver for Node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/mongodb/-/mongodb-2.2.36.tgz">https://registry.npmjs.org/mongodb/-/mongodb-2.2.36.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/mongodb/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Tim2023-workshop/NodeGoat/commit/ec15c2097276b1e889c1398c868bc89ec1c82519">ec15c2097276b1e889c1398c868bc89ec1c82519</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (mongodb version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-7610](https://www.mend.io/vulnerability-database/CVE-2020-7610) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | bson-1.0.9.tgz | Transitive | 3.1.3 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7610</summary> ### Vulnerable Library - <b>bson-1.0.9.tgz</b></p> <p>A bson parser for node.js and the browser</p> <p>Library home page: <a href="https://registry.npmjs.org/bson/-/bson-1.0.9.tgz">https://registry.npmjs.org/bson/-/bson-1.0.9.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bson/package.json</p> <p> Dependency Hierarchy: - mongodb-2.2.36.tgz (Root Library) - mongodb-core-2.1.20.tgz - :x: **bson-1.0.9.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Tim2023-workshop/NodeGoat/commit/ec15c2097276b1e889c1398c868bc89ec1c82519">ec15c2097276b1e889c1398c868bc89ec1c82519</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> All versions of bson before 1.1.4 are vulnerable to Deserialization of Untrusted Data. The package will ignore an unknown value for an object's _bsotype, leading to cases where an object is serialized as a document rather than the intended BSON type. <p>Publish Date: 2020-03-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7610>CVE-2020-7610</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-04-01</p> <p>Fix Resolution (bson): 1.1.4</p> <p>Direct dependency fix Resolution (mongodb): 3.1.3</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
mongodb-2.2.36.tgz: 1 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mongodb-2.2.36.tgz</b></p></summary> <p>The official MongoDB driver for Node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/mongodb/-/mongodb-2.2.36.tgz">https://registry.npmjs.org/mongodb/-/mongodb-2.2.36.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/mongodb/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/Tim2023-workshop/NodeGoat/commit/ec15c2097276b1e889c1398c868bc89ec1c82519">ec15c2097276b1e889c1398c868bc89ec1c82519</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (mongodb version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2020-7610](https://www.mend.io/vulnerability-database/CVE-2020-7610) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | bson-1.0.9.tgz | Transitive | 3.1.3 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7610</summary> ### Vulnerable Library - <b>bson-1.0.9.tgz</b></p> <p>A bson parser for node.js and the browser</p> <p>Library home page: <a href="https://registry.npmjs.org/bson/-/bson-1.0.9.tgz">https://registry.npmjs.org/bson/-/bson-1.0.9.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bson/package.json</p> <p> Dependency Hierarchy: - mongodb-2.2.36.tgz (Root Library) - mongodb-core-2.1.20.tgz - :x: **bson-1.0.9.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Tim2023-workshop/NodeGoat/commit/ec15c2097276b1e889c1398c868bc89ec1c82519">ec15c2097276b1e889c1398c868bc89ec1c82519</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> All versions of bson before 1.1.4 are vulnerable to Deserialization of Untrusted Data. The package will ignore an unknown value for an object's _bsotype, leading to cases where an object is serialized as a document rather than the intended BSON type. <p>Publish Date: 2020-03-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-7610>CVE-2020-7610</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-04-01</p> <p>Fix Resolution (bson): 1.1.4</p> <p>Direct dependency fix Resolution (mongodb): 3.1.3</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_process
mongodb tgz vulnerabilities highest severity is vulnerable library mongodb tgz the official mongodb driver for node js library home page a href path to dependency file package json path to vulnerable library node modules mongodb package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in mongodb version remediation available high bson tgz transitive details cve vulnerable library bson tgz a bson parser for node js and the browser library home page a href path to dependency file package json path to vulnerable library node modules bson package json dependency hierarchy mongodb tgz root library mongodb core tgz x bson tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of bson before are vulnerable to deserialization of untrusted data the package will ignore an unknown value for an object s bsotype leading to cases where an object is serialized as a document rather than the intended bson type publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution bson direct dependency fix resolution mongodb rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
0
7,535
10,616,922,107
IssuesEvent
2019-10-12 15:22:20
KerasKorea/KerasObjectDetector
https://api.github.com/repos/KerasKorea/KerasObjectDetector
opened
[Preprocessing] Function for creating generators
preprocessing
- [ ] The function supports multiple benchmarks. - [ ] The result (the generator) supports simple photometric distortion.
1.0
[Preprocessing] Function for creating generators - - [ ] The function supports multiple benchmarks. - [ ] The result (the generator) supports simple photometric distortion.
process
function for creating generators the function supports multiple benchmarks the result the generator supports simple photometric distortion
1
20,365
27,022,076,018
IssuesEvent
2023-02-11 05:21:23
MikaylaFischler/cc-mek-scada
https://api.github.com/repos/MikaylaFischler/cc-mek-scada
closed
Process Target Energy Storage
supervisor coordinator process control
Coordinator should Coordinator should provide functionality for the user to request a specific quantity of energy to maintain in induction matrices. It will then attempt to satisfy all input requirements to maintain that charge level. - [x] Calculate moving averages of charge - [x] Adjust burn rate to reach target with closed loop control - [x] Test control loop logic - [x] Tune gains
1.0
Process Target Energy Storage - Coordinator should Coordinator should provide functionality for the user to request a specific quantity of energy to maintain in induction matrices. It will then attempt to satisfy all input requirements to maintain that charge level. - [x] Calculate moving averages of charge - [x] Adjust burn rate to reach target with closed loop control - [x] Test control loop logic - [x] Tune gains
process
process target energy storage coordinator should coordinator should provide functionality for the user to request a specific quantity of energy to maintain in induction matrices it will then attempt to satisfy all input requirements to maintain that charge level calculate moving averages of charge adjust burn rate to reach target with closed loop control test control loop logic tune gains
1
819
3,291,936,362
IssuesEvent
2015-10-30 12:04:47
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
closed
Should we still be shipping the built Cesium Viewer?
cleanup dev process
We currently ship with a built copy of the `Cesium Viewer` and even mention it as "ready for deployment" on our website's download page. This is misleading since the Cesium Viewer contains the default Bing key and can't be deployed as-is without violating usage policy. I'm not even sure how many people actually ever use the built version either (or why you would use it other than for development). Should we just stop shipping the built version? It would probably shrink our overall zip file size a bit as well. If we want to continue to ship the built version, we'll need a way to specify the Bing key via a configuration file that Cesium Viewer looks for.
1.0
Should we still be shipping the built Cesium Viewer? - We currently ship with a built copy of the `Cesium Viewer` and even mention it as "ready for deployment" on our website's download page. This is misleading since the Cesium Viewer contains the default Bing key and can't be deployed as-is without violating usage policy. I'm not even sure how many people actually ever use the built version either (or why you would use it other than for development). Should we just stop shipping the built version? It would probably shrink our overall zip file size a bit as well. If we want to continue to ship the built version, we'll need a way to specify the Bing key via a configuration file that Cesium Viewer looks for.
process
should we still be shipping the built cesium viewer we currently ship with a built copy of the cesium viewer and even mention it as ready for deployment on our website s download page this is misleading since the cesium viewer contains the default bing key and can t be deployed as is without violating usage policy i m not even sure how many people actually ever use the built version either or why you would use it other than for development should we just stop shipping the built version it would probably shrink our overall zip file size a bit as well if we want to continue to ship the built version we ll need a way to specify the bing key via a configuration file that cesium viewer looks for
1
15,877
20,053,171,769
IssuesEvent
2022-02-03 09:15:07
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
opened
Remove the `CanProcessMessage` method from the `MessageHandler` module
area:message-processing breaking-change
**Is your feature request related to a problem? Please describe.** When processing messages, received from message pumps or other systems, we use a `MessageHandler` module (static class) to make sure that we can extract open interfaces from the `IServceProvider`. One of these module functions is called `CanProcessMessage` which validates if a `IMessageHandler` can process based on a messaging context. We have renamed this method to `CanProcessMessageBasedOnContext` because there were also other criteria to filter upon, making the method obsolete. **Describe the solution you'd like** Remove the `CanProcessMessage` method from the `MessageHandler` module.
1.0
Remove the `CanProcessMessage` method from the `MessageHandler` module - **Is your feature request related to a problem? Please describe.** When processing messages, received from message pumps or other systems, we use a `MessageHandler` module (static class) to make sure that we can extract open interfaces from the `IServceProvider`. One of these module functions is called `CanProcessMessage` which validates if a `IMessageHandler` can process based on a messaging context. We have renamed this method to `CanProcessMessageBasedOnContext` because there were also other criteria to filter upon, making the method obsolete. **Describe the solution you'd like** Remove the `CanProcessMessage` method from the `MessageHandler` module.
process
remove the canprocessmessage method from the messagehandler module is your feature request related to a problem please describe when processing messages received from message pumps or other systems we use a messagehandler module static class to make sure that we can extract open interfaces from the iservceprovider one of these module functions is called canprocessmessage which validates if a imessagehandler can process based on a messaging context we have renamed this method to canprocessmessagebasedoncontext because there were also other criteria to filter upon making the method obsolete describe the solution you d like remove the canprocessmessage method from the messagehandler module
1
381
3,413,337,417
IssuesEvent
2015-12-06 16:24:56
spyder-ide/spyder
https://api.github.com/repos/spyder-ide/spyder
closed
Create conda.recipe folder at repo level
Enhancement Maintainability
Most project out there that use conda, have the recipe embedded directly in the repo in a `conda.recipe` folder. I think is a good idea to have the same.
True
Create conda.recipe folder at repo level - Most project out there that use conda, have the recipe embedded directly in the repo in a `conda.recipe` folder. I think is a good idea to have the same.
non_process
create conda recipe folder at repo level most project out there that use conda have the recipe embedded directly in the repo in a conda recipe folder i think is a good idea to have the same
0
337,119
10,210,795,168
IssuesEvent
2019-08-14 15:29:07
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
BigQuery: query_parameters fails if None is bound as parameter
api: bigquery priority: p2 type: bug
OS Type & Version: Ubuntu 19.04 x64 Python version: Python 3.7.3 Packges: latest up to this date: ``` 'google-cloud-bigquery==1.18.0', ``` **Steps to reproduce** 1. Create a query, bind `None` (`NULL`) as parameter 2. Execute it 3. Call query_parameters **Code example** ```py from google.cloud import bigquery client = bigquery.Client.from_service_account_json( <...> ) job = client.query( "SELECT LOWER(@none_value)", job_config=bigquery.QueryJobConfig( query_parameters=[ bigquery.ScalarQueryParameter('none_value', 'STRING', None) ] ) ) result = list(job.result()) query_parameters = job.query_parameters ``` **Stack trace** ``` Traceback (most recent call last): File "test.py", line 16, in <module> query_parameters = job.query_parameters File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/job.py", line 2472, in query_parameters return self._configuration.query_parameters File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/job.py", line 2200, in query_parameters return _from_api_repr_query_parameters(prop) File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/job.py", line 1965, in _from_api_repr_query_parameters return [_query_param_from_api_repr(mapping) for mapping in resource] File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/job.py", line 1965, in <listcomp> return [_query_param_from_api_repr(mapping) for mapping in resource] File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/query.py", line 625, in _query_param_from_api_repr return klass.from_api_repr(resource) File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/query.py", line 129, in from_api_repr value = resource["parameterValue"]["value"] KeyError: 'parameterValue' ``` This is related to https://github.com/googleapis/google-cloud-python/issues/7309
1.0
BigQuery: query_parameters fails if None is bound as parameter - OS Type & Version: Ubuntu 19.04 x64 Python version: Python 3.7.3 Packges: latest up to this date: ``` 'google-cloud-bigquery==1.18.0', ``` **Steps to reproduce** 1. Create a query, bind `None` (`NULL`) as parameter 2. Execute it 3. Call query_parameters **Code example** ```py from google.cloud import bigquery client = bigquery.Client.from_service_account_json( <...> ) job = client.query( "SELECT LOWER(@none_value)", job_config=bigquery.QueryJobConfig( query_parameters=[ bigquery.ScalarQueryParameter('none_value', 'STRING', None) ] ) ) result = list(job.result()) query_parameters = job.query_parameters ``` **Stack trace** ``` Traceback (most recent call last): File "test.py", line 16, in <module> query_parameters = job.query_parameters File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/job.py", line 2472, in query_parameters return self._configuration.query_parameters File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/job.py", line 2200, in query_parameters return _from_api_repr_query_parameters(prop) File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/job.py", line 1965, in _from_api_repr_query_parameters return [_query_param_from_api_repr(mapping) for mapping in resource] File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/job.py", line 1965, in <listcomp> return [_query_param_from_api_repr(mapping) for mapping in resource] File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/query.py", line 625, in _query_param_from_api_repr return klass.from_api_repr(resource) File "/test/venv/lib/python3.7/site-packages/google/cloud/bigquery/query.py", line 129, in from_api_repr value = resource["parameterValue"]["value"] KeyError: 'parameterValue' ``` This is related to https://github.com/googleapis/google-cloud-python/issues/7309
non_process
bigquery query parameters fails if none is bound as parameter os type version ubuntu python version python packges latest up to this date google cloud bigquery steps to reproduce create a query bind none null as parameter execute it call query parameters code example py from google cloud import bigquery client bigquery client from service account json job client query select lower none value job config bigquery queryjobconfig query parameters bigquery scalarqueryparameter none value string none result list job result query parameters job query parameters stack trace traceback most recent call last file test py line in query parameters job query parameters file test venv lib site packages google cloud bigquery job py line in query parameters return self configuration query parameters file test venv lib site packages google cloud bigquery job py line in query parameters return from api repr query parameters prop file test venv lib site packages google cloud bigquery job py line in from api repr query parameters return file test venv lib site packages google cloud bigquery job py line in return file test venv lib site packages google cloud bigquery query py line in query param from api repr return klass from api repr resource file test venv lib site packages google cloud bigquery query py line in from api repr value resource keyerror parametervalue this is related to
0
585,003
17,468,591,098
IssuesEvent
2021-08-06 21:05:41
sonia-auv/proc_image_processing
https://api.github.com/repos/sonia-auv/proc_image_processing
opened
Accelerate the execution of PipeAngleDetector
Priority: Medium Type: Enhancement
## Expected Enhancement The execution of PipeAngleDetector should take less than 0,07 seconds ## Current Behavior PipeAngleDetector sometimes take several seconds to execute (seen up to 9s on my machine) ## Possible Solution - Optimize the implementation - Use parallel_for_ from OpenCV when appropriate
1.0
Accelerate the execution of PipeAngleDetector - ## Expected Enhancement The execution of PipeAngleDetector should take less than 0,07 seconds ## Current Behavior PipeAngleDetector sometimes take several seconds to execute (seen up to 9s on my machine) ## Possible Solution - Optimize the implementation - Use parallel_for_ from OpenCV when appropriate
non_process
accelerate the execution of pipeangledetector expected enhancement the execution of pipeangledetector should take less than seconds current behavior pipeangledetector sometimes take several seconds to execute seen up to on my machine possible solution optimize the implementation use parallel for from opencv when appropriate
0
11,100
13,941,258,851
IssuesEvent
2020-10-22 19:08:03
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Torch import fail
module: multiprocessing triaged
## 🐛 Bug When trying to import torch from a fresh install, it fails with the error message ``` Traceback (most recent call last): File "run_torch.py", line 1, in <module> import torch File "/export/tmp/henwood/anaconda3/lib/python3.8/site-packages/torch/__init__.py", line 480, in <module> import torch.multiprocessing File "/export/tmp/henwood/anaconda3/lib/python3.8/site-packages/torch/multiprocessing/__init__.py", line 18, in <module> from .reductions import init_reductions File "/export/tmp/henwood/anaconda3/lib/python3.8/site-packages/torch/multiprocessing/reductions.py", line 7, in <module> from multiprocessing.util import register_after_fork ModuleNotFoundError: No module named 'multiprocessing.util'; 'multiprocessing' is not a package ``` <!-- A clear and concise description of what the bug is. --> ## To Reproduce Steps to reproduce the behavior: 1. Install and import torch 2. Notice the error <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior The import should success. <!-- A clear and concise description of what you expected to happen. --> ## Environment Collecting environment information... PyTorch version: 1.6.0 Is debug build: No CUDA used to build PyTorch: 10.2 OS: CentOS Linux release 7.8.2003 (Core) GCC version: (Homebrew GCC 5.5.0_7) 5.5.0 CMake version: version 3.18.1 Python version: 3.8 Is CUDA available: Yes CUDA runtime version: 10.2.89 GPU models and configuration: GPU 0: Tesla P100-SXM2-16GB GPU 1: Tesla P100-SXM2-16GB GPU 2: Tesla P100-SXM2-16GB GPU 3: Tesla P100-SXM2-16GB Nvidia driver version: 440.33.01 cuDNN version: Could not collect Versions of relevant libraries: [pip] Could not collect [conda] blas 1.0 mkl [conda] mkl 2020.2 256 [conda] mkl-service 2.3.0 py38he904b0f_0 [conda] mkl_fft 1.2.0 py38h23d657b_0 [conda] mkl_random 1.1.1 py38h0573a6f_0 [conda] pytorch 1.6.0 py3.8_cuda10.2.89_cudnn7.6.5_0 pytorch [conda] torchvision 0.7.0 py38_cu102 pytorch
1.0
Torch import fail - ## 🐛 Bug When trying to import torch from a fresh install, it fails with the error message ``` Traceback (most recent call last): File "run_torch.py", line 1, in <module> import torch File "/export/tmp/henwood/anaconda3/lib/python3.8/site-packages/torch/__init__.py", line 480, in <module> import torch.multiprocessing File "/export/tmp/henwood/anaconda3/lib/python3.8/site-packages/torch/multiprocessing/__init__.py", line 18, in <module> from .reductions import init_reductions File "/export/tmp/henwood/anaconda3/lib/python3.8/site-packages/torch/multiprocessing/reductions.py", line 7, in <module> from multiprocessing.util import register_after_fork ModuleNotFoundError: No module named 'multiprocessing.util'; 'multiprocessing' is not a package ``` <!-- A clear and concise description of what the bug is. --> ## To Reproduce Steps to reproduce the behavior: 1. Install and import torch 2. Notice the error <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior The import should success. <!-- A clear and concise description of what you expected to happen. --> ## Environment Collecting environment information... PyTorch version: 1.6.0 Is debug build: No CUDA used to build PyTorch: 10.2 OS: CentOS Linux release 7.8.2003 (Core) GCC version: (Homebrew GCC 5.5.0_7) 5.5.0 CMake version: version 3.18.1 Python version: 3.8 Is CUDA available: Yes CUDA runtime version: 10.2.89 GPU models and configuration: GPU 0: Tesla P100-SXM2-16GB GPU 1: Tesla P100-SXM2-16GB GPU 2: Tesla P100-SXM2-16GB GPU 3: Tesla P100-SXM2-16GB Nvidia driver version: 440.33.01 cuDNN version: Could not collect Versions of relevant libraries: [pip] Could not collect [conda] blas 1.0 mkl [conda] mkl 2020.2 256 [conda] mkl-service 2.3.0 py38he904b0f_0 [conda] mkl_fft 1.2.0 py38h23d657b_0 [conda] mkl_random 1.1.1 py38h0573a6f_0 [conda] pytorch 1.6.0 py3.8_cuda10.2.89_cudnn7.6.5_0 pytorch [conda] torchvision 0.7.0 py38_cu102 pytorch
process
torch import fail 🐛 bug when trying to import torch from a fresh install it fails with the error message traceback most recent call last file run torch py line in import torch file export tmp henwood lib site packages torch init py line in import torch multiprocessing file export tmp henwood lib site packages torch multiprocessing init py line in from reductions import init reductions file export tmp henwood lib site packages torch multiprocessing reductions py line in from multiprocessing util import register after fork modulenotfounderror no module named multiprocessing util multiprocessing is not a package to reproduce steps to reproduce the behavior install and import torch notice the error expected behavior the import should success environment collecting environment information pytorch version is debug build no cuda used to build pytorch os centos linux release core gcc version homebrew gcc cmake version version python version is cuda available yes cuda runtime version gpu models and configuration gpu tesla gpu tesla gpu tesla gpu tesla nvidia driver version cudnn version could not collect versions of relevant libraries could not collect blas mkl mkl mkl service mkl fft mkl random pytorch pytorch torchvision pytorch
1
5,448
8,307,898,538
IssuesEvent
2018-09-23 14:55:25
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
[Process] Option to disable handling signals allowing custom signal handling.
Feature Process
**Description** A option to disable handling signals allowing custom signal handling. **Example** ```php $process->setSignalHandler(SIGINT, $this->sigIntHandler); ``` **OR** ```php $process->handleSignals(false); ``` **The Concept** So the idea is if a user presses `CTRL + c` it will NOT interrupt the Symfony Process but rather defer to a custom handler. Example: ```php public function __construct() { pcntl_signal(SIGINT, [$this, 'sigIntHandler']); } protected function command($command) { $process = new SymfonyProcess($command); $process->handleSignals(false); $process->start(); ... $process->wait(function ($type, $buffer) use ($bar) { // ... DO STUFF $bar->advance(); usleep(200000); }); // we wait until the current process finishes and then dispatch any signals pcntl_signal_dispatch(); } protected function sigIntHandler() { if ($this->io->confirm('Are you sure you want to abort', false)) { die(); } } ```
1.0
[Process] Option to disable handling signals allowing custom signal handling. - **Description** A option to disable handling signals allowing custom signal handling. **Example** ```php $process->setSignalHandler(SIGINT, $this->sigIntHandler); ``` **OR** ```php $process->handleSignals(false); ``` **The Concept** So the idea is if a user presses `CTRL + c` it will NOT interrupt the Symfony Process but rather defer to a custom handler. Example: ```php public function __construct() { pcntl_signal(SIGINT, [$this, 'sigIntHandler']); } protected function command($command) { $process = new SymfonyProcess($command); $process->handleSignals(false); $process->start(); ... $process->wait(function ($type, $buffer) use ($bar) { // ... DO STUFF $bar->advance(); usleep(200000); }); // we wait until the current process finishes and then dispatch any signals pcntl_signal_dispatch(); } protected function sigIntHandler() { if ($this->io->confirm('Are you sure you want to abort', false)) { die(); } } ```
process
option to disable handling signals allowing custom signal handling description a option to disable handling signals allowing custom signal handling example php process setsignalhandler sigint this siginthandler or php process handlesignals false the concept so the idea is if a user presses ctrl c it will not interrupt the symfony process but rather defer to a custom handler example php public function construct pcntl signal sigint protected function command command process new symfonyprocess command process handlesignals false process start process wait function type buffer use bar do stuff bar advance usleep we wait until the current process finishes and then dispatch any signals pcntl signal dispatch protected function siginthandler if this io confirm are you sure you want to abort false die
1
19,684
26,034,182,560
IssuesEvent
2022-12-22 01:59:36
fluent/fluent-bit
https://api.github.com/repos/fluent/fluent-bit
closed
AWS Metadata Filter Plugin for EC2/ECS Metadata
work-in-process AWS Stale
I am considering building an 'AWS Metadata' plugin. It'd query the EC2 and ECS metadata services and add useful metadata to log records. For EC2, it'd use the new IMDSv2, since it is more secure if you're running applications on your instance that are exposed on the public internet. Basically, replicate the functionality of these Fluentd plugins: - https://github.com/takus/fluent-plugin-ec2-metadata - https://github.com/michaelgruber/fluent-plugin-ecs-metadata-filter EC2 metadata will be prioritized and implemented first. Do you think this feature would be useful? Please plus one this issue. Do you have requests for what it will support or how it will work? Please comment.
1.0
AWS Metadata Filter Plugin for EC2/ECS Metadata - I am considering building an 'AWS Metadata' plugin. It'd query the EC2 and ECS metadata services and add useful metadata to log records. For EC2, it'd use the new IMDSv2, since it is more secure if you're running applications on your instance that are exposed on the public internet. Basically, replicate the functionality of these Fluentd plugins: - https://github.com/takus/fluent-plugin-ec2-metadata - https://github.com/michaelgruber/fluent-plugin-ecs-metadata-filter EC2 metadata will be prioritized and implemented first. Do you think this feature would be useful? Please plus one this issue. Do you have requests for what it will support or how it will work? Please comment.
process
aws metadata filter plugin for ecs metadata i am considering building an aws metadata plugin it d query the and ecs metadata services and add useful metadata to log records for it d use the new since it is more secure if you re running applications on your instance that are exposed on the public internet basically replicate the functionality of these fluentd plugins metadata will be prioritized and implemented first do you think this feature would be useful please plus one this issue do you have requests for what it will support or how it will work please comment
1
17,543
23,355,978,400
IssuesEvent
2022-08-10 07:27:09
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
db pull --url with an invalid url produces a misleading error message
bug/2-confirmed kind/bug process/candidate topic: introspection tech/typescript tech/engines/introspection engine
When you use `prisma db pull --url` and provide an invalid URL, the error message is the following: ``` ✖ Introspecting Error: P1012 Introspection failed as your current Prisma schema file is invalid Please fix your current schema manually, use prisma validate to confirm it is valid and then run this command again. Or run this command with the --force flag to ignore your current schema and overwrite it. All local modifications will be lost. ``` Even if your local Prisma schema is valid. The error message should instead say that the url passed through the `--url` argument is not valid. ### Steps to reproduce In a project with a valid SQLite Prisma schema: - Run `prisma validate` to confirm that the schema is valid - Run `prisma db pull --url dev.db`. Observe the error above. - Run `prisma db pull --url file:dev.db`. Observe introspection working fine.
1.0
db pull --url with an invalid url produces a misleading error message - When you use `prisma db pull --url` and provide an invalid URL, the error message is the following: ``` ✖ Introspecting Error: P1012 Introspection failed as your current Prisma schema file is invalid Please fix your current schema manually, use prisma validate to confirm it is valid and then run this command again. Or run this command with the --force flag to ignore your current schema and overwrite it. All local modifications will be lost. ``` Even if your local Prisma schema is valid. The error message should instead say that the url passed through the `--url` argument is not valid. ### Steps to reproduce In a project with a valid SQLite Prisma schema: - Run `prisma validate` to confirm that the schema is valid - Run `prisma db pull --url dev.db`. Observe the error above. - Run `prisma db pull --url file:dev.db`. Observe introspection working fine.
process
db pull url with an invalid url produces a misleading error message when you use prisma db pull url and provide an invalid url the error message is the following ✖ introspecting error introspection failed as your current prisma schema file is invalid please fix your current schema manually use prisma validate to confirm it is valid and then run this command again or run this command with the force flag to ignore your current schema and overwrite it all local modifications will be lost even if your local prisma schema is valid the error message should instead say that the url passed through the url argument is not valid steps to reproduce in a project with a valid sqlite prisma schema run prisma validate to confirm that the schema is valid run prisma db pull url dev db observe the error above run prisma db pull url file dev db observe introspection working fine
1
17,383
23,200,958,748
IssuesEvent
2022-08-01 21:25:03
googleapis/google-cloud-cpp
https://api.github.com/repos/googleapis/google-cloud-cpp
closed
[testing] track impact of presubmit coverage being < CI coverage
type: process
presubmit only runs a subset of the full CI tests. This was an intentional tradeoff the team made between build time and coverage. However, it does mean that the CI run can fail even on "green" PRs. Our tooling identifies this as a "flake" since it's a build break, although it's not actually a flake, it's deterministic. However, it is still an issue - it makes our repo "red" until a fix is submitted, which potentially impacts customers who are trying to use the repo at `HEAD`, and too many CI failures may be interpreted by users to indicate "low-quality" or process problems. We had a team discussion about this last week; this PR is to track occurrences and discuss potential mitigation strategies. Recent occurrences: * today (Nov 9) in #5454 (fixed in #5455) * last week (Nov 6) in #5406 (fixed in #5435). I didn't look for prior examples, but I will look into it and add data here if it's not too time consuming.
1.0
[testing] track impact of presubmit coverage being < CI coverage - presubmit only runs a subset of the full CI tests. This was an intentional tradeoff the team made between build time and coverage. However, it does mean that the CI run can fail even on "green" PRs. Our tooling identifies this as a "flake" since it's a build break, although it's not actually a flake, it's deterministic. However, it is still an issue - it makes our repo "red" until a fix is submitted, which potentially impacts customers who are trying to use the repo at `HEAD`, and too many CI failures may be interpreted by users to indicate "low-quality" or process problems. We had a team discussion about this last week; this PR is to track occurrences and discuss potential mitigation strategies. Recent occurrences: * today (Nov 9) in #5454 (fixed in #5455) * last week (Nov 6) in #5406 (fixed in #5435). I didn't look for prior examples, but I will look into it and add data here if it's not too time consuming.
process
track impact of presubmit coverage being ci coverage presubmit only runs a subset of the full ci tests this was an intentional tradeoff the team made between build time and coverage however it does mean that the ci run can fail even on green prs our tooling identifies this as a flake since it s a build break although it s not actually a flake it s deterministic however it is still an issue it makes our repo red until a fix is submitted which potentially impacts customers who are trying to use the repo at head and too many ci failures may be interpreted by users to indicate low quality or process problems we had a team discussion about this last week this pr is to track occurrences and discuss potential mitigation strategies recent occurrences today nov in fixed in last week nov in fixed in i didn t look for prior examples but i will look into it and add data here if it s not too time consuming
1
253,449
8,056,260,104
IssuesEvent
2018-08-02 12:07:59
openshiftio/openshift.io
https://api.github.com/repos/openshiftio/openshift.io
closed
Inconsistent output from the license analysis service on prod-preview
SEV2-high area/analytics priority/P2 status/completed team/analytics type/bug
# Problem description The license analysis service returns incorrect data with the status set to `failure`. The output is different from one used in documentation for the dependency editor - so either the service has a bug or the documentation is not consistent w.r.o. the request/expected response. ## Documentation https://docs.google.com/document/d/1PIbir8p6IE1fLYpEpP9EoHRpjZ1MwR4j8luHFKar8XI/edit ## Input Post the following payload taken from the *Request and Response doc for dependency editor* to the analytics_license service on `prod-preview`: ```{ "_resolved": [ { "package": "com.googlecode.xmemcached:xmemcached", "version": "2.3.2" }, { "package": "commons-fileupload:commons-fileupload", "version": "1.3" }, { "package": "org.springframework.boot:spring-boot-starter-web", "version": "1.4.1.RELEASE" }, { "package": "com.h2database:h2", "version": "1.4.192" }, { "package": "org.springframework.boot:spring-boot-starter-data-jpa", "version": "1.4.1.RELEASE" } ], "ecosystem": "maven" } ``` ## Output from the service ``` { "conflict_packages": [], "distinct_licenses": [ "apache 2.0" ], "message": "No declared licenses found for 3 component(s).", "outlier_packages": [], "packages": [ { "license_analysis": { "_message": "Representative license found", "_representative_licenses": "apache 2.0", "conflict_licenses": [], "outlier_licenses": [], "status": "Successful", "synonyms": { "Apache License, Version 2.0": "apache 2.0" }, "unknown_licenses": [] }, "licenses": [ "Apache License, Version 2.0" ], "package": "com.googlecode.xmemcached:xmemcached", "version": "2.3.2" }, { "license_analysis": { "_message": "Representative license found", "_representative_licenses": "apache 2.0", "conflict_licenses": [], "outlier_licenses": [], "status": "Successful", "synonyms": { "Apache 2.0": "apache 2.0" }, "unknown_licenses": [] }, "licenses": [ "Apache 2.0" ], "package": "commons-fileupload:commons-fileupload", "version": "1.3" }, { "license_analysis": { "_message": "Input is invalid", "_representative_licenses": null, "conflict_licenses": [], "outlier_licenses": [], "status": "Failure", "synonyms": {}, "unknown_licenses": [] }, "licenses": [], "package": "org.springframework.boot:spring-boot-starter-web", "version": "1.4.1.RELEASE" }, { "license_analysis": { "_message": "Input is invalid", "_representative_licenses": null, "conflict_licenses": [], "outlier_licenses": [], "status": "Failure", "synonyms": {}, "unknown_licenses": [] }, "licenses": [], "package": "com.h2database:h2", "version": "1.4.192" }, { "license_analysis": { "_message": "Input is invalid", "_representative_licenses": null, "conflict_licenses": [], "outlier_licenses": [], "status": "Failure", "synonyms": {}, "unknown_licenses": [] }, "licenses": [], "package": "org.springframework.boot:spring-boot-starter-data-jpa", "version": "1.4.1.RELEASE" } ], "stack_license": null, "status": "Failure", "unknown_licenses": { "component_conflict": [], "really_unknown": [] } } ``` ## Expected output ``` { "conflict_packages": [], "license_filter": { "alternate_packages": { "compatible_packages": [], "conflict_packages": [], "unknown_license_packages": [] }, "companion_packages": { "compatible_packages": [], "conflict_packages": [], "unknown_license_packages": [] } }, "outlier_packages": {}, "packages": [ { "license_analysis": { "_message": "Representative license found", "_representative_licenses": "epl 1.0", "conflict_licenses": [], "outlier_licenses": [], "status": "Successful", "synonyms": { "APACHE": "apache 2.0", "Eclipse Public License": "epl 1.0" }, "unknown_licenses": [] }, "licenses": [ "APACHE", "Eclipse Public License" ], "package": "p1", "version": "1.1" }, { "license_analysis": { "_message": "Representative license found", "_representative_licenses": "gplv2", "conflict_licenses": [], "outlier_licenses": [], "status": "Successful", "synonyms": { "BSD": "bsd-new", "GPL V2": "gplv2" }, "unknown_licenses": [] }, "licenses": [ "BSD", "GPL V2" ], "package": "p2", "version": "1.1" } ], "stack_license": "gplv2", "status": "Successful" } ```
1.0
Inconsistent output from the license analysis service on prod-preview - # Problem description The license analysis service returns incorrect data with the status set to `failure`. The output is different from one used in documentation for the dependency editor - so either the service has a bug or the documentation is not consistent w.r.o. the request/expected response. ## Documentation https://docs.google.com/document/d/1PIbir8p6IE1fLYpEpP9EoHRpjZ1MwR4j8luHFKar8XI/edit ## Input Post the following payload taken from the *Request and Response doc for dependency editor* to the analytics_license service on `prod-preview`: ```{ "_resolved": [ { "package": "com.googlecode.xmemcached:xmemcached", "version": "2.3.2" }, { "package": "commons-fileupload:commons-fileupload", "version": "1.3" }, { "package": "org.springframework.boot:spring-boot-starter-web", "version": "1.4.1.RELEASE" }, { "package": "com.h2database:h2", "version": "1.4.192" }, { "package": "org.springframework.boot:spring-boot-starter-data-jpa", "version": "1.4.1.RELEASE" } ], "ecosystem": "maven" } ``` ## Output from the service ``` { "conflict_packages": [], "distinct_licenses": [ "apache 2.0" ], "message": "No declared licenses found for 3 component(s).", "outlier_packages": [], "packages": [ { "license_analysis": { "_message": "Representative license found", "_representative_licenses": "apache 2.0", "conflict_licenses": [], "outlier_licenses": [], "status": "Successful", "synonyms": { "Apache License, Version 2.0": "apache 2.0" }, "unknown_licenses": [] }, "licenses": [ "Apache License, Version 2.0" ], "package": "com.googlecode.xmemcached:xmemcached", "version": "2.3.2" }, { "license_analysis": { "_message": "Representative license found", "_representative_licenses": "apache 2.0", "conflict_licenses": [], "outlier_licenses": [], "status": "Successful", "synonyms": { "Apache 2.0": "apache 2.0" }, "unknown_licenses": [] }, "licenses": [ "Apache 2.0" ], "package": "commons-fileupload:commons-fileupload", "version": "1.3" }, { "license_analysis": { "_message": "Input is invalid", "_representative_licenses": null, "conflict_licenses": [], "outlier_licenses": [], "status": "Failure", "synonyms": {}, "unknown_licenses": [] }, "licenses": [], "package": "org.springframework.boot:spring-boot-starter-web", "version": "1.4.1.RELEASE" }, { "license_analysis": { "_message": "Input is invalid", "_representative_licenses": null, "conflict_licenses": [], "outlier_licenses": [], "status": "Failure", "synonyms": {}, "unknown_licenses": [] }, "licenses": [], "package": "com.h2database:h2", "version": "1.4.192" }, { "license_analysis": { "_message": "Input is invalid", "_representative_licenses": null, "conflict_licenses": [], "outlier_licenses": [], "status": "Failure", "synonyms": {}, "unknown_licenses": [] }, "licenses": [], "package": "org.springframework.boot:spring-boot-starter-data-jpa", "version": "1.4.1.RELEASE" } ], "stack_license": null, "status": "Failure", "unknown_licenses": { "component_conflict": [], "really_unknown": [] } } ``` ## Expected output ``` { "conflict_packages": [], "license_filter": { "alternate_packages": { "compatible_packages": [], "conflict_packages": [], "unknown_license_packages": [] }, "companion_packages": { "compatible_packages": [], "conflict_packages": [], "unknown_license_packages": [] } }, "outlier_packages": {}, "packages": [ { "license_analysis": { "_message": "Representative license found", "_representative_licenses": "epl 1.0", "conflict_licenses": [], "outlier_licenses": [], "status": "Successful", "synonyms": { "APACHE": "apache 2.0", "Eclipse Public License": "epl 1.0" }, "unknown_licenses": [] }, "licenses": [ "APACHE", "Eclipse Public License" ], "package": "p1", "version": "1.1" }, { "license_analysis": { "_message": "Representative license found", "_representative_licenses": "gplv2", "conflict_licenses": [], "outlier_licenses": [], "status": "Successful", "synonyms": { "BSD": "bsd-new", "GPL V2": "gplv2" }, "unknown_licenses": [] }, "licenses": [ "BSD", "GPL V2" ], "package": "p2", "version": "1.1" } ], "stack_license": "gplv2", "status": "Successful" } ```
non_process
inconsistent output from the license analysis service on prod preview problem description the license analysis service returns incorrect data with the status set to failure the output is different from one used in documentation for the dependency editor so either the service has a bug or the documentation is not consistent w r o the request expected response documentation input post the following payload taken from the request and response doc for dependency editor to the analytics license service on prod preview resolved package com googlecode xmemcached xmemcached version package commons fileupload commons fileupload version package org springframework boot spring boot starter web version release package com version package org springframework boot spring boot starter data jpa version release ecosystem maven output from the service conflict packages distinct licenses apache message no declared licenses found for component s outlier packages packages license analysis message representative license found representative licenses apache conflict licenses outlier licenses status successful synonyms apache license version apache unknown licenses licenses apache license version package com googlecode xmemcached xmemcached version license analysis message representative license found representative licenses apache conflict licenses outlier licenses status successful synonyms apache apache unknown licenses licenses apache package commons fileupload commons fileupload version license analysis message input is invalid representative licenses null conflict licenses outlier licenses status failure synonyms unknown licenses licenses package org springframework boot spring boot starter web version release license analysis message input is invalid representative licenses null conflict licenses outlier licenses status failure synonyms unknown licenses licenses package com version license analysis message input is invalid representative licenses null conflict licenses outlier licenses status failure synonyms unknown licenses licenses package org springframework boot spring boot starter data jpa version release stack license null status failure unknown licenses component conflict really unknown expected output conflict packages license filter alternate packages compatible packages conflict packages unknown license packages companion packages compatible packages conflict packages unknown license packages outlier packages packages license analysis message representative license found representative licenses epl conflict licenses outlier licenses status successful synonyms apache apache eclipse public license epl unknown licenses licenses apache eclipse public license package version license analysis message representative license found representative licenses conflict licenses outlier licenses status successful synonyms bsd bsd new gpl unknown licenses licenses bsd gpl package version stack license status successful
0
570,417
17,023,108,394
IssuesEvent
2021-07-03 00:24:47
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Support for other renderers
Component: josm Priority: major Resolution: invalid Type: enhancement
**[Submitted to the original trac issue database at 10.11am, Wednesday, 12th April 2006]** FrankM: line width should be user defined. Also something like the '+'/'-' in the applet to change line/point (also raw points) thickness Imi: JOSM is no beauty renderer. Maybe include other renderers like osmarender or freemap renderer (or even the applet renderer) could be an option. But I really like to keep the standard renderer as fast and clear as possible.
1.0
Support for other renderers - **[Submitted to the original trac issue database at 10.11am, Wednesday, 12th April 2006]** FrankM: line width should be user defined. Also something like the '+'/'-' in the applet to change line/point (also raw points) thickness Imi: JOSM is no beauty renderer. Maybe include other renderers like osmarender or freemap renderer (or even the applet renderer) could be an option. But I really like to keep the standard renderer as fast and clear as possible.
non_process
support for other renderers frankm line width should be user defined also something like the in the applet to change line point also raw points thickness imi josm is no beauty renderer maybe include other renderers like osmarender or freemap renderer or even the applet renderer could be an option but i really like to keep the standard renderer as fast and clear as possible
0
16,664
21,765,569,092
IssuesEvent
2022-05-13 01:09:40
aws/sagemaker-python-sdk
https://api.github.com/repos/aws/sagemaker-python-sdk
closed
S3 manifest file cannot be used for SageMaker Processing job defined by SageMaker Pipeline
component: processing component: pipelines
**Describe the bug** I am trying to use SageMaker Pipeline to define a Processing job whose inputs are s3 manifest files. Here is my ProcessingInput : ``` inputs=[ ProcessingInput( input_name="raw_train_data", s3_input=S3Input( s3_uri="s3://watch-next-ml-ranker-prototype-data/20210717/data/manifest/train.manifest", local_path="/opt/ml/processing/raw_train", s3_data_type="ManifestFile", )), ProcessingInput( input_name="raw_test_data", s3_input=S3Input( s3_uri="s3://watch-next-ml-ranker-prototype-data/20210717/data/manifest/test.manifest", local_path="/opt/ml/processing/raw_test", s3_data_type="ManifestFile", )) ] ``` When I execute the SageMaker pipeline using Python SDK 2.51.0, an exception is thrown: ``` ClientError: Input raw_train_data missing one or more required fields for SageMaker-managed S3Input: "LocalPath", "S3DownloadMode", "S3InputMode" ``` But, it looks like the S3Input in the SageMaker Python SDK has no attribute of S3DownloadMode. Do you know if it is a SDK bug, and how to fix this issue? **To reproduce** You can reproduce the issue using the above code in the Processing Job step. **Expected behavior** The pipeline should be defined and executed without throwing exception. The S3 manifest file can be used for specifying the input data of Processing job. **Screenshots or logs** It is provided in the above sections. **System information** A description of your system. Please provide: - **SageMaker Python SDK version**: 2.51.0 - **Framework name (eg. PyTorch) or algorithm (eg. KMeans)**: N/A - **Framework version**: N/A - **Python version**: 3.8 - **CPU or GPU**: N/A - **Custom Docker image (Y/N)**: N/A **Additional context** Add any other context about the problem here.
1.0
S3 manifest file cannot be used for SageMaker Processing job defined by SageMaker Pipeline - **Describe the bug** I am trying to use SageMaker Pipeline to define a Processing job whose inputs are s3 manifest files. Here is my ProcessingInput : ``` inputs=[ ProcessingInput( input_name="raw_train_data", s3_input=S3Input( s3_uri="s3://watch-next-ml-ranker-prototype-data/20210717/data/manifest/train.manifest", local_path="/opt/ml/processing/raw_train", s3_data_type="ManifestFile", )), ProcessingInput( input_name="raw_test_data", s3_input=S3Input( s3_uri="s3://watch-next-ml-ranker-prototype-data/20210717/data/manifest/test.manifest", local_path="/opt/ml/processing/raw_test", s3_data_type="ManifestFile", )) ] ``` When I execute the SageMaker pipeline using Python SDK 2.51.0, an exception is thrown: ``` ClientError: Input raw_train_data missing one or more required fields for SageMaker-managed S3Input: "LocalPath", "S3DownloadMode", "S3InputMode" ``` But, it looks like the S3Input in the SageMaker Python SDK has no attribute of S3DownloadMode. Do you know if it is a SDK bug, and how to fix this issue? **To reproduce** You can reproduce the issue using the above code in the Processing Job step. **Expected behavior** The pipeline should be defined and executed without throwing exception. The S3 manifest file can be used for specifying the input data of Processing job. **Screenshots or logs** It is provided in the above sections. **System information** A description of your system. Please provide: - **SageMaker Python SDK version**: 2.51.0 - **Framework name (eg. PyTorch) or algorithm (eg. KMeans)**: N/A - **Framework version**: N/A - **Python version**: 3.8 - **CPU or GPU**: N/A - **Custom Docker image (Y/N)**: N/A **Additional context** Add any other context about the problem here.
process
manifest file cannot be used for sagemaker processing job defined by sagemaker pipeline describe the bug i am trying to use sagemaker pipeline to define a processing job whose inputs are manifest files here is my processinginput inputs processinginput input name raw train data input uri watch next ml ranker prototype data data manifest train manifest local path opt ml processing raw train data type manifestfile processinginput input name raw test data input uri watch next ml ranker prototype data data manifest test manifest local path opt ml processing raw test data type manifestfile when i execute the sagemaker pipeline using python sdk an exception is thrown clienterror input raw train data missing one or more required fields for sagemaker managed localpath but it looks like the in the sagemaker python sdk has no attribute of do you know if it is a sdk bug and how to fix this issue to reproduce you can reproduce the issue using the above code in the processing job step expected behavior the pipeline should be defined and executed without throwing exception the manifest file can be used for specifying the input data of processing job screenshots or logs it is provided in the above sections system information a description of your system please provide sagemaker python sdk version framework name eg pytorch or algorithm eg kmeans n a framework version n a python version cpu or gpu n a custom docker image y n n a additional context add any other context about the problem here
1
12,779
15,163,216,950
IssuesEvent
2021-02-12 11:51:14
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Internal: Send Slack message for each CLI release
kind/feature process/candidate team/client topic: ci/cd
Right now we are not notified of CLI releases in any way, you have to watch the build pipeline. and go into the logs to find out the version name (for `dev` releases for example). We want to send a Slack message to a `#feed-prisma-releases` Slack channel including the version name and release pipeline link. It should probably be implemented as the final step of the current "Publish" job.
1.0
Internal: Send Slack message for each CLI release - Right now we are not notified of CLI releases in any way, you have to watch the build pipeline. and go into the logs to find out the version name (for `dev` releases for example). We want to send a Slack message to a `#feed-prisma-releases` Slack channel including the version name and release pipeline link. It should probably be implemented as the final step of the current "Publish" job.
process
internal send slack message for each cli release right now we are not notified of cli releases in any way you have to watch the build pipeline and go into the logs to find out the version name for dev releases for example we want to send a slack message to a feed prisma releases slack channel including the version name and release pipeline link it should probably be implemented as the final step of the current publish job
1
132,947
18,278,499,637
IssuesEvent
2021-10-04 22:10:46
ghc-dev/Michael-Hicks
https://api.github.com/repos/ghc-dev/Michael-Hicks
opened
CVE-2017-16138 (High) detected in mime-1.3.4.tgz
security vulnerability
## CVE-2017-16138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.3.4.tgz</b></p></summary> <p>A comprehensive library for mime-type mapping</p> <p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.3.4.tgz">https://registry.npmjs.org/mime/-/mime-1.3.4.tgz</a></p> <p>Path to dependency file: Michael-Hicks/package.json</p> <p>Path to vulnerable library: Michael-Hicks/node_modules/mime/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-connect-0.10.1.tgz (Root Library) - connect-2.30.2.tgz - serve-static-1.10.3.tgz - send-0.13.2.tgz - :x: **mime-1.3.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Michael-Hicks/commit/057013f0e6c6f3c2e3d4b4a2ae5197624ad577d3">057013f0e6c6f3c2e3d4b4a2ae5197624ad577d3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 1.4.1,2.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.3.4","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-contrib-connect:0.10.1;connect:2.30.2;serve-static:1.10.3;send:0.13.2;mime:1.3.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-16138 (High) detected in mime-1.3.4.tgz - ## CVE-2017-16138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.3.4.tgz</b></p></summary> <p>A comprehensive library for mime-type mapping</p> <p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.3.4.tgz">https://registry.npmjs.org/mime/-/mime-1.3.4.tgz</a></p> <p>Path to dependency file: Michael-Hicks/package.json</p> <p>Path to vulnerable library: Michael-Hicks/node_modules/mime/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-connect-0.10.1.tgz (Root Library) - connect-2.30.2.tgz - serve-static-1.10.3.tgz - send-0.13.2.tgz - :x: **mime-1.3.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Michael-Hicks/commit/057013f0e6c6f3c2e3d4b4a2ae5197624ad577d3">057013f0e6c6f3c2e3d4b4a2ae5197624ad577d3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 1.4.1,2.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.3.4","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-contrib-connect:0.10.1;connect:2.30.2;serve-static:1.10.3;send:0.13.2;mime:1.3.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in mime tgz cve high severity vulnerability vulnerable library mime tgz a comprehensive library for mime type mapping library home page a href path to dependency file michael hicks package json path to vulnerable library michael hicks node modules mime package json dependency hierarchy grunt contrib connect tgz root library connect tgz serve static tgz send tgz x mime tgz vulnerable library found in head commit a href found in base branch master vulnerability details the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt contrib connect connect serve static send mime isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input vulnerabilityurl
0
101,274
21,639,014,677
IssuesEvent
2022-05-05 16:45:51
github/vscode-codeql
https://api.github.com/repos/github/vscode-codeql
closed
When losing connection to a remote SSH connection, the query history is getting deleted
bug VSCode
**Describe the bug** When I lose connection/reconnect to my remote ssh machine (which holds a CodeQL DB and runs queries), I also lose all my queries in the query history window. I believe this happens because the vs-code window is getting "refreshed" while trying to reconnect to the server, and for some reason there is no way I know of to backup the query history window to file (or something else persistent). This behavior can cause lots of frustration and loss of time or even loss of data in case the original query is no longer available. **Version** The CodeQL and VS Code version in which the bug occurs. <!-- To copy version information for the CodeQL extension, click "CodeQL CLI vX.X.X" in the status bar at the bottom of the screen. To copy detailed version information for VS Code itself, see https://code.visualstudio.com/docs/supporting/FAQ#_how-do-i-find-the-version. --> CodeQL extension version: 1.5.9 CodeQL CLI version: 2.6.0 Platform: linux x64 (Be aware that this is related to the remote machine) VSCODE version (this is my local version, on a Mac OS X): Version: 1.63.2 (Universal) Commit: 899d46d82c4c95423fb7e10e68eba52050e30ba3 Date: 2021-12-15T09:37:28.172Z (1 mo ago) Electron: 13.5.2 Chromium: 91.0.4472.164 Node.js: 14.16.0 V8: 9.1.269.39-electron.0 OS: Darwin x64 21.1.0 **To reproduce** 1. Remote connect using vscode's Remote SSH connection to machine having codeql db (or create a small one) 2. Run any kind of query on that db; Make sure the query was added to the query history 3. Reconnect to that machine (by losing the network connection for example) 4. The query history is empty... **Expected behavior** It would be great to at least hold the history-panel's data in case of a reconnection, though the best behavior can be to hold the history also when disconnecting and reconnecting later to the related machine. **Additional context** I'm not sure what is the behavior of the history-panel when not using a remote connection because I use CodeQL only on remote machines, but I believe it should at least "survive" a connection loss because it can be pretty frequent when using remote connections. If there anything I can do to help fix this issue I will be glad to help. Thank you very much!
1.0
When losing connection to a remote SSH connection, the query history is getting deleted - **Describe the bug** When I lose connection/reconnect to my remote ssh machine (which holds a CodeQL DB and runs queries), I also lose all my queries in the query history window. I believe this happens because the vs-code window is getting "refreshed" while trying to reconnect to the server, and for some reason there is no way I know of to backup the query history window to file (or something else persistent). This behavior can cause lots of frustration and loss of time or even loss of data in case the original query is no longer available. **Version** The CodeQL and VS Code version in which the bug occurs. <!-- To copy version information for the CodeQL extension, click "CodeQL CLI vX.X.X" in the status bar at the bottom of the screen. To copy detailed version information for VS Code itself, see https://code.visualstudio.com/docs/supporting/FAQ#_how-do-i-find-the-version. --> CodeQL extension version: 1.5.9 CodeQL CLI version: 2.6.0 Platform: linux x64 (Be aware that this is related to the remote machine) VSCODE version (this is my local version, on a Mac OS X): Version: 1.63.2 (Universal) Commit: 899d46d82c4c95423fb7e10e68eba52050e30ba3 Date: 2021-12-15T09:37:28.172Z (1 mo ago) Electron: 13.5.2 Chromium: 91.0.4472.164 Node.js: 14.16.0 V8: 9.1.269.39-electron.0 OS: Darwin x64 21.1.0 **To reproduce** 1. Remote connect using vscode's Remote SSH connection to machine having codeql db (or create a small one) 2. Run any kind of query on that db; Make sure the query was added to the query history 3. Reconnect to that machine (by losing the network connection for example) 4. The query history is empty... **Expected behavior** It would be great to at least hold the history-panel's data in case of a reconnection, though the best behavior can be to hold the history also when disconnecting and reconnecting later to the related machine. **Additional context** I'm not sure what is the behavior of the history-panel when not using a remote connection because I use CodeQL only on remote machines, but I believe it should at least "survive" a connection loss because it can be pretty frequent when using remote connections. If there anything I can do to help fix this issue I will be glad to help. Thank you very much!
non_process
when losing connection to a remote ssh connection the query history is getting deleted describe the bug when i lose connection reconnect to my remote ssh machine which holds a codeql db and runs queries i also lose all my queries in the query history window i believe this happens because the vs code window is getting refreshed while trying to reconnect to the server and for some reason there is no way i know of to backup the query history window to file or something else persistent this behavior can cause lots of frustration and loss of time or even loss of data in case the original query is no longer available version the codeql and vs code version in which the bug occurs to copy version information for the codeql extension click codeql cli vx x x in the status bar at the bottom of the screen to copy detailed version information for vs code itself see codeql extension version codeql cli version platform linux be aware that this is related to the remote machine vscode version this is my local version on a mac os x version universal commit date mo ago electron chromium node js electron os darwin to reproduce remote connect using vscode s remote ssh connection to machine having codeql db or create a small one run any kind of query on that db make sure the query was added to the query history reconnect to that machine by losing the network connection for example the query history is empty expected behavior it would be great to at least hold the history panel s data in case of a reconnection though the best behavior can be to hold the history also when disconnecting and reconnecting later to the related machine additional context i m not sure what is the behavior of the history panel when not using a remote connection because i use codeql only on remote machines but i believe it should at least survive a connection loss because it can be pretty frequent when using remote connections if there anything i can do to help fix this issue i will be glad to help thank you very much
0
769,710
27,017,156,176
IssuesEvent
2023-02-10 20:34:52
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
opened
[trusted-contribution] Add option for organization members as trustedContributors
type: feature request priority: p3
It would be wonderful if in addition to users, [trusted-contribution](https://github.com/googleapis/repo-automation-bots/tree/main/packages/trusted-contribution) could support teams and especially [organization members](https://docs.github.com/en/rest/orgs/members?apiVersion=2022-11-28#list-public-organization-members) as trustedContributors.
1.0
[trusted-contribution] Add option for organization members as trustedContributors - It would be wonderful if in addition to users, [trusted-contribution](https://github.com/googleapis/repo-automation-bots/tree/main/packages/trusted-contribution) could support teams and especially [organization members](https://docs.github.com/en/rest/orgs/members?apiVersion=2022-11-28#list-public-organization-members) as trustedContributors.
non_process
add option for organization members as trustedcontributors it would be wonderful if in addition to users could support teams and especially as trustedcontributors
0
69,649
7,156,654,094
IssuesEvent
2018-01-26 16:59:41
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed tests on master: Jepsen/Jepsen: JepsenRegister: JepsenRegister/majority-ring+start-kill-2
Robot test-failure
The following tests appear to have failed: [#492212](https://teamcity.cockroachdb.com/viewLog.html?buildId=492212): ``` --- FAIL: Jepsen/Jepsen: JepsenRegister: JepsenRegister/majority-ring+start-kill-2 (18.327s) None ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed tests on master: Jepsen/Jepsen: JepsenRegister: JepsenRegister/majority-ring+start-kill-2 - The following tests appear to have failed: [#492212](https://teamcity.cockroachdb.com/viewLog.html?buildId=492212): ``` --- FAIL: Jepsen/Jepsen: JepsenRegister: JepsenRegister/majority-ring+start-kill-2 (18.327s) None ``` Please assign, take a look and update the issue accordingly.
non_process
teamcity failed tests on master jepsen jepsen jepsenregister jepsenregister majority ring start kill the following tests appear to have failed fail jepsen jepsen jepsenregister jepsenregister majority ring start kill none please assign take a look and update the issue accordingly
0
16,307
20,960,721,759
IssuesEvent
2022-03-27 19:05:27
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add a film by kirk
suggested title in process
Please add as much of the following info as you can: Title: a film by kirk Type (film/tv show): film Film or show in which it appears: Gilmore Girls Is the parent film/show streaming anywhere? Netflix About when in the parent film/show does it appear? Season 2, Episode 19 Actual footage of the film/show can be seen (yes/no)? yes
1.0
Add a film by kirk - Please add as much of the following info as you can: Title: a film by kirk Type (film/tv show): film Film or show in which it appears: Gilmore Girls Is the parent film/show streaming anywhere? Netflix About when in the parent film/show does it appear? Season 2, Episode 19 Actual footage of the film/show can be seen (yes/no)? yes
process
add a film by kirk please add as much of the following info as you can title a film by kirk type film tv show film film or show in which it appears gilmore girls is the parent film show streaming anywhere netflix about when in the parent film show does it appear season episode actual footage of the film show can be seen yes no yes
1
356,214
10,590,138,614
IssuesEvent
2019-10-09 08:06:31
bedita/bedita
https://api.github.com/repos/bedita/bedita
closed
`GET /object?filter[notexistingfield]=string` => `500` with memory exhausted
Priority - Low Status - Still valid? Type - Bug
### Objects filter Calling `GET /object?filter[notexistingfield]=string` we'd expect a response 200 with no data. Instead we got a 500, and memory exhausted error. Sample stack trace: ``` { "error": { "status": "500", "title": "Allowed memory size of 536870912 bytes exhausted (tried to allocate 72 bytes)", "meta": { "trace": [ "#0 \/var\/www\/workspace\/bedita4\/vendor\/cakephp\/cakephp\/src\/Error\/BaseErrorHandler.php(105): Cake\\Error\\BaseErrorHandler-\u003EhandleFatalError(1, \u0027Allowed memory ...\u0027, \u0027\/var\/www\/worksp...\u0027, 341)", "#1 [internal function]: Cake\\Error\\BaseErrorHandler-\u003ECake\\Error\\{closure}()", "#2 {main}" ] } }, "links": { "self": "http:\/\/bedita4.lcl\/objects?page=1\u0026page_size=10\u0026filter%5Bstatusssss%5D=off", "home": "http:\/\/bedita4.lcl\/home" } } ```
1.0
`GET /object?filter[notexistingfield]=string` => `500` with memory exhausted - ### Objects filter Calling `GET /object?filter[notexistingfield]=string` we'd expect a response 200 with no data. Instead we got a 500, and memory exhausted error. Sample stack trace: ``` { "error": { "status": "500", "title": "Allowed memory size of 536870912 bytes exhausted (tried to allocate 72 bytes)", "meta": { "trace": [ "#0 \/var\/www\/workspace\/bedita4\/vendor\/cakephp\/cakephp\/src\/Error\/BaseErrorHandler.php(105): Cake\\Error\\BaseErrorHandler-\u003EhandleFatalError(1, \u0027Allowed memory ...\u0027, \u0027\/var\/www\/worksp...\u0027, 341)", "#1 [internal function]: Cake\\Error\\BaseErrorHandler-\u003ECake\\Error\\{closure}()", "#2 {main}" ] } }, "links": { "self": "http:\/\/bedita4.lcl\/objects?page=1\u0026page_size=10\u0026filter%5Bstatusssss%5D=off", "home": "http:\/\/bedita4.lcl\/home" } } ```
non_process
get object filter string with memory exhausted objects filter calling get object filter string we d expect a response with no data instead we got a and memory exhausted error sample stack trace error status title allowed memory size of bytes exhausted tried to allocate bytes meta trace var www workspace vendor cakephp cakephp src error baseerrorhandler php cake error baseerrorhandler memory var www worksp cake error baseerrorhandler error closure main links self http lcl objects page size off home http lcl home
0
130,968
12,466,827,687
IssuesEvent
2020-05-28 16:03:58
rte-france/Grid2Op
https://api.github.com/repos/rte-france/Grid2Op
closed
Environment Document for L2RPN Competition
documentation
## Documentation issue description Hi, I think it is better to describe what environment class is used for WCCI Competition in detail. For example the `actionClass`, `rewardClass` and `legalActClass` ect. The description on CodaLab website is too simple. If we made an environment using `env = grid2op.make("l2rpn_wcci_2020")`, we are not sure what are the default arguments and whether they are the same as the platform in Codalab. ## Suggested modifications ``` Documentation issues are low priority. Please provide your suggested modifications to increase processing speed. Thanks for your understanding. ``` Add more explanations in Codalab website about what is the default the `actionClass`, `rewardClass` and `legalActClass` setting. ## Additional context Add any other context here.
1.0
Environment Document for L2RPN Competition - ## Documentation issue description Hi, I think it is better to describe what environment class is used for WCCI Competition in detail. For example the `actionClass`, `rewardClass` and `legalActClass` ect. The description on CodaLab website is too simple. If we made an environment using `env = grid2op.make("l2rpn_wcci_2020")`, we are not sure what are the default arguments and whether they are the same as the platform in Codalab. ## Suggested modifications ``` Documentation issues are low priority. Please provide your suggested modifications to increase processing speed. Thanks for your understanding. ``` Add more explanations in Codalab website about what is the default the `actionClass`, `rewardClass` and `legalActClass` setting. ## Additional context Add any other context here.
non_process
environment document for competition documentation issue description hi i think it is better to describe what environment class is used for wcci competition in detail for example the actionclass rewardclass and legalactclass ect the description on codalab website is too simple if we made an environment using env make wcci we are not sure what are the default arguments and whether they are the same as the platform in codalab suggested modifications documentation issues are low priority please provide your suggested modifications to increase processing speed thanks for your understanding add more explanations in codalab website about what is the default the actionclass rewardclass and legalactclass setting additional context add any other context here
0
5,657
8,528,261,714
IssuesEvent
2018-11-02 22:48:05
GoogleCloudPlatform/google-cloud-cpp
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-cpp
closed
Create a new release and document the steps.
type: process
We need to create a new release and document the steps to do so.
1.0
Create a new release and document the steps. - We need to create a new release and document the steps to do so.
process
create a new release and document the steps we need to create a new release and document the steps to do so
1
4,181
7,114,527,052
IssuesEvent
2018-01-18 01:13:42
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
closed
Create a new variable to kill long running transactions
CONNECTION POOL MYSQL PROTOCOL QUERY PROCESSOR
## WHY Variable `mysql-max_transaction_time` currently has a misleading name. It defines the maximum time a transaction can be idle. As such, there is no way to kill a long running transaction no matter if idle or not. ## WHAT - [ ] Rename current `mysql-max_transaction_time` to `mysql-max_transaction_idle_time` - [ ] track when a transaction is started - [ ] create a new variable ( reuse `mysql-max_transaction_time`) to define when a transaction needs to be killed no matter if idle or not
1.0
Create a new variable to kill long running transactions - ## WHY Variable `mysql-max_transaction_time` currently has a misleading name. It defines the maximum time a transaction can be idle. As such, there is no way to kill a long running transaction no matter if idle or not. ## WHAT - [ ] Rename current `mysql-max_transaction_time` to `mysql-max_transaction_idle_time` - [ ] track when a transaction is started - [ ] create a new variable ( reuse `mysql-max_transaction_time`) to define when a transaction needs to be killed no matter if idle or not
process
create a new variable to kill long running transactions why variable mysql max transaction time currently has a misleading name it defines the maximum time a transaction can be idle as such there is no way to kill a long running transaction no matter if idle or not what rename current mysql max transaction time to mysql max transaction idle time track when a transaction is started create a new variable reuse mysql max transaction time to define when a transaction needs to be killed no matter if idle or not
1
3,432
6,530,885,530
IssuesEvent
2017-08-30 16:35:50
pwittchen/ReactiveNetwork
https://api.github.com/repos/pwittchen/ReactiveNetwork
closed
Release 0.12.0
release process
**Initial release notes**: RxJava1.x: - Fixed NPE occuring when `ConnectivityManager` is `null` in `ReactiveNetwork.observeNetworkConnectivity()` method - issue #209 - Added new method to the API for checking Internet connectivity - issue #205 - `Observable<Boolean> observeInternetConnectivity(strategy, host)` - Added to documentation comment about monitoring Internet connectivity with custom host - issue #204 - Classes which implement InternetObservingStrategy handle custom hosts with and without `http://` or `https://` prefix gracefully - issue #206 - organized packages with unit tests - made the library more hermetic - changed visibility of `SocketInternetObservingStrategy#isConnected(String host, int port, int timeoutInMs, ErrorHandler handler)` method from `public` to `protected` - changed visibility of `SocketInternetObservingStrategy#isConnected(Socket socket, String host, int port, int timeoutInMs, ErrorHandler errorHandler)` method from `public` to `protected` - changed visibility of `Connectivity#create(Context, ConnectivityManager)` method from `public` to `protected` - changed visibility of `WalledGardenInternetObservingStrategy#isConnected(String host, int port, int timeoutInMs, ErrorHandler errorHandler)` method from `public` to `protected` - changed visibility of `WalledGardenInternetObservingStrategy#createHttpUrlConnection(String host, int port, int timeoutInMs)` method from `public` to `protected` RxJava2.x: - Fixed NPE occuring when `ConnectivityManager` is `null` in `ReactiveNetwork.observeNetworkConnectivity()` method - issue #209 - Added new methods to the API for checking Internet connectivity - issue #205 - `Observable<Boolean> observeInternetConnectivity(strategy, host)` - `Single<Boolean> checkInternetConnectivity(strategy, host)` - Added to documentation comment about monitoring Internet connectivity with custom host - issue #204 - Classes which implement InternetObservingStrategy handle custom hosts with and without `http://` or `https://` prefix gracefully - issue #206 - organized packages with unit tests - made the library more hermetic - changed visibility of `SocketInternetObservingStrategy#isConnected(String host, int port, int timeoutInMs, ErrorHandler handler)` method from `public` to `protected` - changed visibility of `SocketInternetObservingStrategy#isConnected(Socket socket, String host, int port, int timeoutInMs, ErrorHandler errorHandler)` method from `public` to `protected` - changed visibility of `Connectivity#create(Context, ConnectivityManager)` method from `public` to `protected` - changed visibility of `WalledGardenInternetObservingStrategy#isConnected(String host, int port, int timeoutInMs, ErrorHandler errorHandler)` method from `public` to `protected` - changed visibility of `WalledGardenInternetObservingStrategy#createHttpUrlConnection(String host, int port, int timeoutInMs)` method from `public` to `protected` **Things to do**: - [x] RxJava1.x branch: - [x] update JavaDoc on `gh-pages` - [x] update release notes (more technical details about hermetic API) - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release - [x] RxJava2.x branch: - [x] update JavaDoc on `gh-pages` - [x] update release notes (more technical details about hermetic API) - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
1.0
Release 0.12.0 - **Initial release notes**: RxJava1.x: - Fixed NPE occuring when `ConnectivityManager` is `null` in `ReactiveNetwork.observeNetworkConnectivity()` method - issue #209 - Added new method to the API for checking Internet connectivity - issue #205 - `Observable<Boolean> observeInternetConnectivity(strategy, host)` - Added to documentation comment about monitoring Internet connectivity with custom host - issue #204 - Classes which implement InternetObservingStrategy handle custom hosts with and without `http://` or `https://` prefix gracefully - issue #206 - organized packages with unit tests - made the library more hermetic - changed visibility of `SocketInternetObservingStrategy#isConnected(String host, int port, int timeoutInMs, ErrorHandler handler)` method from `public` to `protected` - changed visibility of `SocketInternetObservingStrategy#isConnected(Socket socket, String host, int port, int timeoutInMs, ErrorHandler errorHandler)` method from `public` to `protected` - changed visibility of `Connectivity#create(Context, ConnectivityManager)` method from `public` to `protected` - changed visibility of `WalledGardenInternetObservingStrategy#isConnected(String host, int port, int timeoutInMs, ErrorHandler errorHandler)` method from `public` to `protected` - changed visibility of `WalledGardenInternetObservingStrategy#createHttpUrlConnection(String host, int port, int timeoutInMs)` method from `public` to `protected` RxJava2.x: - Fixed NPE occuring when `ConnectivityManager` is `null` in `ReactiveNetwork.observeNetworkConnectivity()` method - issue #209 - Added new methods to the API for checking Internet connectivity - issue #205 - `Observable<Boolean> observeInternetConnectivity(strategy, host)` - `Single<Boolean> checkInternetConnectivity(strategy, host)` - Added to documentation comment about monitoring Internet connectivity with custom host - issue #204 - Classes which implement InternetObservingStrategy handle custom hosts with and without `http://` or `https://` prefix gracefully - issue #206 - organized packages with unit tests - made the library more hermetic - changed visibility of `SocketInternetObservingStrategy#isConnected(String host, int port, int timeoutInMs, ErrorHandler handler)` method from `public` to `protected` - changed visibility of `SocketInternetObservingStrategy#isConnected(Socket socket, String host, int port, int timeoutInMs, ErrorHandler errorHandler)` method from `public` to `protected` - changed visibility of `Connectivity#create(Context, ConnectivityManager)` method from `public` to `protected` - changed visibility of `WalledGardenInternetObservingStrategy#isConnected(String host, int port, int timeoutInMs, ErrorHandler errorHandler)` method from `public` to `protected` - changed visibility of `WalledGardenInternetObservingStrategy#createHttpUrlConnection(String host, int port, int timeoutInMs)` method from `public` to `protected` **Things to do**: - [x] RxJava1.x branch: - [x] update JavaDoc on `gh-pages` - [x] update release notes (more technical details about hermetic API) - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release - [x] RxJava2.x branch: - [x] update JavaDoc on `gh-pages` - [x] update release notes (more technical details about hermetic API) - [x] bump library version - [x] upload archives to Maven Central - [x] close and release artifact on Maven Central - [x] update `CHANGELOG.md` after Maven Sync - [x] bump library version in `README.md` - [x] create new GitHub release
process
release initial release notes x fixed npe occuring when connectivitymanager is null in reactivenetwork observenetworkconnectivity method issue added new method to the api for checking internet connectivity issue observable observeinternetconnectivity strategy host added to documentation comment about monitoring internet connectivity with custom host issue classes which implement internetobservingstrategy handle custom hosts with and without or prefix gracefully issue organized packages with unit tests made the library more hermetic changed visibility of socketinternetobservingstrategy isconnected string host int port int timeoutinms errorhandler handler method from public to protected changed visibility of socketinternetobservingstrategy isconnected socket socket string host int port int timeoutinms errorhandler errorhandler method from public to protected changed visibility of connectivity create context connectivitymanager method from public to protected changed visibility of walledgardeninternetobservingstrategy isconnected string host int port int timeoutinms errorhandler errorhandler method from public to protected changed visibility of walledgardeninternetobservingstrategy createhttpurlconnection string host int port int timeoutinms method from public to protected x fixed npe occuring when connectivitymanager is null in reactivenetwork observenetworkconnectivity method issue added new methods to the api for checking internet connectivity issue observable observeinternetconnectivity strategy host single checkinternetconnectivity strategy host added to documentation comment about monitoring internet connectivity with custom host issue classes which implement internetobservingstrategy handle custom hosts with and without or prefix gracefully issue organized packages with unit tests made the library more hermetic changed visibility of socketinternetobservingstrategy isconnected string host int port int timeoutinms errorhandler handler method from public to protected changed visibility of socketinternetobservingstrategy isconnected socket socket string host int port int timeoutinms errorhandler errorhandler method from public to protected changed visibility of connectivity create context connectivitymanager method from public to protected changed visibility of walledgardeninternetobservingstrategy isconnected string host int port int timeoutinms errorhandler errorhandler method from public to protected changed visibility of walledgardeninternetobservingstrategy createhttpurlconnection string host int port int timeoutinms method from public to protected things to do x branch update javadoc on gh pages update release notes more technical details about hermetic api bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release x branch update javadoc on gh pages update release notes more technical details about hermetic api bump library version upload archives to maven central close and release artifact on maven central update changelog md after maven sync bump library version in readme md create new github release
1
2,755
3,851,246,090
IssuesEvent
2016-04-06 00:42:01
proudcity/wp-proudcity
https://api.github.com/repos/proudcity/wp-proudcity
closed
Rm Black origin widgets?
infrastructure
"you got it to allow us to use tinymce instances in our widget settings page [3:27] not sure if its currently being used"
1.0
Rm Black origin widgets? - "you got it to allow us to use tinymce instances in our widget settings page [3:27] not sure if its currently being used"
non_process
rm black origin widgets you got it to allow us to use tinymce instances in our widget settings page not sure if its currently being used
0
4,200
7,157,612,994
IssuesEvent
2018-01-26 20:33:11
IIIF/api
https://api.github.com/repos/IIIF/api
opened
Define process for accepting entries for registries
notes process
#1310 created several registries, including for extensions, services, profiles, types, motivations and others. We need a process defined for: * How are entries suggested for inclusion? * What are the possible features for each entry? * What makes an appropriate / inappropriate entry? * How is that judgement arrived at in a transparent and fair fashion?
1.0
Define process for accepting entries for registries - #1310 created several registries, including for extensions, services, profiles, types, motivations and others. We need a process defined for: * How are entries suggested for inclusion? * What are the possible features for each entry? * What makes an appropriate / inappropriate entry? * How is that judgement arrived at in a transparent and fair fashion?
process
define process for accepting entries for registries created several registries including for extensions services profiles types motivations and others we need a process defined for how are entries suggested for inclusion what are the possible features for each entry what makes an appropriate inappropriate entry how is that judgement arrived at in a transparent and fair fashion
1