Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
177
| 2,587,552,739
|
IssuesEvent
|
2015-02-17 19:11:57
|
GsDevKit/gsDevKitHome
|
https://api.github.com/repos/GsDevKit/gsDevKitHome
|
closed
|
1.0.0 to 2.0.0 compat ideas
|
in process
|
- [ ] symbolic link from tode/sys/default/clients to tode/clients (similar link for server(?))
- [ ] composition that includes tode/home for home directories
- [ ] directory map of tode/sys documenting the directory structure
- [ ] anything else that preserves changes users made in the old tode structure
- [ ] tag 1.0.0 before merge ... next release 2.0.0
- [ ] script for building /sys structure for exsiting stones
|
1.0
|
1.0.0 to 2.0.0 compat ideas - - [ ] symbolic link from tode/sys/default/clients to tode/clients (similar link for server(?))
- [ ] composition that includes tode/home for home directories
- [ ] directory map of tode/sys documenting the directory structure
- [ ] anything else that preserves changes users made in the old tode structure
- [ ] tag 1.0.0 before merge ... next release 2.0.0
- [ ] script for building /sys structure for exsiting stones
|
process
|
to compat ideas symbolic link from tode sys default clients to tode clients similar link for server composition that includes tode home for home directories directory map of tode sys documenting the directory structure anything else that preserves changes users made in the old tode structure tag before merge next release script for building sys structure for exsiting stones
| 1
|
111
| 3,460,514,519
|
IssuesEvent
|
2015-12-19 07:09:52
|
Automattic/wp-calypso
|
https://api.github.com/repos/Automattic/wp-calypso
|
closed
|
People: Redirect followers to frontend of site after accepting follow invite
|
People Management [Type] Task
|
Based on [this line](https://github.com/Automattic/wp-calypso/blob/master/client/accept-invite/logged-out-invite/signup-form.jsx#L53), we would redirect a logged out follow that signs up to the reader.
But, we'll also need to handle the case for users that only want to follow via email. For these users, let's redirect them to the front-end of the site that they followed.
|
1.0
|
People: Redirect followers to frontend of site after accepting follow invite - Based on [this line](https://github.com/Automattic/wp-calypso/blob/master/client/accept-invite/logged-out-invite/signup-form.jsx#L53), we would redirect a logged out follow that signs up to the reader.
But, we'll also need to handle the case for users that only want to follow via email. For these users, let's redirect them to the front-end of the site that they followed.
|
non_process
|
people redirect followers to frontend of site after accepting follow invite based on we would redirect a logged out follow that signs up to the reader but we ll also need to handle the case for users that only want to follow via email for these users let s redirect them to the front end of the site that they followed
| 0
|
14,488
| 17,602,756,093
|
IssuesEvent
|
2021-08-17 13:45:12
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Changes to viral term 'GO:0039723 suppression by virus of host TBK1 activity'
|
multi-species process
|
KW-1223 defines this keyword as 'Viral protein involved in the evasion of host innate defenses by inhibiting the TBK1 kinase. Upon viral infection, Toll-like receptors (TLRs) or DNA recognition receptors recognize foreign material and transmit the signal to TBK1 that in turn phosphorylates and activates IRF3 and IRF7. Once phosphorylated, IRF3 and/or IRF7 translocate into the nucleus to drive transcription of interferons. Several viruses including herpes simplex virus 1 or vaccinia virus interact directly with and inhibit TBK1 to prevent IRFs activation.' (https://www.uniprot.org/keywords/KW-1223)
Should we keep this term and move it, or merge it into 'suppression by virus of host toll-like receptor signaling pathway',
AND annotate the proteins annotated to the KW to MF 'protein serine/threonine kinase inhibitor activity' ?
If we keep the term, there are two possible parents for this term:
- 'evasion of host immune response', defined as 'A process by which an organism avoids the effects of the host organism's immune response.
- 'suppression by virus of host toll-like receptor signaling pathway', which is not 'evasion' but 'suppression'
- [ ] remove comment that refers to obsolete term: "Comment: When TBK1 acts as part of a TBK1-IKBKE-DDX3 complex, consider also annotating to: suppression by virus of host TBK1-IKBKE-DDX3 complex activity ; GO:0039659."
@pmasson55 what do you think is the best solution ?
Thanks, Pascale
|
1.0
|
Changes to viral term 'GO:0039723 suppression by virus of host TBK1 activity' - KW-1223 defines this keyword as 'Viral protein involved in the evasion of host innate defenses by inhibiting the TBK1 kinase. Upon viral infection, Toll-like receptors (TLRs) or DNA recognition receptors recognize foreign material and transmit the signal to TBK1 that in turn phosphorylates and activates IRF3 and IRF7. Once phosphorylated, IRF3 and/or IRF7 translocate into the nucleus to drive transcription of interferons. Several viruses including herpes simplex virus 1 or vaccinia virus interact directly with and inhibit TBK1 to prevent IRFs activation.' (https://www.uniprot.org/keywords/KW-1223)
Should we keep this term and move it, or merge it into 'suppression by virus of host toll-like receptor signaling pathway',
AND annotate the proteins annotated to the KW to MF 'protein serine/threonine kinase inhibitor activity' ?
If we keep the term, there are two possible parents for this term:
- 'evasion of host immune response', defined as 'A process by which an organism avoids the effects of the host organism's immune response.
- 'suppression by virus of host toll-like receptor signaling pathway', which is not 'evasion' but 'suppression'
- [ ] remove comment that refers to obsolete term: "Comment: When TBK1 acts as part of a TBK1-IKBKE-DDX3 complex, consider also annotating to: suppression by virus of host TBK1-IKBKE-DDX3 complex activity ; GO:0039659."
@pmasson55 what do you think is the best solution ?
Thanks, Pascale
|
process
|
changes to viral term go suppression by virus of host activity kw defines this keyword as viral protein involved in the evasion of host innate defenses by inhibiting the kinase upon viral infection toll like receptors tlrs or dna recognition receptors recognize foreign material and transmit the signal to that in turn phosphorylates and activates and once phosphorylated and or translocate into the nucleus to drive transcription of interferons several viruses including herpes simplex virus or vaccinia virus interact directly with and inhibit to prevent irfs activation should we keep this term and move it or merge it into suppression by virus of host toll like receptor signaling pathway and annotate the proteins annotated to the kw to mf protein serine threonine kinase inhibitor activity if we keep the term there are two possible parents for this term evasion of host immune response defined as a process by which an organism avoids the effects of the host organism s immune response suppression by virus of host toll like receptor signaling pathway which is not evasion but suppression remove comment that refers to obsolete term comment when acts as part of a ikbke complex consider also annotating to suppression by virus of host ikbke complex activity go what do you think is the best solution thanks pascale
| 1
|
5,502
| 8,368,990,172
|
IssuesEvent
|
2018-10-04 16:03:59
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
tail -f access.log,when logrotate run, real-time-html will not update
|
log-processing question
|
when i use cmd
`tail -f /var/log/nginx/access.log | /usr/local/bin/goaccess -o /home/wwwroot/default/report.html --real-time-html --log-format='%^ %^[%d:%t %^] "%r" %s %b "%R" "%u" "%h"' --date-format=%d/%b/%Y --time-format=%T --keep-db-files --load-from-disk --db-path=/home/wwwroot/default/ `
work normal.
but when logrotate run (daily run). **report.html** data will not update. so i have to rerun above cmd `tailf -f ....`
|
1.0
|
tail -f access.log,when logrotate run, real-time-html will not update - when i use cmd
`tail -f /var/log/nginx/access.log | /usr/local/bin/goaccess -o /home/wwwroot/default/report.html --real-time-html --log-format='%^ %^[%d:%t %^] "%r" %s %b "%R" "%u" "%h"' --date-format=%d/%b/%Y --time-format=%T --keep-db-files --load-from-disk --db-path=/home/wwwroot/default/ `
work normal.
but when logrotate run (daily run). **report.html** data will not update. so i have to rerun above cmd `tailf -f ....`
|
process
|
tail f access log when logrotate run real time html will not update when i use cmd tail f var log nginx access log usr local bin goaccess o home wwwroot default report html real time html log format r s b r u h date format d b y time format t keep db files load from disk db path home wwwroot default work normal but when logrotate run daily run report html data will not update so i have to rerun above cmd tailf f
| 1
|
8,994
| 12,103,679,101
|
IssuesEvent
|
2020-04-20 18:50:06
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Version 0.35 regression : column not found error after custom column
|
Priority:P2 Querying/Notebook Querying/Processor Type:Bug
|
**Describe the bug**
When creating a question with a summurize and a Custom Column based on the summurize the error "column not found" appears
**To Reproduce**
you just have to create a query in 0.35.0 or 0.35-RC2 like this one to reproduce the problem (the same query works in 0.34.3 :


Thanks
|
1.0
|
Version 0.35 regression : column not found error after custom column - **Describe the bug**
When creating a question with a summurize and a Custom Column based on the summurize the error "column not found" appears
**To Reproduce**
you just have to create a query in 0.35.0 or 0.35-RC2 like this one to reproduce the problem (the same query works in 0.34.3 :


Thanks
|
process
|
version regression column not found error after custom column describe the bug when creating a question with a summurize and a custom column based on the summurize the error column not found appears to reproduce you just have to create a query in or like this one to reproduce the problem the same query works in thanks
| 1
|
7,760
| 10,879,828,895
|
IssuesEvent
|
2019-11-17 05:37:47
|
google/ground-android
|
https://api.github.com/repos/google/ground-android
|
opened
|
[Feature] Code styling check
|
priority: p2 type: cleanup type: feature request type: process
|
- [ ] Checkstyle gradle plugin for code styling
- [ ] Run check on GCB with every commit and fail if necessary
- [ ] Generate reports
|
1.0
|
[Feature] Code styling check - - [ ] Checkstyle gradle plugin for code styling
- [ ] Run check on GCB with every commit and fail if necessary
- [ ] Generate reports
|
process
|
code styling check checkstyle gradle plugin for code styling run check on gcb with every commit and fail if necessary generate reports
| 1
|
374,793
| 11,095,582,851
|
IssuesEvent
|
2019-12-16 09:24:48
|
unep-grid/map-x-mgl
|
https://api.github.com/repos/unep-grid/map-x-mgl
|
closed
|
Problems with the "download" function
|
bug done priority 2
|
When downloading a dataset, the zip folder is not downloaded automatically as it used to be.
Users must now click on the "click here" green sentence at the bottom of the download window at the end of the download process. It is a problem because:
1. the content of the download window doesn't scroll down automatically to the end of the text where the user could see the sentence "click here to download". The window displays only the beginning of the text which may not be user-friendly.
2. it is much less intuitive
_Suggestion_: the dataset should be downloaded automatically at the end of the download process without clicking on anything.
Also, users receive an email containing a link to download the data, but the link doesn't work and says (for me) "Cette page ne fonctionne pas".
Link: http://api.mapx.org:443/download/mx_dl_5x1d2_rL6mL_FMGfj_iTwKb.zip
Console: there is nothing written in the console
|
1.0
|
Problems with the "download" function - When downloading a dataset, the zip folder is not downloaded automatically as it used to be.
Users must now click on the "click here" green sentence at the bottom of the download window at the end of the download process. It is a problem because:
1. the content of the download window doesn't scroll down automatically to the end of the text where the user could see the sentence "click here to download". The window displays only the beginning of the text which may not be user-friendly.
2. it is much less intuitive
_Suggestion_: the dataset should be downloaded automatically at the end of the download process without clicking on anything.
Also, users receive an email containing a link to download the data, but the link doesn't work and says (for me) "Cette page ne fonctionne pas".
Link: http://api.mapx.org:443/download/mx_dl_5x1d2_rL6mL_FMGfj_iTwKb.zip
Console: there is nothing written in the console
|
non_process
|
problems with the download function when downloading a dataset the zip folder is not downloaded automatically as it used to be users must now click on the click here green sentence at the bottom of the download window at the end of the download process it is a problem because the content of the download window doesn t scroll down automatically to the end of the text where the user could see the sentence click here to download the window displays only the beginning of the text which may not be user friendly it is much less intuitive suggestion the dataset should be downloaded automatically at the end of the download process without clicking on anything also users receive an email containing a link to download the data but the link doesn t work and says for me cette page ne fonctionne pas link console there is nothing written in the console
| 0
|
6,570
| 9,654,194,092
|
IssuesEvent
|
2019-05-19 12:08:06
|
brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
|
https://api.github.com/repos/brandon1roadgears/Interpreter-of-programming-language-of-Turing-Machine
|
closed
|
Проверка программы на работоспособность.
|
Testing process Work in process bug
|
### Необходимо проверить правильность выдаваемых программой ответов.
Для этого необходимо найти в интернете несколько задач с ответами и вписать правила в программы сравнить ответы ресурса и ответы полученные в нашей программе.
Думаю этого ресурса будет достаточно [https://docplayer.ru/32397731-Mashina-tyuringa-i-algoritmy-markova-reshenie-zadach.html](url)
|
2.0
|
Проверка программы на работоспособность. - ### Необходимо проверить правильность выдаваемых программой ответов.
Для этого необходимо найти в интернете несколько задач с ответами и вписать правила в программы сравнить ответы ресурса и ответы полученные в нашей программе.
Думаю этого ресурса будет достаточно [https://docplayer.ru/32397731-Mashina-tyuringa-i-algoritmy-markova-reshenie-zadach.html](url)
|
process
|
проверка программы на работоспособность необходимо проверить правильность выдаваемых программой ответов для этого необходимо найти в интернете несколько задач с ответами и вписать правила в программы сравнить ответы ресурса и ответы полученные в нашей программе думаю этого ресурса будет достаточно url
| 1
|
260,076
| 22,589,691,410
|
IssuesEvent
|
2022-06-28 18:32:31
|
danbudris/vulnerabilityProcessor
|
https://api.github.com/repos/danbudris/vulnerabilityProcessor
|
closed
|
LOW vulnerability in arn:aws:ecr:us-east-1:555555555555:repository/myrepo/sha256:7308b29228bde15a52a49b2f4a4cf95d5e2610e5ca67cdae32430e4b18effd91
|
hey there test severity/LOW
|
Issue auto cut by Vulnerability Processor version `v0.0.1-dev`
Message Source: `EventBridge`
Finding Source: `inspectorV2`
LOW Vulnerability CVE-2016-9085 detected in arn:aws:ecr:us-east-1:555555555555:repository/myrepo/sha256:7308b29228bde15a52a49b2f4a4cf95d5e2610e5ca67cdae32430e4b18effd91
Associated Pull Requests:
- https://github.com/danbudris/vulnerabilityProcessor/pull/106
|
1.0
|
LOW vulnerability in arn:aws:ecr:us-east-1:555555555555:repository/myrepo/sha256:7308b29228bde15a52a49b2f4a4cf95d5e2610e5ca67cdae32430e4b18effd91 - Issue auto cut by Vulnerability Processor version `v0.0.1-dev`
Message Source: `EventBridge`
Finding Source: `inspectorV2`
LOW Vulnerability CVE-2016-9085 detected in arn:aws:ecr:us-east-1:555555555555:repository/myrepo/sha256:7308b29228bde15a52a49b2f4a4cf95d5e2610e5ca67cdae32430e4b18effd91
Associated Pull Requests:
- https://github.com/danbudris/vulnerabilityProcessor/pull/106
|
non_process
|
low vulnerability in arn aws ecr us east repository myrepo issue auto cut by vulnerability processor version dev message source eventbridge finding source low vulnerability cve detected in arn aws ecr us east repository myrepo associated pull requests
| 0
|
54
| 2,515,308,901
|
IssuesEvent
|
2015-01-15 17:44:19
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
reopened
|
Validation for @cols attribute too drastic [DITA OT 2.0]
|
feature P2 preprocess won't fix
|
The Java method "ValidationFilter.validateCols" should not try to validate a `@cols` attribute value if it is set on a tgroup from a table element which has a conref.
For example this topic could be output to HTML using an older DITA OT:
<?xml version='1.0' encoding='UTF-8'?>
<!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "http://docs.oasis-open.org/dita/v1.1/OS/dtd/topic.dtd">
<topic id="introduction">
<title>Introduction</title>
<body>
<p>
<table conref="#introduction/table_ndw_gjy_qq" id="table_dkx_hjy_qq">
<tgroup cols="cols_rgx_hjy_qq">
<tbody>
<row>
<entry/>
</row>
</tbody>
</tgroup>
</table>
<table frame="all" rowsep="1" colsep="1" id="table_ndw_gjy_qq">
<title>abc</title>
<tgroup cols="2">
<colspec colname="c1" colnum="1" colwidth="1.0*"/>
<colspec colname="c2" colnum="2" colwidth="1.0*"/>
<thead>
<row>
<entry/>
<entry/>
</row>
</thead>
<tbody>
<row>
<entry/>
<entry/>
</row>
<row>
<entry/>
<entry/>
</row>
<row>
<entry/>
<entry/>
</row>
</tbody>
</tgroup>
</table></p>
</body>
</topic>
|
1.0
|
Validation for @cols attribute too drastic [DITA OT 2.0] - The Java method "ValidationFilter.validateCols" should not try to validate a `@cols` attribute value if it is set on a tgroup from a table element which has a conref.
For example this topic could be output to HTML using an older DITA OT:
<?xml version='1.0' encoding='UTF-8'?>
<!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "http://docs.oasis-open.org/dita/v1.1/OS/dtd/topic.dtd">
<topic id="introduction">
<title>Introduction</title>
<body>
<p>
<table conref="#introduction/table_ndw_gjy_qq" id="table_dkx_hjy_qq">
<tgroup cols="cols_rgx_hjy_qq">
<tbody>
<row>
<entry/>
</row>
</tbody>
</tgroup>
</table>
<table frame="all" rowsep="1" colsep="1" id="table_ndw_gjy_qq">
<title>abc</title>
<tgroup cols="2">
<colspec colname="c1" colnum="1" colwidth="1.0*"/>
<colspec colname="c2" colnum="2" colwidth="1.0*"/>
<thead>
<row>
<entry/>
<entry/>
</row>
</thead>
<tbody>
<row>
<entry/>
<entry/>
</row>
<row>
<entry/>
<entry/>
</row>
<row>
<entry/>
<entry/>
</row>
</tbody>
</tgroup>
</table></p>
</body>
</topic>
|
process
|
validation for cols attribute too drastic the java method validationfilter validatecols should not try to validate a cols attribute value if it is set on a tgroup from a table element which has a conref for example this topic could be output to html using an older dita ot doctype topic public oasis dtd dita topic en introduction abc
| 1
|
248,328
| 7,929,370,897
|
IssuesEvent
|
2018-07-06 14:52:06
|
containous/traefik
|
https://api.github.com/repos/containous/traefik
|
closed
|
Allow configuration of auth.headerField via labels
|
area/authentication kind/enhancement priority/P2
|
<!--
DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
The issue tracker is for reporting bugs and feature requests only.
For end-user related support questions, refer to one of the following:
- Stack Overflow (using the "traefik" tag): https://stackoverflow.com/questions/tagged/traefik
- the Traefik community Slack channel: https://traefik.herokuapp.com
-->
This is a feature request. In configuring authentication, I can pass the authenticated user to the frontend via an HTTP header using configuration like this:
```toml
# traefik.toml
[entryPoints]
[entryPoints.http]
address = ":80"
[entryPoints.http.auth.basic]
users = ["test:$apr1$H6uskkkW$IgXLP6ewTrSuBkTrqE8wj/"]
[entryPoints.http.auth]
headerField = "X-Webauth-User"
```
This works great when I specify the basic auth users in the entrypoint, but fails when I remove the `[entryPoints.http.auth.basic]` block in order to move that configuration back into a label specified in docker-compose.yml. I'd like to be able to set a custom headerField for each application using a docker label such as:
```yml
# docker-compose.yml
myservice:
image: myimage
labels:
traefik.frontend.auth.headerField: "X-Webauth-User"
traefik.frontend.auth.basic: "test:$$apr1$$H6uskkkW$$IgXLP6ewTrSuBkTrqE8wj/"
```
### Output of `traefik version`:
```
Version: v1.3.8
Codename: raclette
Go version: go1.8.3
Built: 2017-09-07_08:46:19PM
OS/Arch: linux/amd64
```
### What is your environment & configuration (arguments, toml, provider, platform, ...)?
```toml
defaultEntryPoints = ["http", "https"]
[entryPoints]
[entryPoints.http]
address = ":80"
[entryPoints.http.redirect]
entryPoint= "https"
[entryPoints.https]
address = ":8011"
[entryPoints.https.auth.basic]
#[entryPoints.https.tls]
[web]
address=":8080"
[docker]
domain="docker.local"
watch=true
exposedbydefault = false
```
|
1.0
|
Allow configuration of auth.headerField via labels - <!--
DO NOT FILE ISSUES FOR GENERAL SUPPORT QUESTIONS.
The issue tracker is for reporting bugs and feature requests only.
For end-user related support questions, refer to one of the following:
- Stack Overflow (using the "traefik" tag): https://stackoverflow.com/questions/tagged/traefik
- the Traefik community Slack channel: https://traefik.herokuapp.com
-->
This is a feature request. In configuring authentication, I can pass the authenticated user to the frontend via an HTTP header using configuration like this:
```toml
# traefik.toml
[entryPoints]
[entryPoints.http]
address = ":80"
[entryPoints.http.auth.basic]
users = ["test:$apr1$H6uskkkW$IgXLP6ewTrSuBkTrqE8wj/"]
[entryPoints.http.auth]
headerField = "X-Webauth-User"
```
This works great when I specify the basic auth users in the entrypoint, but fails when I remove the `[entryPoints.http.auth.basic]` block in order to move that configuration back into a label specified in docker-compose.yml. I'd like to be able to set a custom headerField for each application using a docker label such as:
```yml
# docker-compose.yml
myservice:
image: myimage
labels:
traefik.frontend.auth.headerField: "X-Webauth-User"
traefik.frontend.auth.basic: "test:$$apr1$$H6uskkkW$$IgXLP6ewTrSuBkTrqE8wj/"
```
### Output of `traefik version`:
```
Version: v1.3.8
Codename: raclette
Go version: go1.8.3
Built: 2017-09-07_08:46:19PM
OS/Arch: linux/amd64
```
### What is your environment & configuration (arguments, toml, provider, platform, ...)?
```toml
defaultEntryPoints = ["http", "https"]
[entryPoints]
[entryPoints.http]
address = ":80"
[entryPoints.http.redirect]
entryPoint= "https"
[entryPoints.https]
address = ":8011"
[entryPoints.https.auth.basic]
#[entryPoints.https.tls]
[web]
address=":8080"
[docker]
domain="docker.local"
watch=true
exposedbydefault = false
```
|
non_process
|
allow configuration of auth headerfield via labels do not file issues for general support questions the issue tracker is for reporting bugs and feature requests only for end user related support questions refer to one of the following stack overflow using the traefik tag the traefik community slack channel this is a feature request in configuring authentication i can pass the authenticated user to the frontend via an http header using configuration like this toml traefik toml address users headerfield x webauth user this works great when i specify the basic auth users in the entrypoint but fails when i remove the block in order to move that configuration back into a label specified in docker compose yml i d like to be able to set a custom headerfield for each application using a docker label such as yml docker compose yml myservice image myimage labels traefik frontend auth headerfield x webauth user traefik frontend auth basic test output of traefik version version codename raclette go version built os arch linux what is your environment configuration arguments toml provider platform toml defaultentrypoints address entrypoint https address address domain docker local watch true exposedbydefault false
| 0
|
11,130
| 13,957,688,580
|
IssuesEvent
|
2020-10-24 08:09:34
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
RO: Question related to harvest results with id INSPIRE-7edbed58-ddbc-11e4-b469-52540004b857_20190610-084632 - Issue while initializing the Discover Matadata Operation
|
Geoportal Harvesting process RO - Romania
|
Dear Angelo,
Dear Davide,
After the harvesting process we have an issue while initializing the Discovery Metadata Operation.
Could you help us to fix this error : The interaction with the remote service at "http://geoportal.gov.ro/Geoportal_INIS/csw202/discovery?service=csw&version=2.0.2&request=GetCapabilities" ended with the following error "javax.xml.stream.FactoryConfigurationError"
Result of the interaction with the Discovery Service:
Resources available for discovery: 159, Expected Resource Count: 140, Actual Resource Count : 140.
Link Evaluation Report : http://inspire-geoportal.ec.europa.eu/sandbox/resources/INSPIRE-7edbed58-ddbc-11e4-b469-52540004b857_20190610-084632/services/1/PullResults/.
Best regards,
Simona Bunea
|
1.0
|
RO: Question related to harvest results with id INSPIRE-7edbed58-ddbc-11e4-b469-52540004b857_20190610-084632 - Issue while initializing the Discover Matadata Operation - Dear Angelo,
Dear Davide,
After the harvesting process we have an issue while initializing the Discovery Metadata Operation.
Could you help us to fix this error : The interaction with the remote service at "http://geoportal.gov.ro/Geoportal_INIS/csw202/discovery?service=csw&version=2.0.2&request=GetCapabilities" ended with the following error "javax.xml.stream.FactoryConfigurationError"
Result of the interaction with the Discovery Service:
Resources available for discovery: 159, Expected Resource Count: 140, Actual Resource Count : 140.
Link Evaluation Report : http://inspire-geoportal.ec.europa.eu/sandbox/resources/INSPIRE-7edbed58-ddbc-11e4-b469-52540004b857_20190610-084632/services/1/PullResults/.
Best regards,
Simona Bunea
|
process
|
ro question related to harvest results with id inspire ddbc issue while initializing the discover matadata operation dear angelo dear davide after the harvesting process we have an issue while initializing the discovery metadata operation could you help us to fix this error the interaction with the remote service at quot ended with the following error quot javax xml stream factoryconfigurationerror quot result of the interaction with the discovery service resources available for discovery expected resource count actual resource count link evaluation report best regards simona bunea
| 1
|
15,693
| 11,661,944,684
|
IssuesEvent
|
2020-03-03 08:04:57
|
ampproject/amp.dev
|
https://api.github.com/repos/ampproject/amp.dev
|
opened
|
Upgrade uuid in amp-analytics API to v7
|
Category: Infrastructure P2: Medium Type: Update
|
uuid introduced breaking changes with v7 we need to adapt to prior updating. See https://github.com/ampproject/amp.dev/pull/3635 for details.
|
1.0
|
Upgrade uuid in amp-analytics API to v7 - uuid introduced breaking changes with v7 we need to adapt to prior updating. See https://github.com/ampproject/amp.dev/pull/3635 for details.
|
non_process
|
upgrade uuid in amp analytics api to uuid introduced breaking changes with we need to adapt to prior updating see for details
| 0
|
18,970
| 11,102,232,773
|
IssuesEvent
|
2019-12-16 23:21:55
|
Azure/azure-cli
|
https://api.github.com/repos/Azure/azure-cli
|
closed
|
azure-cli-acs module requires some strict type checking for all the input parameters
|
AKS Service Attention
|
**Describe the bug**
azure-cli\src\command_modules\azure-cli-acs\azure\cli\command_modules\acs\custom.py
has functions which type cast to int, this needs to have a stricter type checking during input parameter itself.
example: count=int(node_count),
**To Reproduce**
Code improvement to have strict type checking.
**Expected behavior**
Not lead to run time errors due to casting
**Environment summary**
Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows)
**Additional context**
|
1.0
|
azure-cli-acs module requires some strict type checking for all the input parameters - **Describe the bug**
azure-cli\src\command_modules\azure-cli-acs\azure\cli\command_modules\acs\custom.py
has functions which type cast to int, this needs to have a stricter type checking during input parameter itself.
example: count=int(node_count),
**To Reproduce**
Code improvement to have strict type checking.
**Expected behavior**
Not lead to run time errors due to casting
**Environment summary**
Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows)
**Additional context**
|
non_process
|
azure cli acs module requires some strict type checking for all the input parameters describe the bug azure cli src command modules azure cli acs azure cli command modules acs custom py has functions which type cast to int this needs to have a stricter type checking during input parameter itself example count int node count to reproduce code improvement to have strict type checking expected behavior not lead to run time errors due to casting environment summary install method e g pip interactive script apt get docker msi edge build cli version az version os version shell type e g bash cmd exe bash on windows additional context
| 0
|
201,302
| 22,948,163,991
|
IssuesEvent
|
2022-07-19 03:39:36
|
elikkatzgit/TestingPOM
|
https://api.github.com/repos/elikkatzgit/TestingPOM
|
reopened
|
CVE-2020-36187 (High) detected in jackson-databind-2.7.2.jar
|
security vulnerability
|
## CVE-2020-36187 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.2.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36187>CVE-2020-36187</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: 2.9.10.8</p>
</p>
</details>
<p></p>
|
True
|
CVE-2020-36187 (High) detected in jackson-databind-2.7.2.jar - ## CVE-2020-36187 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.2.jar** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36187>CVE-2020-36187</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: 2.9.10.8</p>
</p>
</details>
<p></p>
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href dependency hierarchy x jackson databind jar vulnerable library found in base branch dev vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution
| 0
|
86,801
| 3,733,664,717
|
IssuesEvent
|
2016-03-08 01:30:23
|
openshift/origin
|
https://api.github.com/repos/openshift/origin
|
opened
|
Builder tags should be sorted as numbers when possible in UI
|
component/web kind/bug priority/P3
|
Builder wildfly:10.0 should appear after wildfly:9.0. Tags aren't always numbers, but we should compare them as numbers when sorting if possible.
<img width="380" alt="openshift_web_console" src="https://cloud.githubusercontent.com/assets/1167259/13589056/f8bb137e-e4a2-11e5-9ca9-318f7f320c25.png">
/cc @bparees @jwforres
|
1.0
|
Builder tags should be sorted as numbers when possible in UI - Builder wildfly:10.0 should appear after wildfly:9.0. Tags aren't always numbers, but we should compare them as numbers when sorting if possible.
<img width="380" alt="openshift_web_console" src="https://cloud.githubusercontent.com/assets/1167259/13589056/f8bb137e-e4a2-11e5-9ca9-318f7f320c25.png">
/cc @bparees @jwforres
|
non_process
|
builder tags should be sorted as numbers when possible in ui builder wildfly should appear after wildfly tags aren t always numbers but we should compare them as numbers when sorting if possible img width alt openshift web console src cc bparees jwforres
| 0
|
11,289
| 9,081,903,389
|
IssuesEvent
|
2019-02-17 06:55:45
|
amethyst/amethyst
|
https://api.github.com/repos/amethyst/amethyst
|
closed
|
CI does not check json feature
|
diff: easy pri: important status: ready team: engine team: infrastructure type: bug
|
We're currently not testing the `json` feature, and it turns out it doesn't compile:
https://github.com/amethyst/amethyst/blob/38877ed47c5bb9a7869276fdea545085ade0d81e/amethyst_assets/src/formats.rs#L48-L49
`chain_err` is not defined. Is there any convention for how to migrate to the new error handling @udoprog? I can't find it in the doc.
|
1.0
|
CI does not check json feature - We're currently not testing the `json` feature, and it turns out it doesn't compile:
https://github.com/amethyst/amethyst/blob/38877ed47c5bb9a7869276fdea545085ade0d81e/amethyst_assets/src/formats.rs#L48-L49
`chain_err` is not defined. Is there any convention for how to migrate to the new error handling @udoprog? I can't find it in the doc.
|
non_process
|
ci does not check json feature we re currently not testing the json feature and it turns out it doesn t compile chain err is not defined is there any convention for how to migrate to the new error handling udoprog i can t find it in the doc
| 0
|
526,050
| 15,278,867,378
|
IssuesEvent
|
2021-02-23 02:32:11
|
space-wizards/space-station-14
|
https://api.github.com/repos/space-wizards/space-station-14
|
closed
|
MedicalScanner needs collision fixed
|
Difficulty: 2 - Medium Priority: 2-medium Size: 1 - Tiny Type: Improvement
|
Ideally it's passable to mobs but impassable to walls etc (if it gets moved around) so do with the mask and layers what needs doing.
Need it because if a player moves out of the scanner they will immediately clip.
|
1.0
|
MedicalScanner needs collision fixed - Ideally it's passable to mobs but impassable to walls etc (if it gets moved around) so do with the mask and layers what needs doing.
Need it because if a player moves out of the scanner they will immediately clip.
|
non_process
|
medicalscanner needs collision fixed ideally it s passable to mobs but impassable to walls etc if it gets moved around so do with the mask and layers what needs doing need it because if a player moves out of the scanner they will immediately clip
| 0
|
22,412
| 31,142,293,567
|
IssuesEvent
|
2023-08-16 01:44:54
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Flaky test: <OpenBrowserList /> throws when activeBrowser is null
|
OS: linux process: flaky test topic: flake ❄️ stage: flake stale
|
### Current behavior
This is our top flaky test on Linux, having flaked 440 times in the last 30 days at time of writing. See [Dashboard Insights](https://app.circleci.com/insights/github/cypress-io/cypress/workflows/linux-x64/tests?branch=develop&reporting-window=last-30-days)
We're trying to test an error state, but the error itself is causing the test to flake:
<img width="1423" alt="Screen Shot 2022-08-03 at 2 55 03 PM" src="https://user-images.githubusercontent.com/26726429/182719393-dfaca960-7315-4245-bf39-8406be88ffa5.png">
```
cy.once('uncaught:exception', (err) => {
expect(err.message).to.include('Missing activeBrowser in selectedBrowserId')
done()
})
```
Looks like an easy fix, just need to tweak the logic for asserting the error
### Desired behavior
No more flake 😃
### Test code to reproduce
https://github.com/cypress-io/cypress/blob/develop/packages/launchpad/src/setup/OpenBrowserList.cy.tsx#L134
### Cypress Version
10.4.0
### Other
https://github.com/cypress-io/cypress/issues/23099
|
1.0
|
Flaky test: <OpenBrowserList /> throws when activeBrowser is null - ### Current behavior
This is our top flaky test on Linux, having flaked 440 times in the last 30 days at time of writing. See [Dashboard Insights](https://app.circleci.com/insights/github/cypress-io/cypress/workflows/linux-x64/tests?branch=develop&reporting-window=last-30-days)
We're trying to test an error state, but the error itself is causing the test to flake:
<img width="1423" alt="Screen Shot 2022-08-03 at 2 55 03 PM" src="https://user-images.githubusercontent.com/26726429/182719393-dfaca960-7315-4245-bf39-8406be88ffa5.png">
```
cy.once('uncaught:exception', (err) => {
expect(err.message).to.include('Missing activeBrowser in selectedBrowserId')
done()
})
```
Looks like an easy fix, just need to tweak the logic for asserting the error
### Desired behavior
No more flake 😃
### Test code to reproduce
https://github.com/cypress-io/cypress/blob/develop/packages/launchpad/src/setup/OpenBrowserList.cy.tsx#L134
### Cypress Version
10.4.0
### Other
https://github.com/cypress-io/cypress/issues/23099
|
process
|
flaky test throws when activebrowser is null current behavior this is our top flaky test on linux having flaked times in the last days at time of writing see we re trying to test an error state but the error itself is causing the test to flake img width alt screen shot at pm src cy once uncaught exception err expect err message to include missing activebrowser in selectedbrowserid done looks like an easy fix just need to tweak the logic for asserting the error desired behavior no more flake 😃 test code to reproduce cypress version other
| 1
|
7,327
| 10,468,917,986
|
IssuesEvent
|
2019-09-22 17:02:24
|
produvia/ai-platform
|
https://api.github.com/repos/produvia/ai-platform
|
closed
|
Sentiment Analysis
|
natural-language-processing task wontfix
|
# Goal(s)
- Classify the polarity of a given text
# Input(s)
- Text
# Output(s)
- Text
# Objective Function(s)
- TBD
|
1.0
|
Sentiment Analysis - # Goal(s)
- Classify the polarity of a given text
# Input(s)
- Text
# Output(s)
- Text
# Objective Function(s)
- TBD
|
process
|
sentiment analysis goal s classify the polarity of a given text input s text output s text objective function s tbd
| 1
|
74,274
| 9,011,520,865
|
IssuesEvent
|
2019-02-05 14:53:22
|
iagodahlem/tiempo
|
https://api.github.com/repos/iagodahlem/tiempo
|
opened
|
Design Debits
|
design help wanted
|
- Optimize it to look better on desktop view.
- Find better icons to keep the look n' feel of the app.
- Avoid content jumping on the timer numbers.
|
1.0
|
Design Debits - - Optimize it to look better on desktop view.
- Find better icons to keep the look n' feel of the app.
- Avoid content jumping on the timer numbers.
|
non_process
|
design debits optimize it to look better on desktop view find better icons to keep the look n feel of the app avoid content jumping on the timer numbers
| 0
|
369,449
| 25,847,981,276
|
IssuesEvent
|
2022-12-13 08:19:48
|
ophub/amlogic-s9xxx-armbian
|
https://api.github.com/repos/ophub/amlogic-s9xxx-armbian
|
closed
|
M401A 制作了启动盘,但是黑屏无法启动
|
documentation
|
**Device Information | 设备信息**
- SOC: s905l3a (问了卖家)
- Model 魔百和 M401A
**Armbian Version | 系统版本**
- Kernel Version: 5.15.82
- Release: jammy
**Describe the bug | 问题描述**
下载和烧录以下镜像:
Armbian_23.02.0_amlogic_s905l3a_jammy_5.15.82_server_2022.12.09.img.gz
Armbian_23.02.0_amlogic_s905l3a_jammy_6.0.12_server_2022.12.09.img.gz
用adb connect IP
adb shell reboot update
然后插入优盘,屏幕提示无视频输入,然后一直卡在这里。
如果不插入优盘,会提示update失败和重启菜单。
请问有什么调试和判断方法?
|
1.0
|
M401A 制作了启动盘,但是黑屏无法启动 -
**Device Information | 设备信息**
- SOC: s905l3a (问了卖家)
- Model 魔百和 M401A
**Armbian Version | 系统版本**
- Kernel Version: 5.15.82
- Release: jammy
**Describe the bug | 问题描述**
下载和烧录以下镜像:
Armbian_23.02.0_amlogic_s905l3a_jammy_5.15.82_server_2022.12.09.img.gz
Armbian_23.02.0_amlogic_s905l3a_jammy_6.0.12_server_2022.12.09.img.gz
用adb connect IP
adb shell reboot update
然后插入优盘,屏幕提示无视频输入,然后一直卡在这里。
如果不插入优盘,会提示update失败和重启菜单。
请问有什么调试和判断方法?
|
non_process
|
制作了启动盘,但是黑屏无法启动 device information 设备信息 soc (问了卖家) model 魔百和 armbian version 系统版本 kernel version release jammy describe the bug 问题描述 下载和烧录以下镜像: armbian amlogic jammy server img gz armbian amlogic jammy server img gz 用adb connect ip adb shell reboot update 然后插入优盘,屏幕提示无视频输入,然后一直卡在这里。 如果不插入优盘,会提示update失败和重启菜单。 请问有什么调试和判断方法?
| 0
|
303,261
| 22,961,376,628
|
IssuesEvent
|
2022-07-19 15:38:23
|
JanssenProject/jans
|
https://api.github.com/repos/JanssenProject/jans
|
opened
|
docs: add artifact link
|
area-documentation
|
- [x] Add artifact link for janssen to the main readme
- [x] add latest release link to the main readme
|
1.0
|
docs: add artifact link - - [x] Add artifact link for janssen to the main readme
- [x] add latest release link to the main readme
|
non_process
|
docs add artifact link add artifact link for janssen to the main readme add latest release link to the main readme
| 0
|
235,078
| 18,041,333,645
|
IssuesEvent
|
2021-09-18 04:52:46
|
girlscript/winter-of-contributing
|
https://api.github.com/repos/girlscript/winter-of-contributing
|
opened
|
Python: sets
|
documentation GWOC21 Python Video Audio
|
### Description :
Briefly explain sets in python
### Note :
- Changes should be made inside the `Python/` directory & `Python` branch.
- Issue only for `GWOC'21` contributors.
- Issue will be assigned on a **first come first serve basis**, **1 Issue == 1 PR.**
- In the PR, keep this issue's title as your PR's title.
#### Must Follow : [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of Conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) before start Contributing.
|
1.0
|
Python: sets - ### Description :
Briefly explain sets in python
### Note :
- Changes should be made inside the `Python/` directory & `Python` branch.
- Issue only for `GWOC'21` contributors.
- Issue will be assigned on a **first come first serve basis**, **1 Issue == 1 PR.**
- In the PR, keep this issue's title as your PR's title.
#### Must Follow : [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of Conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) before start Contributing.
|
non_process
|
python sets description briefly explain sets in python note changes should be made inside the python directory python branch issue only for gwoc contributors issue will be assigned on a first come first serve basis issue pr in the pr keep this issue s title as your pr s title must follow before start contributing
| 0
|
14,316
| 17,333,624,174
|
IssuesEvent
|
2021-07-28 07:27:31
|
2i2c-org/team-compass
|
https://api.github.com/repos/2i2c-org/team-compass
|
closed
|
Pilot Hubs should have a single non-2i2c admin user manually specified
|
:label: team-process type: enhancement
|
# Summary
We have run into some confusion with setting admin users on some of our hubs, where people didn't know whether they were admins and were asking 2i2c for admin users status.
This is some extra noise that we have to deal with, and it would be simpler if there were a single, repeatable policy for who gets admin user status on a hub.
I propose that we adopt the following practice:
- All 2i2c engineers get admin user status on hubs via **manual configuration** (ideally this would be automatically interpolated from config somewhere)
- The Community Representative gets admin user status on hubs via **manual configuration**
- All other Hub Administrators must be added via the JupyterHub UI **by the Community Representative**.
# Important information
- Some discussion in slack: https://2i2c.slack.com/archives/C01DB2JRP8W/p1626768120154300
- An issue that led to this conversation https://github.com/2i2c-org/pilot-hubs/issues/529
# Tasks to complete
- [ ] Discuss and/or agree with this policy
- [ ] Write it up into the pilot hubs documentation
- [ ] Done!
Note: I think we can leave our *current* hubs as-is, and adopt this for any new hubs.
|
1.0
|
Pilot Hubs should have a single non-2i2c admin user manually specified - # Summary
We have run into some confusion with setting admin users on some of our hubs, where people didn't know whether they were admins and were asking 2i2c for admin users status.
This is some extra noise that we have to deal with, and it would be simpler if there were a single, repeatable policy for who gets admin user status on a hub.
I propose that we adopt the following practice:
- All 2i2c engineers get admin user status on hubs via **manual configuration** (ideally this would be automatically interpolated from config somewhere)
- The Community Representative gets admin user status on hubs via **manual configuration**
- All other Hub Administrators must be added via the JupyterHub UI **by the Community Representative**.
# Important information
- Some discussion in slack: https://2i2c.slack.com/archives/C01DB2JRP8W/p1626768120154300
- An issue that led to this conversation https://github.com/2i2c-org/pilot-hubs/issues/529
# Tasks to complete
- [ ] Discuss and/or agree with this policy
- [ ] Write it up into the pilot hubs documentation
- [ ] Done!
Note: I think we can leave our *current* hubs as-is, and adopt this for any new hubs.
|
process
|
pilot hubs should have a single non admin user manually specified summary we have run into some confusion with setting admin users on some of our hubs where people didn t know whether they were admins and were asking for admin users status this is some extra noise that we have to deal with and it would be simpler if there were a single repeatable policy for who gets admin user status on a hub i propose that we adopt the following practice all engineers get admin user status on hubs via manual configuration ideally this would be automatically interpolated from config somewhere the community representative gets admin user status on hubs via manual configuration all other hub administrators must be added via the jupyterhub ui by the community representative important information some discussion in slack an issue that led to this conversation tasks to complete discuss and or agree with this policy write it up into the pilot hubs documentation done note i think we can leave our current hubs as is and adopt this for any new hubs
| 1
|
18,392
| 24,529,890,469
|
IssuesEvent
|
2022-10-11 15:37:05
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
incompatible_enforce_config_setting_visibility
|
P2 type: process team-Configurability incompatible-change migration-ready breaking-change-6.0
|
Visibility on `config_setting` isn't historically enforced. This is purely for legacy reasons. There's no philosophical reason to distinguish them.
This flag starts the process of removing the distinction.
Values:
* `--incompatible_enforce_config_setting_visibility=off`: every `config_setting` is visible to every target, regardless of visibility settings
* `--incompatible_enforce_config_setting_visibility=on`: `config_setting` follows the policy set by `--incompatible_config_setting_private_default_visibility` (https://github.com/bazelbuild/bazel/issues/12933).
**Incompatibility error:**
`ERROR: myapp/BUILD:4:1: in config_setting rule //myapp:my_config: target 'myapp:my_config' is not visible from target '//some:other_target. Check the visibility declaration of the former target if you think the dependency is legitimate`
**Migration:**
Treat all `config_setting`s as if they follow standard visibility logic at https://docs.bazel.build/versions/master/visibility.html: have them set visibility explicitly if they'll be used anywhere outside their own package. The ultimate goal of this migration is to fully enforce that expectation.
|
1.0
|
incompatible_enforce_config_setting_visibility - Visibility on `config_setting` isn't historically enforced. This is purely for legacy reasons. There's no philosophical reason to distinguish them.
This flag starts the process of removing the distinction.
Values:
* `--incompatible_enforce_config_setting_visibility=off`: every `config_setting` is visible to every target, regardless of visibility settings
* `--incompatible_enforce_config_setting_visibility=on`: `config_setting` follows the policy set by `--incompatible_config_setting_private_default_visibility` (https://github.com/bazelbuild/bazel/issues/12933).
**Incompatibility error:**
`ERROR: myapp/BUILD:4:1: in config_setting rule //myapp:my_config: target 'myapp:my_config' is not visible from target '//some:other_target. Check the visibility declaration of the former target if you think the dependency is legitimate`
**Migration:**
Treat all `config_setting`s as if they follow standard visibility logic at https://docs.bazel.build/versions/master/visibility.html: have them set visibility explicitly if they'll be used anywhere outside their own package. The ultimate goal of this migration is to fully enforce that expectation.
|
process
|
incompatible enforce config setting visibility visibility on config setting isn t historically enforced this is purely for legacy reasons there s no philosophical reason to distinguish them this flag starts the process of removing the distinction values incompatible enforce config setting visibility off every config setting is visible to every target regardless of visibility settings incompatible enforce config setting visibility on config setting follows the policy set by incompatible config setting private default visibility incompatibility error error myapp build in config setting rule myapp my config target myapp my config is not visible from target some other target check the visibility declaration of the former target if you think the dependency is legitimate migration treat all config setting s as if they follow standard visibility logic at have them set visibility explicitly if they ll be used anywhere outside their own package the ultimate goal of this migration is to fully enforce that expectation
| 1
|
445,437
| 31,238,102,623
|
IssuesEvent
|
2023-08-20 14:20:17
|
KrisztanZero/terra-custos
|
https://api.github.com/repos/KrisztanZero/terra-custos
|
closed
|
Internationalization and Localization
|
documentation
|
### If the application operates in multiple languages, design the internationalization and localization processes. Handling text and date formatting based on different cultural preferences ensures widespread usability.
- [ ] Internationalization and Localization
|
1.0
|
Internationalization and Localization - ### If the application operates in multiple languages, design the internationalization and localization processes. Handling text and date formatting based on different cultural preferences ensures widespread usability.
- [ ] Internationalization and Localization
|
non_process
|
internationalization and localization if the application operates in multiple languages design the internationalization and localization processes handling text and date formatting based on different cultural preferences ensures widespread usability internationalization and localization
| 0
|
22,356
| 31,032,990,460
|
IssuesEvent
|
2023-08-10 13:37:11
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
[MLv2] Column not found errors when joining cards
|
.Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
We have two failing E2E tests on the [joins FE branch](https://github.com/metabase/metabase/pull/32912) when a BE now returns an error when running a query with joined cards:
```
Column "Question 5 - Products → Created At: Month.CREATED_AT" not found;
SQL statement: ...
```
**Failing tests**
- [should join two saved questions with the same implicit/explicit grouped field (metabase#18512)](https://github.com/metabase/metabase/blob/master/e2e/test/scenarios/joins/reproductions/18512-cannot-join-two-saved-questions-with-same-implicit-explicit-grouped-field.cy.spec.js)
- [should join saved questions that themselves contain joins (metabase#12928)](https://github.com/metabase/metabase/blob/master/e2e/test/scenarios/joins/joins.cy.spec.js)
|
1.0
|
[MLv2] Column not found errors when joining cards - We have two failing E2E tests on the [joins FE branch](https://github.com/metabase/metabase/pull/32912) when a BE now returns an error when running a query with joined cards:
```
Column "Question 5 - Products → Created At: Month.CREATED_AT" not found;
SQL statement: ...
```
**Failing tests**
- [should join two saved questions with the same implicit/explicit grouped field (metabase#18512)](https://github.com/metabase/metabase/blob/master/e2e/test/scenarios/joins/reproductions/18512-cannot-join-two-saved-questions-with-same-implicit-explicit-grouped-field.cy.spec.js)
- [should join saved questions that themselves contain joins (metabase#12928)](https://github.com/metabase/metabase/blob/master/e2e/test/scenarios/joins/joins.cy.spec.js)
|
process
|
column not found errors when joining cards we have two failing tests on the when a be now returns an error when running a query with joined cards column question products → created at month created at not found sql statement failing tests
| 1
|
5,752
| 8,597,470,669
|
IssuesEvent
|
2018-11-15 18:46:57
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
opened
|
Firestore: support features blacklisted in conformance tests
|
api: firestore type: process
|
PR #6290 blacklisted a number of conformance tests because we do not currently support the usecases they support:
- `get-*` tests (because we use `BatchGetDocuments` API rather than the `GetDocument` API.
- `listen-*` tests exercise the "watch" features (since landed in PR #6191).
- `update_paths-*` tests (they've been excluded forever, with a note that Python lacked the support).
- Tests involving unimplemented / incorrectly implemented "transforms" (`DELETE`, `ARRAY_REMOVE`, `ARRAY_UNION`).
- `query-*` tests have been (inadvertently) skipped since being copied in from `google-cloud-common` in PR #5351).
This is a tracking issue for removing those skips / blacklist entries:
- [ ] Update `Document.get` to use the `GetDocument` API, and re-enable the (one) `get-basic.textproto` test.
- [ ] Enable the `listen-*` tests and make changes needed for them to pass.
- [ ] Enable the `query-*` tests and make changes needed for them to pass.
- [ ] Figure how to support the "update with paths" use case, and enable those tests.
- [ ] Fix the support for the `DELETE` transform (one failing test) and enable those tests.
- [ ] Figure out how to support the `ARRAY_REMOVE` transform, and enable those tests.
- [ ] Figure out how to support the `ARRAY_UNION` transform, and enable those tests.
|
1.0
|
Firestore: support features blacklisted in conformance tests - PR #6290 blacklisted a number of conformance tests because we do not currently support the usecases they support:
- `get-*` tests (because we use `BatchGetDocuments` API rather than the `GetDocument` API.
- `listen-*` tests exercise the "watch" features (since landed in PR #6191).
- `update_paths-*` tests (they've been excluded forever, with a note that Python lacked the support).
- Tests involving unimplemented / incorrectly implemented "transforms" (`DELETE`, `ARRAY_REMOVE`, `ARRAY_UNION`).
- `query-*` tests have been (inadvertently) skipped since being copied in from `google-cloud-common` in PR #5351).
This is a tracking issue for removing those skips / blacklist entries:
- [ ] Update `Document.get` to use the `GetDocument` API, and re-enable the (one) `get-basic.textproto` test.
- [ ] Enable the `listen-*` tests and make changes needed for them to pass.
- [ ] Enable the `query-*` tests and make changes needed for them to pass.
- [ ] Figure how to support the "update with paths" use case, and enable those tests.
- [ ] Fix the support for the `DELETE` transform (one failing test) and enable those tests.
- [ ] Figure out how to support the `ARRAY_REMOVE` transform, and enable those tests.
- [ ] Figure out how to support the `ARRAY_UNION` transform, and enable those tests.
|
process
|
firestore support features blacklisted in conformance tests pr blacklisted a number of conformance tests because we do not currently support the usecases they support get tests because we use batchgetdocuments api rather than the getdocument api listen tests exercise the watch features since landed in pr update paths tests they ve been excluded forever with a note that python lacked the support tests involving unimplemented incorrectly implemented transforms delete array remove array union query tests have been inadvertently skipped since being copied in from google cloud common in pr this is a tracking issue for removing those skips blacklist entries update document get to use the getdocument api and re enable the one get basic textproto test enable the listen tests and make changes needed for them to pass enable the query tests and make changes needed for them to pass figure how to support the update with paths use case and enable those tests fix the support for the delete transform one failing test and enable those tests figure out how to support the array remove transform and enable those tests figure out how to support the array union transform and enable those tests
| 1
|
18,587
| 24,567,937,441
|
IssuesEvent
|
2022-10-13 05:58:32
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Android] Comprehension test is not getting failed in the updated consent flow
|
Bug P0 Android Process: Fixed Process: Tested QA Process: Tested dev
|
**Steps:**
1. Sign in and enroll to the study
2. Now, Go to SB and update the consent for enrolled participants
3. Open the mobile app and complete the eligibility section
4. Navigate to comprehension section
5. Fail the comprehension section and Observe
**AR:** Comprehension test is not getting failed in the updated consent flow
**ER:** Comprehension test should be failed in the updated consent flow
https://user-images.githubusercontent.com/86007179/188439152-cb26d4dd-671f-40c5-9895-50dbd016ac70.mp4
|
3.0
|
[Android] Comprehension test is not getting failed in the updated consent flow - **Steps:**
1. Sign in and enroll to the study
2. Now, Go to SB and update the consent for enrolled participants
3. Open the mobile app and complete the eligibility section
4. Navigate to comprehension section
5. Fail the comprehension section and Observe
**AR:** Comprehension test is not getting failed in the updated consent flow
**ER:** Comprehension test should be failed in the updated consent flow
https://user-images.githubusercontent.com/86007179/188439152-cb26d4dd-671f-40c5-9895-50dbd016ac70.mp4
|
process
|
comprehension test is not getting failed in the updated consent flow steps sign in and enroll to the study now go to sb and update the consent for enrolled participants open the mobile app and complete the eligibility section navigate to comprehension section fail the comprehension section and observe ar comprehension test is not getting failed in the updated consent flow er comprehension test should be failed in the updated consent flow
| 1
|
20,586
| 27,245,929,762
|
IssuesEvent
|
2023-02-22 02:00:09
|
lizhihao6/get-daily-arxiv-noti
|
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
|
opened
|
New submissions for Wed, 22 Feb 23
|
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
|
## Keyword: events
There is no result
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### ViGU: Vision GNN U-Net for Fast MRI
- **Authors:** Jiahao Huang, Angelica Aviles-Rivero, Carola-Bibiane Schonlieb, Guang Yang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10273
- **Pdf link:** https://arxiv.org/pdf/2302.10273
- **Abstract**
Deep learning models have been widely applied for fast MRI. The majority of existing deep learning models, e.g., convolutional neural networks, work on data with Euclidean or regular grids structures. However, high-dimensional features extracted from MR data could be encapsulated in non-Euclidean manifolds. This disparity between the go-to assumption of existing models and data requirements limits the flexibility to capture irregular anatomical features in MR data. In this work, we introduce a novel Vision GNN type network for fast MRI called Vision GNN U-Net (ViGU). More precisely, the pixel array is first embedded into patches and then converted into a graph. Secondly, a U-shape network is developed using several graph blocks in symmetrical encoder and decoder paths. Moreover, we show that the proposed ViGU can also benefit from Generative Adversarial Networks yielding to its variant ViGU-GAN. We demonstrate, through numerical and visual experiments, that the proposed ViGU and GAN variant outperform existing CNN and GAN-based methods. Moreover, we show that the proposed network readily competes with approaches based on Transformers while requiring a fraction of the computational cost. More importantly, the graph structure of the network reveals how the network extracts features from MR images, providing intuitive explainability.
### Semantic Feature Integration network for Fine-grained Visual Classification
- **Authors:** Hui Wang, Yueyang li, Haichi Luo
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10275
- **Pdf link:** https://arxiv.org/pdf/2302.10275
- **Abstract**
Fine-Grained Visual Classification (FGVC) is known as a challenging task due to subtle differences among subordinate categories. Many current FGVC approaches focus on identifying and locating discriminative regions by using the attention mechanism, but neglect the presence of unnecessary features that hinder the understanding of object structure. These unnecessary features, including 1) ambiguous parts resulting from the visual similarity in object appearances and 2) noninformative parts (e.g., background noise), can have a significant adverse impact on classification results. In this paper, we propose the Semantic Feature Integration network (SFI-Net) to address the above difficulties. By eliminating unnecessary features and reconstructing the semantic relations among discriminative features, our SFI-Net has achieved satisfying performance. The network consists of two modules: 1) the multi-level feature filter (MFF) module is proposed to remove unnecessary features with different receptive field, and then concatenate the preserved features on pixel level for subsequent disposal; 2) the semantic information reconstitution (SIR) module is presented to further establish semantic relations among discriminative features obtained from the MFF module. These two modules are carefully designed to be light-weighted and can be trained end-to-end in a weakly-supervised way. Extensive experiments on four challenging fine-grained benchmarks demonstrate that our proposed SFI-Net achieves the state-of-the-arts performance. Especially, the classification accuracy of our model on CUB-200-2011 and Stanford Dogs reaches 92.64% and 93.03%, respectively.
### Combining Blockchain and Biometrics: A Survey on Technical Aspects and a First Legal Analysis
- **Authors:** Mahdi Ghafourian, Bilgesu Sumer, Ruben Vera-Rodriguez, Julian Fierrez, Ruben Tolosana, Aythami Moralez, Els Kindt
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Distributed, Parallel, and Cluster Computing (cs.DC); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2302.10883
- **Pdf link:** https://arxiv.org/pdf/2302.10883
- **Abstract**
Biometric recognition as a unique, hard-to-forge, and efficient way of identification and verification has become an indispensable part of the current digital world. The fast evolution of this technology has been a strong incentive for integrating it into many applications. Meanwhile, blockchain, the very attractive decentralized ledger technology, has been widely received both by the research and industry in the past years and it is being increasingly deployed nowadays in many different applications, such as money transfer, IoT, healthcare, or logistics. Recently, researchers have started to speculate what would be the pros and cons and what would be the best applications when these two technologies cross paths. This paper provides a survey of technical literature research on the combination of blockchain and biometrics and includes a first legal analysis of this integration to shed light on challenges and potentials. While this combination is still in its infancy and a growing body of literature discusses specific blockchain applications and solutions in an advanced technological set-up, this paper presents a holistic understanding of blockchains applicability in the biometric sector. This study demonstrates that combining blockchain and biometrics would be beneficial for novel applications in biometrics such as the PKI mechanism, distributed trusted service, and identity management. However, blockchain networks at their current stage are not efficient and economical for real-time applications. From a legal point of view, the allocation of accountability remains a main issue, while other difficulties remain, such as conducting a proper Data Protection Impact Assessment. Finally, it supplies technical and legal recommendations to reap the benefits and mitigate the risks of the combination.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### On Interpretable Approaches to Cluster, Classify and Represent Multi-Subspace Data via Minimum Lossy Coding Length based on Rate-Distortion Theory
- **Authors:** Kai-Liang Lu, Avraham Chapman
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10383
- **Pdf link:** https://arxiv.org/pdf/2302.10383
- **Abstract**
To cluster, classify and represent are three fundamental objectives of learning from high-dimensional data with intrinsic structure. To this end, this paper introduces three interpretable approaches, i.e., segmentation (clustering) via the Minimum Lossy Coding Length criterion, classification via the Minimum Incremental Coding Length criterion and representation via the Maximal Coding Rate Reduction criterion. These are derived based on the lossy data coding and compression framework from the principle of rate distortion in information theory. These algorithms are particularly suitable for dealing with finite-sample data (allowed to be sparse or almost degenerate) of mixed Gaussian distributions or subspaces. The theoretical value and attractive features of these methods are summarized by comparison with other learning methods or evaluation criteria. This summary note aims to provide a theoretical guide to researchers (also engineers) interested in understanding 'white-box' machine (deep) learning methods.
### Device Tuning for Multi-Task Large Model
- **Authors:** Penghao Jiang, Xuanchen Hou, Yinsi Zhou
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10820
- **Pdf link:** https://arxiv.org/pdf/2302.10820
- **Abstract**
Unsupervised pre-training approaches have achieved great success in many fields such as Computer Vision (CV), Natural Language Processing (NLP) and so on. However, compared to typical deep learning models, pre-training or even fine-tuning the state-of-the-art self-attention models is extremely expensive, as they require much more computational and memory resources. It severely limits their applications and success in a variety of domains, especially for multi-task learning. To improve the efficiency, we propose Device Tuning for the efficient multi-task model, which is a massively multitask framework across the cloud and device and is designed to encourage learning of representations that generalize better to many different tasks. Specifically, we design Device Tuning architecture of a multi-task model that benefits both cloud modelling and device modelling, which reduces the communication between device and cloud by representation compression. Experimental results demonstrate the effectiveness of our proposed method.
## Keyword: RAW
### Automotive RADAR sub-sampling via object detection networks: Leveraging prior signal information
- **Authors:** Madhumitha Sakthi, Ahmed Tewfik, Marius Arvinte, Haris Vikalo
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Signal Processing (eess.SP)
- **Arxiv link:** https://arxiv.org/abs/2302.10450
- **Pdf link:** https://arxiv.org/pdf/2302.10450
- **Abstract**
Automotive radar has increasingly attracted attention due to growing interest in autonomous driving technologies. Acquiring situational awareness using multimodal data collected at high sampling rates by various sensing devices including cameras, LiDAR, and radar requires considerable power, memory and compute resources which are often limited at an edge device. In this paper, we present a novel adaptive radar sub-sampling algorithm designed to identify regions that require more detailed/accurate reconstruction based on prior environmental conditions' knowledge, enabling near-optimal performance at considerably lower effective sampling rates. Designed to robustly perform under variable weather conditions, the algorithm was shown on the Oxford raw radar and RADIATE dataset to achieve accurate reconstruction utilizing only 10% of the original samples in good weather and 20% in extreme (snow, fog) weather conditions. A further modification of the algorithm incorporates object motion to enable reliable identification of important regions. This includes monitoring possible future occlusions caused by objects detected in the present frame. Finally, we train a YOLO network on the RADIATE dataset to perform object detection directly on RADAR data and obtain a 6.6% AP50 improvement over the baseline Faster R-CNN network.
### Bokeh Rendering Based on Adaptive Depth Calibration Network
- **Authors:** Lu Liu, Lei Zhou, Yuhan Dong
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10808
- **Pdf link:** https://arxiv.org/pdf/2302.10808
- **Abstract**
Bokeh rendering is a popular and effective technique used in photography to create an aesthetically pleasing effect. It is widely used to blur the background and highlight the subject in the foreground, thereby drawing the viewer's attention to the main focus of the image. In traditional digital single-lens reflex cameras (DSLRs), this effect is achieved through the use of a large aperture lens. This allows the camera to capture images with shallow depth-of-field, in which only a small area of the image is in sharp focus, while the rest of the image is blurred. However, the hardware embedded in mobile phones is typically much smaller and more limited than that found in DSLRs. Consequently, mobile phones are not able to capture natural shallow depth-of-field photos, which can be a significant limitation for mobile photography. To address this challenge, in this paper, we propose a novel method for bokeh rendering using the Vision Transformer, a recent and powerful deep learning architecture. Our approach employs an adaptive depth calibration network that acts as a confidence level to compensate for errors in monocular depth estimation. This network is used to supervise the rendering process in conjunction with depth information, allowing for the generation of high-quality bokeh images at high resolutions. Our experiments demonstrate that our proposed method outperforms state-of-the-art methods, achieving about 24.7% improvements on LPIPS and obtaining higher PSNR scores.
## Keyword: raw image
There is no result
|
2.0
|
New submissions for Wed, 22 Feb 23 - ## Keyword: events
There is no result
## Keyword: event camera
There is no result
## Keyword: events camera
There is no result
## Keyword: white balance
There is no result
## Keyword: color contrast
There is no result
## Keyword: AWB
There is no result
## Keyword: ISP
### ViGU: Vision GNN U-Net for Fast MRI
- **Authors:** Jiahao Huang, Angelica Aviles-Rivero, Carola-Bibiane Schonlieb, Guang Yang
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10273
- **Pdf link:** https://arxiv.org/pdf/2302.10273
- **Abstract**
Deep learning models have been widely applied for fast MRI. The majority of existing deep learning models, e.g., convolutional neural networks, work on data with Euclidean or regular grids structures. However, high-dimensional features extracted from MR data could be encapsulated in non-Euclidean manifolds. This disparity between the go-to assumption of existing models and data requirements limits the flexibility to capture irregular anatomical features in MR data. In this work, we introduce a novel Vision GNN type network for fast MRI called Vision GNN U-Net (ViGU). More precisely, the pixel array is first embedded into patches and then converted into a graph. Secondly, a U-shape network is developed using several graph blocks in symmetrical encoder and decoder paths. Moreover, we show that the proposed ViGU can also benefit from Generative Adversarial Networks yielding to its variant ViGU-GAN. We demonstrate, through numerical and visual experiments, that the proposed ViGU and GAN variant outperform existing CNN and GAN-based methods. Moreover, we show that the proposed network readily competes with approaches based on Transformers while requiring a fraction of the computational cost. More importantly, the graph structure of the network reveals how the network extracts features from MR images, providing intuitive explainability.
### Semantic Feature Integration network for Fine-grained Visual Classification
- **Authors:** Hui Wang, Yueyang li, Haichi Luo
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10275
- **Pdf link:** https://arxiv.org/pdf/2302.10275
- **Abstract**
Fine-Grained Visual Classification (FGVC) is known as a challenging task due to subtle differences among subordinate categories. Many current FGVC approaches focus on identifying and locating discriminative regions by using the attention mechanism, but neglect the presence of unnecessary features that hinder the understanding of object structure. These unnecessary features, including 1) ambiguous parts resulting from the visual similarity in object appearances and 2) noninformative parts (e.g., background noise), can have a significant adverse impact on classification results. In this paper, we propose the Semantic Feature Integration network (SFI-Net) to address the above difficulties. By eliminating unnecessary features and reconstructing the semantic relations among discriminative features, our SFI-Net has achieved satisfying performance. The network consists of two modules: 1) the multi-level feature filter (MFF) module is proposed to remove unnecessary features with different receptive field, and then concatenate the preserved features on pixel level for subsequent disposal; 2) the semantic information reconstitution (SIR) module is presented to further establish semantic relations among discriminative features obtained from the MFF module. These two modules are carefully designed to be light-weighted and can be trained end-to-end in a weakly-supervised way. Extensive experiments on four challenging fine-grained benchmarks demonstrate that our proposed SFI-Net achieves the state-of-the-arts performance. Especially, the classification accuracy of our model on CUB-200-2011 and Stanford Dogs reaches 92.64% and 93.03%, respectively.
### Combining Blockchain and Biometrics: A Survey on Technical Aspects and a First Legal Analysis
- **Authors:** Mahdi Ghafourian, Bilgesu Sumer, Ruben Vera-Rodriguez, Julian Fierrez, Ruben Tolosana, Aythami Moralez, Els Kindt
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Cryptography and Security (cs.CR); Distributed, Parallel, and Cluster Computing (cs.DC); Machine Learning (cs.LG)
- **Arxiv link:** https://arxiv.org/abs/2302.10883
- **Pdf link:** https://arxiv.org/pdf/2302.10883
- **Abstract**
Biometric recognition as a unique, hard-to-forge, and efficient way of identification and verification has become an indispensable part of the current digital world. The fast evolution of this technology has been a strong incentive for integrating it into many applications. Meanwhile, blockchain, the very attractive decentralized ledger technology, has been widely received both by the research and industry in the past years and it is being increasingly deployed nowadays in many different applications, such as money transfer, IoT, healthcare, or logistics. Recently, researchers have started to speculate what would be the pros and cons and what would be the best applications when these two technologies cross paths. This paper provides a survey of technical literature research on the combination of blockchain and biometrics and includes a first legal analysis of this integration to shed light on challenges and potentials. While this combination is still in its infancy and a growing body of literature discusses specific blockchain applications and solutions in an advanced technological set-up, this paper presents a holistic understanding of blockchains applicability in the biometric sector. This study demonstrates that combining blockchain and biometrics would be beneficial for novel applications in biometrics such as the PKI mechanism, distributed trusted service, and identity management. However, blockchain networks at their current stage are not efficient and economical for real-time applications. From a legal point of view, the allocation of accountability remains a main issue, while other difficulties remain, such as conducting a proper Data Protection Impact Assessment. Finally, it supplies technical and legal recommendations to reap the benefits and mitigate the risks of the combination.
## Keyword: image signal processing
There is no result
## Keyword: image signal process
There is no result
## Keyword: compression
### On Interpretable Approaches to Cluster, Classify and Represent Multi-Subspace Data via Minimum Lossy Coding Length based on Rate-Distortion Theory
- **Authors:** Kai-Liang Lu, Avraham Chapman
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10383
- **Pdf link:** https://arxiv.org/pdf/2302.10383
- **Abstract**
To cluster, classify and represent are three fundamental objectives of learning from high-dimensional data with intrinsic structure. To this end, this paper introduces three interpretable approaches, i.e., segmentation (clustering) via the Minimum Lossy Coding Length criterion, classification via the Minimum Incremental Coding Length criterion and representation via the Maximal Coding Rate Reduction criterion. These are derived based on the lossy data coding and compression framework from the principle of rate distortion in information theory. These algorithms are particularly suitable for dealing with finite-sample data (allowed to be sparse or almost degenerate) of mixed Gaussian distributions or subspaces. The theoretical value and attractive features of these methods are summarized by comparison with other learning methods or evaluation criteria. This summary note aims to provide a theoretical guide to researchers (also engineers) interested in understanding 'white-box' machine (deep) learning methods.
### Device Tuning for Multi-Task Large Model
- **Authors:** Penghao Jiang, Xuanchen Hou, Yinsi Zhou
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10820
- **Pdf link:** https://arxiv.org/pdf/2302.10820
- **Abstract**
Unsupervised pre-training approaches have achieved great success in many fields such as Computer Vision (CV), Natural Language Processing (NLP) and so on. However, compared to typical deep learning models, pre-training or even fine-tuning the state-of-the-art self-attention models is extremely expensive, as they require much more computational and memory resources. It severely limits their applications and success in a variety of domains, especially for multi-task learning. To improve the efficiency, we propose Device Tuning for the efficient multi-task model, which is a massively multitask framework across the cloud and device and is designed to encourage learning of representations that generalize better to many different tasks. Specifically, we design Device Tuning architecture of a multi-task model that benefits both cloud modelling and device modelling, which reduces the communication between device and cloud by representation compression. Experimental results demonstrate the effectiveness of our proposed method.
## Keyword: RAW
### Automotive RADAR sub-sampling via object detection networks: Leveraging prior signal information
- **Authors:** Madhumitha Sakthi, Ahmed Tewfik, Marius Arvinte, Haris Vikalo
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Signal Processing (eess.SP)
- **Arxiv link:** https://arxiv.org/abs/2302.10450
- **Pdf link:** https://arxiv.org/pdf/2302.10450
- **Abstract**
Automotive radar has increasingly attracted attention due to growing interest in autonomous driving technologies. Acquiring situational awareness using multimodal data collected at high sampling rates by various sensing devices including cameras, LiDAR, and radar requires considerable power, memory and compute resources which are often limited at an edge device. In this paper, we present a novel adaptive radar sub-sampling algorithm designed to identify regions that require more detailed/accurate reconstruction based on prior environmental conditions' knowledge, enabling near-optimal performance at considerably lower effective sampling rates. Designed to robustly perform under variable weather conditions, the algorithm was shown on the Oxford raw radar and RADIATE dataset to achieve accurate reconstruction utilizing only 10% of the original samples in good weather and 20% in extreme (snow, fog) weather conditions. A further modification of the algorithm incorporates object motion to enable reliable identification of important regions. This includes monitoring possible future occlusions caused by objects detected in the present frame. Finally, we train a YOLO network on the RADIATE dataset to perform object detection directly on RADAR data and obtain a 6.6% AP50 improvement over the baseline Faster R-CNN network.
### Bokeh Rendering Based on Adaptive Depth Calibration Network
- **Authors:** Lu Liu, Lei Zhou, Yuhan Dong
- **Subjects:** Computer Vision and Pattern Recognition (cs.CV)
- **Arxiv link:** https://arxiv.org/abs/2302.10808
- **Pdf link:** https://arxiv.org/pdf/2302.10808
- **Abstract**
Bokeh rendering is a popular and effective technique used in photography to create an aesthetically pleasing effect. It is widely used to blur the background and highlight the subject in the foreground, thereby drawing the viewer's attention to the main focus of the image. In traditional digital single-lens reflex cameras (DSLRs), this effect is achieved through the use of a large aperture lens. This allows the camera to capture images with shallow depth-of-field, in which only a small area of the image is in sharp focus, while the rest of the image is blurred. However, the hardware embedded in mobile phones is typically much smaller and more limited than that found in DSLRs. Consequently, mobile phones are not able to capture natural shallow depth-of-field photos, which can be a significant limitation for mobile photography. To address this challenge, in this paper, we propose a novel method for bokeh rendering using the Vision Transformer, a recent and powerful deep learning architecture. Our approach employs an adaptive depth calibration network that acts as a confidence level to compensate for errors in monocular depth estimation. This network is used to supervise the rendering process in conjunction with depth information, allowing for the generation of high-quality bokeh images at high resolutions. Our experiments demonstrate that our proposed method outperforms state-of-the-art methods, achieving about 24.7% improvements on LPIPS and obtaining higher PSNR scores.
## Keyword: raw image
There is no result
|
process
|
new submissions for wed feb keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb there is no result keyword isp vigu vision gnn u net for fast mri authors jiahao huang angelica aviles rivero carola bibiane schonlieb guang yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deep learning models have been widely applied for fast mri the majority of existing deep learning models e g convolutional neural networks work on data with euclidean or regular grids structures however high dimensional features extracted from mr data could be encapsulated in non euclidean manifolds this disparity between the go to assumption of existing models and data requirements limits the flexibility to capture irregular anatomical features in mr data in this work we introduce a novel vision gnn type network for fast mri called vision gnn u net vigu more precisely the pixel array is first embedded into patches and then converted into a graph secondly a u shape network is developed using several graph blocks in symmetrical encoder and decoder paths moreover we show that the proposed vigu can also benefit from generative adversarial networks yielding to its variant vigu gan we demonstrate through numerical and visual experiments that the proposed vigu and gan variant outperform existing cnn and gan based methods moreover we show that the proposed network readily competes with approaches based on transformers while requiring a fraction of the computational cost more importantly the graph structure of the network reveals how the network extracts features from mr images providing intuitive explainability semantic feature integration network for fine grained visual classification authors hui wang yueyang li haichi luo subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract fine grained visual classification fgvc is known as a challenging task due to subtle differences among subordinate categories many current fgvc approaches focus on identifying and locating discriminative regions by using the attention mechanism but neglect the presence of unnecessary features that hinder the understanding of object structure these unnecessary features including ambiguous parts resulting from the visual similarity in object appearances and noninformative parts e g background noise can have a significant adverse impact on classification results in this paper we propose the semantic feature integration network sfi net to address the above difficulties by eliminating unnecessary features and reconstructing the semantic relations among discriminative features our sfi net has achieved satisfying performance the network consists of two modules the multi level feature filter mff module is proposed to remove unnecessary features with different receptive field and then concatenate the preserved features on pixel level for subsequent disposal the semantic information reconstitution sir module is presented to further establish semantic relations among discriminative features obtained from the mff module these two modules are carefully designed to be light weighted and can be trained end to end in a weakly supervised way extensive experiments on four challenging fine grained benchmarks demonstrate that our proposed sfi net achieves the state of the arts performance especially the classification accuracy of our model on cub and stanford dogs reaches and respectively combining blockchain and biometrics a survey on technical aspects and a first legal analysis authors mahdi ghafourian bilgesu sumer ruben vera rodriguez julian fierrez ruben tolosana aythami moralez els kindt subjects computer vision and pattern recognition cs cv cryptography and security cs cr distributed parallel and cluster computing cs dc machine learning cs lg arxiv link pdf link abstract biometric recognition as a unique hard to forge and efficient way of identification and verification has become an indispensable part of the current digital world the fast evolution of this technology has been a strong incentive for integrating it into many applications meanwhile blockchain the very attractive decentralized ledger technology has been widely received both by the research and industry in the past years and it is being increasingly deployed nowadays in many different applications such as money transfer iot healthcare or logistics recently researchers have started to speculate what would be the pros and cons and what would be the best applications when these two technologies cross paths this paper provides a survey of technical literature research on the combination of blockchain and biometrics and includes a first legal analysis of this integration to shed light on challenges and potentials while this combination is still in its infancy and a growing body of literature discusses specific blockchain applications and solutions in an advanced technological set up this paper presents a holistic understanding of blockchains applicability in the biometric sector this study demonstrates that combining blockchain and biometrics would be beneficial for novel applications in biometrics such as the pki mechanism distributed trusted service and identity management however blockchain networks at their current stage are not efficient and economical for real time applications from a legal point of view the allocation of accountability remains a main issue while other difficulties remain such as conducting a proper data protection impact assessment finally it supplies technical and legal recommendations to reap the benefits and mitigate the risks of the combination keyword image signal processing there is no result keyword image signal process there is no result keyword compression on interpretable approaches to cluster classify and represent multi subspace data via minimum lossy coding length based on rate distortion theory authors kai liang lu avraham chapman subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract to cluster classify and represent are three fundamental objectives of learning from high dimensional data with intrinsic structure to this end this paper introduces three interpretable approaches i e segmentation clustering via the minimum lossy coding length criterion classification via the minimum incremental coding length criterion and representation via the maximal coding rate reduction criterion these are derived based on the lossy data coding and compression framework from the principle of rate distortion in information theory these algorithms are particularly suitable for dealing with finite sample data allowed to be sparse or almost degenerate of mixed gaussian distributions or subspaces the theoretical value and attractive features of these methods are summarized by comparison with other learning methods or evaluation criteria this summary note aims to provide a theoretical guide to researchers also engineers interested in understanding white box machine deep learning methods device tuning for multi task large model authors penghao jiang xuanchen hou yinsi zhou subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract unsupervised pre training approaches have achieved great success in many fields such as computer vision cv natural language processing nlp and so on however compared to typical deep learning models pre training or even fine tuning the state of the art self attention models is extremely expensive as they require much more computational and memory resources it severely limits their applications and success in a variety of domains especially for multi task learning to improve the efficiency we propose device tuning for the efficient multi task model which is a massively multitask framework across the cloud and device and is designed to encourage learning of representations that generalize better to many different tasks specifically we design device tuning architecture of a multi task model that benefits both cloud modelling and device modelling which reduces the communication between device and cloud by representation compression experimental results demonstrate the effectiveness of our proposed method keyword raw automotive radar sub sampling via object detection networks leveraging prior signal information authors madhumitha sakthi ahmed tewfik marius arvinte haris vikalo subjects computer vision and pattern recognition cs cv signal processing eess sp arxiv link pdf link abstract automotive radar has increasingly attracted attention due to growing interest in autonomous driving technologies acquiring situational awareness using multimodal data collected at high sampling rates by various sensing devices including cameras lidar and radar requires considerable power memory and compute resources which are often limited at an edge device in this paper we present a novel adaptive radar sub sampling algorithm designed to identify regions that require more detailed accurate reconstruction based on prior environmental conditions knowledge enabling near optimal performance at considerably lower effective sampling rates designed to robustly perform under variable weather conditions the algorithm was shown on the oxford raw radar and radiate dataset to achieve accurate reconstruction utilizing only of the original samples in good weather and in extreme snow fog weather conditions a further modification of the algorithm incorporates object motion to enable reliable identification of important regions this includes monitoring possible future occlusions caused by objects detected in the present frame finally we train a yolo network on the radiate dataset to perform object detection directly on radar data and obtain a improvement over the baseline faster r cnn network bokeh rendering based on adaptive depth calibration network authors lu liu lei zhou yuhan dong subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract bokeh rendering is a popular and effective technique used in photography to create an aesthetically pleasing effect it is widely used to blur the background and highlight the subject in the foreground thereby drawing the viewer s attention to the main focus of the image in traditional digital single lens reflex cameras dslrs this effect is achieved through the use of a large aperture lens this allows the camera to capture images with shallow depth of field in which only a small area of the image is in sharp focus while the rest of the image is blurred however the hardware embedded in mobile phones is typically much smaller and more limited than that found in dslrs consequently mobile phones are not able to capture natural shallow depth of field photos which can be a significant limitation for mobile photography to address this challenge in this paper we propose a novel method for bokeh rendering using the vision transformer a recent and powerful deep learning architecture our approach employs an adaptive depth calibration network that acts as a confidence level to compensate for errors in monocular depth estimation this network is used to supervise the rendering process in conjunction with depth information allowing for the generation of high quality bokeh images at high resolutions our experiments demonstrate that our proposed method outperforms state of the art methods achieving about improvements on lpips and obtaining higher psnr scores keyword raw image there is no result
| 1
|
3,966
| 15,049,751,154
|
IssuesEvent
|
2021-02-03 11:56:04
|
mozilla-mobile/firefox-ios
|
https://api.github.com/repos/mozilla-mobile/firefox-ios
|
opened
|
[UITests] The UI test suite does compile since the latest A-S update
|
eng:automation
|
All the other test suites are working fine but this one is failing with error:
`~/Client.xcodeproj Building for iOS Simulator, but the linked and embedded framework 'KIF.framework' was built for iOS + iOS Simulator.`
Please see [logs](https://app.bitrise.io/build/6a0538bdffb4c4de#?tab=log).
❌ `error: Building for iOS Simulator, but the linked and embedded framework 'KIF.framework' was built for iOS + iOS Simulator. (in target 'UITests' from project 'Client')`
|
1.0
|
[UITests] The UI test suite does compile since the latest A-S update - All the other test suites are working fine but this one is failing with error:
`~/Client.xcodeproj Building for iOS Simulator, but the linked and embedded framework 'KIF.framework' was built for iOS + iOS Simulator.`
Please see [logs](https://app.bitrise.io/build/6a0538bdffb4c4de#?tab=log).
❌ `error: Building for iOS Simulator, but the linked and embedded framework 'KIF.framework' was built for iOS + iOS Simulator. (in target 'UITests' from project 'Client')`
|
non_process
|
the ui test suite does compile since the latest a s update all the other test suites are working fine but this one is failing with error client xcodeproj building for ios simulator but the linked and embedded framework kif framework was built for ios ios simulator please see ❌ error building for ios simulator but the linked and embedded framework kif framework was built for ios ios simulator in target uitests from project client
| 0
|
84,764
| 10,560,580,860
|
IssuesEvent
|
2019-10-04 14:10:36
|
kabanero-io/kabanero-operator
|
https://api.github.com/repos/kabanero-io/kabanero-operator
|
closed
|
Discuss the scope of the Appsody operator deployed by the Kabanero operator
|
design
|
I'd like to continue the discussion started in #94 about the scoping of the Appsody operator deployed by the Kabanero operator. Currently for each `Kabanero` object instance, an Appsody operator is deployed in that namespace, which only watches that namespace. There can only be one `Kabanero` instance per namespace.
@arthurdm @navidsh @kyleschlosser
Going forward, there is a desire to have the Appsody operator watch multiple namespaces. I've had conversations with several different people with different opinions so this is really going to be a brain dump which can hopefully evolve into a discussion about how this should look going forward.
1) There is a desire to only deploy the Appsody operator if there are collections in the Kabanero instance that require it. So for example if there were a Kabanero instance which only contained collections that promote code from dev to test, there would be no need for Appsody in that instance, and therefore the Appsody operator would not be installed there. I think this was with the understanding that Appsody only watches a single namespace, which is no longer a restriction.
2) There is a desire to have the Appsody operator watch more than one namespace. My understanding of the next rev of Kabanero is that there will be a Kabanero Instance which contains a new concept called a Workspace, where a workspace is a pair (or more) of Namespaces. Under this arrangement the Appsody operator would need to watch all of the namespaces contained in all of the Workspaces. It seems at this point that Appsody should be a cluster scoped operator. At that point it becomes difficult for the Kabanero operator to manage, since it would need to keep track of whether it installed an Appsody operator to this cluster or not (any Kabanero instance on the cluster might have done it) and coordinate when it's safe to remove the Appsody operator. It would become a pre-req dependency like Knative and Tekton Pipelines, that is pre-installed on the cluster and is not managed by the Kabanero operator.
3) It was mentioned in #94 that it might be nice to give the user the choice as to what to watch. I just worry about the operator managing that in an intelligent way, such that we don't deploy two operators that are both watching the same namespace and both trying to react to what is happening there. Maybe there is no concern here but it seems like a bad idea to me.
|
1.0
|
Discuss the scope of the Appsody operator deployed by the Kabanero operator - I'd like to continue the discussion started in #94 about the scoping of the Appsody operator deployed by the Kabanero operator. Currently for each `Kabanero` object instance, an Appsody operator is deployed in that namespace, which only watches that namespace. There can only be one `Kabanero` instance per namespace.
@arthurdm @navidsh @kyleschlosser
Going forward, there is a desire to have the Appsody operator watch multiple namespaces. I've had conversations with several different people with different opinions so this is really going to be a brain dump which can hopefully evolve into a discussion about how this should look going forward.
1) There is a desire to only deploy the Appsody operator if there are collections in the Kabanero instance that require it. So for example if there were a Kabanero instance which only contained collections that promote code from dev to test, there would be no need for Appsody in that instance, and therefore the Appsody operator would not be installed there. I think this was with the understanding that Appsody only watches a single namespace, which is no longer a restriction.
2) There is a desire to have the Appsody operator watch more than one namespace. My understanding of the next rev of Kabanero is that there will be a Kabanero Instance which contains a new concept called a Workspace, where a workspace is a pair (or more) of Namespaces. Under this arrangement the Appsody operator would need to watch all of the namespaces contained in all of the Workspaces. It seems at this point that Appsody should be a cluster scoped operator. At that point it becomes difficult for the Kabanero operator to manage, since it would need to keep track of whether it installed an Appsody operator to this cluster or not (any Kabanero instance on the cluster might have done it) and coordinate when it's safe to remove the Appsody operator. It would become a pre-req dependency like Knative and Tekton Pipelines, that is pre-installed on the cluster and is not managed by the Kabanero operator.
3) It was mentioned in #94 that it might be nice to give the user the choice as to what to watch. I just worry about the operator managing that in an intelligent way, such that we don't deploy two operators that are both watching the same namespace and both trying to react to what is happening there. Maybe there is no concern here but it seems like a bad idea to me.
|
non_process
|
discuss the scope of the appsody operator deployed by the kabanero operator i d like to continue the discussion started in about the scoping of the appsody operator deployed by the kabanero operator currently for each kabanero object instance an appsody operator is deployed in that namespace which only watches that namespace there can only be one kabanero instance per namespace arthurdm navidsh kyleschlosser going forward there is a desire to have the appsody operator watch multiple namespaces i ve had conversations with several different people with different opinions so this is really going to be a brain dump which can hopefully evolve into a discussion about how this should look going forward there is a desire to only deploy the appsody operator if there are collections in the kabanero instance that require it so for example if there were a kabanero instance which only contained collections that promote code from dev to test there would be no need for appsody in that instance and therefore the appsody operator would not be installed there i think this was with the understanding that appsody only watches a single namespace which is no longer a restriction there is a desire to have the appsody operator watch more than one namespace my understanding of the next rev of kabanero is that there will be a kabanero instance which contains a new concept called a workspace where a workspace is a pair or more of namespaces under this arrangement the appsody operator would need to watch all of the namespaces contained in all of the workspaces it seems at this point that appsody should be a cluster scoped operator at that point it becomes difficult for the kabanero operator to manage since it would need to keep track of whether it installed an appsody operator to this cluster or not any kabanero instance on the cluster might have done it and coordinate when it s safe to remove the appsody operator it would become a pre req dependency like knative and tekton pipelines that is pre installed on the cluster and is not managed by the kabanero operator it was mentioned in that it might be nice to give the user the choice as to what to watch i just worry about the operator managing that in an intelligent way such that we don t deploy two operators that are both watching the same namespace and both trying to react to what is happening there maybe there is no concern here but it seems like a bad idea to me
| 0
|
295,443
| 9,086,524,254
|
IssuesEvent
|
2019-02-18 11:10:20
|
FOSSEE/online_test
|
https://api.github.com/repos/FOSSEE/online_test
|
closed
|
Link of download file for a question is broken.
|
Priority: Critical
|
After clicking the link it does not download the file and redirects to the same page.
|
1.0
|
Link of download file for a question is broken. - After clicking the link it does not download the file and redirects to the same page.
|
non_process
|
link of download file for a question is broken after clicking the link it does not download the file and redirects to the same page
| 0
|
415,846
| 12,135,512,866
|
IssuesEvent
|
2020-04-23 12:39:40
|
celo-org/celo-monorepo
|
https://api.github.com/repos/celo-org/celo-monorepo
|
closed
|
Minor Support Flow Improvements
|
Priority: P2 enhancement wallet
|
### Expected Behavior
1) On successful send of a report, redirect them back to the app on the settings page and flash a toast with the following success message:
"Your issue has been successfully reported! Expect an email response soon!"
2) add testnet name and address to the JSON blob in the email text
### Current Behavior
What is the current behavior?
1) we redirect back to the Contact form
2) no testnet name or address in JSON blob
|
1.0
|
Minor Support Flow Improvements - ### Expected Behavior
1) On successful send of a report, redirect them back to the app on the settings page and flash a toast with the following success message:
"Your issue has been successfully reported! Expect an email response soon!"
2) add testnet name and address to the JSON blob in the email text
### Current Behavior
What is the current behavior?
1) we redirect back to the Contact form
2) no testnet name or address in JSON blob
|
non_process
|
minor support flow improvements expected behavior on successful send of a report redirect them back to the app on the settings page and flash a toast with the following success message your issue has been successfully reported expect an email response soon add testnet name and address to the json blob in the email text current behavior what is the current behavior we redirect back to the contact form no testnet name or address in json blob
| 0
|
6,306
| 7,565,052,258
|
IssuesEvent
|
2018-04-21 04:56:48
|
Trumeet/MiPushFramework
|
https://api.github.com/repos/Trumeet/MiPushFramework
|
closed
|
[feature request] 在一些 ROM 上利用提权漏洞给 Push Service 提权
|
feature request push service
|
在一些官改、第三方 ROM 上,ROMer 会以 Google 提供的测试签名给 platform 签名。我们可以利用该签名达到 system uid 的效果。
TODO
- [x] 对 Push 增加 system flavor,在该 flavor 的 manifest 中设置 shared uid 以及在 gradle 中设置签名
- [x] 更改 Push 提供的权限 protection level,以便使非 platform 签名的 manager 得以使用
- [x] 在下载 service 的提示方面加入判断
- [x] 在 manager 主界面提醒用户可以使用 system 版本 service
- [x] 使用 platform 权限执行一些操作,比如常驻
|
1.0
|
[feature request] 在一些 ROM 上利用提权漏洞给 Push Service 提权 - 在一些官改、第三方 ROM 上,ROMer 会以 Google 提供的测试签名给 platform 签名。我们可以利用该签名达到 system uid 的效果。
TODO
- [x] 对 Push 增加 system flavor,在该 flavor 的 manifest 中设置 shared uid 以及在 gradle 中设置签名
- [x] 更改 Push 提供的权限 protection level,以便使非 platform 签名的 manager 得以使用
- [x] 在下载 service 的提示方面加入判断
- [x] 在 manager 主界面提醒用户可以使用 system 版本 service
- [x] 使用 platform 权限执行一些操作,比如常驻
|
non_process
|
在一些 rom 上利用提权漏洞给 push service 提权 在一些官改、第三方 rom 上,romer 会以 google 提供的测试签名给 platform 签名。我们可以利用该签名达到 system uid 的效果。 todo 对 push 增加 system flavor,在该 flavor 的 manifest 中设置 shared uid 以及在 gradle 中设置签名 更改 push 提供的权限 protection level,以便使非 platform 签名的 manager 得以使用 在下载 service 的提示方面加入判断 在 manager 主界面提醒用户可以使用 system 版本 service 使用 platform 权限执行一些操作,比如常驻
| 0
|
4,315
| 7,203,329,742
|
IssuesEvent
|
2018-02-06 08:49:19
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE][processing] New extract by expression algorithm
|
Automatic new feature Processing
|
Original commit: https://github.com/qgis/QGIS/commit/74e64645e410dff89685f029ea0169eb825c7689 by nyalldawson
Filters an input layer by expression
|
1.0
|
[FEATURE][processing] New extract by expression algorithm - Original commit: https://github.com/qgis/QGIS/commit/74e64645e410dff89685f029ea0169eb825c7689 by nyalldawson
Filters an input layer by expression
|
process
|
new extract by expression algorithm original commit by nyalldawson filters an input layer by expression
| 1
|
15,648
| 19,846,472,919
|
IssuesEvent
|
2022-01-21 07:10:41
|
ooi-data/CE02SHSP-SP001-06-SPKIRJ000-recovered_cspp-spkir_abj_cspp_instrument_recovered
|
https://api.github.com/repos/ooi-data/CE02SHSP-SP001-06-SPKIRJ000-recovered_cspp-spkir_abj_cspp_instrument_recovered
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:10:41.257749.
## Details
Flow name: `CE02SHSP-SP001-06-SPKIRJ000-recovered_cspp-spkir_abj_cspp_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:10:41.257749.
## Details
Flow name: `CE02SHSP-SP001-06-SPKIRJ000-recovered_cspp-spkir_abj_cspp_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered cspp spkir abj cspp instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
14,610
| 17,754,399,314
|
IssuesEvent
|
2021-08-28 13:07:39
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
Set process priority
|
Feature Process Stalled
|
It would be nice to be able to set the priority on a process before or after starting it.
Relevant info:
https://www.php.net/manual/en/function.proc-nice.php (available also on windows since php 7.2)
https://www.php.net/manual/en/function.pcntl-setpriority.php
https://www.nixtutor.com/linux/changing-priority-on-linux-processes/
https://superuser.com/questions/285840/in-windows-7-run-command-from-batch-file-with-reduced-priority
https://superuser.com/questions/620724/changing-windows-process-priority-via-command-line
|
1.0
|
Set process priority - It would be nice to be able to set the priority on a process before or after starting it.
Relevant info:
https://www.php.net/manual/en/function.proc-nice.php (available also on windows since php 7.2)
https://www.php.net/manual/en/function.pcntl-setpriority.php
https://www.nixtutor.com/linux/changing-priority-on-linux-processes/
https://superuser.com/questions/285840/in-windows-7-run-command-from-batch-file-with-reduced-priority
https://superuser.com/questions/620724/changing-windows-process-priority-via-command-line
|
process
|
set process priority it would be nice to be able to set the priority on a process before or after starting it relevant info available also on windows since php
| 1
|
201,332
| 15,802,210,798
|
IssuesEvent
|
2021-04-03 08:36:47
|
Rye98/ped
|
https://api.github.com/repos/Rye98/ped
|
opened
|
Use of square brackets in UG
|
severity.Low type.DocumentationBug
|
The use of `[ ]` seems to denote an optional parameter by the user, but this is not stated in the UG.

<!--session: 1617437739916-16e9ed70-5759-4d38-9b06-8fec6f89714b-->
|
1.0
|
Use of square brackets in UG - The use of `[ ]` seems to denote an optional parameter by the user, but this is not stated in the UG.

<!--session: 1617437739916-16e9ed70-5759-4d38-9b06-8fec6f89714b-->
|
non_process
|
use of square brackets in ug the use of seems to denote an optional parameter by the user but this is not stated in the ug
| 0
|
4,704
| 7,544,068,470
|
IssuesEvent
|
2018-04-17 17:18:02
|
UnbFeelings/unb-feelings-docs
|
https://api.github.com/repos/UnbFeelings/unb-feelings-docs
|
opened
|
[Não Conformidade] Métricas de Código
|
Desenvolvimento Processo Qualidade invalid
|
@UnbFeelings/devel
@UnbFeelings/process
Perante critérios definidos para as [Auditorias](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria) fora auditada a [Métricas de Código](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Auditoria-M%C3%A9tricas-de-C%C3%B3digo-Ciclo-1).
### Descrição
Não foi encontrada nenhuma evidência do resultado da coleta das métricas de código, das quais foram definidas anteriormente .
#### Recomendações
É recomendável que o time de processo se reúna com o time de desenvolvimento para discutir a importância da coleta de métricas e como fazê-la.
Com base na Política de Não Conformidades utilizando a matriz GUT, obteve-se uma pontuação de 64 pontos, o que se encaixa em problema mediano, assim o prazo para resolução da Não conformidade é de 3 dias
#### Detalhes
**Auditor**: Igor Gabriel
**Técnica de Audição**: Checklist
**Tipo:** Desenvolvimento e Processo
**Prazo:** 20/04/2018
|
1.0
|
[Não Conformidade] Métricas de Código - @UnbFeelings/devel
@UnbFeelings/process
Perante critérios definidos para as [Auditorias](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria) fora auditada a [Métricas de Código](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Auditoria-M%C3%A9tricas-de-C%C3%B3digo-Ciclo-1).
### Descrição
Não foi encontrada nenhuma evidência do resultado da coleta das métricas de código, das quais foram definidas anteriormente .
#### Recomendações
É recomendável que o time de processo se reúna com o time de desenvolvimento para discutir a importância da coleta de métricas e como fazê-la.
Com base na Política de Não Conformidades utilizando a matriz GUT, obteve-se uma pontuação de 64 pontos, o que se encaixa em problema mediano, assim o prazo para resolução da Não conformidade é de 3 dias
#### Detalhes
**Auditor**: Igor Gabriel
**Técnica de Audição**: Checklist
**Tipo:** Desenvolvimento e Processo
**Prazo:** 20/04/2018
|
process
|
métricas de código unbfeelings devel unbfeelings process perante critérios definidos para as fora auditada a descrição não foi encontrada nenhuma evidência do resultado da coleta das métricas de código das quais foram definidas anteriormente recomendações é recomendável que o time de processo se reúna com o time de desenvolvimento para discutir a importância da coleta de métricas e como fazê la com base na política de não conformidades utilizando a matriz gut obteve se uma pontuação de pontos o que se encaixa em problema mediano assim o prazo para resolução da não conformidade é de dias detalhes auditor igor gabriel técnica de audição checklist tipo desenvolvimento e processo prazo
| 1
|
3,528
| 6,568,988,939
|
IssuesEvent
|
2017-09-09 00:46:44
|
MikaylaFischler/dorm-leds
|
https://api.github.com/repos/MikaylaFischler/dorm-leds
|
closed
|
Implement Button Control System
|
control device process
|
Listen for button presses and set strip animations.
Should display status on LCD panel (and memory usage).
|
1.0
|
Implement Button Control System - Listen for button presses and set strip animations.
Should display status on LCD panel (and memory usage).
|
process
|
implement button control system listen for button presses and set strip animations should display status on lcd panel and memory usage
| 1
|
14,186
| 17,090,733,302
|
IssuesEvent
|
2021-07-08 17:04:56
|
MicrosoftDocs/architecture-center
|
https://api.github.com/repos/MicrosoftDocs/architecture-center
|
closed
|
blob_service_client should be replaced by blob_service_client_instance
|
Pri1 doc-enhancement machine-learning/svc team-data-science-process/subsvc triaged
|
[Enter feedback here]
blob_client_instance = **blob_service_client**.get_blob_client(CONTAINERNAME, BLOBNAME, snapshot=None)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 647ed505-3e6e-3db7-63ee-025edf6e63f4
* Version Independent ID: d910ba75-6c86-6172-64be-e45ba2bf5a3d
* Content: [Explore data in Azure Blob Storage with pandas - Team Data Science Process - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/data-science-process/explore-data-blob)
* Content Source: [docs/data-science-process/explore-data-blob.md](https://github.com/microsoftdocs/architecture-center/blob/master/docs/data-science-process/explore-data-blob.md)
* Service: **machine-learning**
* Sub-service: **team-data-science-process**
* GitHub Login: @marktab
* Microsoft Alias: **tdsp**
|
1.0
|
blob_service_client should be replaced by blob_service_client_instance -
[Enter feedback here]
blob_client_instance = **blob_service_client**.get_blob_client(CONTAINERNAME, BLOBNAME, snapshot=None)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 647ed505-3e6e-3db7-63ee-025edf6e63f4
* Version Independent ID: d910ba75-6c86-6172-64be-e45ba2bf5a3d
* Content: [Explore data in Azure Blob Storage with pandas - Team Data Science Process - Azure Architecture Center](https://docs.microsoft.com/en-us/azure/architecture/data-science-process/explore-data-blob)
* Content Source: [docs/data-science-process/explore-data-blob.md](https://github.com/microsoftdocs/architecture-center/blob/master/docs/data-science-process/explore-data-blob.md)
* Service: **machine-learning**
* Sub-service: **team-data-science-process**
* GitHub Login: @marktab
* Microsoft Alias: **tdsp**
|
process
|
blob service client should be replaced by blob service client instance blob client instance blob service client get blob client containername blobname snapshot none document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service machine learning sub service team data science process github login marktab microsoft alias tdsp
| 1
|
12,796
| 15,180,570,246
|
IssuesEvent
|
2021-02-15 00:27:32
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
Drawn & parametric masks for lens correction
|
difficulty: average feature: enhancement no-issue-activity scope: image processing scope: software support
|
**Is your feature request related to a problem? Please describe.**
Currently, it is not possible to control which area are being brighten within the lens correction module. The vignetting option approach of leads to blown out highlights which are not being blown out before applying this correction option.
**Describe the solution you'd like**
It would be very useful to have an option similar to the drawn & parametric masks to select which parts of the image should be brighten up via the vignetting options. With the option brush, circle etc. in addition to the parametric mask it would be possible to avoid the blown out highlands.
|
1.0
|
Drawn & parametric masks for lens correction - **Is your feature request related to a problem? Please describe.**
Currently, it is not possible to control which area are being brighten within the lens correction module. The vignetting option approach of leads to blown out highlights which are not being blown out before applying this correction option.
**Describe the solution you'd like**
It would be very useful to have an option similar to the drawn & parametric masks to select which parts of the image should be brighten up via the vignetting options. With the option brush, circle etc. in addition to the parametric mask it would be possible to avoid the blown out highlands.
|
process
|
drawn parametric masks for lens correction is your feature request related to a problem please describe currently it is not possible to control which area are being brighten within the lens correction module the vignetting option approach of leads to blown out highlights which are not being blown out before applying this correction option describe the solution you d like it would be very useful to have an option similar to the drawn parametric masks to select which parts of the image should be brighten up via the vignetting options with the option brush circle etc in addition to the parametric mask it would be possible to avoid the blown out highlands
| 1
|
10,582
| 3,409,384,306
|
IssuesEvent
|
2015-12-04 15:33:14
|
ufal/lindat-dspace
|
https://api.github.com/repos/ufal/lindat-dspace
|
opened
|
Bitstream only embargoes
|
documentation maybe-bug
|
Show same error message/picture as with "item level" embargoes.
Check the functionality. How do I delete it? After setting it to no value I was not able to put there a date again, on save it looked ok, but when comming back it was empty again...
|
1.0
|
Bitstream only embargoes - Show same error message/picture as with "item level" embargoes.
Check the functionality. How do I delete it? After setting it to no value I was not able to put there a date again, on save it looked ok, but when comming back it was empty again...
|
non_process
|
bitstream only embargoes show same error message picture as with item level embargoes check the functionality how do i delete it after setting it to no value i was not able to put there a date again on save it looked ok but when comming back it was empty again
| 0
|
75,187
| 9,828,075,436
|
IssuesEvent
|
2019-06-15 08:01:18
|
input-output-hk/haskell.nix
|
https://api.github.com/repos/input-output-hk/haskell.nix
|
closed
|
Overview of the system for nix developers
|
documentation
|
It would be good to have some overview of how the pieces fit together. This would be for nix developers who would like to contribute to the maintenance and improvement of this system.
For example:
- the difference between `config.packages` vs `config.hsPkgs`.
- more description of what each module does in new-package-set.
- what are hackage.nix and stackage.nix?
- what are nix-tools?
- ...
|
1.0
|
Overview of the system for nix developers - It would be good to have some overview of how the pieces fit together. This would be for nix developers who would like to contribute to the maintenance and improvement of this system.
For example:
- the difference between `config.packages` vs `config.hsPkgs`.
- more description of what each module does in new-package-set.
- what are hackage.nix and stackage.nix?
- what are nix-tools?
- ...
|
non_process
|
overview of the system for nix developers it would be good to have some overview of how the pieces fit together this would be for nix developers who would like to contribute to the maintenance and improvement of this system for example the difference between config packages vs config hspkgs more description of what each module does in new package set what are hackage nix and stackage nix what are nix tools
| 0
|
14,307
| 17,294,877,866
|
IssuesEvent
|
2021-07-25 14:18:47
|
sct-pipeline/ukbiobank-spinalcord-csa
|
https://api.github.com/repos/sct-pipeline/ukbiobank-spinalcord-csa
|
closed
|
Compute distance between PMJ and C2-C3 disc
|
process_data
|
## Description
Compute distance from PMJ along extrapolated centerline from C2-C3 disc in processing pipeline.
|
1.0
|
Compute distance between PMJ and C2-C3 disc - ## Description
Compute distance from PMJ along extrapolated centerline from C2-C3 disc in processing pipeline.
|
process
|
compute distance between pmj and disc description compute distance from pmj along extrapolated centerline from disc in processing pipeline
| 1
|
33,481
| 27,497,820,469
|
IssuesEvent
|
2023-03-05 10:55:28
|
lablup/backend.ai
|
https://api.github.com/repos/lablup/backend.ai
|
opened
|
Fix conda-pack workflow and adjust the workflow step dependencies
|
bug infrastructure enhance
|
https://github.com/lablup/backend.ai/actions/runs/4334482265/jobs/7568438532#step:8:22
Currently, the default workflow is like:
<img width="682" alt="image" src="https://user-images.githubusercontent.com/555156/222956183-67c0b1a1-967c-44ec-908c-57f92016e350.png">
Let's change it like:
<img width="685" alt="image" src="https://user-images.githubusercontent.com/555156/222956200-c15ac656-d056-443f-8988-9fef63f1dd0f.png">
so that any failures in the conda-pack step does not interrupt the original Linux wheel build and deployment.
|
1.0
|
Fix conda-pack workflow and adjust the workflow step dependencies - https://github.com/lablup/backend.ai/actions/runs/4334482265/jobs/7568438532#step:8:22
Currently, the default workflow is like:
<img width="682" alt="image" src="https://user-images.githubusercontent.com/555156/222956183-67c0b1a1-967c-44ec-908c-57f92016e350.png">
Let's change it like:
<img width="685" alt="image" src="https://user-images.githubusercontent.com/555156/222956200-c15ac656-d056-443f-8988-9fef63f1dd0f.png">
so that any failures in the conda-pack step does not interrupt the original Linux wheel build and deployment.
|
non_process
|
fix conda pack workflow and adjust the workflow step dependencies currently the default workflow is like img width alt image src let s change it like img width alt image src so that any failures in the conda pack step does not interrupt the original linux wheel build and deployment
| 0
|
267,935
| 28,565,185,044
|
IssuesEvent
|
2023-04-21 01:01:40
|
amaybaum-reachability2/vprofile-project
|
https://api.github.com/repos/amaybaum-reachability2/vprofile-project
|
closed
|
hibernate-validator-5.2.1.Final.jar: 2 vulnerabilities (highest severity is: 7.0) reachable - autoclosed
|
Mend: dependency security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-validator-5.2.1.Final.jar</b></p></summary>
<p>Hibernate's Bean Validation (JSR-303) reference implementation.</p>
<p>Library home page: <a href="http://hibernate.org/validator">http://hibernate.org/validator</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/hibernate/hibernate-validator/5.2.1.Final/hibernate-validator-5.2.1.Final.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (hibernate-validator version) | Remediation Available | Reachability |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | --- |
| [CVE-2017-7536](https://www.mend.io/vulnerability-database/CVE-2017-7536) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.0 | hibernate-validator-5.2.1.Final.jar | Direct | 5.3.0.Alpha1 | ✅|<p align="center"><a href="#">[<img src='https://whitesource-resources.whitesourcesoftware.com/viaRed.png' width=19 height=20>](## 'The vulnerability is likely to be reachable.')</a></p> |
| [CVE-2020-10693](https://www.mend.io/vulnerability-database/CVE-2020-10693) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.3 | hibernate-validator-5.2.1.Final.jar | Direct | 6.0.0.Alpha1 | ✅|<p align="center"><a href="#">[<img src='https://whitesource-resources.whitesourcesoftware.com/viaRed.png' width=19 height=20>](## 'The vulnerability is likely to be reachable.')</a></p> |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> <img src='https://whitesource-resources.whitesourcesoftware.com/viaRed.png' width=19 height=20> CVE-2017-7536</summary>
### Vulnerable Library - <b>hibernate-validator-5.2.1.Final.jar</b></p>
<p>Hibernate's Bean Validation (JSR-303) reference implementation.</p>
<p>Library home page: <a href="http://hibernate.org/validator">http://hibernate.org/validator</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/hibernate/hibernate-validator/5.2.1.Final/hibernate-validator-5.2.1.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-validator-5.2.1.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>vp-rem</b></p>
</p>
<p></p>
### Reachability Analysis
<p>
This vulnerability is potentially used
```
com.visualpathit.account.validator.UserValidator (Application)
-> org.springframework.validation.beanvalidation.LocalValidatorFactoryBean (Extension)
-> org.springframework.validation.beanvalidation.LocalValidatorFactoryBean$HibernateValidatorDelegate (Extension)
-> org.hibernate.validator.internal.util.privilegedactions.LoadClass (Extension)
...
-> org.hibernate.validator.internal.engine.ValidatorFactoryImpl (Extension)
-> org.hibernate.validator.internal.engine.ValidatorImpl (Extension)
-> ❌ org.hibernate.validator.internal.util.privilegedactions.GetDeclaredField (Vulnerable Component)
```
</p>
<p></p>
### Vulnerability Details
<p>
In Hibernate Validator 5.2.x before 5.2.5 final, 5.3.x, and 5.4.x, it was found that when the security manager's reflective permissions, which allows it to access the private members of the class, are granted to Hibernate Validator, a potential privilege escalation can occur. By allowing the calling code to access those private members without the permission an attacker may be able to validate an invalid instance and access the private member value via ConstraintViolation#getInvalidValue().
<p>Publish Date: 2018-01-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-7536>CVE-2017-7536</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.0</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7536">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7536</a></p>
<p>Release Date: 2017-06-27</p>
<p>Fix Resolution: 5.3.0.Alpha1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> <img src='https://whitesource-resources.whitesourcesoftware.com/viaRed.png' width=19 height=20> CVE-2020-10693</summary>
### Vulnerable Library - <b>hibernate-validator-5.2.1.Final.jar</b></p>
<p>Hibernate's Bean Validation (JSR-303) reference implementation.</p>
<p>Library home page: <a href="http://hibernate.org/validator">http://hibernate.org/validator</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/hibernate/hibernate-validator/5.2.1.Final/hibernate-validator-5.2.1.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-validator-5.2.1.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>vp-rem</b></p>
</p>
<p></p>
### Reachability Analysis
<p>
This vulnerability is potentially used
```
com.visualpathit.account.validator.UserValidator (Application)
-> org.springframework.validation.beanvalidation.LocalValidatorFactoryBean (Extension)
-> org.springframework.validation.beanvalidation.LocalValidatorFactoryBean$HibernateValidatorDelegate (Extension)
-> org.hibernate.validator.messageinterpolation.ResourceBundleMessageInterpolator (Extension)
...
-> org.hibernate.validator.internal.engine.messageinterpolation.parser.TokenCollector (Extension)
-> org.hibernate.validator.internal.engine.messageinterpolation.parser.BeginState (Extension)
-> ❌ org.hibernate.validator.internal.engine.messageinterpolation.parser.MessageState (Vulnerable Component)
```
</p>
<p></p>
### Vulnerability Details
<p>
A flaw was found in Hibernate Validator version 6.1.2.Final. A bug in the message interpolation processor enables invalid EL expressions to be evaluated as if they were valid. This flaw allows attackers to bypass input sanitation (escaping, stripping) controls that developers may have put in place when handling user-controlled data in error messages.
<p>Publish Date: 2020-05-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-10693>CVE-2020-10693</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://in.relation.to/2020/05/07/hibernate-validator-615-6020-released/">https://in.relation.to/2020/05/07/hibernate-validator-615-6020-released/</a></p>
<p>Release Date: 2020-05-06</p>
<p>Fix Resolution: 6.0.0.Alpha1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
True
|
hibernate-validator-5.2.1.Final.jar: 2 vulnerabilities (highest severity is: 7.0) reachable - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-validator-5.2.1.Final.jar</b></p></summary>
<p>Hibernate's Bean Validation (JSR-303) reference implementation.</p>
<p>Library home page: <a href="http://hibernate.org/validator">http://hibernate.org/validator</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/hibernate/hibernate-validator/5.2.1.Final/hibernate-validator-5.2.1.Final.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (hibernate-validator version) | Remediation Available | Reachability |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | --- |
| [CVE-2017-7536](https://www.mend.io/vulnerability-database/CVE-2017-7536) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.0 | hibernate-validator-5.2.1.Final.jar | Direct | 5.3.0.Alpha1 | ✅|<p align="center"><a href="#">[<img src='https://whitesource-resources.whitesourcesoftware.com/viaRed.png' width=19 height=20>](## 'The vulnerability is likely to be reachable.')</a></p> |
| [CVE-2020-10693](https://www.mend.io/vulnerability-database/CVE-2020-10693) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 5.3 | hibernate-validator-5.2.1.Final.jar | Direct | 6.0.0.Alpha1 | ✅|<p align="center"><a href="#">[<img src='https://whitesource-resources.whitesourcesoftware.com/viaRed.png' width=19 height=20>](## 'The vulnerability is likely to be reachable.')</a></p> |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> <img src='https://whitesource-resources.whitesourcesoftware.com/viaRed.png' width=19 height=20> CVE-2017-7536</summary>
### Vulnerable Library - <b>hibernate-validator-5.2.1.Final.jar</b></p>
<p>Hibernate's Bean Validation (JSR-303) reference implementation.</p>
<p>Library home page: <a href="http://hibernate.org/validator">http://hibernate.org/validator</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/hibernate/hibernate-validator/5.2.1.Final/hibernate-validator-5.2.1.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-validator-5.2.1.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>vp-rem</b></p>
</p>
<p></p>
### Reachability Analysis
<p>
This vulnerability is potentially used
```
com.visualpathit.account.validator.UserValidator (Application)
-> org.springframework.validation.beanvalidation.LocalValidatorFactoryBean (Extension)
-> org.springframework.validation.beanvalidation.LocalValidatorFactoryBean$HibernateValidatorDelegate (Extension)
-> org.hibernate.validator.internal.util.privilegedactions.LoadClass (Extension)
...
-> org.hibernate.validator.internal.engine.ValidatorFactoryImpl (Extension)
-> org.hibernate.validator.internal.engine.ValidatorImpl (Extension)
-> ❌ org.hibernate.validator.internal.util.privilegedactions.GetDeclaredField (Vulnerable Component)
```
</p>
<p></p>
### Vulnerability Details
<p>
In Hibernate Validator 5.2.x before 5.2.5 final, 5.3.x, and 5.4.x, it was found that when the security manager's reflective permissions, which allows it to access the private members of the class, are granted to Hibernate Validator, a potential privilege escalation can occur. By allowing the calling code to access those private members without the permission an attacker may be able to validate an invalid instance and access the private member value via ConstraintViolation#getInvalidValue().
<p>Publish Date: 2018-01-10
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-7536>CVE-2017-7536</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.0</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7536">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7536</a></p>
<p>Release Date: 2017-06-27</p>
<p>Fix Resolution: 5.3.0.Alpha1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> <img src='https://whitesource-resources.whitesourcesoftware.com/viaRed.png' width=19 height=20> CVE-2020-10693</summary>
### Vulnerable Library - <b>hibernate-validator-5.2.1.Final.jar</b></p>
<p>Hibernate's Bean Validation (JSR-303) reference implementation.</p>
<p>Library home page: <a href="http://hibernate.org/validator">http://hibernate.org/validator</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/hibernate/hibernate-validator/5.2.1.Final/hibernate-validator-5.2.1.Final.jar</p>
<p>
Dependency Hierarchy:
- :x: **hibernate-validator-5.2.1.Final.jar** (Vulnerable Library)
<p>Found in base branch: <b>vp-rem</b></p>
</p>
<p></p>
### Reachability Analysis
<p>
This vulnerability is potentially used
```
com.visualpathit.account.validator.UserValidator (Application)
-> org.springframework.validation.beanvalidation.LocalValidatorFactoryBean (Extension)
-> org.springframework.validation.beanvalidation.LocalValidatorFactoryBean$HibernateValidatorDelegate (Extension)
-> org.hibernate.validator.messageinterpolation.ResourceBundleMessageInterpolator (Extension)
...
-> org.hibernate.validator.internal.engine.messageinterpolation.parser.TokenCollector (Extension)
-> org.hibernate.validator.internal.engine.messageinterpolation.parser.BeginState (Extension)
-> ❌ org.hibernate.validator.internal.engine.messageinterpolation.parser.MessageState (Vulnerable Component)
```
</p>
<p></p>
### Vulnerability Details
<p>
A flaw was found in Hibernate Validator version 6.1.2.Final. A bug in the message interpolation processor enables invalid EL expressions to be evaluated as if they were valid. This flaw allows attackers to bypass input sanitation (escaping, stripping) controls that developers may have put in place when handling user-controlled data in error messages.
<p>Publish Date: 2020-05-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-10693>CVE-2020-10693</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://in.relation.to/2020/05/07/hibernate-validator-615-6020-released/">https://in.relation.to/2020/05/07/hibernate-validator-615-6020-released/</a></p>
<p>Release Date: 2020-05-06</p>
<p>Fix Resolution: 6.0.0.Alpha1</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
non_process
|
hibernate validator final jar vulnerabilities highest severity is reachable autoclosed vulnerable library hibernate validator final jar hibernate s bean validation jsr reference implementation library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org hibernate hibernate validator final hibernate validator final jar vulnerabilities cve severity cvss dependency type fixed in hibernate validator version remediation available reachability high hibernate validator final jar direct the vulnerability is likely to be reachable medium hibernate validator final jar direct the vulnerability is likely to be reachable details cve vulnerable library hibernate validator final jar hibernate s bean validation jsr reference implementation library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org hibernate hibernate validator final hibernate validator final jar dependency hierarchy x hibernate validator final jar vulnerable library found in base branch vp rem reachability analysis this vulnerability is potentially used com visualpathit account validator uservalidator application org springframework validation beanvalidation localvalidatorfactorybean extension org springframework validation beanvalidation localvalidatorfactorybean hibernatevalidatordelegate extension org hibernate validator internal util privilegedactions loadclass extension org hibernate validator internal engine validatorfactoryimpl extension org hibernate validator internal engine validatorimpl extension ❌ org hibernate validator internal util privilegedactions getdeclaredfield vulnerable component vulnerability details in hibernate validator x before final x and x it was found that when the security manager s reflective permissions which allows it to access the private members of the class are granted to hibernate validator a potential privilege escalation can occur by allowing the calling code to access those private members without the permission an attacker may be able to validate an invalid instance and access the private member value via constraintviolation getinvalidvalue publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library hibernate validator final jar hibernate s bean validation jsr reference implementation library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org hibernate hibernate validator final hibernate validator final jar dependency hierarchy x hibernate validator final jar vulnerable library found in base branch vp rem reachability analysis this vulnerability is potentially used com visualpathit account validator uservalidator application org springframework validation beanvalidation localvalidatorfactorybean extension org springframework validation beanvalidation localvalidatorfactorybean hibernatevalidatordelegate extension org hibernate validator messageinterpolation resourcebundlemessageinterpolator extension org hibernate validator internal engine messageinterpolation parser tokencollector extension org hibernate validator internal engine messageinterpolation parser beginstate extension ❌ org hibernate validator internal engine messageinterpolation parser messagestate vulnerable component vulnerability details a flaw was found in hibernate validator version final a bug in the message interpolation processor enables invalid el expressions to be evaluated as if they were valid this flaw allows attackers to bypass input sanitation escaping stripping controls that developers may have put in place when handling user controlled data in error messages publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
| 0
|
368,920
| 25,814,947,835
|
IssuesEvent
|
2022-12-12 03:38:52
|
facebookresearch/detectron2
|
https://api.github.com/repos/facebookresearch/detectron2
|
opened
|
The PointSup does not upload their pretrained model.
|
documentation
|
## 📚 Documentation Issue
This issue category is for problems about existing documentation, not for asking how-to questions.
* Provide a link to an existing documentation/comment/tutorial:
https://github.com/facebookresearch/detectron2/blob/main/projects/PointSup/README.md
* How should the above documentation/comment/tutorial improve:
Upload the pretrained model.
|
1.0
|
The PointSup does not upload their pretrained model. - ## 📚 Documentation Issue
This issue category is for problems about existing documentation, not for asking how-to questions.
* Provide a link to an existing documentation/comment/tutorial:
https://github.com/facebookresearch/detectron2/blob/main/projects/PointSup/README.md
* How should the above documentation/comment/tutorial improve:
Upload the pretrained model.
|
non_process
|
the pointsup does not upload their pretrained model 📚 documentation issue this issue category is for problems about existing documentation not for asking how to questions provide a link to an existing documentation comment tutorial how should the above documentation comment tutorial improve upload the pretrained model
| 0
|
1,221
| 3,750,103,500
|
IssuesEvent
|
2016-03-11 04:08:05
|
HPENetworking/topology_docker
|
https://api.github.com/repos/HPENetworking/topology_docker
|
closed
|
Place node-specific code in separate repos
|
enhancement in process
|
The implementation of nodes for this platform engine resides in this very same repo, they should be moved to completely separate the node development from this.
|
1.0
|
Place node-specific code in separate repos - The implementation of nodes for this platform engine resides in this very same repo, they should be moved to completely separate the node development from this.
|
process
|
place node specific code in separate repos the implementation of nodes for this platform engine resides in this very same repo they should be moved to completely separate the node development from this
| 1
|
410,815
| 27,803,491,164
|
IssuesEvent
|
2023-03-17 17:40:51
|
quarto-dev/quarto-cli
|
https://api.github.com/repos/quarto-dev/quarto-cli
|
closed
|
Fix Preview for PRs on quarto-web
|
bug documentation
|
**Use case:** It's nice to get feedback on docs/blog from people outside of the dev team (e.g. #os-dev-rel, or #quarto), but it's a big ask to have people outside the team render the docs locally to review a PR.
Related issue in quarto-dev/quarto-actions#33
|
1.0
|
Fix Preview for PRs on quarto-web - **Use case:** It's nice to get feedback on docs/blog from people outside of the dev team (e.g. #os-dev-rel, or #quarto), but it's a big ask to have people outside the team render the docs locally to review a PR.
Related issue in quarto-dev/quarto-actions#33
|
non_process
|
fix preview for prs on quarto web use case it s nice to get feedback on docs blog from people outside of the dev team e g os dev rel or quarto but it s a big ask to have people outside the team render the docs locally to review a pr related issue in quarto dev quarto actions
| 0
|
76,399
| 15,496,001,305
|
IssuesEvent
|
2021-03-11 01:53:06
|
yhuangsh/50pm
|
https://api.github.com/repos/yhuangsh/50pm
|
opened
|
CVE-2021-27515 (Medium) detected in url-parse-1.4.4.tgz
|
security vulnerability
|
## CVE-2021-27515 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.4.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.4.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.4.tgz</a></p>
<p>Path to dependency file: /50pm/frontend/50pm/package.json</p>
<p>Path to vulnerable library: 50pm/frontend/50pm/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.8.tgz (Root Library)
- react-dev-utils-8.0.0.tgz
- sockjs-client-1.3.0.tgz
- :x: **url-parse-1.4.4.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path.
<p>Publish Date: 2021-02-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p>
<p>Release Date: 2021-02-22</p>
<p>Fix Resolution: 1.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-27515 (Medium) detected in url-parse-1.4.4.tgz - ## CVE-2021-27515 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.4.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.4.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.4.tgz</a></p>
<p>Path to dependency file: /50pm/frontend/50pm/package.json</p>
<p>Path to vulnerable library: 50pm/frontend/50pm/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-2.1.8.tgz (Root Library)
- react-dev-utils-8.0.0.tgz
- sockjs-client-1.3.0.tgz
- :x: **url-parse-1.4.4.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse before 1.5.0 mishandles certain uses of backslash such as http:\/ and interprets the URI as a relative path.
<p>Publish Date: 2021-02-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515>CVE-2021-27515</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-27515</a></p>
<p>Release Date: 2021-02-22</p>
<p>Fix Resolution: 1.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file frontend package json path to vulnerable library frontend node modules url parse package json dependency hierarchy react scripts tgz root library react dev utils tgz sockjs client tgz x url parse tgz vulnerable library vulnerability details url parse before mishandles certain uses of backslash such as http and interprets the uri as a relative path publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
29,132
| 13,960,727,806
|
IssuesEvent
|
2020-10-24 22:39:17
|
TurboWarp/scratch-vm
|
https://api.github.com/repos/TurboWarp/scratch-vm
|
closed
|
Track variable types
|
performance
|
### Expected Behavior
The compiler should be able to reason about the types of variables.
### Actual Behavior
The compiler is unable to reason about variable types -- everything is TYPE_UNKNOWN
|
True
|
Track variable types - ### Expected Behavior
The compiler should be able to reason about the types of variables.
### Actual Behavior
The compiler is unable to reason about variable types -- everything is TYPE_UNKNOWN
|
non_process
|
track variable types expected behavior the compiler should be able to reason about the types of variables actual behavior the compiler is unable to reason about variable types everything is type unknown
| 0
|
433,408
| 12,505,750,659
|
IssuesEvent
|
2020-06-02 11:19:12
|
gitcoinco/web
|
https://api.github.com/repos/gitcoinco/web
|
closed
|
metamask/mascara support
|
Gitcoin.co priority: backlog
|
For users who can't/don't install the metamask extension,
we can point them to [1]; if the web browser has advanced service worker support.
See [2] for reference.
[1]https://wallet.metamask.io/
[2]https://github.com/MetaMask/mascara
|
1.0
|
metamask/mascara support - For users who can't/don't install the metamask extension,
we can point them to [1]; if the web browser has advanced service worker support.
See [2] for reference.
[1]https://wallet.metamask.io/
[2]https://github.com/MetaMask/mascara
|
non_process
|
metamask mascara support for users who can t don t install the metamask extension we can point them to if the web browser has advanced service worker support see for reference
| 0
|
15,779
| 10,347,960,823
|
IssuesEvent
|
2019-09-04 18:38:08
|
microsoft/botframework-solutions
|
https://api.github.com/repos/microsoft/botframework-solutions
|
closed
|
Running local instance of VA
|
Bot Services customer-replied-to customer-reported
|
Hi,
we have a use case in which we need to run a local instance of the VA (on premise), meaning that the client application shall connect to the bot using local network (and not via channel services).
Can this be done already with the current version? I think there were discussions about the BotAdapter to allow direct connection (similar to REST service endpoint) to the bot. Is that implemented already?
Ideally we would like to wrap the bot in a docker container
Looking forward for your feedback!
|
1.0
|
Running local instance of VA - Hi,
we have a use case in which we need to run a local instance of the VA (on premise), meaning that the client application shall connect to the bot using local network (and not via channel services).
Can this be done already with the current version? I think there were discussions about the BotAdapter to allow direct connection (similar to REST service endpoint) to the bot. Is that implemented already?
Ideally we would like to wrap the bot in a docker container
Looking forward for your feedback!
|
non_process
|
running local instance of va hi we have a use case in which we need to run a local instance of the va on premise meaning that the client application shall connect to the bot using local network and not via channel services can this be done already with the current version i think there were discussions about the botadapter to allow direct connection similar to rest service endpoint to the bot is that implemented already ideally we would like to wrap the bot in a docker container looking forward for your feedback
| 0
|
726,188
| 24,991,260,813
|
IssuesEvent
|
2022-11-02 18:59:43
|
dondi/GRNsight
|
https://api.github.com/repos/dondi/GRNsight
|
closed
|
Demo indicators appear to out of sync
|
bug priority 0
|
In current v6.0.0, choosing a demo from the sidebar dropdown menu does not update the UI consistently.
Choosing from the menubar continues to work as expected.
|
1.0
|
Demo indicators appear to out of sync - In current v6.0.0, choosing a demo from the sidebar dropdown menu does not update the UI consistently.
Choosing from the menubar continues to work as expected.
|
non_process
|
demo indicators appear to out of sync in current choosing a demo from the sidebar dropdown menu does not update the ui consistently choosing from the menubar continues to work as expected
| 0
|
9,357
| 12,368,215,268
|
IssuesEvent
|
2020-05-18 13:30:55
|
kubeflow/kubeflow
|
https://api.github.com/repos/kubeflow/kubeflow
|
closed
|
Stale Bot closes triaged and lifecycle/frozen issues
|
area/engprod kind/process priority/p1
|
/kind bug
**What steps did you take and what happened:**
Stale Bot closes triaged and lifecycle/frozen issues
https://github.com/kubeflow/kubeflow/issues/4068
**What did you expect to happen:**
I wouldn't expect lifecycle/frozen issues to be closed.
Not sure who to ping for this one :/
/cc @jlewi
|
1.0
|
Stale Bot closes triaged and lifecycle/frozen issues - /kind bug
**What steps did you take and what happened:**
Stale Bot closes triaged and lifecycle/frozen issues
https://github.com/kubeflow/kubeflow/issues/4068
**What did you expect to happen:**
I wouldn't expect lifecycle/frozen issues to be closed.
Not sure who to ping for this one :/
/cc @jlewi
|
process
|
stale bot closes triaged and lifecycle frozen issues kind bug what steps did you take and what happened stale bot closes triaged and lifecycle frozen issues what did you expect to happen i wouldn t expect lifecycle frozen issues to be closed not sure who to ping for this one cc jlewi
| 1
|
24,989
| 6,618,248,762
|
IssuesEvent
|
2017-09-21 07:18:54
|
RemcoTjuna/CodeValidator
|
https://api.github.com/repos/RemcoTjuna/CodeValidator
|
opened
|
Ik wil als gebruiker een e-mail ontvangen met daarin mijn prijs
|
code email feature laravel
|
Zodra ik als gebruiker een valide code heb ingevoerd, wil ik hiervan een e-mail ontvangen met daarin ook meteen beschreven wat ik gewonnen heb.
|
1.0
|
Ik wil als gebruiker een e-mail ontvangen met daarin mijn prijs - Zodra ik als gebruiker een valide code heb ingevoerd, wil ik hiervan een e-mail ontvangen met daarin ook meteen beschreven wat ik gewonnen heb.
|
non_process
|
ik wil als gebruiker een e mail ontvangen met daarin mijn prijs zodra ik als gebruiker een valide code heb ingevoerd wil ik hiervan een e mail ontvangen met daarin ook meteen beschreven wat ik gewonnen heb
| 0
|
66,642
| 8,956,209,432
|
IssuesEvent
|
2019-01-26 15:37:53
|
sylabs/cri
|
https://api.github.com/repos/sylabs/cri
|
closed
|
Document SIF-specific behavior when there are multiple tags on same image
|
documentation
|
We need to provide nice description for issue #15.
|
1.0
|
Document SIF-specific behavior when there are multiple tags on same image - We need to provide nice description for issue #15.
|
non_process
|
document sif specific behavior when there are multiple tags on same image we need to provide nice description for issue
| 0
|
57,505
| 14,165,127,440
|
IssuesEvent
|
2020-11-12 06:37:36
|
LalithK90/phonesAndAccessories
|
https://api.github.com/repos/LalithK90/phonesAndAccessories
|
opened
|
CVE-2020-10683 (High) detected in dom4j-2.1.1.jar
|
security vulnerability
|
## CVE-2020-10683 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dom4j-2.1.1.jar</b></p></summary>
<p>flexible XML framework for Java</p>
<p>Library home page: <a href="http://dom4j.github.io/">http://dom4j.github.io/</a></p>
<p>Path to dependency file: phonesAndAccessories/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.dom4j/dom4j/2.1.1/3dce5dbb3571aa820c677fadd8349bfa8f00c199/dom4j-2.1.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-jpa-2.2.4.RELEASE.jar (Root Library)
- hibernate-core-5.4.10.Final.jar
- :x: **dom4j-2.1.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LalithK90/phonesAndAccessories/commit/de989f1378965819f8d8272bf75dbb1a92607a7c">de989f1378965819f8d8272bf75dbb1a92607a7c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
dom4j before 2.0.3 and 2.1.x before 2.1.3 allows external DTDs and External Entities by default, which might enable XXE attacks. However, there is popular external documentation from OWASP showing how to enable the safe, non-default behavior in any application that uses dom4j.
<p>Publish Date: 2020-05-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10683>CVE-2020-10683</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/dom4j/dom4j/tree/version-2.1.3,https://github.com/dom4j/dom4j/tree/version-2.0.3">https://github.com/dom4j/dom4j/tree/version-2.1.3,https://github.com/dom4j/dom4j/tree/version-2.0.3</a></p>
<p>Release Date: 2020-05-01</p>
<p>Fix Resolution: org.dom4j:dom4j:2.1.3,org.dom4j:dom4j:2.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-10683 (High) detected in dom4j-2.1.1.jar - ## CVE-2020-10683 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>dom4j-2.1.1.jar</b></p></summary>
<p>flexible XML framework for Java</p>
<p>Library home page: <a href="http://dom4j.github.io/">http://dom4j.github.io/</a></p>
<p>Path to dependency file: phonesAndAccessories/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.dom4j/dom4j/2.1.1/3dce5dbb3571aa820c677fadd8349bfa8f00c199/dom4j-2.1.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-jpa-2.2.4.RELEASE.jar (Root Library)
- hibernate-core-5.4.10.Final.jar
- :x: **dom4j-2.1.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/LalithK90/phonesAndAccessories/commit/de989f1378965819f8d8272bf75dbb1a92607a7c">de989f1378965819f8d8272bf75dbb1a92607a7c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
dom4j before 2.0.3 and 2.1.x before 2.1.3 allows external DTDs and External Entities by default, which might enable XXE attacks. However, there is popular external documentation from OWASP showing how to enable the safe, non-default behavior in any application that uses dom4j.
<p>Publish Date: 2020-05-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10683>CVE-2020-10683</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/dom4j/dom4j/tree/version-2.1.3,https://github.com/dom4j/dom4j/tree/version-2.0.3">https://github.com/dom4j/dom4j/tree/version-2.1.3,https://github.com/dom4j/dom4j/tree/version-2.0.3</a></p>
<p>Release Date: 2020-05-01</p>
<p>Fix Resolution: org.dom4j:dom4j:2.1.3,org.dom4j:dom4j:2.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in jar cve high severity vulnerability vulnerable library jar flexible xml framework for java library home page a href path to dependency file phonesandaccessories build gradle path to vulnerable library home wss scanner gradle caches modules files org jar dependency hierarchy spring boot starter data jpa release jar root library hibernate core final jar x jar vulnerable library found in head commit a href found in base branch master vulnerability details before and x before allows external dtds and external entities by default which might enable xxe attacks however there is popular external documentation from owasp showing how to enable the safe non default behavior in any application that uses publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org org step up your open source security game with whitesource
| 0
|
332,982
| 29,504,725,692
|
IssuesEvent
|
2023-06-03 06:41:13
|
gitpod-io/website
|
https://api.github.com/repos/gitpod-io/website
|
closed
|
Provide example / contextual URL on `/#getting started`
|
usability testing team: website
|
### Page URL
https://www.gitpod.io/#get-started
### User feedback
Provide a full example link for the "prefix any URL". The feedback comes from [this usability test (internal link)](https://www.notion.so/gitpod/Vinny-Ruia-f687beea5c79434aa3c3bbc66dc16f9a) where a user tried to paste without the https, suggesting a full URL example here could help.

|
1.0
|
Provide example / contextual URL on `/#getting started` - ### Page URL
https://www.gitpod.io/#get-started
### User feedback
Provide a full example link for the "prefix any URL". The feedback comes from [this usability test (internal link)](https://www.notion.so/gitpod/Vinny-Ruia-f687beea5c79434aa3c3bbc66dc16f9a) where a user tried to paste without the https, suggesting a full URL example here could help.

|
non_process
|
provide example contextual url on getting started page url user feedback provide a full example link for the prefix any url the feedback comes from where a user tried to paste without the https suggesting a full url example here could help
| 0
|
58,463
| 14,401,321,052
|
IssuesEvent
|
2020-12-03 13:34:51
|
tensorflow/tensorflow
|
https://api.github.com/repos/tensorflow/tensorflow
|
closed
|
No matching distribution found for Tensorflow for pip-20.3
|
stat:awaiting response subtype: ubuntu/linux type:build/install
|
<em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution: Arch Linux
- TensorFlow version: N/A (pip-20.3 can't install any version of Tensorflow)
- Python version: 3.9.0
- CUDA/cuDNN version: N/A (running on CPU)
- GPU model and memory: N/A (running on CPU)
**Describe the problem**
When trying to install Tensorflow from pip-20.3, pip says:
> ERROR: Could not find a version that satisfies the requirement tensorflow (from versions: none)
> ERROR: No matching distribution found for tensorflow
**Provide the exact sequence of commands / steps that you executed before running into the problem**
> pip install tensorflow
> pip3 install tensorflow
> pip3.9 install tensorflow
(All of these commands yield the same error message)
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
1.0
|
No matching distribution found for Tensorflow for pip-20.3 - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution: Arch Linux
- TensorFlow version: N/A (pip-20.3 can't install any version of Tensorflow)
- Python version: 3.9.0
- CUDA/cuDNN version: N/A (running on CPU)
- GPU model and memory: N/A (running on CPU)
**Describe the problem**
When trying to install Tensorflow from pip-20.3, pip says:
> ERROR: Could not find a version that satisfies the requirement tensorflow (from versions: none)
> ERROR: No matching distribution found for tensorflow
**Provide the exact sequence of commands / steps that you executed before running into the problem**
> pip install tensorflow
> pip3 install tensorflow
> pip3.9 install tensorflow
(All of these commands yield the same error message)
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
|
non_process
|
no matching distribution found for tensorflow for pip please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution arch linux tensorflow version n a pip can t install any version of tensorflow python version cuda cudnn version n a running on cpu gpu model and memory n a running on cpu describe the problem when trying to install tensorflow from pip pip says error could not find a version that satisfies the requirement tensorflow from versions none error no matching distribution found for tensorflow provide the exact sequence of commands steps that you executed before running into the problem pip install tensorflow install tensorflow install tensorflow all of these commands yield the same error message any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached
| 0
|
72,788
| 7,311,951,828
|
IssuesEvent
|
2018-02-28 19:24:13
|
datafolklabs/cement
|
https://api.github.com/repos/datafolklabs/cement
|
closed
|
Rewrite / Refactor All Tests on PyTest (Cement 3)
|
portland testing
|
Testing is a wreck... there is 100% coverage, but a lot of it is messy and clumsy. Would like a complete top down rewrite once Cement 3 nears stable:
* http://doc.pytest.org/
<!---
@huboard:{"milestone_order":192.06720864048003}
-->
|
1.0
|
Rewrite / Refactor All Tests on PyTest (Cement 3) - Testing is a wreck... there is 100% coverage, but a lot of it is messy and clumsy. Would like a complete top down rewrite once Cement 3 nears stable:
* http://doc.pytest.org/
<!---
@huboard:{"milestone_order":192.06720864048003}
-->
|
non_process
|
rewrite refactor all tests on pytest cement testing is a wreck there is coverage but a lot of it is messy and clumsy would like a complete top down rewrite once cement nears stable huboard milestone order
| 0
|
20,565
| 27,225,245,563
|
IssuesEvent
|
2023-02-21 09:15:05
|
oasis-tcs/csaf
|
https://api.github.com/repos/oasis-tcs/csaf
|
closed
|
Register path in .well-known
|
csaf 2.0 external oasis_tc_process
|
We should register `.well-known/csaf/` ([Secition 7.1.9](https://github.com/oasis-tcs/csaf/blob/master/csaf_2.0/prose/csaf-v2-editor-draft.md#719-requirement-9-well-known-url-for-provider-metadatajson)) and `.well-known/csaf-aggregator/` ([Section 7.1.21](https://github.com/oasis-tcs/csaf/blob/master/csaf_2.0/prose/csaf-v2-editor-draft.md#7121-requirement-21-list-of-csaf-providers)).
|
1.0
|
Register path in .well-known - We should register `.well-known/csaf/` ([Secition 7.1.9](https://github.com/oasis-tcs/csaf/blob/master/csaf_2.0/prose/csaf-v2-editor-draft.md#719-requirement-9-well-known-url-for-provider-metadatajson)) and `.well-known/csaf-aggregator/` ([Section 7.1.21](https://github.com/oasis-tcs/csaf/blob/master/csaf_2.0/prose/csaf-v2-editor-draft.md#7121-requirement-21-list-of-csaf-providers)).
|
process
|
register path in well known we should register well known csaf and well known csaf aggregator
| 1
|
18,302
| 24,415,697,775
|
IssuesEvent
|
2022-10-05 15:41:28
|
googleapis/gax-java
|
https://api.github.com/repos/googleapis/gax-java
|
closed
|
(Re-)Start publishing SNAPSHOT builds nightly
|
type: process semver: patch needs work
|
This will help us detect downstream breakage before releasing the library
|
1.0
|
(Re-)Start publishing SNAPSHOT builds nightly - This will help us detect downstream breakage before releasing the library
|
process
|
re start publishing snapshot builds nightly this will help us detect downstream breakage before releasing the library
| 1
|
60,924
| 14,935,241,276
|
IssuesEvent
|
2021-01-25 11:39:48
|
feelpp/feelpp
|
https://api.github.com/repos/feelpp/feelpp
|
closed
|
Add debian build support
|
module:build priority:High
|
Add debian scripts to Feel++ so that we can build debian/ubuntu packages more easily and systematically
|
1.0
|
Add debian build support - Add debian scripts to Feel++ so that we can build debian/ubuntu packages more easily and systematically
|
non_process
|
add debian build support add debian scripts to feel so that we can build debian ubuntu packages more easily and systematically
| 0
|
111,045
| 17,009,696,787
|
IssuesEvent
|
2021-07-02 01:08:21
|
tamirdahan/keycloak
|
https://api.github.com/repos/tamirdahan/keycloak
|
opened
|
CVE-2019-11358 (Medium) detected in multiple libraries
|
security vulnerability
|
## CVE-2019-11358 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.11.1.min.js</b>, <b>jquery-2.2.0.min.js</b>, <b>jquery-1.11.3.min.js</b>, <b>jquery-1.11.0.min.js</b>, <b>jquery-2.1.1.min.js</b>, <b>jquery-3.1.1.min.js</b>, <b>jquery-3.2.1.tgz</b></p></summary>
<p>
<details><summary><b>jquery-1.11.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/bootstrap-touchspin/demo/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-2.2.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.0/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html,keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.2.0.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.11.3.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/custom_theme/base.html,keycloak/common/resources/node_modules/bootstrap-select/docs/custom_theme/base.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/test.html,keycloak/common/resources/node_modules/bootstrap-select/test.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.3.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.11.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/jquery-match-height/test/page/test.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.0.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-2.1.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html,keycloak/themes/target/classes/theme/keycloak/common/resources/node_modules/bootstrap-slider/dependencies/js/jquery.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-3.1.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-switch/docs/documentation-2.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-switch/docs/documentation-2.html,keycloak/common/resources/node_modules/bootstrap-switch/docs/documentation-2.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.1.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-3.2.1.tgz</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://registry.npmjs.org/jquery/-/jquery-3.2.1.tgz">https://registry.npmjs.org/jquery/-/jquery-3.2.1.tgz</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/package.json</p>
<p>Path to vulnerable library: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/jquery/package.json,keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery/package.json</p>
<p>
Dependency Hierarchy:
- rcue-3.48.1.tgz (Root Library)
- :x: **jquery-3.2.1.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/c3/htdocs/samples/element.html","/themes/target/classes/theme/keycloak/common/resources/node_modules/c3/htdocs/samples/element.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/c3/htdocs/samples/element.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-touchspin/demo/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.2.0","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.3","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/custom_theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/custom_theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/test.html","/themes/target/classes/theme/keycloak/common/resources/node_modules/bootstrap-select/test.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.0","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/jquery-match-height/test/page/test.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery-match-height/test/page/test.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"3.1.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-switch/docs/documentation-2.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-switch/docs/documentation-2.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:3.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"javascript/Node.js","packageName":"jquery","packageVersion":"3.2.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/package.json"],"isTransitiveDependency":true,"dependencyTree":"rcue:3.48.1;jquery:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-11358","vulnerabilityDetails":"jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-11358 (Medium) detected in multiple libraries - ## CVE-2019-11358 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.11.1.min.js</b>, <b>jquery-2.2.0.min.js</b>, <b>jquery-1.11.3.min.js</b>, <b>jquery-1.11.0.min.js</b>, <b>jquery-2.1.1.min.js</b>, <b>jquery-3.1.1.min.js</b>, <b>jquery-3.2.1.tgz</b></p></summary>
<p>
<details><summary><b>jquery-1.11.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/bootstrap-touchspin/demo/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-2.2.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.0/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html,keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.2.0.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.11.3.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/custom_theme/base.html,keycloak/common/resources/node_modules/bootstrap-select/docs/custom_theme/base.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/test.html,keycloak/common/resources/node_modules/bootstrap-select/test.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.3.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.11.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/jquery-match-height/test/page/test.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.11.0.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-2.1.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html,keycloak/themes/target/classes/theme/keycloak/common/resources/node_modules/bootstrap-slider/dependencies/js/jquery.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-3.1.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.1.1/jquery.min.js</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-switch/docs/documentation-2.html</p>
<p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-switch/docs/documentation-2.html,keycloak/common/resources/node_modules/bootstrap-switch/docs/documentation-2.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.1.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-3.2.1.tgz</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://registry.npmjs.org/jquery/-/jquery-3.2.1.tgz">https://registry.npmjs.org/jquery/-/jquery-3.2.1.tgz</a></p>
<p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/package.json</p>
<p>Path to vulnerable library: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/jquery/package.json,keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery/package.json</p>
<p>
Dependency Hierarchy:
- rcue-3.48.1.tgz (Root Library)
- :x: **jquery-3.2.1.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/c3/htdocs/samples/element.html","/themes/target/classes/theme/keycloak/common/resources/node_modules/c3/htdocs/samples/element.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/c3/htdocs/samples/element.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-touchspin/demo/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.2.0","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.3","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/custom_theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/custom_theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/test.html","/themes/target/classes/theme/keycloak/common/resources/node_modules/bootstrap-select/test.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.0","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/jquery-match-height/test/page/test.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery-match-height/test/page/test.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"3.1.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-switch/docs/documentation-2.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-switch/docs/documentation-2.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:3.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"javascript/Node.js","packageName":"jquery","packageVersion":"3.2.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/package.json"],"isTransitiveDependency":true,"dependencyTree":"rcue:3.48.1;jquery:3.2.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-11358","vulnerabilityDetails":"jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js jquery min js jquery min js jquery min js jquery tgz jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules rcue node modules patternfly node modules bootstrap touchspin demo index html path to vulnerable library keycloak common resources node modules rcue node modules patternfly node modules bootstrap touchspin demo index html keycloak common resources node modules rcue node modules patternfly node modules htdocs samples element html keycloak common resources node modules htdocs samples element html keycloak common resources node modules htdocs samples element html keycloak common resources node modules bootstrap touchspin demo index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules rcue node modules patternfly node modules bootstrap select docs docs playground index html path to vulnerable library keycloak common resources node modules rcue node modules patternfly node modules bootstrap select docs docs playground index html keycloak common resources node modules bootstrap select docs docs playground index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules bootstrap select docs docs playground test html path to vulnerable library keycloak common resources node modules bootstrap select docs docs playground test html keycloak common resources node modules rcue node modules patternfly node modules bootstrap select docs custom theme base html keycloak common resources node modules bootstrap select docs custom theme base html keycloak common resources node modules rcue node modules patternfly node modules bootstrap select test html keycloak common resources node modules bootstrap select test html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules jquery match height test page test html path to vulnerable library keycloak common resources node modules jquery match height test page lib jquery min js keycloak common resources node modules rcue node modules patternfly node modules jquery match height test page lib jquery min js dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules eonasdan bootstrap datetimepicker docs theme base html path to vulnerable library keycloak common resources node modules eonasdan bootstrap datetimepicker docs theme base html keycloak common resources node modules rcue node modules patternfly node modules eonasdan bootstrap datetimepicker docs theme base html keycloak themes target classes theme keycloak common resources node modules bootstrap slider dependencies js jquery min js dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules rcue node modules patternfly node modules bootstrap switch docs documentation html path to vulnerable library keycloak common resources node modules rcue node modules patternfly node modules bootstrap switch docs documentation html keycloak common resources node modules bootstrap switch docs documentation html dependency hierarchy x jquery min js vulnerable library jquery tgz javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources package json path to vulnerable library keycloak themes src main resources theme keycloak common resources node modules rcue node modules jquery package json keycloak themes src main resources theme keycloak common resources node modules rcue node modules patternfly node modules jquery package json dependency hierarchy rcue tgz root library x jquery tgz vulnerable library found in base branch master vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion packagetype javascript node js packagename jquery packageversion packagefilepaths istransitivedependency true dependencytree rcue jquery isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype vulnerabilityurl
| 0
|
6,886
| 10,025,932,452
|
IssuesEvent
|
2019-07-17 04:32:45
|
CATcher-org/CATcher
|
https://api.github.com/repos/CATcher-org/CATcher
|
closed
|
Multiple Sessions Capability
|
aspect-Process category.Epic p.Medium
|
Allow the application to host multiple sessions by referring to a different set of repositories for each phase. For which the following features must be implemented.
- The repositories are to be referenced through a remote file in a data repository that can be setup by an admin / org and the application simply reads the location of that particular data repo to start the available phase. #85
- A button at the header that indicates the currently opened phase (only 1 in production) that when clicked will allow us to switch between open phases (multiple in development) #100
|
1.0
|
Multiple Sessions Capability - Allow the application to host multiple sessions by referring to a different set of repositories for each phase. For which the following features must be implemented.
- The repositories are to be referenced through a remote file in a data repository that can be setup by an admin / org and the application simply reads the location of that particular data repo to start the available phase. #85
- A button at the header that indicates the currently opened phase (only 1 in production) that when clicked will allow us to switch between open phases (multiple in development) #100
|
process
|
multiple sessions capability allow the application to host multiple sessions by referring to a different set of repositories for each phase for which the following features must be implemented the repositories are to be referenced through a remote file in a data repository that can be setup by an admin org and the application simply reads the location of that particular data repo to start the available phase a button at the header that indicates the currently opened phase only in production that when clicked will allow us to switch between open phases multiple in development
| 1
|
249
| 2,671,030,843
|
IssuesEvent
|
2015-03-24 01:15:29
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
reopened
|
create a feature for calculating volume within a given cord extension
|
enhancement priority: medium sct_process_segmentation
|
possible features:
- per vertebral level
- per slice range
bit overlap with sct_extract_metric --> use code instead of duplicating (see with Simon). Maybe create a module or so.
|
1.0
|
create a feature for calculating volume within a given cord extension - possible features:
- per vertebral level
- per slice range
bit overlap with sct_extract_metric --> use code instead of duplicating (see with Simon). Maybe create a module or so.
|
process
|
create a feature for calculating volume within a given cord extension possible features per vertebral level per slice range bit overlap with sct extract metric use code instead of duplicating see with simon maybe create a module or so
| 1
|
11,141
| 13,957,692,267
|
IssuesEvent
|
2020-10-24 08:10:38
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
Harvesting Request
|
Geoportal Harvesting process SE - Sweden
|
Hi!
What is the status on weekly harvesting from the Swedish node? Is it running or is it per request?
Anyway, we would like to have a new harvesting from the Swedish node if it is posible.
Kind regards
Björn Olofsson, The Swedish Geodataportal
|
1.0
|
Harvesting Request - Hi!
What is the status on weekly harvesting from the Swedish node? Is it running or is it per request?
Anyway, we would like to have a new harvesting from the Swedish node if it is posible.
Kind regards
Björn Olofsson, The Swedish Geodataportal
|
process
|
harvesting request hi what is the status on weekly harvesting from the swedish node is it running or is it per request anyway we would like to have a new harvesting from the swedish node if it is posible kind regards bj ouml rn olofsson the swedish geodataportal
| 1
|
43,160
| 5,579,220,503
|
IssuesEvent
|
2017-03-28 14:11:45
|
AyranIsTheNewRaki/Herodot
|
https://api.github.com/repos/AyranIsTheNewRaki/Herodot
|
closed
|
high level architecture formation
|
design implementation meeting request research
|
meeting on 16 march 2017 thursday after class to decide on the high level architecture of the system, a milestone is also created for this important date. @AyranIsTheNewRaki/ayranisthenewraki
if we cannot achieve all of our agenda items there is another meeting set tentatively for 17 march 2017 after class
**please add the agenda items here:**
1. it is advisable for members to read about the research wiki before coming to the meeting
2- starting with ERDs we will try to form an high level design of the system components
3- finalize the project plan on ms project
|
1.0
|
high level architecture formation - meeting on 16 march 2017 thursday after class to decide on the high level architecture of the system, a milestone is also created for this important date. @AyranIsTheNewRaki/ayranisthenewraki
if we cannot achieve all of our agenda items there is another meeting set tentatively for 17 march 2017 after class
**please add the agenda items here:**
1. it is advisable for members to read about the research wiki before coming to the meeting
2- starting with ERDs we will try to form an high level design of the system components
3- finalize the project plan on ms project
|
non_process
|
high level architecture formation meeting on march thursday after class to decide on the high level architecture of the system a milestone is also created for this important date ayranisthenewraki ayranisthenewraki if we cannot achieve all of our agenda items there is another meeting set tentatively for march after class please add the agenda items here it is advisable for members to read about the research wiki before coming to the meeting starting with erds we will try to form an high level design of the system components finalize the project plan on ms project
| 0
|
15,082
| 18,788,867,728
|
IssuesEvent
|
2021-11-08 14:54:17
|
redwoodjs/redwood
|
https://api.github.com/repos/redwoodjs/redwood
|
closed
|
[Patch] Firebase Auth setup Fails in v0.38
|
triage/processing
|
Running `yarn rw setup auth firebase` fails:
```
✖ Generating auth lib...
→ 24 | { _event, _context }
Adding auth config to web...
Adding auth config to GraphQL API...
Adding required web packages...
Adding required api packages...
Installing packages...
One more thing...
/Users/price/Repos/tdp-redwood-tutorial-test/api/src/lib/auth.js: Unexpected token, expected "," (21:1)
19 | * @see https://github.com/redwoodjs/redwood/tree/main/packages/auth for examples
20 | */
> 21 | export const getCurrentUser = async (
| ^
22 | decoded,
23 | { _token, _type },
24 | { _event, _context }
error Command failed with exit code 1.
```
Looks like an issue with the auth.js template.
@dac09 possible for you to take a look at this?
|
1.0
|
[Patch] Firebase Auth setup Fails in v0.38 - Running `yarn rw setup auth firebase` fails:
```
✖ Generating auth lib...
→ 24 | { _event, _context }
Adding auth config to web...
Adding auth config to GraphQL API...
Adding required web packages...
Adding required api packages...
Installing packages...
One more thing...
/Users/price/Repos/tdp-redwood-tutorial-test/api/src/lib/auth.js: Unexpected token, expected "," (21:1)
19 | * @see https://github.com/redwoodjs/redwood/tree/main/packages/auth for examples
20 | */
> 21 | export const getCurrentUser = async (
| ^
22 | decoded,
23 | { _token, _type },
24 | { _event, _context }
error Command failed with exit code 1.
```
Looks like an issue with the auth.js template.
@dac09 possible for you to take a look at this?
|
process
|
firebase auth setup fails in running yarn rw setup auth firebase fails ✖ generating auth lib → event context adding auth config to web adding auth config to graphql api adding required web packages adding required api packages installing packages one more thing users price repos tdp redwood tutorial test api src lib auth js unexpected token expected see for examples export const getcurrentuser async decoded token type event context error command failed with exit code looks like an issue with the auth js template possible for you to take a look at this
| 1
|
77,528
| 14,877,451,792
|
IssuesEvent
|
2021-01-20 03:11:27
|
OpenEugene/openboard
|
https://api.github.com/repos/OpenEugene/openboard
|
closed
|
Pathways exist for collaborators
|
note: no code org: define
|
Blocked by #89
Based on this conversation: https://eugenetech.slack.com/archives/CKWAJL9MM/p1596660601237400
Once more proven procedures and relevant docs are in place, ensure pathways exist for new collaborators.
AC
TBD
|
1.0
|
Pathways exist for collaborators - Blocked by #89
Based on this conversation: https://eugenetech.slack.com/archives/CKWAJL9MM/p1596660601237400
Once more proven procedures and relevant docs are in place, ensure pathways exist for new collaborators.
AC
TBD
|
non_process
|
pathways exist for collaborators blocked by based on this conversation once more proven procedures and relevant docs are in place ensure pathways exist for new collaborators ac tbd
| 0
|
9,328
| 12,339,860,085
|
IssuesEvent
|
2020-05-14 18:52:30
|
GetTerminus/terminus-ui
|
https://api.github.com/repos/GetTerminus/terminus-ui
|
closed
|
Build: Demo application not updating when library styles changed
|
Goal: Process Improvement Needs: exploration Type: chore
|
Current the Angular CLI watcher for the demo app does not reflect any updates on animation, scss and sometimes component files, which reduces development work efficiency. We should see whether this can be fixed.
|
1.0
|
Build: Demo application not updating when library styles changed - Current the Angular CLI watcher for the demo app does not reflect any updates on animation, scss and sometimes component files, which reduces development work efficiency. We should see whether this can be fixed.
|
process
|
build demo application not updating when library styles changed current the angular cli watcher for the demo app does not reflect any updates on animation scss and sometimes component files which reduces development work efficiency we should see whether this can be fixed
| 1
|
13,689
| 16,447,386,698
|
IssuesEvent
|
2021-05-20 21:25:12
|
GoogleChrome/chromium-dashboard
|
https://api.github.com/repos/GoogleChrome/chromium-dashboard
|
opened
|
Clearly highlight required artifacts/outputs/gates through the process.
|
UX process guidance
|
Clearly highlight required artifacts/outputs/gates through the process. That way, even if one doesn’t read through all the text, they can use this as a checklist to ensure they are following the necessary steps.
|
1.0
|
Clearly highlight required artifacts/outputs/gates through the process. - Clearly highlight required artifacts/outputs/gates through the process. That way, even if one doesn’t read through all the text, they can use this as a checklist to ensure they are following the necessary steps.
|
process
|
clearly highlight required artifacts outputs gates through the process clearly highlight required artifacts outputs gates through the process that way even if one doesn’t read through all the text they can use this as a checklist to ensure they are following the necessary steps
| 1
|
120,300
| 17,644,083,018
|
IssuesEvent
|
2021-08-20 01:38:51
|
DavidSpek/pipelines
|
https://api.github.com/repos/DavidSpek/pipelines
|
opened
|
CVE-2021-37666 (High) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl
|
security vulnerability
|
## CVE-2021-37666 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt</p>
<p>Path to vulnerable library: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt,pipelines/samples/core/ai_platform/training</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. In affected versions an attacker can cause undefined behavior via binding a reference to null pointer in `tf.raw_ops.RaggedTensorToVariant`. The [implementation](https://github.com/tensorflow/tensorflow/blob/460e000de3a83278fb00b61a16d161b1964f15f4/tensorflow/core/kernels/ragged_tensor_to_variant_op.cc#L129) has an incomplete validation of the splits values, missing the case when the argument would be empty. We have patched the issue in GitHub commit be7a4de6adfbd303ce08be4332554dff70362612. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37666>CVE-2021-37666</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-w4xf-2pqw-5mq7">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-w4xf-2pqw-5mq7</a></p>
<p>Release Date: 2021-08-12</p>
<p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-37666 (High) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl - ## CVE-2021-37666 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt</p>
<p>Path to vulnerable library: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt,pipelines/samples/core/ai_platform/training</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an end-to-end open source platform for machine learning. In affected versions an attacker can cause undefined behavior via binding a reference to null pointer in `tf.raw_ops.RaggedTensorToVariant`. The [implementation](https://github.com/tensorflow/tensorflow/blob/460e000de3a83278fb00b61a16d161b1964f15f4/tensorflow/core/kernels/ragged_tensor_to_variant_op.cc#L129) has an incomplete validation of the splits values, missing the case when the argument would be empty. We have patched the issue in GitHub commit be7a4de6adfbd303ce08be4332554dff70362612. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37666>CVE-2021-37666</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-w4xf-2pqw-5mq7">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-w4xf-2pqw-5mq7</a></p>
<p>Release Date: 2021-08-12</p>
<p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file pipelines contrib components openvino ovms deployer containers requirements txt path to vulnerable library pipelines contrib components openvino ovms deployer containers requirements txt pipelines samples core ai platform training dependency hierarchy x tensorflow whl vulnerable library found in base branch master vulnerability details tensorflow is an end to end open source platform for machine learning in affected versions an attacker can cause undefined behavior via binding a reference to null pointer in tf raw ops raggedtensortovariant the has an incomplete validation of the splits values missing the case when the argument would be empty we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
| 0
|
4,156
| 7,103,715,535
|
IssuesEvent
|
2018-01-16 06:53:35
|
log2timeline/plaso
|
https://api.github.com/repos/log2timeline/plaso
|
closed
|
Deprecate usage of knowledge_base.platform in favor of per source path specification system configuration
|
needs closer look preprocessing
|
Deprecate usage of knowledge_base.platform in favor of per source path specification system configuration
* [x] ~~Deprecate usage of knowledge_base.platform in favor of operating_system value~~
* ~~https://codereview.appspot.com/331540043/~~
|
1.0
|
Deprecate usage of knowledge_base.platform in favor of per source path specification system configuration - Deprecate usage of knowledge_base.platform in favor of per source path specification system configuration
* [x] ~~Deprecate usage of knowledge_base.platform in favor of operating_system value~~
* ~~https://codereview.appspot.com/331540043/~~
|
process
|
deprecate usage of knowledge base platform in favor of per source path specification system configuration deprecate usage of knowledge base platform in favor of per source path specification system configuration deprecate usage of knowledge base platform in favor of operating system value
| 1
|
92,137
| 3,867,253,118
|
IssuesEvent
|
2016-04-09 04:55:43
|
SpongePowered/Ore
|
https://api.github.com/repos/SpongePowered/Ore
|
closed
|
Establish color template across Ore
|
frontend low priority
|
Maybe it is just me but even if yellow is our brand coloring, I have issues with reading the words that are yellow...they bleed together and I tend to have to re-read some names to make sure I'm reading this right.
This would be fixed if Ore had a black theme (YES YES YES YES YES YES) and the yellow on the black would be so so sexy but if we want to keep the light theme then we should consider different shades of yellow.
@Tzk , maybe you could help too?
|
1.0
|
Establish color template across Ore - Maybe it is just me but even if yellow is our brand coloring, I have issues with reading the words that are yellow...they bleed together and I tend to have to re-read some names to make sure I'm reading this right.
This would be fixed if Ore had a black theme (YES YES YES YES YES YES) and the yellow on the black would be so so sexy but if we want to keep the light theme then we should consider different shades of yellow.
@Tzk , maybe you could help too?
|
non_process
|
establish color template across ore maybe it is just me but even if yellow is our brand coloring i have issues with reading the words that are yellow they bleed together and i tend to have to re read some names to make sure i m reading this right this would be fixed if ore had a black theme yes yes yes yes yes yes and the yellow on the black would be so so sexy but if we want to keep the light theme then we should consider different shades of yellow tzk maybe you could help too
| 0
|
7,492
| 10,580,363,017
|
IssuesEvent
|
2019-10-08 06:29:53
|
didi/mpx
|
https://api.github.com/repos/didi/mpx
|
closed
|
mpx 不支持 weui-miniprogram组件库 跨平台编译
|
processing
|
实践:微信小程序代码跨平台编译为字节跳动小程序,源项目中用到了微信官方的 weui-miniprogram组件库,编译成字节跳动小程序,编译时没有报错,在字节跳动开发者工具中打开,报
Error:
in file ./components/mpx26b204e2/src/components/weui/cells/cells
parent.append is not a function
weui-miniprogram,文档在这:
https://developers.weixin.qq.com/miniprogram/dev/extended/weui/quickstart.html
weui-miniprogram,github在这:
https://github.com/wechat-miniprogram/weui-miniprogram
|
1.0
|
mpx 不支持 weui-miniprogram组件库 跨平台编译 - 实践:微信小程序代码跨平台编译为字节跳动小程序,源项目中用到了微信官方的 weui-miniprogram组件库,编译成字节跳动小程序,编译时没有报错,在字节跳动开发者工具中打开,报
Error:
in file ./components/mpx26b204e2/src/components/weui/cells/cells
parent.append is not a function
weui-miniprogram,文档在这:
https://developers.weixin.qq.com/miniprogram/dev/extended/weui/quickstart.html
weui-miniprogram,github在这:
https://github.com/wechat-miniprogram/weui-miniprogram
|
process
|
mpx 不支持 weui miniprogram组件库 跨平台编译 实践:微信小程序代码跨平台编译为字节跳动小程序,源项目中用到了微信官方的 weui miniprogram组件库,编译成字节跳动小程序,编译时没有报错,在字节跳动开发者工具中打开,报 error in file components src components weui cells cells parent append is not a function weui miniprogram,文档在这: weui miniprogram,github在这:
| 1
|
32,157
| 15,244,276,164
|
IssuesEvent
|
2021-02-19 12:32:12
|
getsentry/sentry-java
|
https://api.github.com/repos/getsentry/sentry-java
|
closed
|
SentryTransaction.getStatus throws NPE
|
bug performance
|
SentryTransaction.getStatus throws NPE if the transaction is not finished yet.
https://github.com/getsentry/sentry-java/blob/b016da47681036b854fc62d3d141751d8c249a2a/sentry/src/main/java/io/sentry/SentryTransaction.java#L238-L242
it should probably be `context.getStatus()` instead as `this.getContexts().setTrace(this.context);` only happens on `finish`
|
True
|
SentryTransaction.getStatus throws NPE - SentryTransaction.getStatus throws NPE if the transaction is not finished yet.
https://github.com/getsentry/sentry-java/blob/b016da47681036b854fc62d3d141751d8c249a2a/sentry/src/main/java/io/sentry/SentryTransaction.java#L238-L242
it should probably be `context.getStatus()` instead as `this.getContexts().setTrace(this.context);` only happens on `finish`
|
non_process
|
sentrytransaction getstatus throws npe sentrytransaction getstatus throws npe if the transaction is not finished yet it should probably be context getstatus instead as this getcontexts settrace this context only happens on finish
| 0
|
10,460
| 13,238,224,856
|
IssuesEvent
|
2020-08-18 23:42:07
|
googleapis/repo-automation-bots
|
https://api.github.com/repos/googleapis/repo-automation-bots
|
closed
|
Investigate limiting bot deployments for only bots that have changed
|
type: process
|
Ideally leverage some existing tool (e.g. bazel, terraform) that knows how to resolve changes rather than rolling our own.
|
1.0
|
Investigate limiting bot deployments for only bots that have changed - Ideally leverage some existing tool (e.g. bazel, terraform) that knows how to resolve changes rather than rolling our own.
|
process
|
investigate limiting bot deployments for only bots that have changed ideally leverage some existing tool e g bazel terraform that knows how to resolve changes rather than rolling our own
| 1
|
60,375
| 17,023,409,221
|
IssuesEvent
|
2021-07-03 01:52:52
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
Problem with upload on edit with save Potlatch v1.0
|
Component: potlatch (flash editor) Priority: critical Resolution: fixed Type: defect
|
**[Submitted to the original trac issue database at 10.41am, Tuesday, 26th May 2009]**
I've just lost more than an hour's work due to this bug :'( There are two different behaviours with the same effect seen this morning:
1) When you try to save, the upload box pops up but it never finished. Eventually you see the red triangle in the bottom left. After cancelling the upload (cos nothing's happened in a very long time!) Clicking on the red triangle, and clicking retry, then clicking save again, it flashed up "deleting POIs" (I didn't delete any POIs!) and then says that it's uploaded ok. Problem is, it hasn't uploaded anything. Clearing the cache and reloading the edit page shows this. DOH!
2) When you try to save, again, the upload box appears but never finishes. Eventually you cancel and try again, and it says there's nothing to upload. Again, nothing has been uploaded.
I understand these problems are probably related to heavy server load, but we need a better way of dealing with this.
|
1.0
|
Problem with upload on edit with save Potlatch v1.0 - **[Submitted to the original trac issue database at 10.41am, Tuesday, 26th May 2009]**
I've just lost more than an hour's work due to this bug :'( There are two different behaviours with the same effect seen this morning:
1) When you try to save, the upload box pops up but it never finished. Eventually you see the red triangle in the bottom left. After cancelling the upload (cos nothing's happened in a very long time!) Clicking on the red triangle, and clicking retry, then clicking save again, it flashed up "deleting POIs" (I didn't delete any POIs!) and then says that it's uploaded ok. Problem is, it hasn't uploaded anything. Clearing the cache and reloading the edit page shows this. DOH!
2) When you try to save, again, the upload box appears but never finishes. Eventually you cancel and try again, and it says there's nothing to upload. Again, nothing has been uploaded.
I understand these problems are probably related to heavy server load, but we need a better way of dealing with this.
|
non_process
|
problem with upload on edit with save potlatch i ve just lost more than an hour s work due to this bug there are two different behaviours with the same effect seen this morning when you try to save the upload box pops up but it never finished eventually you see the red triangle in the bottom left after cancelling the upload cos nothing s happened in a very long time clicking on the red triangle and clicking retry then clicking save again it flashed up deleting pois i didn t delete any pois and then says that it s uploaded ok problem is it hasn t uploaded anything clearing the cache and reloading the edit page shows this doh when you try to save again the upload box appears but never finishes eventually you cancel and try again and it says there s nothing to upload again nothing has been uploaded i understand these problems are probably related to heavy server load but we need a better way of dealing with this
| 0
|
113,941
| 4,581,820,495
|
IssuesEvent
|
2016-09-19 07:50:15
|
tom139/BeaconStrips-CLIPS
|
https://api.github.com/repos/tom139/BeaconStrips-CLIPS
|
closed
|
Add gestione errori login
|
high priority viewcontroller
|
Aggiungere la gestione degli errori per il listener della chiamata di login
|
1.0
|
Add gestione errori login - Aggiungere la gestione degli errori per il listener della chiamata di login
|
non_process
|
add gestione errori login aggiungere la gestione degli errori per il listener della chiamata di login
| 0
|
7,341
| 10,476,200,400
|
IssuesEvent
|
2019-09-23 18:03:06
|
stekylsha/CISC210Lab
|
https://api.github.com/repos/stekylsha/CISC210Lab
|
reopened
|
Waterfall
|
Software Process
|
### Story
As a future software developer
I want to understand the software process that was
So I will understand what process evolved from and why.
### Acceptance Criteria
Demonstrate:
- [x] Understanding the stages of the Waterfall process (SDLC)
- [x] _System Definition (not covered as it is business and not engineering)_
- [x] Responsible parties for each stage
- [x] Artifacts for each stage
|
1.0
|
Waterfall - ### Story
As a future software developer
I want to understand the software process that was
So I will understand what process evolved from and why.
### Acceptance Criteria
Demonstrate:
- [x] Understanding the stages of the Waterfall process (SDLC)
- [x] _System Definition (not covered as it is business and not engineering)_
- [x] Responsible parties for each stage
- [x] Artifacts for each stage
|
process
|
waterfall story as a future software developer i want to understand the software process that was so i will understand what process evolved from and why acceptance criteria demonstrate understanding the stages of the waterfall process sdlc system definition not covered as it is business and not engineering responsible parties for each stage artifacts for each stage
| 1
|
214,369
| 24,069,540,701
|
IssuesEvent
|
2022-09-18 01:02:21
|
Gal-Doron/operator-registry
|
https://api.github.com/repos/Gal-Doron/operator-registry
|
closed
|
CVE-2018-20677 (Medium) detected in github.com/bugsnag/bugsnag-go-v1.5.3 - autoclosed
|
security vulnerability
|
## CVE-2018-20677 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/bugsnag/bugsnag-go-v1.5.3</b></p></summary>
<p>Automatic panic monitoring for Go and Go web frameworks, like negroni, gin, and revel</p>
<p>
Dependency Hierarchy:
- github.com/docker/distribution-v2.8.0-beta.1 (Root Library)
- :x: **github.com/bugsnag/bugsnag-go-v1.5.3** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677>CVE-2018-20677</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
</details>
<p></p>
|
True
|
CVE-2018-20677 (Medium) detected in github.com/bugsnag/bugsnag-go-v1.5.3 - autoclosed - ## CVE-2018-20677 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/bugsnag/bugsnag-go-v1.5.3</b></p></summary>
<p>Automatic panic monitoring for Go and Go web frameworks, like negroni, gin, and revel</p>
<p>
Dependency Hierarchy:
- github.com/docker/distribution-v2.8.0-beta.1 (Root Library)
- :x: **github.com/bugsnag/bugsnag-go-v1.5.3** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20677>CVE-2018-20677</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p>
</p>
</details>
<p></p>
|
non_process
|
cve medium detected in github com bugsnag bugsnag go autoclosed cve medium severity vulnerability vulnerable library github com bugsnag bugsnag go automatic panic monitoring for go and go web frameworks like negroni gin and revel dependency hierarchy github com docker distribution beta root library x github com bugsnag bugsnag go vulnerable library found in base branch master vulnerability details in bootstrap before xss is possible in the affix configuration target property publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter
| 0
|
439,273
| 30,688,256,320
|
IssuesEvent
|
2023-07-26 13:41:43
|
FgForrest/evitaDB
|
https://api.github.com/repos/FgForrest/evitaDB
|
opened
|
Remove `ignoreTest` in documentation tests
|
documentation
|
Currently there are a few documentation tests that are disabled by the `ignoreTest` keyword because the GraphQL / REST requires special treatment - such as domain / url specification and environment preparation that is currently not happening properly.
This issue should investigate and fix all related issues so that the `ignoreTest` can be eliminated.
|
1.0
|
Remove `ignoreTest` in documentation tests - Currently there are a few documentation tests that are disabled by the `ignoreTest` keyword because the GraphQL / REST requires special treatment - such as domain / url specification and environment preparation that is currently not happening properly.
This issue should investigate and fix all related issues so that the `ignoreTest` can be eliminated.
|
non_process
|
remove ignoretest in documentation tests currently there are a few documentation tests that are disabled by the ignoretest keyword because the graphql rest requires special treatment such as domain url specification and environment preparation that is currently not happening properly this issue should investigate and fix all related issues so that the ignoretest can be eliminated
| 0
|
358,117
| 10,618,383,413
|
IssuesEvent
|
2019-10-13 03:57:12
|
k8smeetup/website-tasks
|
https://api.github.com/repos/k8smeetup/website-tasks
|
closed
|
/blog/_posts/2016-07-00-Kubernetes-1-3-Bridging-Cloud-Native-And-Enterprise-Workloads.md
|
lang/zh priority/P0 priority/P1 sync/new version/1.16 welcome
|
Source File: [/blog/_posts/2016-07-00-Kubernetes-1-3-Bridging-Cloud-Native-And-Enterprise-Workloads.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/blog/_posts/2016-07-00-Kubernetes-1-3-Bridging-Cloud-Native-And-Enterprise-Workloads.md)
|
2.0
|
/blog/_posts/2016-07-00-Kubernetes-1-3-Bridging-Cloud-Native-And-Enterprise-Workloads.md - Source File: [/blog/_posts/2016-07-00-Kubernetes-1-3-Bridging-Cloud-Native-And-Enterprise-Workloads.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/blog/_posts/2016-07-00-Kubernetes-1-3-Bridging-Cloud-Native-And-Enterprise-Workloads.md)
|
non_process
|
blog posts kubernetes bridging cloud native and enterprise workloads md source file
| 0
|
5,893
| 8,709,849,427
|
IssuesEvent
|
2018-12-06 15:01:37
|
Open-EO/openeo-api
|
https://api.github.com/repos/Open-EO/openeo-api
|
closed
|
Avoid name in favor of id and title
|
accepted data discovery process discovery
|
In the specification it is defined that each process and collection has a name. This name is expected to be a unique identifier. I'd propose to change this to an id due to the following reasons:
* id and title are unambiguous, name is not. Throughout software industry name is used for both titles and identifiers. Just by changing name to id the documentation would be much clearer what is expected there.
* Consistency. name is used for processes and datasets, all other use id. As WFS/STAC will change to id, we should also apply this ton the processes. It would also make the conection to process_id in the process graph more clear.
Related issues: https://github.com/opengeospatial/WFS_FES/issues/171 + https://github.com/radiantearth/stac-spec/pull/262
|
1.0
|
Avoid name in favor of id and title - In the specification it is defined that each process and collection has a name. This name is expected to be a unique identifier. I'd propose to change this to an id due to the following reasons:
* id and title are unambiguous, name is not. Throughout software industry name is used for both titles and identifiers. Just by changing name to id the documentation would be much clearer what is expected there.
* Consistency. name is used for processes and datasets, all other use id. As WFS/STAC will change to id, we should also apply this ton the processes. It would also make the conection to process_id in the process graph more clear.
Related issues: https://github.com/opengeospatial/WFS_FES/issues/171 + https://github.com/radiantearth/stac-spec/pull/262
|
process
|
avoid name in favor of id and title in the specification it is defined that each process and collection has a name this name is expected to be a unique identifier i d propose to change this to an id due to the following reasons id and title are unambiguous name is not throughout software industry name is used for both titles and identifiers just by changing name to id the documentation would be much clearer what is expected there consistency name is used for processes and datasets all other use id as wfs stac will change to id we should also apply this ton the processes it would also make the conection to process id in the process graph more clear related issues
| 1
|
642,377
| 20,886,432,532
|
IssuesEvent
|
2022-03-23 06:01:02
|
uranussolutions/minhdu-fontend
|
https://api.github.com/repos/uranussolutions/minhdu-fontend
|
closed
|
Cho phép sửa thông tin phiếu lương
|
HR Refactor high priority medium
|
Bấm vào icon edit sẽ hiển thị lên popup gồm
- Ngày tạo
- Đơn vị: Load list đơn vị từ api -> gửi lên id
- Chức vụ: Load list chức vụ từ api dựa trên đơn vị được chọn -> gửi lên id
- Công thức lương
- Ngày công chuẩn: ràng buộc < 31
-> Gửi lên api: update payroll

|
1.0
|
Cho phép sửa thông tin phiếu lương - Bấm vào icon edit sẽ hiển thị lên popup gồm
- Ngày tạo
- Đơn vị: Load list đơn vị từ api -> gửi lên id
- Chức vụ: Load list chức vụ từ api dựa trên đơn vị được chọn -> gửi lên id
- Công thức lương
- Ngày công chuẩn: ràng buộc < 31
-> Gửi lên api: update payroll

|
non_process
|
cho phép sửa thông tin phiếu lương bấm vào icon edit sẽ hiển thị lên popup gồm ngày tạo đơn vị load list đơn vị từ api gửi lên id chức vụ load list chức vụ từ api dựa trên đơn vị được chọn gửi lên id công thức lương ngày công chuẩn ràng buộc gửi lên api update payroll
| 0
|
21,095
| 28,045,402,011
|
IssuesEvent
|
2023-03-28 22:18:03
|
Azure/azure-sdk-tools
|
https://api.github.com/repos/Azure/azure-sdk-tools
|
closed
|
Actions need sparse checkout
|
Central-EngSys GitHub Event Processor
|
The github-event-processor yml files all use actions/checkout@v3. While this does sync the repository with a depth of 1, it still takes around 1m 30seconds or longer to sync up. The rest of the steps have roughly the following times
1. Setup the Actions job ~ 6 seconds
2. Install github-event-processor ~ 10 seconds.
3. github-event-processor processing the event ~ 5 seconds
//CC @jsquire @weshaggard @benbp @kurtzeborn
|
1.0
|
Actions need sparse checkout - The github-event-processor yml files all use actions/checkout@v3. While this does sync the repository with a depth of 1, it still takes around 1m 30seconds or longer to sync up. The rest of the steps have roughly the following times
1. Setup the Actions job ~ 6 seconds
2. Install github-event-processor ~ 10 seconds.
3. github-event-processor processing the event ~ 5 seconds
//CC @jsquire @weshaggard @benbp @kurtzeborn
|
process
|
actions need sparse checkout the github event processor yml files all use actions checkout while this does sync the repository with a depth of it still takes around or longer to sync up the rest of the steps have roughly the following times setup the actions job seconds install github event processor seconds github event processor processing the event seconds cc jsquire weshaggard benbp kurtzeborn
| 1
|
21,819
| 30,316,668,066
|
IssuesEvent
|
2023-07-10 16:01:30
|
tdwg/dwc
|
https://api.github.com/repos/tdwg/dwc
|
closed
|
Change term - pointRadiusSpatialFit
|
Term - change Class - Location non-normative Process - complete
|
## Term change
* Submitter: John Wieczorek
* Efficacy Justification (why is this change necessary?): Clarification
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term):
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: None
Current Term definition: https://dwc.tdwg.org/list/#dwc_pointRadiusSpatialFit
"The ratio of the area of the point-radius (decimalLatitude, decimalLongitude, coordinateUncertaintyInMeters) to the area of the true (original, or most specific) spatial representation of the Location. Legal values are 0, greater than or equal to 1, or undefined. A value of 1 is an exact match or 100% overlap. A value of 0 should be used if the given point-radius does not completely contain the original representation. The pointRadiusSpatialFit is undefined (and should be left empty) if the original representation is a point without uncertainty and the given georeference is not that same point (without uncertainty). If both the original and the given georeference are the same point, the pointRadiusSpatialFit is 1."
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): pointRadiusSpatialFit
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Location
* Definition of the term (normative): "The ratio of the area of the point-radius (decimalLatitude, decimalLongitude, coordinateUncertaintyInMeters) to the area of the true (original, or most specific) spatial representation of the Location. Legal values are 0, greater than or equal to 1, or undefined. A value of 1 is an exact match or 100% overlap. A value of 0 should be used if the given point-radius does not completely contain the original representation. The pointRadiusSpatialFit is undefined (and should be left empty) if the original representation is **any geometry without area (e.g., a point or polyline) and**~~a point~~ without uncertainty and the given georeference is not that same **geometry**~~point~~ (without uncertainty). If both the original and the given georeference are the same point, the pointRadiusSpatialFit is 1."
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative):
* Refines (identifier of the broader term this term refines; normative):
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): **http://rs.tdwg.org/dwc/terms/version/pointRadiusSpatialFit-2020-08-20**
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/Gathering/SiteCoordinateSets/SiteCoordinates/PointRadiusSpatialFit
This definition change only clarifies what to do in non-area representations other than a point. It does not change the meaning or the usage.
|
1.0
|
Change term - pointRadiusSpatialFit - ## Term change
* Submitter: John Wieczorek
* Efficacy Justification (why is this change necessary?): Clarification
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term):
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: None
Current Term definition: https://dwc.tdwg.org/list/#dwc_pointRadiusSpatialFit
"The ratio of the area of the point-radius (decimalLatitude, decimalLongitude, coordinateUncertaintyInMeters) to the area of the true (original, or most specific) spatial representation of the Location. Legal values are 0, greater than or equal to 1, or undefined. A value of 1 is an exact match or 100% overlap. A value of 0 should be used if the given point-radius does not completely contain the original representation. The pointRadiusSpatialFit is undefined (and should be left empty) if the original representation is a point without uncertainty and the given georeference is not that same point (without uncertainty). If both the original and the given georeference are the same point, the pointRadiusSpatialFit is 1."
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): pointRadiusSpatialFit
* Organized in Class (e.g., Occurrence, Event, Location, Taxon): Location
* Definition of the term (normative): "The ratio of the area of the point-radius (decimalLatitude, decimalLongitude, coordinateUncertaintyInMeters) to the area of the true (original, or most specific) spatial representation of the Location. Legal values are 0, greater than or equal to 1, or undefined. A value of 1 is an exact match or 100% overlap. A value of 0 should be used if the given point-radius does not completely contain the original representation. The pointRadiusSpatialFit is undefined (and should be left empty) if the original representation is **any geometry without area (e.g., a point or polyline) and**~~a point~~ without uncertainty and the given georeference is not that same **geometry**~~point~~ (without uncertainty). If both the original and the given georeference are the same point, the pointRadiusSpatialFit is 1."
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative):
* Refines (identifier of the broader term this term refines; normative):
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): **http://rs.tdwg.org/dwc/terms/version/pointRadiusSpatialFit-2020-08-20**
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/Gathering/SiteCoordinateSets/SiteCoordinates/PointRadiusSpatialFit
This definition change only clarifies what to do in non-area representations other than a point. It does not change the meaning or the usage.
|
process
|
change term pointradiusspatialfit term change submitter john wieczorek efficacy justification why is this change necessary clarification demand justification if the change is semantic in nature name at least two organizations that independently need this term stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version none current term definition the ratio of the area of the point radius decimallatitude decimallongitude coordinateuncertaintyinmeters to the area of the true original or most specific spatial representation of the location legal values are greater than or equal to or undefined a value of is an exact match or overlap a value of should be used if the given point radius does not completely contain the original representation the pointradiusspatialfit is undefined and should be left empty if the original representation is a point without uncertainty and the given georeference is not that same point without uncertainty if both the original and the given georeference are the same point the pointradiusspatialfit is proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes pointradiusspatialfit organized in class e g occurrence event location taxon location definition of the term normative the ratio of the area of the point radius decimallatitude decimallongitude coordinateuncertaintyinmeters to the area of the true original or most specific spatial representation of the location legal values are greater than or equal to or undefined a value of is an exact match or overlap a value of should be used if the given point radius does not completely contain the original representation the pointradiusspatialfit is undefined and should be left empty if the original representation is any geometry without area e g a point or polyline and a point without uncertainty and the given georeference is not that same geometry point without uncertainty if both the original and the given georeference are the same point the pointradiusspatialfit is usage comments recommendations regarding content etc not normative examples not normative refines identifier of the broader term this term refines normative replaces identifier of the existing term that would be deprecated and replaced by this term normative abcd xpath of the equivalent term in abcd or efg not normative datasets dataset units unit gathering sitecoordinatesets sitecoordinates pointradiusspatialfit this definition change only clarifies what to do in non area representations other than a point it does not change the meaning or the usage
| 1
|
12,475
| 14,942,700,278
|
IssuesEvent
|
2021-01-25 21:43:30
|
ORNL-AMO/AMO-Tools-Suite
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Suite
|
opened
|
User input HHV directly
|
Process Heating enhancement
|
Issue overview
--------------
In the suite:
Will have to update bindings and GasCompositions method to optionally take in a heatingValue, if the heatingValue is passed would skip the calculation for heating value.
Is there anything for the solid/liquid that would need to be specially done?
What other values would need to be manually entered (stoichemetric Air? density/specific volume?)
|
1.0
|
User input HHV directly - Issue overview
--------------
In the suite:
Will have to update bindings and GasCompositions method to optionally take in a heatingValue, if the heatingValue is passed would skip the calculation for heating value.
Is there anything for the solid/liquid that would need to be specially done?
What other values would need to be manually entered (stoichemetric Air? density/specific volume?)
|
process
|
user input hhv directly issue overview in the suite will have to update bindings and gascompositions method to optionally take in a heatingvalue if the heatingvalue is passed would skip the calculation for heating value is there anything for the solid liquid that would need to be specially done what other values would need to be manually entered stoichemetric air density specific volume
| 1
|
97,976
| 4,015,565,107
|
IssuesEvent
|
2016-05-15 00:54:19
|
8fold/html5-generator-php
|
https://api.github.com/repos/8fold/html5-generator-php
|
closed
|
Confirm optional attributes are in OPTIONAL_KEYS
|
[priority] high
|
All the attributes beyond the global attributes should be an optional key.
Verify all elements are properly defined regarding optional keys.
|
1.0
|
Confirm optional attributes are in OPTIONAL_KEYS - All the attributes beyond the global attributes should be an optional key.
Verify all elements are properly defined regarding optional keys.
|
non_process
|
confirm optional attributes are in optional keys all the attributes beyond the global attributes should be an optional key verify all elements are properly defined regarding optional keys
| 0
|
215,647
| 7,296,262,436
|
IssuesEvent
|
2018-02-26 10:10:36
|
jcavat/LibellID
|
https://api.github.com/repos/jcavat/LibellID
|
closed
|
Pop-up at first use of the app
|
effort: easy priority: medium type: improvement
|
Il faudrait ajouter le logo de l'application et le texte du pdf ci-joints avec la même mise en page
<img width="999" alt="logo_libellulid_carre" src="https://user-images.githubusercontent.com/32202706/32438829-479652c8-c2ec-11e7-819c-0e874c3e859f.png">
dans le pop-up d'avertissement qui s'ouvre à la première utilisation de l'application.
[1 Contenu avertissement.pdf](https://github.com/jcavat/LibellID/files/1445995/1.Contenu.avertissement.pdf)
|
1.0
|
Pop-up at first use of the app - Il faudrait ajouter le logo de l'application et le texte du pdf ci-joints avec la même mise en page
<img width="999" alt="logo_libellulid_carre" src="https://user-images.githubusercontent.com/32202706/32438829-479652c8-c2ec-11e7-819c-0e874c3e859f.png">
dans le pop-up d'avertissement qui s'ouvre à la première utilisation de l'application.
[1 Contenu avertissement.pdf](https://github.com/jcavat/LibellID/files/1445995/1.Contenu.avertissement.pdf)
|
non_process
|
pop up at first use of the app il faudrait ajouter le logo de l application et le texte du pdf ci joints avec la même mise en page img width alt logo libellulid carre src dans le pop up d avertissement qui s ouvre à la première utilisation de l application
| 0
|
156,303
| 5,966,989,222
|
IssuesEvent
|
2017-05-30 15:06:02
|
zeit/now-cli
|
https://api.github.com/repos/zeit/now-cli
|
closed
|
Determine deployment type automatically
|
high priority version-minor
|
- package.json and dockerfile => docker
- dockerfile => docker
- package.json => npm
- no meta file => static
- item is a file, not a folder => static
It will never prompt and just let the user define with "--docker", "--static" and "--npm" and to the remaining determination part (if no flag defined) automatically by itself.
|
1.0
|
Determine deployment type automatically - - package.json and dockerfile => docker
- dockerfile => docker
- package.json => npm
- no meta file => static
- item is a file, not a folder => static
It will never prompt and just let the user define with "--docker", "--static" and "--npm" and to the remaining determination part (if no flag defined) automatically by itself.
|
non_process
|
determine deployment type automatically package json and dockerfile docker dockerfile docker package json npm no meta file static item is a file not a folder static it will never prompt and just let the user define with docker static and npm and to the remaining determination part if no flag defined automatically by itself
| 0
|
13,032
| 15,382,621,301
|
IssuesEvent
|
2021-03-03 01:01:05
|
googleapis/synthtool
|
https://api.github.com/repos/googleapis/synthtool
|
opened
|
Shard the Python autosynth jobs
|
type: process
|
Look into sharding the Python autosynth jobs like NodeJS and Java. See https://github.com/googleapis/synthtool/pull/744 for how this was done for Java.
CC @crwilcox @craiglabenz
|
1.0
|
Shard the Python autosynth jobs - Look into sharding the Python autosynth jobs like NodeJS and Java. See https://github.com/googleapis/synthtool/pull/744 for how this was done for Java.
CC @crwilcox @craiglabenz
|
process
|
shard the python autosynth jobs look into sharding the python autosynth jobs like nodejs and java see for how this was done for java cc crwilcox craiglabenz
| 1
|
6,645
| 9,762,073,413
|
IssuesEvent
|
2019-06-05 10:20:13
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
opened
|
Processing/gdalwarp returns wrong error message
|
Bug Processing
|
<!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error -->
**QGIS and OS versions**
<!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here -->
**Additional context**
<!-- Add any other context about the problem here. -->
|
1.0
|
Processing/gdalwarp returns wrong error message - <!--
Bug fixing and feature development is a community responsibility, and not the responsibility of the QGIS project alone.
If this bug report or feature request is high-priority for you, we suggest engaging a QGIS developer or support organisation and financially sponsoring a fix
Checklist before submitting
- [ ] Search through existing issue reports and gis.stackexchange.com to check whether the issue already exists
- [ ] Test with a [clean new user profile](https://docs.qgis.org/testing/en/docs/user_manual/introduction/qgis_configuration.html?highlight=profile#working-with-user-profiles).
- [ ] Create a light and self-contained sample dataset and project file which demonstrates the issue
If the issue concerns a **third party plugin**, then it **cannot** be fixed by the QGIS team. Please raise your issue in the dedicated bug tracker for that specific plugin (as listed in the plugin's description). -->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
**How to Reproduce**
<!-- Steps, sample datasets and qgis project file to reproduce the behavior. Screencasts or screenshots welcome
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error -->
**QGIS and OS versions**
<!-- In the QGIS menu help/about, click in the dialog, Ctrl+A and then Ctrl+C. Finally paste here -->
**Additional context**
<!-- Add any other context about the problem here. -->
|
process
|
processing gdalwarp returns wrong error message bug fixing and feature development is a community responsibility and not the responsibility of the qgis project alone if this bug report or feature request is high priority for you we suggest engaging a qgis developer or support organisation and financially sponsoring a fix checklist before submitting search through existing issue reports and gis stackexchange com to check whether the issue already exists test with a create a light and self contained sample dataset and project file which demonstrates the issue if the issue concerns a third party plugin then it cannot be fixed by the qgis team please raise your issue in the dedicated bug tracker for that specific plugin as listed in the plugin s description describe the bug how to reproduce steps sample datasets and qgis project file to reproduce the behavior screencasts or screenshots welcome go to click on scroll down to see error qgis and os versions additional context
| 1
|
21,083
| 28,037,564,989
|
IssuesEvent
|
2023-03-28 16:02:45
|
joeynmt/joeynmt
|
https://api.github.com/repos/joeynmt/joeynmt
|
closed
|
Running build_vocab.py for wmt17_bpe with or without --joint?
|
work in process
|
**Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. task description: running the default wmt17_ende_bpe
2. configuration file : wmt17_ende_bpe.py in this github
3. code call : python build_vocab.py configs/wmt17_ende_bpe.py
**Logged output**
### Training bpe...
100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████| 32000/32000 [01:31<00:00, 349.71it/s]
### Applying bpe...
### Building vocab...
### Done.
### Vocab file subwords/wmt17_bpe.vocab will be overwritten.
### Codes file subwords/wmt17_bpe.codes will be overwitten.
### Training bpe...
100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████| 32000/32000 [03:04<00:00, 173.51it/s]
### Applying bpe...
### Building vocab...
### Done.
**Expected behavior**
de and en should bpe should both be there, no overwritten.
**System (please complete the following information):**
- OS: ubuntu 22.04
- CPU / GPU: 9900k+4090
- Python Version: 3.9
**Additional context**
if I use: build_vocab.py configs/wmt17_ende_bpe.py --joint, an assert triggered in line 312 in build_vocab.py
if args.joint:
for s, t in zip(src_tuple[1:], trg_tuple[1:]):
assert s == t
|
1.0
|
Running build_vocab.py for wmt17_bpe with or without --joint? - **Describe the bug**
A clear and concise description of what the bug is.
**To Reproduce**
Steps to reproduce the behavior:
1. task description: running the default wmt17_ende_bpe
2. configuration file : wmt17_ende_bpe.py in this github
3. code call : python build_vocab.py configs/wmt17_ende_bpe.py
**Logged output**
### Training bpe...
100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████| 32000/32000 [01:31<00:00, 349.71it/s]
### Applying bpe...
### Building vocab...
### Done.
### Vocab file subwords/wmt17_bpe.vocab will be overwritten.
### Codes file subwords/wmt17_bpe.codes will be overwitten.
### Training bpe...
100%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████| 32000/32000 [03:04<00:00, 173.51it/s]
### Applying bpe...
### Building vocab...
### Done.
**Expected behavior**
de and en should bpe should both be there, no overwritten.
**System (please complete the following information):**
- OS: ubuntu 22.04
- CPU / GPU: 9900k+4090
- Python Version: 3.9
**Additional context**
if I use: build_vocab.py configs/wmt17_ende_bpe.py --joint, an assert triggered in line 312 in build_vocab.py
if args.joint:
for s, t in zip(src_tuple[1:], trg_tuple[1:]):
assert s == t
|
process
|
running build vocab py for bpe with or without joint describe the bug a clear and concise description of what the bug is to reproduce steps to reproduce the behavior task description running the default ende bpe configuration file ende bpe py in this github code call python build vocab py configs ende bpe py logged output training bpe ██████████████████████████████████████████████████████████████████████████████████████████████████████████████ applying bpe building vocab done vocab file subwords bpe vocab will be overwritten codes file subwords bpe codes will be overwitten training bpe ██████████████████████████████████████████████████████████████████████████████████████████████████████████████ applying bpe building vocab done expected behavior de and en should bpe should both be there no overwritten system please complete the following information os ubuntu cpu gpu python version additional context if i use build vocab py configs ende bpe py joint an assert triggered in line in build vocab py if args joint for s t in zip src tuple trg tuple assert s t
| 1
|
330,618
| 10,053,265,888
|
IssuesEvent
|
2019-07-21 15:21:05
|
ticket721/web-app
|
https://api.github.com/repos/ticket721/web-app
|
closed
|
d.feat: integrate ethvtx update
|
[priority] [➖➖➖ ] high [status] to do [type] feat
|
- Support ERC20 balances
- Support refresh rates customization
- Lower all refresh rates (server and node)
|
1.0
|
d.feat: integrate ethvtx update - - Support ERC20 balances
- Support refresh rates customization
- Lower all refresh rates (server and node)
|
non_process
|
d feat integrate ethvtx update support balances support refresh rates customization lower all refresh rates server and node
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.