Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
343,050
| 30,650,980,658
|
IssuesEvent
|
2023-07-25 08:53:33
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
closed
|
Fix device.test_gpu_is_available
|
Sub Task Failing Test
|
| | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|
1.0
|
Fix device.test_gpu_is_available - | | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
|
non_process
|
fix device test gpu is available jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
| 0
|
586
| 3,062,240,628
|
IssuesEvent
|
2015-08-16 11:34:49
|
yeoman/yeoman-app
|
https://api.github.com/repos/yeoman/yeoman-app
|
closed
|
Radio buttons not working for me
|
bug renderer process
|
All of the radio buttons generated from prompts `type: list` are not working with the following JS error in the console:
```
[6016:0726/125628:INFO:CONSOLE(28)] "Uncaught TypeError: Cannot read property 'value' of null", source: .../yeoman-app/src/renderer/components/prompts/list.jsx (28)
```
|
1.0
|
Radio buttons not working for me - All of the radio buttons generated from prompts `type: list` are not working with the following JS error in the console:
```
[6016:0726/125628:INFO:CONSOLE(28)] "Uncaught TypeError: Cannot read property 'value' of null", source: .../yeoman-app/src/renderer/components/prompts/list.jsx (28)
```
|
process
|
radio buttons not working for me all of the radio buttons generated from prompts type list are not working with the following js error in the console uncaught typeerror cannot read property value of null source yeoman app src renderer components prompts list jsx
| 1
|
107
| 2,545,536,557
|
IssuesEvent
|
2015-01-29 17:44:28
|
tinkerpop/tinkerpop3
|
https://api.github.com/repos/tinkerpop/tinkerpop3
|
opened
|
Remove Neo4jGraphTraversal.
|
enhancement neo4j process
|
The only reason we need Neo4jGraphTraversal is cause of mid-traversal `cypher()`. Because of this, we have lots of overhead -- source code generation, some tests can't run because it method counts are off, etc. Also, lots of Neo4j code will go away .... e.g. Neo4jElementTraversal, Neo4jVertexTraversal, Neo4jGraphTraversals, Neo4jDefualGraphTraversal, Neo4jEdgeTravaersal, Neo4jVertexPropertyTraversal....
I'm wondering if we get rid of `g.V().out().id().cypher('MATCH...').select('a').outE()` in favor of JUST g.cypher(MATCH).select('a'). In short, cypher() is a method off `Neo4jGraph` and yields a `GraphTraversal`.
Thoughts?
|
1.0
|
Remove Neo4jGraphTraversal. - The only reason we need Neo4jGraphTraversal is cause of mid-traversal `cypher()`. Because of this, we have lots of overhead -- source code generation, some tests can't run because it method counts are off, etc. Also, lots of Neo4j code will go away .... e.g. Neo4jElementTraversal, Neo4jVertexTraversal, Neo4jGraphTraversals, Neo4jDefualGraphTraversal, Neo4jEdgeTravaersal, Neo4jVertexPropertyTraversal....
I'm wondering if we get rid of `g.V().out().id().cypher('MATCH...').select('a').outE()` in favor of JUST g.cypher(MATCH).select('a'). In short, cypher() is a method off `Neo4jGraph` and yields a `GraphTraversal`.
Thoughts?
|
process
|
remove the only reason we need is cause of mid traversal cypher because of this we have lots of overhead source code generation some tests can t run because it method counts are off etc also lots of code will go away e g i m wondering if we get rid of g v out id cypher match select a oute in favor of just g cypher match select a in short cypher is a method off and yields a graphtraversal thoughts
| 1
|
19,618
| 25,970,985,497
|
IssuesEvent
|
2022-12-19 11:11:55
|
googleapis/gapic-generator-csharp
|
https://api.github.com/repos/googleapis/gapic-generator-csharp
|
closed
|
Security Policy violation Binary Artifacts
|
type: process allstar
|
_This issue was automatically created by [Allstar](https://github.com/ossf/allstar/)._
**Security Policy Violation**
Project is out of compliance with Binary Artifacts policy: binaries present in source code
**Rule Description**
Binary Artifacts are an increased security risk in your repository. Binary artifacts cannot be reviewed, allowing the introduction of possibly obsolete or maliciously subverted executables. For more information see the [Security Scorecards Documentation](https://github.com/ossf/scorecard/blob/main/docs/checks.md#binary-artifacts) for Binary Artifacts.
**Remediation Steps**
To remediate, remove the generated executable artifacts from the repository.
**Artifacts Found**
- tools/grpc_csharp_plugin
- tools/grpc_csharp_plugin.exe
- tools/protoc
- tools/protoc.exe
**Additional Information**
This policy is drawn from [Security Scorecards](https://github.com/ossf/scorecard/), which is a tool that scores a project's adherence to security best practices. You may wish to run a Scorecards scan directly on this repository for more details.
---
<!-- Edit section #updates --><!-- Current result text hash: 1fb9f938741aaa770502d7faa1836bd83fd751f88f55304b5deaf90bc8fae266 --><!-- Edit section #updates -->
Allstar has been installed on all Google managed GitHub orgs. Policies are gradually being rolled out and enforced by the GOSST and OSPO teams. Learn more at http://go/allstar
This issue will auto resolve when the policy is in compliance.
Issue created by Allstar. See https://github.com/ossf/allstar/ for more information. For questions specific to the repository, please contact the owner or maintainer.
|
1.0
|
Security Policy violation Binary Artifacts - _This issue was automatically created by [Allstar](https://github.com/ossf/allstar/)._
**Security Policy Violation**
Project is out of compliance with Binary Artifacts policy: binaries present in source code
**Rule Description**
Binary Artifacts are an increased security risk in your repository. Binary artifacts cannot be reviewed, allowing the introduction of possibly obsolete or maliciously subverted executables. For more information see the [Security Scorecards Documentation](https://github.com/ossf/scorecard/blob/main/docs/checks.md#binary-artifacts) for Binary Artifacts.
**Remediation Steps**
To remediate, remove the generated executable artifacts from the repository.
**Artifacts Found**
- tools/grpc_csharp_plugin
- tools/grpc_csharp_plugin.exe
- tools/protoc
- tools/protoc.exe
**Additional Information**
This policy is drawn from [Security Scorecards](https://github.com/ossf/scorecard/), which is a tool that scores a project's adherence to security best practices. You may wish to run a Scorecards scan directly on this repository for more details.
---
<!-- Edit section #updates --><!-- Current result text hash: 1fb9f938741aaa770502d7faa1836bd83fd751f88f55304b5deaf90bc8fae266 --><!-- Edit section #updates -->
Allstar has been installed on all Google managed GitHub orgs. Policies are gradually being rolled out and enforced by the GOSST and OSPO teams. Learn more at http://go/allstar
This issue will auto resolve when the policy is in compliance.
Issue created by Allstar. See https://github.com/ossf/allstar/ for more information. For questions specific to the repository, please contact the owner or maintainer.
|
process
|
security policy violation binary artifacts this issue was automatically created by security policy violation project is out of compliance with binary artifacts policy binaries present in source code rule description binary artifacts are an increased security risk in your repository binary artifacts cannot be reviewed allowing the introduction of possibly obsolete or maliciously subverted executables for more information see the for binary artifacts remediation steps to remediate remove the generated executable artifacts from the repository artifacts found tools grpc csharp plugin tools grpc csharp plugin exe tools protoc tools protoc exe additional information this policy is drawn from which is a tool that scores a project s adherence to security best practices you may wish to run a scorecards scan directly on this repository for more details allstar has been installed on all google managed github orgs policies are gradually being rolled out and enforced by the gosst and ospo teams learn more at this issue will auto resolve when the policy is in compliance issue created by allstar see for more information for questions specific to the repository please contact the owner or maintainer
| 1
|
18,704
| 24,598,953,416
|
IssuesEvent
|
2022-10-14 10:44:30
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
[postgresqlExtensions] consider a destructive change check for dropped-and-recreated extensions in Migrate
|
process/candidate kind/improvement topic: migrate tech/engines/migration engine team/schema topic: postgresqlExtensions
|
Extensions create schema items. Dropping an extension can cause data loss. We drop extensions when we want to relocate non-relocatable extensions.
However, this is safe for _most_ extensions, that do not create tables where users are supposed to insert data.
|
1.0
|
[postgresqlExtensions] consider a destructive change check for dropped-and-recreated extensions in Migrate - Extensions create schema items. Dropping an extension can cause data loss. We drop extensions when we want to relocate non-relocatable extensions.
However, this is safe for _most_ extensions, that do not create tables where users are supposed to insert data.
|
process
|
consider a destructive change check for dropped and recreated extensions in migrate extensions create schema items dropping an extension can cause data loss we drop extensions when we want to relocate non relocatable extensions however this is safe for most extensions that do not create tables where users are supposed to insert data
| 1
|
4,983
| 7,816,432,955
|
IssuesEvent
|
2018-06-13 04:39:59
|
h-a-m-r-time/grace-shopper
|
https://api.github.com/repos/h-a-m-r-time/grace-shopper
|
closed
|
Deploy Research/Discussion
|
process
|
Discuss who's responsiblity to keep deploys going forward, maybe setting up a rolling automated overnight deploy plus allowing us to kick off a deploy whenever necessary
|
1.0
|
Deploy Research/Discussion - Discuss who's responsiblity to keep deploys going forward, maybe setting up a rolling automated overnight deploy plus allowing us to kick off a deploy whenever necessary
|
process
|
deploy research discussion discuss who s responsiblity to keep deploys going forward maybe setting up a rolling automated overnight deploy plus allowing us to kick off a deploy whenever necessary
| 1
|
48,667
| 25,750,301,468
|
IssuesEvent
|
2022-12-08 12:46:57
|
AutomatedProcessImprovement/Simod
|
https://api.github.com/repos/AutomatedProcessImprovement/Simod
|
closed
|
Remove time.sleep from parallel processing in the utilities module
|
performance experiment
|
This line, https://github.com/AutomatedProcessImprovement/Simod/blob/master/src/simod/utilities.py#L143, contributes to about 15% of running time of Simod.
|
True
|
Remove time.sleep from parallel processing in the utilities module - This line, https://github.com/AutomatedProcessImprovement/Simod/blob/master/src/simod/utilities.py#L143, contributes to about 15% of running time of Simod.
|
non_process
|
remove time sleep from parallel processing in the utilities module this line contributes to about of running time of simod
| 0
|
17,978
| 23,995,303,975
|
IssuesEvent
|
2022-09-14 07:00:46
|
ankidroid/Anki-Android
|
https://api.github.com/repos/ankidroid/Anki-Android
|
closed
|
Tests: Use hamcrest.io matchers for files
|
Priority-Low Good First Issue! Stale Test process
|
http://hamcrest.org/JavaHamcrest/javadoc/2.2/org/hamcrest/io/FileMatchers.html is now usable in the codebase. Let's use it to improve tests.
<details><summary>old</summary>
I'd really like to see more custom Hamcrest matchers in the codebase
For example: "file exists" would be useful, so we can reduce the verbosity of error messages.
And something along the lines of `hasSingleElementEqualTo`
_Originally posted by @david-allison in https://github.com/ankidroid/Anki-Android/issues/10385#issuecomment-1049199502_
</details>
|
1.0
|
Tests: Use hamcrest.io matchers for files - http://hamcrest.org/JavaHamcrest/javadoc/2.2/org/hamcrest/io/FileMatchers.html is now usable in the codebase. Let's use it to improve tests.
<details><summary>old</summary>
I'd really like to see more custom Hamcrest matchers in the codebase
For example: "file exists" would be useful, so we can reduce the verbosity of error messages.
And something along the lines of `hasSingleElementEqualTo`
_Originally posted by @david-allison in https://github.com/ankidroid/Anki-Android/issues/10385#issuecomment-1049199502_
</details>
|
process
|
tests use hamcrest io matchers for files is now usable in the codebase let s use it to improve tests old i d really like to see more custom hamcrest matchers in the codebase for example file exists would be useful so we can reduce the verbosity of error messages and something along the lines of hassingleelementequalto originally posted by david allison in
| 1
|
18,122
| 24,158,121,879
|
IssuesEvent
|
2022-09-22 09:22:07
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Processing algorithm "Split with lines" does not split polygon
|
Feedback Processing Bug
|
### What is the bug or the crash?
When using the processing algorithm "Split with lines" to split poylgon with certain line data, it does not split polygon at all.
Here is a simple (topologically correct) line layer to test the bug with: [lines_simplified.zip](https://github.com/qgis/QGIS/files/9581708/lines_simplified.zip)
### Steps to reproduce the issue
1. Add lines_simplified.json to the project

2. Create polygon from layer's extent (with alg "Polygon from layer extent")

3. Run algorithm "Split with lines" using the extent layer and the line layer and get the following output

The output has actually two polyons. One large one, and one tiny one of the upper right corner

### Versions
QGISin versio
3.22.9-Białowieża
QGIS-koodin versio
a8e9e6fa
Qt:n versio
5.15.3
Python-versio
3.9.5
GDAL/OGR:n versio
3.5.1
PROJ-versio
9.0.1
EPSG-rekisteritietokannan versio
v10.064 (2022-05-19)
GEOS-versio
3.10.3-CAPI-1.16.1
SQLite-versio
3.38.1
PDAL-versio
2.3.0
PostgreSQL:n asiakkaan versio
14.3
SpatiaLite-versio
5.0.1
QWT-versio
6.1.6
QScintilla2-versio
2.13.1
Käyttöjärjestelmän versio
Windows 10 Version 2009
Aktiiviset python-lisäosat
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
plugin_reloader
0.9.1
Ilre
0.1.0
Also tested with QGIS 3.26.2 on Arch Linux
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
The problem might lie with the dataset used. The same dataset causes the problems in #50226
|
1.0
|
Processing algorithm "Split with lines" does not split polygon - ### What is the bug or the crash?
When using the processing algorithm "Split with lines" to split poylgon with certain line data, it does not split polygon at all.
Here is a simple (topologically correct) line layer to test the bug with: [lines_simplified.zip](https://github.com/qgis/QGIS/files/9581708/lines_simplified.zip)
### Steps to reproduce the issue
1. Add lines_simplified.json to the project

2. Create polygon from layer's extent (with alg "Polygon from layer extent")

3. Run algorithm "Split with lines" using the extent layer and the line layer and get the following output

The output has actually two polyons. One large one, and one tiny one of the upper right corner

### Versions
QGISin versio
3.22.9-Białowieża
QGIS-koodin versio
a8e9e6fa
Qt:n versio
5.15.3
Python-versio
3.9.5
GDAL/OGR:n versio
3.5.1
PROJ-versio
9.0.1
EPSG-rekisteritietokannan versio
v10.064 (2022-05-19)
GEOS-versio
3.10.3-CAPI-1.16.1
SQLite-versio
3.38.1
PDAL-versio
2.3.0
PostgreSQL:n asiakkaan versio
14.3
SpatiaLite-versio
5.0.1
QWT-versio
6.1.6
QScintilla2-versio
2.13.1
Käyttöjärjestelmän versio
Windows 10 Version 2009
Aktiiviset python-lisäosat
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
plugin_reloader
0.9.1
Ilre
0.1.0
Also tested with QGIS 3.26.2 on Arch Linux
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
The problem might lie with the dataset used. The same dataset causes the problems in #50226
|
process
|
processing algorithm split with lines does not split polygon what is the bug or the crash when using the processing algorithm split with lines to split poylgon with certain line data it does not split polygon at all here is a simple topologically correct line layer to test the bug with steps to reproduce the issue add lines simplified json to the project create polygon from layer s extent with alg polygon from layer extent run algorithm split with lines using the extent layer and the line layer and get the following output the output has actually two polyons one large one and one tiny one of the upper right corner versions qgisin versio białowieża qgis koodin versio qt n versio python versio gdal ogr n versio proj versio epsg rekisteritietokannan versio geos versio capi sqlite versio pdal versio postgresql n asiakkaan versio spatialite versio qwt versio versio käyttöjärjestelmän versio windows version aktiiviset python lisäosat db manager grassprovider metasearch processing sagaprovider plugin reloader ilre also tested with qgis on arch linux supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context the problem might lie with the dataset used the same dataset causes the problems in
| 1
|
12,133
| 14,740,951,908
|
IssuesEvent
|
2021-01-07 09:52:32
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Customer / Client Portal Management
|
anc-process anp-not prioritized ant-child/secondary ant-enhancement customer portal
|
In GitLab by @kdjstudios on Dec 14, 2018, 08:43
**Submitted by:** Kyle
**Helpdesk:** NA
**Server:** ALL
**Client/Site:** All
**Account:** NA
**Issue:**
We need a better way to manage and maintain client portal logins.
- Currently if a user needs to switch from managing one customer to another there is no way for operations to remove a login from the UI.
- Also merging multiple customer logins into one user portal is difficult too and needs to be improved. #576 #1065 (Mock up #597)
- There is no easy way for a customer/client to change their password if they forgot it. (Currently a new invite has to be sent) #1100
|
1.0
|
Customer / Client Portal Management - In GitLab by @kdjstudios on Dec 14, 2018, 08:43
**Submitted by:** Kyle
**Helpdesk:** NA
**Server:** ALL
**Client/Site:** All
**Account:** NA
**Issue:**
We need a better way to manage and maintain client portal logins.
- Currently if a user needs to switch from managing one customer to another there is no way for operations to remove a login from the UI.
- Also merging multiple customer logins into one user portal is difficult too and needs to be improved. #576 #1065 (Mock up #597)
- There is no easy way for a customer/client to change their password if they forgot it. (Currently a new invite has to be sent) #1100
|
process
|
customer client portal management in gitlab by kdjstudios on dec submitted by kyle helpdesk na server all client site all account na issue we need a better way to manage and maintain client portal logins currently if a user needs to switch from managing one customer to another there is no way for operations to remove a login from the ui also merging multiple customer logins into one user portal is difficult too and needs to be improved mock up there is no easy way for a customer client to change their password if they forgot it currently a new invite has to be sent
| 1
|
249,025
| 26,875,573,761
|
IssuesEvent
|
2023-02-05 01:05:28
|
gabriel-milan/nginx-ingress-django-auth
|
https://api.github.com/repos/gabriel-milan/nginx-ingress-django-auth
|
opened
|
CVE-2023-23969 (Medium) detected in Django-3.2.13-py3-none-any.whl
|
security vulnerability
|
## CVE-2023-23969 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-3.2.13-py3-none-any.whl</b></p></summary>
<p>A high-level Python web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl">https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl</a></p>
<p>
Dependency Hierarchy:
- :x: **Django-3.2.13-py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Django 3.2 before 3.2.17, 4.0 before 4.0.9, and 4.1 before 4.1.6, the parsed values of Accept-Language headers are cached in order to avoid repetitive parsing. This leads to a potential denial-of-service vector via excessive memory usage if the raw value of Accept-Language headers is very large.
<p>Publish Date: 2023-02-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23969>CVE-2023-23969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.djangoproject.com/weblog/2023/feb/01/security-releases/">https://www.djangoproject.com/weblog/2023/feb/01/security-releases/</a></p>
<p>Release Date: 2023-02-01</p>
<p>Fix Resolution: django - 3.2.17, 4.0.9, 4.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-23969 (Medium) detected in Django-3.2.13-py3-none-any.whl - ## CVE-2023-23969 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-3.2.13-py3-none-any.whl</b></p></summary>
<p>A high-level Python web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl">https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl</a></p>
<p>
Dependency Hierarchy:
- :x: **Django-3.2.13-py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Django 3.2 before 3.2.17, 4.0 before 4.0.9, and 4.1 before 4.1.6, the parsed values of Accept-Language headers are cached in order to avoid repetitive parsing. This leads to a potential denial-of-service vector via excessive memory usage if the raw value of Accept-Language headers is very large.
<p>Publish Date: 2023-02-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23969>CVE-2023-23969</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.djangoproject.com/weblog/2023/feb/01/security-releases/">https://www.djangoproject.com/weblog/2023/feb/01/security-releases/</a></p>
<p>Release Date: 2023-02-01</p>
<p>Fix Resolution: django - 3.2.17, 4.0.9, 4.1.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in django none any whl cve medium severity vulnerability vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href dependency hierarchy x django none any whl vulnerable library found in base branch master vulnerability details in django before before and before the parsed values of accept language headers are cached in order to avoid repetitive parsing this leads to a potential denial of service vector via excessive memory usage if the raw value of accept language headers is very large publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution django step up your open source security game with mend
| 0
|
18,313
| 2,615,168,421
|
IssuesEvent
|
2015-03-01 06:49:55
|
chrsmith/html5rocks
|
https://api.github.com/repos/chrsmith/html5rocks
|
closed
|
Cutting v7 for www
|
auto-migrated Milestone-7 Priority-P0 Type-Task
|
```
Revision was:
http://code.google.com/p/html5rocks/source/detail?r=098033a3cfc174cc447581e16256
874afd91570a&name=v4
Only www was pushed.
```
Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 5 Aug 2011 at 7:01
|
1.0
|
Cutting v7 for www - ```
Revision was:
http://code.google.com/p/html5rocks/source/detail?r=098033a3cfc174cc447581e16256
874afd91570a&name=v4
Only www was pushed.
```
Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 5 Aug 2011 at 7:01
|
non_process
|
cutting for www revision was name only www was pushed original issue reported on code google com by ericbide com on aug at
| 0
|
93,575
| 26,988,813,137
|
IssuesEvent
|
2023-02-09 18:07:47
|
bitcoin/bitcoin
|
https://api.github.com/repos/bitcoin/bitcoin
|
closed
|
macos depends build does not cache in the CI
|
Bug Build system
|
At least in the CI, the depends build does not cache. E.g. https://cirrus-ci.com/task/5094611354386432?logs=ci#L789
This causes the build to take an hour when it should be done in less than 10 minutes witch a ccache.
|
1.0
|
macos depends build does not cache in the CI - At least in the CI, the depends build does not cache. E.g. https://cirrus-ci.com/task/5094611354386432?logs=ci#L789
This causes the build to take an hour when it should be done in less than 10 minutes witch a ccache.
|
non_process
|
macos depends build does not cache in the ci at least in the ci the depends build does not cache e g this causes the build to take an hour when it should be done in less than minutes witch a ccache
| 0
|
181,869
| 14,891,370,600
|
IssuesEvent
|
2021-01-21 00:42:24
|
lightningnetwork/lnd
|
https://api.github.com/repos/lightningnetwork/lnd
|
closed
|
Docker image not exists
|
docker documentation
|
The [DOCKER Readme](https://github.com/lightningnetwork/lnd/blob/master/docs/DOCKER.md) specifies to use the [lightninglabs/lnd](https://hub.docker.com/r/lightninglabs/lnd/) image, but it currently does not exist on dockerhub.
|
1.0
|
Docker image not exists - The [DOCKER Readme](https://github.com/lightningnetwork/lnd/blob/master/docs/DOCKER.md) specifies to use the [lightninglabs/lnd](https://hub.docker.com/r/lightninglabs/lnd/) image, but it currently does not exist on dockerhub.
|
non_process
|
docker image not exists the specifies to use the image but it currently does not exist on dockerhub
| 0
|
299,899
| 25,936,193,734
|
IssuesEvent
|
2022-12-16 14:22:14
|
DucTrann1310/FeedbackOnline
|
https://api.github.com/repos/DucTrann1310/FeedbackOnline
|
opened
|
[BugID_34]_FUNC_Quản lý topic_Tạo mới topic_Tạo mới topic thành công khi tên topic đã tồn tại trong hệ thống
|
bug Open Integration Test Fun_Wrong Business logic Priority_Medium Severity_Medium
|
Precondition :
- Admin đang ở màn hình tạo mới Topic
- Tồn tại một record với TenTopic = 'Hoangnhi' trong DB table Topic
Step:
1. Nhập "Hoangnhi" vào [Tên Topic] textbox
2. Click [Hoàn tất] button
Actual output:
1. Quay về màn hình Quản lý topic
2. Topic vừa tạo mới lưu ở DB Topic và refresh grid
Expected output: Hiển thị inline message màu đỏ"Tên Topic đã tồn tại!" phía dưới trường Tên Topic và set focus tại đó
|
1.0
|
[BugID_34]_FUNC_Quản lý topic_Tạo mới topic_Tạo mới topic thành công khi tên topic đã tồn tại trong hệ thống - Precondition :
- Admin đang ở màn hình tạo mới Topic
- Tồn tại một record với TenTopic = 'Hoangnhi' trong DB table Topic
Step:
1. Nhập "Hoangnhi" vào [Tên Topic] textbox
2. Click [Hoàn tất] button
Actual output:
1. Quay về màn hình Quản lý topic
2. Topic vừa tạo mới lưu ở DB Topic và refresh grid
Expected output: Hiển thị inline message màu đỏ"Tên Topic đã tồn tại!" phía dưới trường Tên Topic và set focus tại đó
|
non_process
|
func quản lý topic tạo mới topic tạo mới topic thành công khi tên topic đã tồn tại trong hệ thống precondition admin đang ở màn hình tạo mới topic tồn tại một record với tentopic hoangnhi trong db table topic step nhập hoangnhi vào textbox click button actual output quay về màn hình quản lý topic topic vừa tạo mới lưu ở db topic và refresh grid expected output hiển thị inline message màu đỏ tên topic đã tồn tại phía dưới trường tên topic và set focus tại đó
| 0
|
18,668
| 24,584,819,983
|
IssuesEvent
|
2022-10-13 18:42:51
|
googleapis/google-cloud-ruby
|
https://api.github.com/repos/googleapis/google-cloud-ruby
|
closed
|
Example Rails app
|
type: process
|
It would be useful to have a sample rails app that configures and uses many of the most popular gems, including all the stackdriver gems. The app should be configured to run using forked processes on Puma, and demonstrate how to work around the restrictions on delaying GRPC loading until after the fork.
|
1.0
|
Example Rails app - It would be useful to have a sample rails app that configures and uses many of the most popular gems, including all the stackdriver gems. The app should be configured to run using forked processes on Puma, and demonstrate how to work around the restrictions on delaying GRPC loading until after the fork.
|
process
|
example rails app it would be useful to have a sample rails app that configures and uses many of the most popular gems including all the stackdriver gems the app should be configured to run using forked processes on puma and demonstrate how to work around the restrictions on delaying grpc loading until after the fork
| 1
|
4,375
| 7,260,516,209
|
IssuesEvent
|
2018-02-18 10:54:51
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE] Extra control when smoothing geometries
|
Automatic new feature Processing
|
Original commit: https://github.com/qgis/QGIS/commit/4d60d0cdb54cc75f15c4e841aa87a3fd6994c918 by nyalldawson
Add option to QgsGeometry::smooth to not smooth segments shorter than a certain threshold or sharp corners with an angle exceeding a threshold
Expose the angle threshold to processing smooth algorithm
Also:
- optimise QgsGeometry::smooth for new geometry classes
- Fix smooth does not work with geometries containing Z/M
|
1.0
|
[FEATURE] Extra control when smoothing geometries - Original commit: https://github.com/qgis/QGIS/commit/4d60d0cdb54cc75f15c4e841aa87a3fd6994c918 by nyalldawson
Add option to QgsGeometry::smooth to not smooth segments shorter than a certain threshold or sharp corners with an angle exceeding a threshold
Expose the angle threshold to processing smooth algorithm
Also:
- optimise QgsGeometry::smooth for new geometry classes
- Fix smooth does not work with geometries containing Z/M
|
process
|
extra control when smoothing geometries original commit by nyalldawson add option to qgsgeometry smooth to not smooth segments shorter than a certain threshold or sharp corners with an angle exceeding a threshold expose the angle threshold to processing smooth algorithm also optimise qgsgeometry smooth for new geometry classes fix smooth does not work with geometries containing z m
| 1
|
18,386
| 24,515,230,732
|
IssuesEvent
|
2022-10-11 03:59:58
|
f5devcentral/container-egress-service
|
https://api.github.com/repos/f5devcentral/container-egress-service
|
closed
|
CVE-2021-44716
|
fixed processing
|
[CVE-2021-44716](https://nvd.nist.gov/vuln/detail/CVE-2021-44716) Published: January 01, 2022; 12:15:08 AM -0500 V3.1: 7.5 **HIGH** V2.0: 5.0 MEDIUM
> net/http in Go before 1.16.12 and 1.17.x before 1.17.5 allows uncontrolled memory consumption in the header canonicalization cache via HTTP/2 requests.
**To Reproduce**
https://github.com/f5devcentral/container-egress-service/blob/3e8f64bb9249ae60325fa7cd71e77b078abcfef2/go.sum#L428
```
golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20180906233101-161cd47e91fd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20181023162649-9b4f9f5ad519/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20181201002055-351d144fa1fc/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20190213061140-3a22650c66bd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20190311183353-d8887717615a/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg=
golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg=
golang.org/x/net v0.0.0-20190501004415-9ce7a6920f09/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg=
golang.org/x/net v0.0.0-20190503192946-f4e77d36d62c/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg=
golang.org/x/net v0.0.0-20190603091049-60506f45cf65/go.mod h1:HSz+uSET+XFnRR8LxR5pz3Of3rY3CfYBVs4xY44aLks=
golang.org/x/net v0.0.0-20190613194153-d28f0bde5980/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190628185345-da137c7871d7/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190724013045-ca1201d0de80/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190827160401-ba9fcec4b297/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190923162816-aa69164e4478/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200114155413-6afb5195e5aa/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200202094626-16171245cfb2/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200222125558-5a598a2470a0/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200301022130-244492dfa37a/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200324143707-d3edc9973b7e/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200501053045-e0ff5e5a1de5/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200506145744-7e3656a0809f/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200513185701-a91f0712d120/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200520182314-0ba52f642ac2/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200625001655-4c5254603344/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA=
golang.org/x/net v0.0.0-20200707034311-ab3426394381/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA=
golang.org/x/net v0.0.0-20200822124328-c89045814202/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA=
golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU=
golang.org/x/net v0.0.0-20201031054903-ff519b6c9102/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU=
golang.org/x/net v0.0.0-20201110031124-69a78807bb2b/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU=
golang.org/x/net v0.0.0-20201209123823-ac852fbbde11/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg=
golang.org/x/net v0.0.0-20210119194325-5f4716e94777/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg=
golang.org/x/net v0.0.0-20210226172049-e18ecbb05110/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg=
golang.org/x/net v0.0.0-20210316092652-d523dce5a7f4/go.mod h1:RBQZq4jEuRlivfhVLdyRGr576XBO4/greRjx4P4O3yc=
golang.org/x/net v0.0.0-20210405180319-a5a99cb37ef4/go.mod h1:p54w0d4576C0XHj96bSt6lcn1PtDYWL6XObtHCRCNQM=
golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2 h1:CIJ76btIcR3eFI5EgSo6k1qKw9KJexJuRLI9G7Hp5wE=
golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
```
**Expected behavior**
golang.org/x/net 0.0.0-20211209124913-491a49abca63 or higher
|
1.0
|
CVE-2021-44716 - [CVE-2021-44716](https://nvd.nist.gov/vuln/detail/CVE-2021-44716) Published: January 01, 2022; 12:15:08 AM -0500 V3.1: 7.5 **HIGH** V2.0: 5.0 MEDIUM
> net/http in Go before 1.16.12 and 1.17.x before 1.17.5 allows uncontrolled memory consumption in the header canonicalization cache via HTTP/2 requests.
**To Reproduce**
https://github.com/f5devcentral/container-egress-service/blob/3e8f64bb9249ae60325fa7cd71e77b078abcfef2/go.sum#L428
```
golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20180906233101-161cd47e91fd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20181023162649-9b4f9f5ad519/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20181201002055-351d144fa1fc/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20190213061140-3a22650c66bd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
golang.org/x/net v0.0.0-20190311183353-d8887717615a/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg=
golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg=
golang.org/x/net v0.0.0-20190501004415-9ce7a6920f09/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg=
golang.org/x/net v0.0.0-20190503192946-f4e77d36d62c/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg=
golang.org/x/net v0.0.0-20190603091049-60506f45cf65/go.mod h1:HSz+uSET+XFnRR8LxR5pz3Of3rY3CfYBVs4xY44aLks=
golang.org/x/net v0.0.0-20190613194153-d28f0bde5980/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190628185345-da137c7871d7/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190724013045-ca1201d0de80/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190827160401-ba9fcec4b297/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20190923162816-aa69164e4478/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200114155413-6afb5195e5aa/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200202094626-16171245cfb2/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200222125558-5a598a2470a0/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200301022130-244492dfa37a/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s=
golang.org/x/net v0.0.0-20200324143707-d3edc9973b7e/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200501053045-e0ff5e5a1de5/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200506145744-7e3656a0809f/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200513185701-a91f0712d120/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200520182314-0ba52f642ac2/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A=
golang.org/x/net v0.0.0-20200625001655-4c5254603344/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA=
golang.org/x/net v0.0.0-20200707034311-ab3426394381/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA=
golang.org/x/net v0.0.0-20200822124328-c89045814202/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA=
golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU=
golang.org/x/net v0.0.0-20201031054903-ff519b6c9102/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU=
golang.org/x/net v0.0.0-20201110031124-69a78807bb2b/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU=
golang.org/x/net v0.0.0-20201209123823-ac852fbbde11/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg=
golang.org/x/net v0.0.0-20210119194325-5f4716e94777/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg=
golang.org/x/net v0.0.0-20210226172049-e18ecbb05110/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg=
golang.org/x/net v0.0.0-20210316092652-d523dce5a7f4/go.mod h1:RBQZq4jEuRlivfhVLdyRGr576XBO4/greRjx4P4O3yc=
golang.org/x/net v0.0.0-20210405180319-a5a99cb37ef4/go.mod h1:p54w0d4576C0XHj96bSt6lcn1PtDYWL6XObtHCRCNQM=
golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2 h1:CIJ76btIcR3eFI5EgSo6k1qKw9KJexJuRLI9G7Hp5wE=
golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
```
**Expected behavior**
golang.org/x/net 0.0.0-20211209124913-491a49abca63 or higher
|
process
|
cve published january am high medium net http in go before and x before allows uncontrolled memory consumption in the header canonicalization cache via http requests to reproduce golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod hsz uset golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net golang org x net go mod expected behavior golang org x net or higher
| 1
|
297,743
| 25,759,860,493
|
IssuesEvent
|
2022-12-08 19:30:05
|
rackslab/racksdb
|
https://api.github.com/repos/rackslab/racksdb
|
opened
|
Define unit tests for core generic components
|
tests
|
Unit tests must be defined with a good coverage of core generic components. It should at least cover the following points:
- [ ] Back references in Schema
- [ ] Default properties in Schema
- [ ] Boolean properties in Schema
- [ ] Circular dependencies in schemas
- [ ] Generic DB loads
|
1.0
|
Define unit tests for core generic components - Unit tests must be defined with a good coverage of core generic components. It should at least cover the following points:
- [ ] Back references in Schema
- [ ] Default properties in Schema
- [ ] Boolean properties in Schema
- [ ] Circular dependencies in schemas
- [ ] Generic DB loads
|
non_process
|
define unit tests for core generic components unit tests must be defined with a good coverage of core generic components it should at least cover the following points back references in schema default properties in schema boolean properties in schema circular dependencies in schemas generic db loads
| 0
|
765
| 3,250,945,546
|
IssuesEvent
|
2015-10-19 06:29:26
|
e-government-ua/i
|
https://api.github.com/repos/e-government-ua/i
|
closed
|
На главном портале добавить в объект markers новый валидатор по КВЕД
|
active bug In process of testing test _central-js
|
(в под-объект validate)
id = KVED //вид экономической деятельности по КВЕД.
Логика: две цифры точка две цифры (первые две цифры не могут быть 04, 34, 40, 44, 48, 54, 57, 67, 76, 83, 89)
Сообщение: Такого КВЕД не існує - (ви не можете вписувати літери)
Поля: kved
|
1.0
|
На главном портале добавить в объект markers новый валидатор по КВЕД - (в под-объект validate)
id = KVED //вид экономической деятельности по КВЕД.
Логика: две цифры точка две цифры (первые две цифры не могут быть 04, 34, 40, 44, 48, 54, 57, 67, 76, 83, 89)
Сообщение: Такого КВЕД не існує - (ви не можете вписувати літери)
Поля: kved
|
process
|
на главном портале добавить в объект markers новый валидатор по квед в под объект validate id kved вид экономической деятельности по квед логика две цифры точка две цифры первые две цифры не могут быть сообщение такого квед не існує ви не можете вписувати літери поля kved
| 1
|
17,589
| 23,408,195,825
|
IssuesEvent
|
2022-08-12 14:46:11
|
hashgraph/hedera-json-rpc-relay
|
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
|
closed
|
Add htsPrecompile acceptance tests support for nonFungible token allowance/approval methods
|
enhancement limechain P2 process
|
### Problem
No htsPrecompile acceptance tests support for non fungible token allowance/approval methods exists
### Solution
Add support for the non fungible token allowance/approval verifications:
- getApproved(address token, int64 serialNumber) external returns (int64 responseCode, address approved)
- setApprovalForAll(address token, address operator, bool approved) external returns (int64 responseCode)
- isApprovedForAll(address token, address owner, address operator) external returns (int64 responseCode, bool approved)
### Alternatives
_No response_
|
1.0
|
Add htsPrecompile acceptance tests support for nonFungible token allowance/approval methods - ### Problem
No htsPrecompile acceptance tests support for non fungible token allowance/approval methods exists
### Solution
Add support for the non fungible token allowance/approval verifications:
- getApproved(address token, int64 serialNumber) external returns (int64 responseCode, address approved)
- setApprovalForAll(address token, address operator, bool approved) external returns (int64 responseCode)
- isApprovedForAll(address token, address owner, address operator) external returns (int64 responseCode, bool approved)
### Alternatives
_No response_
|
process
|
add htsprecompile acceptance tests support for nonfungible token allowance approval methods problem no htsprecompile acceptance tests support for non fungible token allowance approval methods exists solution add support for the non fungible token allowance approval verifications getapproved address token serialnumber external returns responsecode address approved setapprovalforall address token address operator bool approved external returns responsecode isapprovedforall address token address owner address operator external returns responsecode bool approved alternatives no response
| 1
|
771,644
| 27,088,149,438
|
IssuesEvent
|
2023-02-14 18:37:38
|
googleapis/google-cloud-ruby
|
https://api.github.com/repos/googleapis/google-cloud-ruby
|
closed
|
[Nightly CI Failures] Failures detected for google-cloud-talent-v4
|
type: bug priority: p1 nightly failure
|
At 2023-02-09 08:59:34 UTC, detected failures in google-cloud-talent-v4 for: rubocop
report_key_0af3fa4a3a8bd1dc69005961b063920e
|
1.0
|
[Nightly CI Failures] Failures detected for google-cloud-talent-v4 - At 2023-02-09 08:59:34 UTC, detected failures in google-cloud-talent-v4 for: rubocop
report_key_0af3fa4a3a8bd1dc69005961b063920e
|
non_process
|
failures detected for google cloud talent at utc detected failures in google cloud talent for rubocop report key
| 0
|
6,275
| 9,231,176,627
|
IssuesEvent
|
2019-03-13 01:09:14
|
EthVM/EthVM
|
https://api.github.com/repos/EthVM/EthVM
|
closed
|
Usage of Buffer correctly
|
bug milestone:2 priority:medium project:api project:explorer project:processing
|
Currently in backend we're storing BigIntegers as Decimal128 which is an special type supported by Mongo.
Whenever we try to send the value to the front-end as Buffer, I've found issues to convert those value back to BN (js). So we opted to send the values as raw strings.
This task should take care of researching and possible fixing this little issue.
|
1.0
|
Usage of Buffer correctly - Currently in backend we're storing BigIntegers as Decimal128 which is an special type supported by Mongo.
Whenever we try to send the value to the front-end as Buffer, I've found issues to convert those value back to BN (js). So we opted to send the values as raw strings.
This task should take care of researching and possible fixing this little issue.
|
process
|
usage of buffer correctly currently in backend we re storing bigintegers as which is an special type supported by mongo whenever we try to send the value to the front end as buffer i ve found issues to convert those value back to bn js so we opted to send the values as raw strings this task should take care of researching and possible fixing this little issue
| 1
|
175,851
| 21,336,915,248
|
IssuesEvent
|
2022-04-18 15:36:27
|
ShaikUsaf/frameworks_native_AOSP10_r33_ShaikUsaf-frameworks_native_AOSP10_r33_CVE-2020-0226
|
https://api.github.com/repos/ShaikUsaf/frameworks_native_AOSP10_r33_ShaikUsaf-frameworks_native_AOSP10_r33_CVE-2020-0226
|
opened
|
CVE-2020-0392 (High) detected in multiple libraries
|
security vulnerability
|
## CVE-2020-0392 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r45</b>, <b>nativeandroid-10.0.0_r45</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In getLayerDebugInfo of SurfaceFlinger.cpp, there is a possible code execution due to a double free. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-9 Android-10 Android-11Android ID: A-150226608
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0392>CVE-2020-0392</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: android-9.0.0_r60,android-10.0.0_r46</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-0392 (High) detected in multiple libraries - ## CVE-2020-0392 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r45</b>, <b>nativeandroid-10.0.0_r45</b></p></summary>
<p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In getLayerDebugInfo of SurfaceFlinger.cpp, there is a possible code execution due to a double free. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-9 Android-10 Android-11Android ID: A-150226608
<p>Publish Date: 2020-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0392>CVE-2020-0392</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: android-9.0.0_r60,android-10.0.0_r46</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries nativeandroid nativeandroid nativeandroid nativeandroid nativeandroid vulnerability details in getlayerdebuginfo of surfaceflinger cpp there is a possible code execution due to a double free this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android step up your open source security game with whitesource
| 0
|
365,395
| 25,533,630,341
|
IssuesEvent
|
2022-11-29 10:17:58
|
SovereignCloudStack/Docs
|
https://api.github.com/repos/SovereignCloudStack/Docs
|
closed
|
Release Notes for R3
|
documentation IaaS Container Ops release3 IAM
|
We have started to collect the central pieces of the release notes.
Please have a look at https://github.com/SovereignCloudStack/Docs/blob/feat/RelNotes3/Release-Notes/Release3.md in branch feat/RelNotes3 of this Docs repository.
Pull requests and comments are more than welcome!
|
1.0
|
Release Notes for R3 - We have started to collect the central pieces of the release notes.
Please have a look at https://github.com/SovereignCloudStack/Docs/blob/feat/RelNotes3/Release-Notes/Release3.md in branch feat/RelNotes3 of this Docs repository.
Pull requests and comments are more than welcome!
|
non_process
|
release notes for we have started to collect the central pieces of the release notes please have a look at in branch feat of this docs repository pull requests and comments are more than welcome
| 0
|
764,307
| 26,794,171,007
|
IssuesEvent
|
2023-02-01 10:38:23
|
unitaryfund/mitiq
|
https://api.github.com/repos/unitaryfund/mitiq
|
closed
|
Improve the handling of negative probabilities in REM
|
priority/mid feature-request rem
|
## Issue Description
At the moment we clip negative probabilities when running REM.
An alternative solution, typically used in the literature, is to return the closest positive probability distribution.
I.e., after getting the adjusted quasi-prob distribution $p' = C^{-1} p$, where $C$ is the confusion matrix, instead of returning
$p_{\rm positive} = clip(p) / |clip(p)|_1$ it is probably better to return something the optimal solution of:
$$ \min_{p_{\rm positive}} \||p' - p_{\rm positive}\||_1$$
See e.g. https://arxiv.org/abs/2006.14044
## Additional comment
In principle, if we are just interested in computing expectation values, one can also work with the non-positive distribution $p' = C^{-1} p$. Actually this would be the optimal approach since if we avoid any post-processing we get an unbiased estimate of the expectation value, similar to PEC. See e.g. https://arxiv.org/abs/2006.14044 . Not sure however how easy is the implementation of this approach in Mitiq.
|
1.0
|
Improve the handling of negative probabilities in REM - ## Issue Description
At the moment we clip negative probabilities when running REM.
An alternative solution, typically used in the literature, is to return the closest positive probability distribution.
I.e., after getting the adjusted quasi-prob distribution $p' = C^{-1} p$, where $C$ is the confusion matrix, instead of returning
$p_{\rm positive} = clip(p) / |clip(p)|_1$ it is probably better to return something the optimal solution of:
$$ \min_{p_{\rm positive}} \||p' - p_{\rm positive}\||_1$$
See e.g. https://arxiv.org/abs/2006.14044
## Additional comment
In principle, if we are just interested in computing expectation values, one can also work with the non-positive distribution $p' = C^{-1} p$. Actually this would be the optimal approach since if we avoid any post-processing we get an unbiased estimate of the expectation value, similar to PEC. See e.g. https://arxiv.org/abs/2006.14044 . Not sure however how easy is the implementation of this approach in Mitiq.
|
non_process
|
improve the handling of negative probabilities in rem issue description at the moment we clip negative probabilities when running rem an alternative solution typically used in the literature is to return the closest positive probability distribution i e after getting the adjusted quasi prob distribution p c p where c is the confusion matrix instead of returning p rm positive clip p clip p it is probably better to return something the optimal solution of min p rm positive p p rm positive see e g additional comment in principle if we are just interested in computing expectation values one can also work with the non positive distribution p c p actually this would be the optimal approach since if we avoid any post processing we get an unbiased estimate of the expectation value similar to pec see e g not sure however how easy is the implementation of this approach in mitiq
| 0
|
21,321
| 28,931,475,805
|
IssuesEvent
|
2023-05-09 00:00:12
|
devinit/DIwebsite-redesign
|
https://api.github.com/repos/devinit/DIwebsite-redesign
|
closed
|
Enhance accessibility on the Short Publication Page
|
Enhancement wontfix Feedback to Process
|
Perform an accessibility audit on the Short Publications Page and fix any flagged accessibility issues.
Resources: [WAVE](https://chrome.google.com/webstore/detail/wave-evaluation-tool/jbbplnpkjmmeebjpijfedlgcdilocofh)
For full coverage, you'll need to create a test page that makes use of every possible StreamField item.
|
1.0
|
Enhance accessibility on the Short Publication Page - Perform an accessibility audit on the Short Publications Page and fix any flagged accessibility issues.
Resources: [WAVE](https://chrome.google.com/webstore/detail/wave-evaluation-tool/jbbplnpkjmmeebjpijfedlgcdilocofh)
For full coverage, you'll need to create a test page that makes use of every possible StreamField item.
|
process
|
enhance accessibility on the short publication page perform an accessibility audit on the short publications page and fix any flagged accessibility issues resources for full coverage you ll need to create a test page that makes use of every possible streamfield item
| 1
|
410,317
| 11,986,254,271
|
IssuesEvent
|
2020-04-07 18:58:16
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
coronavirus.saude.gov.br - Desktop site instead of mobile site on "Release Notes" page
|
browser-firefox engine-gecko priority-normal severity-minor type-covid19
|
<!-- @browser: Firefox Preview Nightly 200405 (🦎: 76.0a1-20200403063228) -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:77.0) Gecko/20100101 Firefox/77.0 -->
<!-- @reported_with: -->
**URL**: https://coronavirus.saude.gov.br/templates/padraogoverno01/releaseNotes.html
**Browser / Version**: Firefox Preview Nightly 200405 (🦎: 76.0a1-20200403063228)
**Operating System**: Samsung Galaxy S6 Edge (Android 7.0) - 1440 x 2560 pixels (~577 ppi pixel density)
**Tested Another Browser**: Yes Chrome
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
1. Navigate to https://coronavirus.saude.gov.br/templates/padraogoverno01/releaseNotes.html
2. Observe the page.
**Expected Behavior:**
Mobile layout is displayed.
**Actual Behavior:**
Desktop layout is displayed.
**Notes:**
1. Screenshot attached.
2. The issue is not reproducible on Chrome 80.0.3987.162.
Watchers:
@softvision-oana-arbuzov
@softvision-sergiulogigan
@cipriansv
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/b87f0add-10fc-4207-8294-c0ce33731ed4.jpg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
coronavirus.saude.gov.br - Desktop site instead of mobile site on "Release Notes" page - <!-- @browser: Firefox Preview Nightly 200405 (🦎: 76.0a1-20200403063228) -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:77.0) Gecko/20100101 Firefox/77.0 -->
<!-- @reported_with: -->
**URL**: https://coronavirus.saude.gov.br/templates/padraogoverno01/releaseNotes.html
**Browser / Version**: Firefox Preview Nightly 200405 (🦎: 76.0a1-20200403063228)
**Operating System**: Samsung Galaxy S6 Edge (Android 7.0) - 1440 x 2560 pixels (~577 ppi pixel density)
**Tested Another Browser**: Yes Chrome
**Problem type**: Desktop site instead of mobile site
**Description**: Desktop site instead of mobile site
**Steps to Reproduce**:
1. Navigate to https://coronavirus.saude.gov.br/templates/padraogoverno01/releaseNotes.html
2. Observe the page.
**Expected Behavior:**
Mobile layout is displayed.
**Actual Behavior:**
Desktop layout is displayed.
**Notes:**
1. Screenshot attached.
2. The issue is not reproducible on Chrome 80.0.3987.162.
Watchers:
@softvision-oana-arbuzov
@softvision-sergiulogigan
@cipriansv
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/b87f0add-10fc-4207-8294-c0ce33731ed4.jpg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
coronavirus saude gov br desktop site instead of mobile site on release notes page url browser version firefox preview nightly 🦎 operating system samsung galaxy edge android x pixels ppi pixel density tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce navigate to observe the page expected behavior mobile layout is displayed actual behavior desktop layout is displayed notes screenshot attached the issue is not reproducible on chrome watchers softvision oana arbuzov softvision sergiulogigan cipriansv view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
18,301
| 24,414,659,728
|
IssuesEvent
|
2022-10-05 14:57:39
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
Prevent new uses of old ztest API
|
Process
|
Discussed in the TSC meeting, we need a way to prevent new uses of the ztest API.
I believe that the best way to do this will be to add a new check for PRs similar to a language filter that checks for inappropriate language. To do this we'd need:
1. Create a new repo something like `zephyrproject-rtos/word-blocklist`
2. Provide an implementation of the GitHub check that scans for new test suites:
a. `ztest_register_test_suite`
b. `ztest_test_suite`
I don't believe that we should block adding more tests to existing suites. That will automatically happen as people migrate to the V2 API.
Once we're done, `zephyrproject-rtos/word-blocklist` can remain and we can use this for other migrations. It's also possible that we can find an existing GitHub CI for profanity filter and simply fork it. I don't think it'll be a bad idea anyway to have a profanity filter.
|
1.0
|
Prevent new uses of old ztest API - Discussed in the TSC meeting, we need a way to prevent new uses of the ztest API.
I believe that the best way to do this will be to add a new check for PRs similar to a language filter that checks for inappropriate language. To do this we'd need:
1. Create a new repo something like `zephyrproject-rtos/word-blocklist`
2. Provide an implementation of the GitHub check that scans for new test suites:
a. `ztest_register_test_suite`
b. `ztest_test_suite`
I don't believe that we should block adding more tests to existing suites. That will automatically happen as people migrate to the V2 API.
Once we're done, `zephyrproject-rtos/word-blocklist` can remain and we can use this for other migrations. It's also possible that we can find an existing GitHub CI for profanity filter and simply fork it. I don't think it'll be a bad idea anyway to have a profanity filter.
|
process
|
prevent new uses of old ztest api discussed in the tsc meeting we need a way to prevent new uses of the ztest api i believe that the best way to do this will be to add a new check for prs similar to a language filter that checks for inappropriate language to do this we d need create a new repo something like zephyrproject rtos word blocklist provide an implementation of the github check that scans for new test suites a ztest register test suite b ztest test suite i don t believe that we should block adding more tests to existing suites that will automatically happen as people migrate to the api once we re done zephyrproject rtos word blocklist can remain and we can use this for other migrations it s also possible that we can find an existing github ci for profanity filter and simply fork it i don t think it ll be a bad idea anyway to have a profanity filter
| 1
|
11,098
| 13,940,133,617
|
IssuesEvent
|
2020-10-22 17:27:31
|
googleapis/java-webrisk
|
https://api.github.com/repos/googleapis/java-webrisk
|
closed
|
Dependency Dashboard
|
api: webrisk type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-webrisk-1.x -->[chore(deps): update dependency com.google.cloud:google-cloud-webrisk to v1.1.1](../pull/218)
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->[chore(deps): update dependency com.google.cloud:libraries-bom to v10](../pull/229)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-webrisk-1.x -->[chore(deps): update dependency com.google.cloud:google-cloud-webrisk to v1.1.1](../pull/218)
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->[chore(deps): update dependency com.google.cloud:libraries-bom to v10](../pull/229)
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any pull pull check this box to trigger a request for renovate to run again on this repository
| 1
|
53,936
| 13,230,144,888
|
IssuesEvent
|
2020-08-18 09:20:33
|
feelpp/feelpp
|
https://api.github.com/repos/feelpp/feelpp
|
opened
|
Re-enable doxygen
|
module:build module:doc type:feature
|
As discussed with @vincentchabannes we should re-enable doxygen to build the API documentation
This shall work as follows
- [ ] add github action to build the api docs and push them to gh-pages
- [ ] update [Doxyfile.in](https://github.com/feelpp/feelpp/blob/develop/doc/api/Doxyfile.in)
- [ ] use grouping to add classes to modules (here module means a subdirectory)
We have to sets of modules
- feelpp modules, the name of the module should be `feelpp/<module name>`
- feelpp-toolboxes modules, the name of the module should be `feelpp-toolboxes/<module name>`
|
1.0
|
Re-enable doxygen - As discussed with @vincentchabannes we should re-enable doxygen to build the API documentation
This shall work as follows
- [ ] add github action to build the api docs and push them to gh-pages
- [ ] update [Doxyfile.in](https://github.com/feelpp/feelpp/blob/develop/doc/api/Doxyfile.in)
- [ ] use grouping to add classes to modules (here module means a subdirectory)
We have to sets of modules
- feelpp modules, the name of the module should be `feelpp/<module name>`
- feelpp-toolboxes modules, the name of the module should be `feelpp-toolboxes/<module name>`
|
non_process
|
re enable doxygen as discussed with vincentchabannes we should re enable doxygen to build the api documentation this shall work as follows add github action to build the api docs and push them to gh pages update use grouping to add classes to modules here module means a subdirectory we have to sets of modules feelpp modules the name of the module should be feelpp feelpp toolboxes modules the name of the module should be feelpp toolboxes
| 0
|
4,474
| 7,341,343,093
|
IssuesEvent
|
2018-03-07 01:34:00
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
How to enable these Preview features in Azure SQL?
|
cxp in-process product-question sql-database triaged
|
How to enable these Preview features in Azure SQL?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9ca439af-510f-303c-e72c-424b78ac0ac7
* Version Independent ID: 616d8cf5-1631-1179-0c59-f8b4f6c1f7bb
* [Content](https://docs.microsoft.com/en-us/azure/sql-database/sql-database-data-discovery-and-classification)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/sql-database/sql-database-data-discovery-and-classification.md)
* Service: sql-database
|
1.0
|
How to enable these Preview features in Azure SQL? - How to enable these Preview features in Azure SQL?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9ca439af-510f-303c-e72c-424b78ac0ac7
* Version Independent ID: 616d8cf5-1631-1179-0c59-f8b4f6c1f7bb
* [Content](https://docs.microsoft.com/en-us/azure/sql-database/sql-database-data-discovery-and-classification)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/sql-database/sql-database-data-discovery-and-classification.md)
* Service: sql-database
|
process
|
how to enable these preview features in azure sql how to enable these preview features in azure sql document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id service sql database
| 1
|
81,794
| 15,630,140,519
|
IssuesEvent
|
2021-03-22 01:28:53
|
TrapTeamCCNZ/webui
|
https://api.github.com/repos/TrapTeamCCNZ/webui
|
closed
|
CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - autoclosed
|
security vulnerability
|
## CVE-2020-7733 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p>
<p>Path to dependency file: webui/AdminLTE-3.0.0-rc.6/AdminLTE-3.0.0-rc.6/package.json</p>
<p>Path to vulnerable library: webui/AdminLTE-3.0.0-rc.6/AdminLTE-3.0.0-rc.6/node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.7.tgz (Root Library)
- :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TrapTeamCCNZ/webui/commit/914b985d163674d879ae2e213efd026abf814948">914b985d163674d879ae2e213efd026abf814948</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA.
<p>Publish Date: 2020-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7733>CVE-2020-7733</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 0.7.22</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - autoclosed - ## CVE-2020-7733 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary>
<p>Lightweight JavaScript-based user-agent string parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p>
<p>Path to dependency file: webui/AdminLTE-3.0.0-rc.6/AdminLTE-3.0.0-rc.6/package.json</p>
<p>Path to vulnerable library: webui/AdminLTE-3.0.0-rc.6/AdminLTE-3.0.0-rc.6/node_modules/ua-parser-js/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.26.7.tgz (Root Library)
- :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TrapTeamCCNZ/webui/commit/914b985d163674d879ae2e213efd026abf814948">914b985d163674d879ae2e213efd026abf814948</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA.
<p>Publish Date: 2020-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7733>CVE-2020-7733</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 0.7.22</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in ua parser js tgz autoclosed cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file webui adminlte rc adminlte rc package json path to vulnerable library webui adminlte rc adminlte rc node modules ua parser js package json dependency hierarchy browser sync tgz root library x ua parser js tgz vulnerable library found in head commit a href vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos via the regex for redmi phones and mi pad tablets ua publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
260,093
| 19,657,469,531
|
IssuesEvent
|
2022-01-10 13:58:54
|
felixstrobel/mailtm
|
https://api.github.com/repos/felixstrobel/mailtm
|
opened
|
Add in-code documentation
|
documentation enhancement
|
We should probably write an in-code documentation in [GoDoc](https://go.dev/blog/godoc) to improve the development speed.
|
1.0
|
Add in-code documentation - We should probably write an in-code documentation in [GoDoc](https://go.dev/blog/godoc) to improve the development speed.
|
non_process
|
add in code documentation we should probably write an in code documentation in to improve the development speed
| 0
|
14,865
| 18,273,498,331
|
IssuesEvent
|
2021-10-04 16:03:25
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Re-introspection support using `db pull` with MongoDB
|
kind/feature process/candidate topic: re-introspection tech/engines tech/typescript team/migrations topic: mongodb
|
Introspection support for MongoDB (Preview) was shipped with Prisma version 3.2.0 without re-introspection.
Meaning that during re-introspection some information will be lost (relations, comments, mapped fields...)
|
1.0
|
Re-introspection support using `db pull` with MongoDB - Introspection support for MongoDB (Preview) was shipped with Prisma version 3.2.0 without re-introspection.
Meaning that during re-introspection some information will be lost (relations, comments, mapped fields...)
|
process
|
re introspection support using db pull with mongodb introspection support for mongodb preview was shipped with prisma version without re introspection meaning that during re introspection some information will be lost relations comments mapped fields
| 1
|
140,605
| 12,943,159,681
|
IssuesEvent
|
2020-07-18 05:29:15
|
Naeemkh/tsprocess
|
https://api.github.com/repos/Naeemkh/tsprocess
|
closed
|
doc: Add doctest
|
documentation
|
Add doctest to make sure that documents and the codebase are getting updated together.
Read more:
https://docs.python.org/3/library/doctest.html
|
1.0
|
doc: Add doctest - Add doctest to make sure that documents and the codebase are getting updated together.
Read more:
https://docs.python.org/3/library/doctest.html
|
non_process
|
doc add doctest add doctest to make sure that documents and the codebase are getting updated together read more
| 0
|
10,429
| 13,218,985,374
|
IssuesEvent
|
2020-08-17 09:40:52
|
ESMValGroup/ESMValCore
|
https://api.github.com/repos/ESMValGroup/ESMValCore
|
closed
|
Standardized anomalies should be unitless.
|
bug preprocessor
|
**Describe the bug**
The title says it all. Found by @hb326 while testing https://github.com/ESMValGroup/ESMValTool/issues/1687.
|
1.0
|
Standardized anomalies should be unitless. - **Describe the bug**
The title says it all. Found by @hb326 while testing https://github.com/ESMValGroup/ESMValTool/issues/1687.
|
process
|
standardized anomalies should be unitless describe the bug the title says it all found by while testing
| 1
|
19,680
| 26,031,908,370
|
IssuesEvent
|
2022-12-21 22:20:22
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
What, exactly, is the set of invalid characters for a build number?
|
doc-enhancement devops/prod Pri1 devops-cicd-process/tech
|
In this documentation, with respect to the token `$(Build.DefinitionName)`, the following note is provided in the table:
> Note: The pipeline name must not contain invalid or whitespace characters.
But this article fails to define what are invalid characters.
For example, I want to set the date _and time_ in my build number. Are colons valid characters?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93
* Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7
* Content: [Run (build) number - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=classic)
* Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/run-number.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
What, exactly, is the set of invalid characters for a build number? -
In this documentation, with respect to the token `$(Build.DefinitionName)`, the following note is provided in the table:
> Note: The pipeline name must not contain invalid or whitespace characters.
But this article fails to define what are invalid characters.
For example, I want to set the date _and time_ in my build number. Are colons valid characters?
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93
* Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7
* Content: [Run (build) number - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=classic)
* Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/run-number.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
what exactly is the set of invalid characters for a build number in this documentation with respect to the token build definitionname the following note is provided in the table note the pipeline name must not contain invalid or whitespace characters but this article fails to define what are invalid characters for example i want to set the date and time in my build number are colons valid characters document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
14,924
| 18,359,529,127
|
IssuesEvent
|
2021-10-09 01:45:46
|
DevExpress/testcafe-hammerhead
|
https://api.github.com/repos/DevExpress/testcafe-hammerhead
|
closed
|
Testcafe version of browser.waitUntil and browser.execute
|
TYPE: bug AREA: client FREQUENCY: level 1 SYSTEM: client side processing STATE: Stale
|
<!--
If you have all reproduction steps with a complete sample app, please share as many details as possible in the sections below.
Make sure that you tried using the latest TestCafe version (https://github.com/DevExpress/testcafe/releases), where this behavior might have been already addressed.
Before submitting an issue, please check CONTRIBUTING.md and existing issues in this repository (https://github.com/DevExpress/testcafe/issues) in case a similar issue exists or was already addressed. This may save your time (and ours).
-->
### What is your Test Scenario?
Replicate the following code written in Selenium to TestCafe framework.
```
browser.waitUntil(
() => {
return browser.execute(() => {
return (
navigator &&
navigator.serviceWorker &&
navigator.serviceWorker.controller &&
navigator.serviceWorker.controller.state &&
// We want to ensure the Service Worker is fully activated,
// other states don't mean success.
navigator.serviceWorker.controller.state === 'activated'
);
}).value;
},
10000,
'Service Worker registration failed!'
);
```
### What is the Current behavior?
Can't figure out how to write the above code using ClientFunctions
### What is the Expected behavior?
Use ClientFunctions to test for Service Workers
### Your Environment details:
* testcafe version: 1.2.1
* node.js version: v10.15.3
* command-line arguments: testcafe chrome -d --skip-js-errors
* browser name and version: Chrome
* platform and version: macOS 10.14
|
1.0
|
Testcafe version of browser.waitUntil and browser.execute - <!--
If you have all reproduction steps with a complete sample app, please share as many details as possible in the sections below.
Make sure that you tried using the latest TestCafe version (https://github.com/DevExpress/testcafe/releases), where this behavior might have been already addressed.
Before submitting an issue, please check CONTRIBUTING.md and existing issues in this repository (https://github.com/DevExpress/testcafe/issues) in case a similar issue exists or was already addressed. This may save your time (and ours).
-->
### What is your Test Scenario?
Replicate the following code written in Selenium to TestCafe framework.
```
browser.waitUntil(
() => {
return browser.execute(() => {
return (
navigator &&
navigator.serviceWorker &&
navigator.serviceWorker.controller &&
navigator.serviceWorker.controller.state &&
// We want to ensure the Service Worker is fully activated,
// other states don't mean success.
navigator.serviceWorker.controller.state === 'activated'
);
}).value;
},
10000,
'Service Worker registration failed!'
);
```
### What is the Current behavior?
Can't figure out how to write the above code using ClientFunctions
### What is the Expected behavior?
Use ClientFunctions to test for Service Workers
### Your Environment details:
* testcafe version: 1.2.1
* node.js version: v10.15.3
* command-line arguments: testcafe chrome -d --skip-js-errors
* browser name and version: Chrome
* platform and version: macOS 10.14
|
process
|
testcafe version of browser waituntil and browser execute if you have all reproduction steps with a complete sample app please share as many details as possible in the sections below make sure that you tried using the latest testcafe version where this behavior might have been already addressed before submitting an issue please check contributing md and existing issues in this repository in case a similar issue exists or was already addressed this may save your time and ours what is your test scenario replicate the following code written in selenium to testcafe framework browser waituntil return browser execute return navigator navigator serviceworker navigator serviceworker controller navigator serviceworker controller state we want to ensure the service worker is fully activated other states don t mean success navigator serviceworker controller state activated value service worker registration failed what is the current behavior can t figure out how to write the above code using clientfunctions what is the expected behavior use clientfunctions to test for service workers your environment details testcafe version node js version command line arguments testcafe chrome d skip js errors browser name and version chrome platform and version macos
| 1
|
303,875
| 23,043,226,940
|
IssuesEvent
|
2022-07-23 13:35:00
|
numpy/numpy
|
https://api.github.com/repos/numpy/numpy
|
opened
|
DOC: einsum examples of diagonalising and writeable view
|
04 - Documentation
|
### Issue with current documentation:
There are two issues I think worth expanding the docs with.
### 1) Diagonalising vectors in tensors
One can diagonalise a simple vector in variety of ways;
```python
>>> a = np.array([1, 2, 3])
>>> a * np.eye(3). # broadcasting
>>> np.multiply(a, np.eye(3)). # method broadcasting
>>> np.diag(a). # targeted method
>>> np.einsum('i, ij -> ij', a, np.eye(3)). # explicit einsum
```
I think it is useful to document `einsum` can do this since it can then be used for larger tensors where broadcasting does not.
My use case was to convert a set of 1d samples to a set of 2d-diagonal samples,
```python
>>> 1d_samples = np.array([1, 2, 3], [10, 20, 30]])
>>> 2d_samples = np.einsum("ki, ij -> kij", 1d_samples, np.eye(3))
>>> 2d_samples[1, :, :]
array([[[ 1., 0., 0.],
[ 0., 2., 0.],
[ 0., 0., 3.]],
[[10., 0., 0.],
[ 0., 20., 0.],
[ 0., 0., 30.]]])
```
### 2) Writable views
In the docs it is written that, for example as of NumPy 1.10
> np.einsum('ii->i', a) will return a writeable view of the diagonal of a 2D array.
I tried to use this the other and couldn't see how it was writeable to the original array. And thus, if the `return` is just a new ndarray isn't it expected that that array should be writeable as per any other array? I find this a confusing part of the docs if the meaning is just to say an array is returned.
### Idea or request for content:
as above
|
1.0
|
DOC: einsum examples of diagonalising and writeable view - ### Issue with current documentation:
There are two issues I think worth expanding the docs with.
### 1) Diagonalising vectors in tensors
One can diagonalise a simple vector in variety of ways;
```python
>>> a = np.array([1, 2, 3])
>>> a * np.eye(3). # broadcasting
>>> np.multiply(a, np.eye(3)). # method broadcasting
>>> np.diag(a). # targeted method
>>> np.einsum('i, ij -> ij', a, np.eye(3)). # explicit einsum
```
I think it is useful to document `einsum` can do this since it can then be used for larger tensors where broadcasting does not.
My use case was to convert a set of 1d samples to a set of 2d-diagonal samples,
```python
>>> 1d_samples = np.array([1, 2, 3], [10, 20, 30]])
>>> 2d_samples = np.einsum("ki, ij -> kij", 1d_samples, np.eye(3))
>>> 2d_samples[1, :, :]
array([[[ 1., 0., 0.],
[ 0., 2., 0.],
[ 0., 0., 3.]],
[[10., 0., 0.],
[ 0., 20., 0.],
[ 0., 0., 30.]]])
```
### 2) Writable views
In the docs it is written that, for example as of NumPy 1.10
> np.einsum('ii->i', a) will return a writeable view of the diagonal of a 2D array.
I tried to use this the other and couldn't see how it was writeable to the original array. And thus, if the `return` is just a new ndarray isn't it expected that that array should be writeable as per any other array? I find this a confusing part of the docs if the meaning is just to say an array is returned.
### Idea or request for content:
as above
|
non_process
|
doc einsum examples of diagonalising and writeable view issue with current documentation there are two issues i think worth expanding the docs with diagonalising vectors in tensors one can diagonalise a simple vector in variety of ways python a np array a np eye broadcasting np multiply a np eye method broadcasting np diag a targeted method np einsum i ij ij a np eye explicit einsum i think it is useful to document einsum can do this since it can then be used for larger tensors where broadcasting does not my use case was to convert a set of samples to a set of diagonal samples python samples np array samples np einsum ki ij kij samples np eye samples array writable views in the docs it is written that for example as of numpy np einsum ii i a will return a writeable view of the diagonal of a array i tried to use this the other and couldn t see how it was writeable to the original array and thus if the return is just a new ndarray isn t it expected that that array should be writeable as per any other array i find this a confusing part of the docs if the meaning is just to say an array is returned idea or request for content as above
| 0
|
10,595
| 13,402,035,804
|
IssuesEvent
|
2020-09-03 18:17:51
|
panther-labs/panther
|
https://api.github.com/repos/panther-labs/panther
|
opened
|
Gluesync tool should search to update partitions at dates before the table creation time
|
enhancement p0 p1 team:data processing
|
### Describe the bug
> **NOTE**: A quickfix for this should be included in the 1.8 to avoid surprises during upgrades
Currently `gluesync` needs a specific date in the past to sync partitions at dates before the Glue table creation time.
This is problematic because historical data and events from misconfigured devices can put partitions at arbitrary points in the past. Especially in 1.8 release we added the `source_id` and `source_label` fields to *all tables*, which requires all Glue tables to be updated and all partitions synced. For tables that have data at some point in the past of table creation time this **will break queries** after deployment and will require manual intervention to fix. Specifically the person deploying should find out which is the oldest date they have data for in the problematic tables and run `gluesync` for these with the appropriate start date.
### Expected behavior
Ideally the deployment should handle the migration transparently to the user
### Proposed fix
After a short discussion with @rleighton we decided on adding a query to the Glue API for table partitions *before table creation time* and also syncing those during a table sync. This will add some delay to the tool but it will ensure that all partitions of the table will be updated to the latest schema.
|
1.0
|
Gluesync tool should search to update partitions at dates before the table creation time - ### Describe the bug
> **NOTE**: A quickfix for this should be included in the 1.8 to avoid surprises during upgrades
Currently `gluesync` needs a specific date in the past to sync partitions at dates before the Glue table creation time.
This is problematic because historical data and events from misconfigured devices can put partitions at arbitrary points in the past. Especially in 1.8 release we added the `source_id` and `source_label` fields to *all tables*, which requires all Glue tables to be updated and all partitions synced. For tables that have data at some point in the past of table creation time this **will break queries** after deployment and will require manual intervention to fix. Specifically the person deploying should find out which is the oldest date they have data for in the problematic tables and run `gluesync` for these with the appropriate start date.
### Expected behavior
Ideally the deployment should handle the migration transparently to the user
### Proposed fix
After a short discussion with @rleighton we decided on adding a query to the Glue API for table partitions *before table creation time* and also syncing those during a table sync. This will add some delay to the tool but it will ensure that all partitions of the table will be updated to the latest schema.
|
process
|
gluesync tool should search to update partitions at dates before the table creation time describe the bug note a quickfix for this should be included in the to avoid surprises during upgrades currently gluesync needs a specific date in the past to sync partitions at dates before the glue table creation time this is problematic because historical data and events from misconfigured devices can put partitions at arbitrary points in the past especially in release we added the source id and source label fields to all tables which requires all glue tables to be updated and all partitions synced for tables that have data at some point in the past of table creation time this will break queries after deployment and will require manual intervention to fix specifically the person deploying should find out which is the oldest date they have data for in the problematic tables and run gluesync for these with the appropriate start date expected behavior ideally the deployment should handle the migration transparently to the user proposed fix after a short discussion with rleighton we decided on adding a query to the glue api for table partitions before table creation time and also syncing those during a table sync this will add some delay to the tool but it will ensure that all partitions of the table will be updated to the latest schema
| 1
|
18,932
| 24,886,933,328
|
IssuesEvent
|
2022-10-28 08:35:53
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
opened
|
Error: [libs/dml/src/render/render_configuration.rs:58:54] called `Option::unwrap()` on a `None` value
|
bug/1-unconfirmed kind/bug process/candidate tech/engines/introspection engine topic: error reporting team/schema
|
<!-- If required, please update the title to be clear and descriptive -->
Command: `prisma db pull`
Version: `4.4.0`
Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6`
Report: https://prisma-errors.netlify.app/report/14391
OS: `x64 darwin 21.6.0`
```
|
1.0
|
Error: [libs/dml/src/render/render_configuration.rs:58:54] called `Option::unwrap()` on a `None` value - <!-- If required, please update the title to be clear and descriptive -->
Command: `prisma db pull`
Version: `4.4.0`
Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6`
Report: https://prisma-errors.netlify.app/report/14391
OS: `x64 darwin 21.6.0`
```
|
process
|
error called option unwrap on a none value command prisma db pull version binary version report os darwin
| 1
|
9,325
| 12,338,986,459
|
IssuesEvent
|
2020-05-14 17:20:55
|
unicode-org/icu4x
|
https://api.github.com/repos/unicode-org/icu4x
|
closed
|
Consider using the RACI framework for responsibility assignment
|
C-process T-docs
|
RACI is described here: https://en.wikipedia.org/wiki/Responsibility_assignment_matrix
tl;dr: this allows a distinction between people who are *accountable* for the outcome, and people who actually do the work (can be the same as accountable, but not necessary); as well as calling out explicitly the folks who can provide info vs people who are informed only.
https://github.com/unicode-org/icu4x/blob/master/docs/triaging.md#assignee talks about a "champion". This would be "accountable". Using prior art in responsibility assingment allows us not to spend time reinventing that.
|
1.0
|
Consider using the RACI framework for responsibility assignment - RACI is described here: https://en.wikipedia.org/wiki/Responsibility_assignment_matrix
tl;dr: this allows a distinction between people who are *accountable* for the outcome, and people who actually do the work (can be the same as accountable, but not necessary); as well as calling out explicitly the folks who can provide info vs people who are informed only.
https://github.com/unicode-org/icu4x/blob/master/docs/triaging.md#assignee talks about a "champion". This would be "accountable". Using prior art in responsibility assingment allows us not to spend time reinventing that.
|
process
|
consider using the raci framework for responsibility assignment raci is described here tl dr this allows a distinction between people who are accountable for the outcome and people who actually do the work can be the same as accountable but not necessary as well as calling out explicitly the folks who can provide info vs people who are informed only talks about a champion this would be accountable using prior art in responsibility assingment allows us not to spend time reinventing that
| 1
|
16,307
| 11,907,355,290
|
IssuesEvent
|
2020-03-30 22:07:07
|
APSIMInitiative/ApsimX
|
https://api.github.com/repos/APSIMInitiative/ApsimX
|
closed
|
Soil downloader doesn't work in German locale
|
bug interface/infrastructure
|
Attempts to use the soil download in a German locale fail because the latitude and longitude values in the request uri get written in a culture-sensitive manner (using a comma as decimal place separator):
```
http://www.asris.csiro.au/ASRISApi/api/APSIM/getClosestApsoil?maxCnt=5&longitude=151,26&latitude=-27,18
```
|
1.0
|
Soil downloader doesn't work in German locale - Attempts to use the soil download in a German locale fail because the latitude and longitude values in the request uri get written in a culture-sensitive manner (using a comma as decimal place separator):
```
http://www.asris.csiro.au/ASRISApi/api/APSIM/getClosestApsoil?maxCnt=5&longitude=151,26&latitude=-27,18
```
|
non_process
|
soil downloader doesn t work in german locale attempts to use the soil download in a german locale fail because the latitude and longitude values in the request uri get written in a culture sensitive manner using a comma as decimal place separator
| 0
|
8,002
| 11,196,615,064
|
IssuesEvent
|
2020-01-03 10:37:22
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
`Clip raster by extension` batch mode: no path for output choosable
|
Bug Feedback High Priority Processing Regression
|
In the ``Clip raster by extent`` of gdal, in the **batch mode** it is not possible to choose an output for the raster files (and having the nice *Fill automatically with... (parameter, numbers..)* options. It seems to be just an empty string. With the ``Clip raster by mask`` always gdal, all seems working fine:
1. Clip raster by extent (not working)

2. Clip raster by mask (working)

QGIS 3.9 on Linux Debian Sid
|
1.0
|
`Clip raster by extension` batch mode: no path for output choosable - In the ``Clip raster by extent`` of gdal, in the **batch mode** it is not possible to choose an output for the raster files (and having the nice *Fill automatically with... (parameter, numbers..)* options. It seems to be just an empty string. With the ``Clip raster by mask`` always gdal, all seems working fine:
1. Clip raster by extent (not working)

2. Clip raster by mask (working)

QGIS 3.9 on Linux Debian Sid
|
process
|
clip raster by extension batch mode no path for output choosable in the clip raster by extent of gdal in the batch mode it is not possible to choose an output for the raster files and having the nice fill automatically with parameter numbers options it seems to be just an empty string with the clip raster by mask always gdal all seems working fine clip raster by extent not working clip raster by mask working qgis on linux debian sid
| 1
|
143,151
| 13,055,525,550
|
IssuesEvent
|
2020-07-30 01:57:20
|
miguelangelsoria/Reclamaciones
|
https://api.github.com/repos/miguelangelsoria/Reclamaciones
|
reopened
|
Actualización a base de datos
|
documentation
|
SE cambia el nombre de la tabla Trabajadores por el de Registros
|
1.0
|
Actualización a base de datos - SE cambia el nombre de la tabla Trabajadores por el de Registros
|
non_process
|
actualización a base de datos se cambia el nombre de la tabla trabajadores por el de registros
| 0
|
5,735
| 8,577,758,676
|
IssuesEvent
|
2018-11-13 01:33:12
|
googleapis/cloud-trace-nodejs
|
https://api.github.com/repos/googleapis/cloud-trace-nodejs
|
closed
|
Move samples from docs-samples
|
type: process
|
Hi there, our docs-samples repo still have some [trace samples](https://github.com/GoogleCloudPlatform/nodejs-docs-samples/tree/master/trace). For Node.js, all samples live with their client library if they have one. Can you move the samples over please. Thanks! cc @ofrobots
|
1.0
|
Move samples from docs-samples - Hi there, our docs-samples repo still have some [trace samples](https://github.com/GoogleCloudPlatform/nodejs-docs-samples/tree/master/trace). For Node.js, all samples live with their client library if they have one. Can you move the samples over please. Thanks! cc @ofrobots
|
process
|
move samples from docs samples hi there our docs samples repo still have some for node js all samples live with their client library if they have one can you move the samples over please thanks cc ofrobots
| 1
|
36,852
| 8,167,589,516
|
IssuesEvent
|
2018-08-26 00:57:07
|
idaholab/raven
|
https://api.github.com/repos/idaholab/raven
|
opened
|
codecoverage is months old
|
defect devel master priority_critical
|
--------
Issue Description
--------
##### What did you expect to see happen?
The code coverage script to be functional.
The code coverage is months old since the check_coverage.sh is broken.
This is causing the master branch to be red since a while.
##### What did you see instead?
The code coverage script is currently broken
##### Do you have a suggested fix for the development team?
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or improvement?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
|
1.0
|
codecoverage is months old - --------
Issue Description
--------
##### What did you expect to see happen?
The code coverage script to be functional.
The code coverage is months old since the check_coverage.sh is broken.
This is causing the master branch to be red since a while.
##### What did you see instead?
The code coverage script is currently broken
##### Do you have a suggested fix for the development team?
##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue.
----------------
For Change Control Board: Issue Review
----------------
This review should occur before any development is performed as a response to this issue.
- [ ] 1. Is it tagged with a type: defect or improvement?
- [ ] 2. Is it tagged with a priority: critical, normal or minor?
- [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements?
- [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users.
- [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.)
-------
For Change Control Board: Issue Closure
-------
This review should occur when the issue is imminently going to be closed.
- [ ] 1. If the issue is a defect, is the defect fixed?
- [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.)
- [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)?
- [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)?
- [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
|
non_process
|
codecoverage is months old issue description what did you expect to see happen the code coverage script to be functional the code coverage is months old since the check coverage sh is broken this is causing the master branch to be red since a while what did you see instead the code coverage script is currently broken do you have a suggested fix for the development team please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
| 0
|
63,643
| 26,478,389,186
|
IssuesEvent
|
2023-01-17 12:58:43
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Upgrade Proof URL Migration
|
Team:AppServicesSv Epic:KBNA-59
|
Currently Kibana URLs can break when our users migrate across minor versions. This can happen for various reasons, including: (1) because app teams change the URL structure, (2) because the ID of a saved object that was referenced in the URL changed; (3) because data structure, say global filters, has changed its shape.
The scope of this project is to identify actual and possible causes why Kibana URLs might break across minor upgrades, followed by implementation of guidelines, best practices and possibly utility libraries which will help app teams ensure their URLs are upgrade proof.
|
1.0
|
Upgrade Proof URL Migration - Currently Kibana URLs can break when our users migrate across minor versions. This can happen for various reasons, including: (1) because app teams change the URL structure, (2) because the ID of a saved object that was referenced in the URL changed; (3) because data structure, say global filters, has changed its shape.
The scope of this project is to identify actual and possible causes why Kibana URLs might break across minor upgrades, followed by implementation of guidelines, best practices and possibly utility libraries which will help app teams ensure their URLs are upgrade proof.
|
non_process
|
upgrade proof url migration currently kibana urls can break when our users migrate across minor versions this can happen for various reasons including because app teams change the url structure because the id of a saved object that was referenced in the url changed because data structure say global filters has changed its shape the scope of this project is to identify actual and possible causes why kibana urls might break across minor upgrades followed by implementation of guidelines best practices and possibly utility libraries which will help app teams ensure their urls are upgrade proof
| 0
|
20,988
| 11,572,818,480
|
IssuesEvent
|
2020-02-21 01:19:35
|
Azure/azure-cli
|
https://api.github.com/repos/Azure/azure-cli
|
closed
|
Could not enable aks dashboard on aks cluster
|
AKS Service Attention
|
## Describe the bug
Was trying to enable the kubernetes dashboard on an existing aks cluster deployed through Terraform.
**Command Name**
`az aks enable-addons`
**Errors:**
```
argument of type 'NoneType' is not iterable
Traceback (most recent call last):
Temp\pip-install-0u_uiwkq\knack\knack\cli.py, ln 206, in invoke
azure\cli\core\commands\__init__.py, ln 608, in execute
azure\cli\core\commands\__init__.py, ln 666, in _run_jobs_serially
azure\cli\core\commands\__init__.py, ln 659, in _run_job
Local\Temp\pip-install-0u_uiwkq\six\six.py, ln 696, in reraise
azure\cli\core\commands\__init__.py, ln 636, in _run_job
azure\cli\core\commands\__init__.py, ln 306, in __call__
azure-cli-core\azure\cli\core\__init__.py, ln 493, in default_command_handler
azure\cli\command_modules\acs\custom.py, ln 1844, in aks_enable_addons
azure\cli\command_modules\acs\custom.py, ln 2487, in _ensure_container_insights_for_monitoring
TypeError: argument of type 'NoneType' is not iterable
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az aks enable-addons --addons {} -n {} -g {}`
## Expected Behavior
## Environment Summary
```
Windows-10-10.0.18362-SP0
Python 3.6.6
Shell: powershell.exe
azure-cli 2.0.80 *
Extensions:
azure-devops 0.17.0
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
|
1.0
|
Could not enable aks dashboard on aks cluster - ## Describe the bug
Was trying to enable the kubernetes dashboard on an existing aks cluster deployed through Terraform.
**Command Name**
`az aks enable-addons`
**Errors:**
```
argument of type 'NoneType' is not iterable
Traceback (most recent call last):
Temp\pip-install-0u_uiwkq\knack\knack\cli.py, ln 206, in invoke
azure\cli\core\commands\__init__.py, ln 608, in execute
azure\cli\core\commands\__init__.py, ln 666, in _run_jobs_serially
azure\cli\core\commands\__init__.py, ln 659, in _run_job
Local\Temp\pip-install-0u_uiwkq\six\six.py, ln 696, in reraise
azure\cli\core\commands\__init__.py, ln 636, in _run_job
azure\cli\core\commands\__init__.py, ln 306, in __call__
azure-cli-core\azure\cli\core\__init__.py, ln 493, in default_command_handler
azure\cli\command_modules\acs\custom.py, ln 1844, in aks_enable_addons
azure\cli\command_modules\acs\custom.py, ln 2487, in _ensure_container_insights_for_monitoring
TypeError: argument of type 'NoneType' is not iterable
```
## To Reproduce:
Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information.
- _Put any pre-requisite steps here..._
- `az aks enable-addons --addons {} -n {} -g {}`
## Expected Behavior
## Environment Summary
```
Windows-10-10.0.18362-SP0
Python 3.6.6
Shell: powershell.exe
azure-cli 2.0.80 *
Extensions:
azure-devops 0.17.0
```
## Additional Context
<!--Please don't remove this:-->
<!--auto-generated-->
|
non_process
|
could not enable aks dashboard on aks cluster describe the bug was trying to enable the kubernetes dashboard on an existing aks cluster deployed through terraform command name az aks enable addons errors argument of type nonetype is not iterable traceback most recent call last temp pip install uiwkq knack knack cli py ln in invoke azure cli core commands init py ln in execute azure cli core commands init py ln in run jobs serially azure cli core commands init py ln in run job local temp pip install uiwkq six six py ln in reraise azure cli core commands init py ln in run job azure cli core commands init py ln in call azure cli core azure cli core init py ln in default command handler azure cli command modules acs custom py ln in aks enable addons azure cli command modules acs custom py ln in ensure container insights for monitoring typeerror argument of type nonetype is not iterable to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az aks enable addons addons n g expected behavior environment summary windows python shell powershell exe azure cli extensions azure devops additional context
| 0
|
36,102
| 7,860,889,754
|
IssuesEvent
|
2018-06-21 21:34:40
|
StrikeNP/trac_test
|
https://api.github.com/repos/StrikeNP/trac_test
|
closed
|
TRAC appears to convert character sets incorrectly (Trac #6)
|
Migrated from Trac defect senkbeil@uwm.edu utilities
|
Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.
Attachments:
[plot_explicit_ta_configs.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_explicit_ta_configs.maff)
[plot_new_pdf_config_1_plot_2.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_config_1_plot_2.maff)
[plot_combo_pdf_run_3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_combo_pdf_run_3.maff)
[plot_input_fields_rtp3_thlp3_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_input_fields_rtp3_thlp3_1.maff)
[plot_new_pdf_20180522_test_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_20180522_test_1.maff)
[plot_attempts_8_10.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempts_8_10.maff)
[plot_attempt_8_only.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempt_8_only.maff)
[plot_beta_1p3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3.maff)
[plot_beta_1p3_all.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3_all.maff)
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/6
```json
{
"status": "closed",
"changetime": "2009-05-13T18:15:45",
"description": "Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.",
"reporter": "dschanen@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1242238545000000",
"component": "utilities",
"summary": "TRAC appears to convert character sets incorrectly",
"priority": "minor",
"keywords": "",
"time": "2009-05-04T20:44:26",
"milestone": "",
"owner": "senkbeil@uwm.edu",
"type": "defect"
}
```
|
1.0
|
TRAC appears to convert character sets incorrectly (Trac #6) - Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.
Attachments:
[plot_explicit_ta_configs.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_explicit_ta_configs.maff)
[plot_new_pdf_config_1_plot_2.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_config_1_plot_2.maff)
[plot_combo_pdf_run_3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_combo_pdf_run_3.maff)
[plot_input_fields_rtp3_thlp3_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_input_fields_rtp3_thlp3_1.maff)
[plot_new_pdf_20180522_test_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_20180522_test_1.maff)
[plot_attempts_8_10.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempts_8_10.maff)
[plot_attempt_8_only.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempt_8_only.maff)
[plot_beta_1p3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3.maff)
[plot_beta_1p3_all.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3_all.maff)
Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/6
```json
{
"status": "closed",
"changetime": "2009-05-13T18:15:45",
"description": "Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.",
"reporter": "dschanen@uwm.edu",
"cc": "",
"resolution": "Verified by V. Larson",
"_ts": "1242238545000000",
"component": "utilities",
"summary": "TRAC appears to convert character sets incorrectly",
"priority": "minor",
"keywords": "",
"time": "2009-05-04T20:44:26",
"milestone": "",
"owner": "senkbeil@uwm.edu",
"type": "defect"
}
```
|
non_process
|
trac appears to convert character sets incorrectly trac some of our source files e g microphys driver contain some non ascii characters according to the trac website it uses unicode for character encoding so i think these should display properly but they do not attachments migrated from json status closed changetime description some of our source files e g microphys driver contain some non ascii characters according to the trac website it uses unicode for character encoding so i think these should display properly but they do not reporter dschanen uwm edu cc resolution verified by v larson ts component utilities summary trac appears to convert character sets incorrectly priority minor keywords time milestone owner senkbeil uwm edu type defect
| 0
|
21,160
| 3,466,366,909
|
IssuesEvent
|
2015-12-22 02:49:09
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
CakePHP 3.1: cakephp-plugins.php has the wrong path?
|
Defect
|
<?php
$baseDir = dirname(dirname(__FILE__));
return [
'plugins' => [
'ADmad/HybridAuth' => 'C:\xampp\htdocs\mysite/vendor/admad/cakephp-hybridauth/',
'Admin' => 'C:\xampp\htdocs\mysite/plugins/Admin/'
]
]
After I changed the name of the project folder, I received a number of errors that occur that the plugin does not exist. The problem is that the paths is written as above, instead:
'ADmad/HybridAuth' => $baseDir . '/vendor/admad/cakephp-hybridauth/',
'Admin' => $baseDir . '/plugins/Admin/',
|
1.0
|
CakePHP 3.1: cakephp-plugins.php has the wrong path? -
<?php
$baseDir = dirname(dirname(__FILE__));
return [
'plugins' => [
'ADmad/HybridAuth' => 'C:\xampp\htdocs\mysite/vendor/admad/cakephp-hybridauth/',
'Admin' => 'C:\xampp\htdocs\mysite/plugins/Admin/'
]
]
After I changed the name of the project folder, I received a number of errors that occur that the plugin does not exist. The problem is that the paths is written as above, instead:
'ADmad/HybridAuth' => $baseDir . '/vendor/admad/cakephp-hybridauth/',
'Admin' => $baseDir . '/plugins/Admin/',
|
non_process
|
cakephp cakephp plugins php has the wrong path php basedir dirname dirname file return plugins admad hybridauth c xampp htdocs mysite vendor admad cakephp hybridauth admin c xampp htdocs mysite plugins admin after i changed the name of the project folder i received a number of errors that occur that the plugin does not exist the problem is that the paths is written as above instead admad hybridauth basedir vendor admad cakephp hybridauth admin basedir plugins admin
| 0
|
16,510
| 21,519,225,157
|
IssuesEvent
|
2022-04-28 12:52:15
|
camunda/zeebe
|
https://api.github.com/repos/camunda/zeebe
|
closed
|
Job of cancelled instance can be activated if an error was thrown on it
|
kind/bug scope/broker severity/high area/reliability team/process-automation
|
**Describe the Bug**
A job can still be activated (that is, Zeebe will return it as part of an ActivateJobsResponse) even after its associated process instance has been cancelled, if an error was previously thrown on that job.
**To Reproduce**
@korthout's method is easier than mine, especially for Camunda folks:
> I was also able to reproduce it through zbctl:
>
> ```
> zbctl deploy ./one_task.bpmn
> sleep 1
> zbctl create instance one_task
> zbctl activate jobs do_something
> zbctl throwError job 2251799813685256 --errorCode FOO --errorMessage foobar
> sleep 1
> zbctl cancel instance 2251799813685251
> sleep 1
> zbctl activate jobs do_something
> ```
>
> This activated the job and returned:
>
> ```
> {
> "jobs": [
> {
> "key": "2251799813685256",
> "type": "do_something",
> "processInstanceKey": "2251799813685251",
> "bpmnProcessId": "one_task",
> "processDefinitionVersion": 1,
> "processDefinitionKey": "2251799813685249",
> "elementId": "do_something",
> "elementInstanceKey": "2251799813685255",
> "customHeaders": "{\"what_to_do\":\"nothing\"}",
> "worker": "zbctl",
> "retries": 2,
> "deadline": "1642166365427",
> "variables": "{}"
> }
> ]
> }
> ```
<details>
<summary>My original method</summary>
ZIP file: [reproduce_bug.zip](https://github.com/camunda-cloud/zeebe/files/7866658/reproduce_bug.zip)
1. Ensure you have ruby installed, as well as the [zeebe_client ruby gem](https://github.com/zeebe-io/zeebe-client-ruby).
2. Download the attached ZIP file and expand it to reveal the `reproduce_bug.rb` and `one_task.bpmn` files, and put them in a directory together.
3. Ensure you have Zeebe running. If you need to change Zeebe's address away from `localhost:26500`, edit the `reproduce_bug.rb` file.
4. Run `ruby reproduce_bug.rb` and observe the output.
The script will deploy the included BPMN, create an instance of that process, and activate jobs for that process instance. If the jobs that Zeebe returns are for a running process instance, it will throw an error on them. It will then cancel that process instance and loop around and create another instance. If the jobs that it receives from Zeebe are from a process instance that was already successfully cancelled, it will write a line of output including "[!!! BUG !!!]". The script will exit after 5 occurrences of the bug or 50 loops, whichever happens first.
</details>
**Expected Behavior**
After Zeebe successfully cancels a process instance, no jobs from that instance should ever be returned to a future ActivateJobs request.
**Actual Behavior**
Jobs which previously had an error thrown on them before their process instance was cancelled are being returned to ActivateJobs requests that are placed after their process instance was cancelled.
**Log/Stacktrace**
<details>
<summary>My original method</summary>
Here is the output of a sample run of the `reproduce_bug.rb` script:
```
$ ruby reproduce_bug.rb
deploying process definition from /usr/src/app/one_task.bpmn
deployed process `one_task` with key 2251799813685249
created instance 2251799813685423
activated job 2251799813685429 for process instance 2251799813685423
threw error on job 2251799813685429
cancelled instance 2251799813685423
created instance 2251799813685432
[!!! BUG !!!] activated job 2251799813685429 for cancelled process instance 2251799813685423 [!!! BUG !!!]
cancelled instance 2251799813685432
created instance 2251799813685439
activated job 2251799813685445 for process instance 2251799813685439
threw error on job 2251799813685445
cancelled instance 2251799813685439
created instance 2251799813685448
[!!! BUG !!!] activated job 2251799813685445 for cancelled process instance 2251799813685439 [!!! BUG !!!]
cancelled instance 2251799813685448
created instance 2251799813685455
activated job 2251799813685461 for process instance 2251799813685455
threw error on job 2251799813685461
cancelled instance 2251799813685455
created instance 2251799813685464
[!!! BUG !!!] activated job 2251799813685461 for cancelled process instance 2251799813685455 [!!! BUG !!!]
cancelled instance 2251799813685464
created instance 2251799813685471
activated job 2251799813685477 for process instance 2251799813685471
threw error on job 2251799813685477
cancelled instance 2251799813685471
created instance 2251799813685480
[!!! BUG !!!] activated job 2251799813685429 for cancelled process instance 2251799813685423 [!!! BUG !!!]
[!!! BUG !!!] activated job 2251799813685445 for cancelled process instance 2251799813685439 [!!! BUG !!!]
[!!! BUG !!!] activated job 2251799813685477 for cancelled process instance 2251799813685471 [!!! BUG !!!]
cancelled instance 2251799813685480
```
</details>
**Environment:**
- OS: Debian GNU/Linux 10 in a Docker container
- Zeebe Version: 1.2.9 and 1.3.0 (verified against both)
- Configuration: zeebe-kafka-exporter v3.0.0
|
1.0
|
Job of cancelled instance can be activated if an error was thrown on it - **Describe the Bug**
A job can still be activated (that is, Zeebe will return it as part of an ActivateJobsResponse) even after its associated process instance has been cancelled, if an error was previously thrown on that job.
**To Reproduce**
@korthout's method is easier than mine, especially for Camunda folks:
> I was also able to reproduce it through zbctl:
>
> ```
> zbctl deploy ./one_task.bpmn
> sleep 1
> zbctl create instance one_task
> zbctl activate jobs do_something
> zbctl throwError job 2251799813685256 --errorCode FOO --errorMessage foobar
> sleep 1
> zbctl cancel instance 2251799813685251
> sleep 1
> zbctl activate jobs do_something
> ```
>
> This activated the job and returned:
>
> ```
> {
> "jobs": [
> {
> "key": "2251799813685256",
> "type": "do_something",
> "processInstanceKey": "2251799813685251",
> "bpmnProcessId": "one_task",
> "processDefinitionVersion": 1,
> "processDefinitionKey": "2251799813685249",
> "elementId": "do_something",
> "elementInstanceKey": "2251799813685255",
> "customHeaders": "{\"what_to_do\":\"nothing\"}",
> "worker": "zbctl",
> "retries": 2,
> "deadline": "1642166365427",
> "variables": "{}"
> }
> ]
> }
> ```
<details>
<summary>My original method</summary>
ZIP file: [reproduce_bug.zip](https://github.com/camunda-cloud/zeebe/files/7866658/reproduce_bug.zip)
1. Ensure you have ruby installed, as well as the [zeebe_client ruby gem](https://github.com/zeebe-io/zeebe-client-ruby).
2. Download the attached ZIP file and expand it to reveal the `reproduce_bug.rb` and `one_task.bpmn` files, and put them in a directory together.
3. Ensure you have Zeebe running. If you need to change Zeebe's address away from `localhost:26500`, edit the `reproduce_bug.rb` file.
4. Run `ruby reproduce_bug.rb` and observe the output.
The script will deploy the included BPMN, create an instance of that process, and activate jobs for that process instance. If the jobs that Zeebe returns are for a running process instance, it will throw an error on them. It will then cancel that process instance and loop around and create another instance. If the jobs that it receives from Zeebe are from a process instance that was already successfully cancelled, it will write a line of output including "[!!! BUG !!!]". The script will exit after 5 occurrences of the bug or 50 loops, whichever happens first.
</details>
**Expected Behavior**
After Zeebe successfully cancels a process instance, no jobs from that instance should ever be returned to a future ActivateJobs request.
**Actual Behavior**
Jobs which previously had an error thrown on them before their process instance was cancelled are being returned to ActivateJobs requests that are placed after their process instance was cancelled.
**Log/Stacktrace**
<details>
<summary>My original method</summary>
Here is the output of a sample run of the `reproduce_bug.rb` script:
```
$ ruby reproduce_bug.rb
deploying process definition from /usr/src/app/one_task.bpmn
deployed process `one_task` with key 2251799813685249
created instance 2251799813685423
activated job 2251799813685429 for process instance 2251799813685423
threw error on job 2251799813685429
cancelled instance 2251799813685423
created instance 2251799813685432
[!!! BUG !!!] activated job 2251799813685429 for cancelled process instance 2251799813685423 [!!! BUG !!!]
cancelled instance 2251799813685432
created instance 2251799813685439
activated job 2251799813685445 for process instance 2251799813685439
threw error on job 2251799813685445
cancelled instance 2251799813685439
created instance 2251799813685448
[!!! BUG !!!] activated job 2251799813685445 for cancelled process instance 2251799813685439 [!!! BUG !!!]
cancelled instance 2251799813685448
created instance 2251799813685455
activated job 2251799813685461 for process instance 2251799813685455
threw error on job 2251799813685461
cancelled instance 2251799813685455
created instance 2251799813685464
[!!! BUG !!!] activated job 2251799813685461 for cancelled process instance 2251799813685455 [!!! BUG !!!]
cancelled instance 2251799813685464
created instance 2251799813685471
activated job 2251799813685477 for process instance 2251799813685471
threw error on job 2251799813685477
cancelled instance 2251799813685471
created instance 2251799813685480
[!!! BUG !!!] activated job 2251799813685429 for cancelled process instance 2251799813685423 [!!! BUG !!!]
[!!! BUG !!!] activated job 2251799813685445 for cancelled process instance 2251799813685439 [!!! BUG !!!]
[!!! BUG !!!] activated job 2251799813685477 for cancelled process instance 2251799813685471 [!!! BUG !!!]
cancelled instance 2251799813685480
```
</details>
**Environment:**
- OS: Debian GNU/Linux 10 in a Docker container
- Zeebe Version: 1.2.9 and 1.3.0 (verified against both)
- Configuration: zeebe-kafka-exporter v3.0.0
|
process
|
job of cancelled instance can be activated if an error was thrown on it describe the bug a job can still be activated that is zeebe will return it as part of an activatejobsresponse even after its associated process instance has been cancelled if an error was previously thrown on that job to reproduce korthout s method is easier than mine especially for camunda folks i was also able to reproduce it through zbctl zbctl deploy one task bpmn sleep zbctl create instance one task zbctl activate jobs do something zbctl throwerror job errorcode foo errormessage foobar sleep zbctl cancel instance sleep zbctl activate jobs do something this activated the job and returned jobs key type do something processinstancekey bpmnprocessid one task processdefinitionversion processdefinitionkey elementid do something elementinstancekey customheaders what to do nothing worker zbctl retries deadline variables my original method zip file ensure you have ruby installed as well as the download the attached zip file and expand it to reveal the reproduce bug rb and one task bpmn files and put them in a directory together ensure you have zeebe running if you need to change zeebe s address away from localhost edit the reproduce bug rb file run ruby reproduce bug rb and observe the output the script will deploy the included bpmn create an instance of that process and activate jobs for that process instance if the jobs that zeebe returns are for a running process instance it will throw an error on them it will then cancel that process instance and loop around and create another instance if the jobs that it receives from zeebe are from a process instance that was already successfully cancelled it will write a line of output including the script will exit after occurrences of the bug or loops whichever happens first expected behavior after zeebe successfully cancels a process instance no jobs from that instance should ever be returned to a future activatejobs request actual behavior jobs which previously had an error thrown on them before their process instance was cancelled are being returned to activatejobs requests that are placed after their process instance was cancelled log stacktrace my original method here is the output of a sample run of the reproduce bug rb script ruby reproduce bug rb deploying process definition from usr src app one task bpmn deployed process one task with key created instance activated job for process instance threw error on job cancelled instance created instance activated job for cancelled process instance cancelled instance created instance activated job for process instance threw error on job cancelled instance created instance activated job for cancelled process instance cancelled instance created instance activated job for process instance threw error on job cancelled instance created instance activated job for cancelled process instance cancelled instance created instance activated job for process instance threw error on job cancelled instance created instance activated job for cancelled process instance activated job for cancelled process instance activated job for cancelled process instance cancelled instance environment os debian gnu linux in a docker container zeebe version and verified against both configuration zeebe kafka exporter
| 1
|
19,723
| 26,073,834,701
|
IssuesEvent
|
2022-12-24 07:07:41
|
pyanodon/pybugreports
|
https://api.github.com/repos/pyanodon/pybugreports
|
closed
|
Incompatable with 248k Mod
|
mod:pypostprocessing crash compatibility
|
### Mod source
PyAE Beta
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [ ] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [X] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [X] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [X] Pypostprocessing failure
- [ ] Other
### What is the problem?
[factorio-current (1).log](https://github.com/pyanodon/pybugreports/files/9637768/factorio-current.1.log)
### Steps to reproduce
Load PYbeta and 248k mods
### Additional context
_No response_
### Log file
_No response_
|
1.0
|
Incompatable with 248k Mod - ### Mod source
PyAE Beta
### Which mod are you having an issue with?
- [ ] pyalienlife
- [ ] pyalternativeenergy
- [ ] pycoalprocessing
- [ ] pyfusionenergy
- [ ] pyhightech
- [ ] pyindustry
- [ ] pypetroleumhandling
- [X] pypostprocessing
- [ ] pyrawores
### Operating system
>=Windows 10
### What kind of issue is this?
- [X] Compatibility
- [ ] Locale (names, descriptions, unknown keys)
- [ ] Graphical
- [ ] Crash
- [ ] Progression
- [ ] Balance
- [X] Pypostprocessing failure
- [ ] Other
### What is the problem?
[factorio-current (1).log](https://github.com/pyanodon/pybugreports/files/9637768/factorio-current.1.log)
### Steps to reproduce
Load PYbeta and 248k mods
### Additional context
_No response_
### Log file
_No response_
|
process
|
incompatable with mod mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem steps to reproduce load pybeta and mods additional context no response log file no response
| 1
|
5,897
| 8,712,368,632
|
IssuesEvent
|
2018-12-06 22:00:28
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
RFC process.on("shutdown") proposal
|
feature request process
|
See this popular question on so https://stackoverflow.com/q/14031763/3748498
By now there is no clean API to do cleanup when node.js exits, in order to catch all situation you need to do following:
```js
for (const event of ["exit", "SIGINT", "SIGUSR1", "SIGUSR2", "uncaughtException", "SIGTERM"]) {
process.on(event, cleanUp)
}
```
This issue is a place for discussion how to fix it.
My proposal is `process.on("shutdown", cleanUp)`.
|
1.0
|
RFC process.on("shutdown") proposal - See this popular question on so https://stackoverflow.com/q/14031763/3748498
By now there is no clean API to do cleanup when node.js exits, in order to catch all situation you need to do following:
```js
for (const event of ["exit", "SIGINT", "SIGUSR1", "SIGUSR2", "uncaughtException", "SIGTERM"]) {
process.on(event, cleanUp)
}
```
This issue is a place for discussion how to fix it.
My proposal is `process.on("shutdown", cleanUp)`.
|
process
|
rfc process on shutdown proposal see this popular question on so by now there is no clean api to do cleanup when node js exits in order to catch all situation you need to do following js for const event of process on event cleanup this issue is a place for discussion how to fix it my proposal is process on shutdown cleanup
| 1
|
117,455
| 11,947,011,036
|
IssuesEvent
|
2020-04-03 09:10:18
|
marmelab/react-admin
|
https://api.github.com/repos/marmelab/react-admin
|
closed
|
Document how to override notifications messages
|
documentation enhancement needs more info old
|
This has already been asked before: #180 but the answer provided was to basically not use the Admin component. I believe this could be a useful feature. Can we discuss some alternative approaches to provide an easier to use and user friendly server messaging?
I propose to add support for a mapping file, similar to the way Translation messages are implemented, such that every error (or even every response) received from the server will be looked up in this server messages mapping file by the Admin on rest saga.
|
1.0
|
Document how to override notifications messages - This has already been asked before: #180 but the answer provided was to basically not use the Admin component. I believe this could be a useful feature. Can we discuss some alternative approaches to provide an easier to use and user friendly server messaging?
I propose to add support for a mapping file, similar to the way Translation messages are implemented, such that every error (or even every response) received from the server will be looked up in this server messages mapping file by the Admin on rest saga.
|
non_process
|
document how to override notifications messages this has already been asked before but the answer provided was to basically not use the admin component i believe this could be a useful feature can we discuss some alternative approaches to provide an easier to use and user friendly server messaging i propose to add support for a mapping file similar to the way translation messages are implemented such that every error or even every response received from the server will be looked up in this server messages mapping file by the admin on rest saga
| 0
|
116,612
| 17,380,518,793
|
IssuesEvent
|
2021-07-31 16:03:11
|
AlexRogalskiy/charts
|
https://api.github.com/repos/AlexRogalskiy/charts
|
opened
|
CVE-2020-7656 (Medium) detected in jquery-1.8.1.min.js
|
security vulnerability
|
## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: charts/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: charts/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/charts/commit/6ec1d9fe297e071640b6fc8bc19516a88e99467a">6ec1d9fe297e071640b6fc8bc19516a88e99467a</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p>
<p>Release Date: 2020-05-28</p>
<p>Fix Resolution: jquery - 1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-7656 (Medium) detected in jquery-1.8.1.min.js - ## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p>
<p>Path to dependency file: charts/node_modules/redeyed/examples/browser/index.html</p>
<p>Path to vulnerable library: charts/node_modules/redeyed/examples/browser/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.8.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/charts/commit/6ec1d9fe297e071640b6fc8bc19516a88e99467a">6ec1d9fe297e071640b6fc8bc19516a88e99467a</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p>
<p>Release Date: 2020-05-28</p>
<p>Fix Resolution: jquery - 1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file charts node modules redeyed examples browser index html path to vulnerable library charts node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
| 0
|
36,426
| 5,051,575,899
|
IssuesEvent
|
2016-12-20 22:19:53
|
DirtyHairy/stella
|
https://api.github.com/repos/DirtyHairy/stella
|
closed
|
Tapper (from AtariAge)
|
regression test
|
Improvement in the new core:
Old stella versions had a small graphical glitch in "Tapper" (circled in yellow in the picture)
Now it's fine with the new core (and in Stellerator too)

|
1.0
|
Tapper (from AtariAge) - Improvement in the new core:
Old stella versions had a small graphical glitch in "Tapper" (circled in yellow in the picture)
Now it's fine with the new core (and in Stellerator too)

|
non_process
|
tapper from atariage improvement in the new core old stella versions had a small graphical glitch in tapper circled in yellow in the picture now it s fine with the new core and in stellerator too
| 0
|
21,049
| 27,993,405,596
|
IssuesEvent
|
2023-03-27 06:36:47
|
TeamAidemy/ds-paper-summaries
|
https://api.github.com/repos/TeamAidemy/ds-paper-summaries
|
opened
|
All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text
|
Natural language processing Evaluation
|
Elizabeth Clark, Tal August, Sofia Serrano, Nikita Haduong, Suchin Gururangan, and Noah A. Smith. 2022. “All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text.” arXiv:2107.00061 [cs.CL]. https://arxiv.org/abs/2107.00061.
- 言語モデルが生成したテキストの評価を人間が行っているが、そもそもそれ('Human' Is Gold)でよいのか?と警鐘を鳴らす論文
- 人間が書いた文章と機械(GPT-2, 3)が書いた文章の識別タスクを評価者に実施してもらった結果、約半数は見分けがつかないことを実証
- 評価者を訓練すれば区別できるようになるのでは? → 訓練しても正答率は微増に留まり、統計的な有意差は確認されなかった
- だから、そもそも評価方法を再考しようねと提言
## Abstract
> Human evaluations are typically considered the gold standard in natural language generation, but as models' fluency improves, how well can evaluators detect and judge machine-generated text? We run a study assessing non-experts' ability to distinguish between human- and machine-authored text (GPT2 and GPT3) in three domains (stories, news articles, and recipes). We find that, without training, evaluators distinguished between GPT3- and human-authored text at random chance level. We explore three approaches for quickly training evaluators to better identify GPT3-authored text (detailed instructions, annotated examples, and paired examples) and find that while evaluators' accuracy improved up to 55%, it did not significantly improve across the three domains. Given the inconsistent results across text domains and the often contradictory reasons evaluators gave for their judgments, we examine the role untrained human evaluations play in NLG evaluation and provide recommendations to NLG researchers for improving human evaluations of text generated from state-of-the-art models.
(DeepL翻訳)
自然言語生成において、人間の評価は一般的にゴールドスタンダードと考えられているが、モデルの流暢性が向上するにつれ、評価者は機械が生成したテキストをどれだけ検出し判断できるのだろうか?我々は、3つのドメイン(ストーリー、ニュース記事、レシピ)において、非専門家が人間が作成したテキストと機械が作成したテキスト(GPT2とGPT3)を区別する能力を評価する研究を実施した。その結果、訓練なしでも、評価者はGPT3と人間作成のテキストをランダムな確率レベルで区別することがわかった。また、GPT3が作成したテキストをより適切に識別するために、3つのアプローチ(詳細な説明、注釈付き例、ペア例)を検討し、評価者の精度が最大55%向上するものの、3つのドメインで有意な向上が見られないことを発見しました。テキストドメイン間で一貫性のない結果と、評価者が判断した理由がしばしば矛盾していたことから、訓練されていない人間の評価がNLG評価において果たす役割を検証し、NLG研究者に最先端のモデルから生成されたテキストの人間評価を改善するための提言を行う。
## コード
まとめ作成時点では無し
## 解決した課題/先行研究との比較
GPT-2, 3 と言語モデルの発展につれて、どんどん流暢な文章が生成可能になってきたが、果たして人間は言語モデルが生成した文章と人間が書いた文章を見分けられるだろうか?という疑問からスタート
<!--
<img width="411" alt="スクリーンショット 2023-03-14 10 40 38" src="https://user-images.githubusercontent.com/68265677/224870206-e1e00ba1-beac-4949-b7f2-15e3ce971d40.png">
*あなた自身は、これを人間が書いた文章なのかAIが書いた文章なのか見抜けますか?(正解は本論文のFig.1参照)* -->
| <img width="500" alt="スクリーンショット 2023-03-14 10 40 38" src="https://user-images.githubusercontent.com/68265677/224870206-e1e00ba1-beac-4949-b7f2-15e3ce971d40.png"> |
|:--:|
| *あなた自身は、これを人間が書いた文章なのかAIが書いた文章なのか見抜けますか?(正解は本論文のFig.1参照)* |
- 人間の言語モデル識別能力に関して調査した以下のような先行研究が存在
- GPT2-L が生成したテキストを、訓練された評価者でも、約30%は「これは人間が書いた」と識別してしまう (Ippolito et al., 2020)
- GPT3-davinci が生成したニュース記事を、評価者は約50%の確率で「これは人間が書いた」と識別してしまう (Brown et al., 2020)
- 本論文は「評価の設定・データ・参加者を変えて調査してみた」という、上記の派生系の1種としての位置づけ。特にいずれかを否定する趣旨ではない
**本論文内で直接解決した課題はなく、今後はこうすべきという提言に留まる(詳細は後述)**
## 評価指標
- 人間が書いた文章と機械(GPT-2, GPT-3)が書いた文章が(当然、各文章を人間と機械のいずれが書いたかは非公開の状態で)計5つ提示されるので、評価者はそれぞれを4段階で評価 + なぜそう思ったかのコメントを残す
- Definitely human-written
- Possibly human-written
- Possibly machine-generated
- Definitely machine-generated
- 評価者は[Amazon Mechanical Turk (AMT)](https://www.mturk.com/)で収集
- 言語面でのアンバランスが生じないように→米国在住の方に限定
- 雑な評価をしないように→1,000以上のAMTタスクを完了&95%以上の合格率を持つ方に限定
- 2つの条件下でそれぞれ実験
- いずれの条件の場合も評価者一人あたりは計30文章を評価する
- 3ドメイン(①物語、②ニュース記事、③料理のレシピ)
- 2モデル(GPT-2, GPT-3)
- 5つの文章が提示される
- ∴ 3 x 2 x 5 = 30
- 評価前の訓練**なし**: 130名
- 評価前の訓練**あり**: 195名(上記130名とは完全に別)
<!-- - 訓練内容① インストラクション
<img width="418" alt="スクリーンショット 2023-02-13 17 11 51" src="https://user-images.githubusercontent.com/68265677/218404698-d315b143-d47a-4d30-89a1-c9fe289ce4e0.png">
- 訓練内容② 例示
<img width="209" alt="スクリーンショット 2023-02-13 17 11 59" src="https://user-images.githubusercontent.com/68265677/218404669-6dc63ba3-207a-40f9-8d30-3f1d29677cd7.png">
- 訓練内容③ 比較
<img width="324" alt="スクリーンショット 2023-02-13 17 12 15" src="https://user-images.githubusercontent.com/68265677/218404645-495998d3-6fa7-4870-9438-d6b46c80a980.png"> -->
| <img width="350" alt="training1" src="https://user-images.githubusercontent.com/68265677/218404698-d315b143-d47a-4d30-89a1-c9fe289ce4e0.png"> | <img width="350" alt="training2" src="https://user-images.githubusercontent.com/68265677/218404669-6dc63ba3-207a-40f9-8d30-3f1d29677cd7.png"> | <img width="350" alt="training3" src="https://user-images.githubusercontent.com/68265677/218404645-495998d3-6fa7-4870-9438-d6b46c80a980.png"> |
| :--: | :--: | :--: |
| *訓練1 インストラクション* | *訓練2 例示* | *訓練3 比較* |
### 結果
- GPT-2の書いた文章と人間の書いた文章の判別: 正答率 57.9%
- GPT-3の書いた文章と人間の書いた文章の判別: 正答率 49.9% **→ 約半数の文章は見分けがついていない**
- このようにGPT-2→GPT-3にモデルが変わると正答率は10pt近く減少するにも関わらず、「自信を持って回答した率」は有意に減少しない **→ 生成モデルの能力を過小評価している**
<!-- - <img width="625" alt="スクリーンショット 2023-02-13 16 58 23" src="https://user-images.githubusercontent.com/68265677/218401956-f36e0032-1c52-4a7f-bedc-1ac7153959fa.png">
- ↑ appendix tab.5 の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧 -->
| <img width="550" alt="tab5" src="https://user-images.githubusercontent.com/68265677/218401956-f36e0032-1c52-4a7f-bedc-1ac7153959fa.png"> |
|:--:|
| *appendix tab.5 の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧* |
- 訓練ありのとき、訓練なしのときよりも正答率はわずかに向上。しかし、統計的な有意差はナシ
- 一応、3つの訓練の中で最も効果的だったのは、訓練内容②の例示
## 残された課題・議論
- タイトルの通り 'Human' Is Not Gold であることが実証されたため、言語モデルの評価方法を再検討すべきと提言。代替の方法として挙げられているのは以下
- TuringAdvice: 有用なアドバイスを生成する能力で評価 (Zellers et al., 2021)
- RoFT: 評価者に推測ゲームを通して、人間と機械が生成したテキストの境界を決定させる (Dugan et al., 2020)
- 生成されたテキストを用いて物語を書かせて評価 (Clark and Smith, 2021. Akoury et al., 2020.)
- **論文内では上記のどれが推奨等は言及なく、その検討自体は残された課題と言えよう**
- どうしても本論文と同じような条件下で人間による評価が必要な場合は、評価者を例題で訓練することを推奨
- 評価者が暗黙のうちに持っている評価基準が、研究者の意図するものとずれている可能性があるため
- 本論文内の実験でも、評価者に対して事前の訓練を施すことで(わずかではあるが)評価精度が向上した
- また、評価者に与えた指示と訓練の詳細を論文に含めることを推奨
## 関連論文
Chmielewski, M., and Kucker, S. C. 2020. "An MTurk Crisis? Shifts in Data Quality and the Impact on Study Results." Social Psychological and Personality Science, 11(4), 464–473. https://doi.org/10.1177/1948550619875149
- 本論文の主張からは少し観点がずれるが、そもそも評価者を集めるときに AMT を使うこと自体がナンセンスなのでは?と提唱している論文
- 日本語解説記事: [MTurkを使った研究はもう信用できないものなのか](https://kuri8ive.hatenablog.com/entry/memo-about-mturk-crisis)
## 参考情報
- [北海道大学調和系工学研究室 平田氏によるサマリー](https://www.slideshare.net/harmonylab/all-thats-human-is-not-gold-evaluating-human-evaluation-of-generated-text) (Slideshare)
|
1.0
|
All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text - Elizabeth Clark, Tal August, Sofia Serrano, Nikita Haduong, Suchin Gururangan, and Noah A. Smith. 2022. “All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text.” arXiv:2107.00061 [cs.CL]. https://arxiv.org/abs/2107.00061.
- 言語モデルが生成したテキストの評価を人間が行っているが、そもそもそれ('Human' Is Gold)でよいのか?と警鐘を鳴らす論文
- 人間が書いた文章と機械(GPT-2, 3)が書いた文章の識別タスクを評価者に実施してもらった結果、約半数は見分けがつかないことを実証
- 評価者を訓練すれば区別できるようになるのでは? → 訓練しても正答率は微増に留まり、統計的な有意差は確認されなかった
- だから、そもそも評価方法を再考しようねと提言
## Abstract
> Human evaluations are typically considered the gold standard in natural language generation, but as models' fluency improves, how well can evaluators detect and judge machine-generated text? We run a study assessing non-experts' ability to distinguish between human- and machine-authored text (GPT2 and GPT3) in three domains (stories, news articles, and recipes). We find that, without training, evaluators distinguished between GPT3- and human-authored text at random chance level. We explore three approaches for quickly training evaluators to better identify GPT3-authored text (detailed instructions, annotated examples, and paired examples) and find that while evaluators' accuracy improved up to 55%, it did not significantly improve across the three domains. Given the inconsistent results across text domains and the often contradictory reasons evaluators gave for their judgments, we examine the role untrained human evaluations play in NLG evaluation and provide recommendations to NLG researchers for improving human evaluations of text generated from state-of-the-art models.
(DeepL翻訳)
自然言語生成において、人間の評価は一般的にゴールドスタンダードと考えられているが、モデルの流暢性が向上するにつれ、評価者は機械が生成したテキストをどれだけ検出し判断できるのだろうか?我々は、3つのドメイン(ストーリー、ニュース記事、レシピ)において、非専門家が人間が作成したテキストと機械が作成したテキスト(GPT2とGPT3)を区別する能力を評価する研究を実施した。その結果、訓練なしでも、評価者はGPT3と人間作成のテキストをランダムな確率レベルで区別することがわかった。また、GPT3が作成したテキストをより適切に識別するために、3つのアプローチ(詳細な説明、注釈付き例、ペア例)を検討し、評価者の精度が最大55%向上するものの、3つのドメインで有意な向上が見られないことを発見しました。テキストドメイン間で一貫性のない結果と、評価者が判断した理由がしばしば矛盾していたことから、訓練されていない人間の評価がNLG評価において果たす役割を検証し、NLG研究者に最先端のモデルから生成されたテキストの人間評価を改善するための提言を行う。
## コード
まとめ作成時点では無し
## 解決した課題/先行研究との比較
GPT-2, 3 と言語モデルの発展につれて、どんどん流暢な文章が生成可能になってきたが、果たして人間は言語モデルが生成した文章と人間が書いた文章を見分けられるだろうか?という疑問からスタート
<!--
<img width="411" alt="スクリーンショット 2023-03-14 10 40 38" src="https://user-images.githubusercontent.com/68265677/224870206-e1e00ba1-beac-4949-b7f2-15e3ce971d40.png">
*あなた自身は、これを人間が書いた文章なのかAIが書いた文章なのか見抜けますか?(正解は本論文のFig.1参照)* -->
| <img width="500" alt="スクリーンショット 2023-03-14 10 40 38" src="https://user-images.githubusercontent.com/68265677/224870206-e1e00ba1-beac-4949-b7f2-15e3ce971d40.png"> |
|:--:|
| *あなた自身は、これを人間が書いた文章なのかAIが書いた文章なのか見抜けますか?(正解は本論文のFig.1参照)* |
- 人間の言語モデル識別能力に関して調査した以下のような先行研究が存在
- GPT2-L が生成したテキストを、訓練された評価者でも、約30%は「これは人間が書いた」と識別してしまう (Ippolito et al., 2020)
- GPT3-davinci が生成したニュース記事を、評価者は約50%の確率で「これは人間が書いた」と識別してしまう (Brown et al., 2020)
- 本論文は「評価の設定・データ・参加者を変えて調査してみた」という、上記の派生系の1種としての位置づけ。特にいずれかを否定する趣旨ではない
**本論文内で直接解決した課題はなく、今後はこうすべきという提言に留まる(詳細は後述)**
## 評価指標
- 人間が書いた文章と機械(GPT-2, GPT-3)が書いた文章が(当然、各文章を人間と機械のいずれが書いたかは非公開の状態で)計5つ提示されるので、評価者はそれぞれを4段階で評価 + なぜそう思ったかのコメントを残す
- Definitely human-written
- Possibly human-written
- Possibly machine-generated
- Definitely machine-generated
- 評価者は[Amazon Mechanical Turk (AMT)](https://www.mturk.com/)で収集
- 言語面でのアンバランスが生じないように→米国在住の方に限定
- 雑な評価をしないように→1,000以上のAMTタスクを完了&95%以上の合格率を持つ方に限定
- 2つの条件下でそれぞれ実験
- いずれの条件の場合も評価者一人あたりは計30文章を評価する
- 3ドメイン(①物語、②ニュース記事、③料理のレシピ)
- 2モデル(GPT-2, GPT-3)
- 5つの文章が提示される
- ∴ 3 x 2 x 5 = 30
- 評価前の訓練**なし**: 130名
- 評価前の訓練**あり**: 195名(上記130名とは完全に別)
<!-- - 訓練内容① インストラクション
<img width="418" alt="スクリーンショット 2023-02-13 17 11 51" src="https://user-images.githubusercontent.com/68265677/218404698-d315b143-d47a-4d30-89a1-c9fe289ce4e0.png">
- 訓練内容② 例示
<img width="209" alt="スクリーンショット 2023-02-13 17 11 59" src="https://user-images.githubusercontent.com/68265677/218404669-6dc63ba3-207a-40f9-8d30-3f1d29677cd7.png">
- 訓練内容③ 比較
<img width="324" alt="スクリーンショット 2023-02-13 17 12 15" src="https://user-images.githubusercontent.com/68265677/218404645-495998d3-6fa7-4870-9438-d6b46c80a980.png"> -->
| <img width="350" alt="training1" src="https://user-images.githubusercontent.com/68265677/218404698-d315b143-d47a-4d30-89a1-c9fe289ce4e0.png"> | <img width="350" alt="training2" src="https://user-images.githubusercontent.com/68265677/218404669-6dc63ba3-207a-40f9-8d30-3f1d29677cd7.png"> | <img width="350" alt="training3" src="https://user-images.githubusercontent.com/68265677/218404645-495998d3-6fa7-4870-9438-d6b46c80a980.png"> |
| :--: | :--: | :--: |
| *訓練1 インストラクション* | *訓練2 例示* | *訓練3 比較* |
### 結果
- GPT-2の書いた文章と人間の書いた文章の判別: 正答率 57.9%
- GPT-3の書いた文章と人間の書いた文章の判別: 正答率 49.9% **→ 約半数の文章は見分けがついていない**
- このようにGPT-2→GPT-3にモデルが変わると正答率は10pt近く減少するにも関わらず、「自信を持って回答した率」は有意に減少しない **→ 生成モデルの能力を過小評価している**
<!-- - <img width="625" alt="スクリーンショット 2023-02-13 16 58 23" src="https://user-images.githubusercontent.com/68265677/218401956-f36e0032-1c52-4a7f-bedc-1ac7153959fa.png">
- ↑ appendix tab.5 の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧 -->
| <img width="550" alt="tab5" src="https://user-images.githubusercontent.com/68265677/218401956-f36e0032-1c52-4a7f-bedc-1ac7153959fa.png"> |
|:--:|
| *appendix tab.5 の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧* |
- 訓練ありのとき、訓練なしのときよりも正答率はわずかに向上。しかし、統計的な有意差はナシ
- 一応、3つの訓練の中で最も効果的だったのは、訓練内容②の例示
## 残された課題・議論
- タイトルの通り 'Human' Is Not Gold であることが実証されたため、言語モデルの評価方法を再検討すべきと提言。代替の方法として挙げられているのは以下
- TuringAdvice: 有用なアドバイスを生成する能力で評価 (Zellers et al., 2021)
- RoFT: 評価者に推測ゲームを通して、人間と機械が生成したテキストの境界を決定させる (Dugan et al., 2020)
- 生成されたテキストを用いて物語を書かせて評価 (Clark and Smith, 2021. Akoury et al., 2020.)
- **論文内では上記のどれが推奨等は言及なく、その検討自体は残された課題と言えよう**
- どうしても本論文と同じような条件下で人間による評価が必要な場合は、評価者を例題で訓練することを推奨
- 評価者が暗黙のうちに持っている評価基準が、研究者の意図するものとずれている可能性があるため
- 本論文内の実験でも、評価者に対して事前の訓練を施すことで(わずかではあるが)評価精度が向上した
- また、評価者に与えた指示と訓練の詳細を論文に含めることを推奨
## 関連論文
Chmielewski, M., and Kucker, S. C. 2020. "An MTurk Crisis? Shifts in Data Quality and the Impact on Study Results." Social Psychological and Personality Science, 11(4), 464–473. https://doi.org/10.1177/1948550619875149
- 本論文の主張からは少し観点がずれるが、そもそも評価者を集めるときに AMT を使うこと自体がナンセンスなのでは?と提唱している論文
- 日本語解説記事: [MTurkを使った研究はもう信用できないものなのか](https://kuri8ive.hatenablog.com/entry/memo-about-mturk-crisis)
## 参考情報
- [北海道大学調和系工学研究室 平田氏によるサマリー](https://www.slideshare.net/harmonylab/all-thats-human-is-not-gold-evaluating-human-evaluation-of-generated-text) (Slideshare)
|
process
|
all that s human is not gold evaluating human evaluation of generated text elizabeth clark tal august sofia serrano nikita haduong suchin gururangan and noah a smith “all that s human is not gold evaluating human evaluation of generated text ” arxiv 言語モデルが生成したテキストの評価を人間が行っているが、そもそもそれ( human is gold)でよいのか?と警鐘を鳴らす論文 人間が書いた文章と機械(gpt )が書いた文章の識別タスクを評価者に実施してもらった結果、約半数は見分けがつかないことを実証 評価者を訓練すれば区別できるようになるのでは? → 訓練しても正答率は微増に留まり、統計的な有意差は確認されなかった だから、そもそも評価方法を再考しようねと提言 abstract human evaluations are typically considered the gold standard in natural language generation but as models fluency improves how well can evaluators detect and judge machine generated text we run a study assessing non experts ability to distinguish between human and machine authored text and in three domains stories news articles and recipes we find that without training evaluators distinguished between and human authored text at random chance level we explore three approaches for quickly training evaluators to better identify authored text detailed instructions annotated examples and paired examples and find that while evaluators accuracy improved up to it did not significantly improve across the three domains given the inconsistent results across text domains and the often contradictory reasons evaluators gave for their judgments we examine the role untrained human evaluations play in nlg evaluation and provide recommendations to nlg researchers for improving human evaluations of text generated from state of the art models deepl翻訳 自然言語生成において、人間の評価は一般的にゴールドスタンダードと考えられているが、モデルの流暢性が向上するにつれ、評価者は機械が生成したテキストをどれだけ検出し判断できるのだろうか?我々は、 (ストーリー、ニュース記事、レシピ)において、非専門家が人間が作成したテキストと機械が作成したテキスト( )を区別する能力を評価する研究を実施した。その結果、訓練なしでも、 。また、 、 (詳細な説明、注釈付き例、ペア例)を検討し、 向上するものの、 。テキストドメイン間で一貫性のない結果と、評価者が判断した理由がしばしば矛盾していたことから、訓練されていない人間の評価がnlg評価において果たす役割を検証し、nlg研究者に最先端のモデルから生成されたテキストの人間評価を改善するための提言を行う。 コード まとめ作成時点では無し 解決した課題 先行研究との比較 gpt と言語モデルの発展につれて、どんどん流暢な文章が生成可能になってきたが、果たして人間は言語モデルが生成した文章と人間が書いた文章を見分けられるだろうか?という疑問からスタート img width alt スクリーンショット src あなた自身は、これを人間が書いた文章なのかaiが書いた文章なのか見抜けますか?(正解は本論文のfig ) img width alt スクリーンショット src あなた自身は、これを人間が書いた文章なのかaiが書いた文章なのか見抜けますか?(正解は本論文のfig ) 人間の言語モデル識別能力に関して調査した以下のような先行研究が存在 l が生成したテキストを、訓練された評価者でも、 は「これは人間が書いた」と識別してしまう ippolito et al davinci が生成したニュース記事を、 の確率で「これは人間が書いた」と識別してしまう brown et al 本論文は「評価の設定・データ・参加者を変えて調査してみた」という、 。特にいずれかを否定する趣旨ではない 本論文内で直接解決した課題はなく、今後はこうすべきという提言に留まる(詳細は後述) 評価指標 人間が書いた文章と機械(gpt gpt )が書いた文章が(当然、各文章を人間と機械のいずれが書いたかは非公開の状態で) 、 なぜそう思ったかのコメントを残す definitely human written possibly human written possibly machine generated definitely machine generated 評価者は 言語面でのアンバランスが生じないように→米国在住の方に限定 雑な評価をしないように→ & 以上の合格率を持つ方に限定 (①物語、②ニュース記事、③料理のレシピ) (gpt gpt ) ∴ x x 評価前の訓練 なし 評価前の訓練 あり ( ) 訓練内容① インストラクション img width alt スクリーンショット src 訓練内容② 例示 img width alt スクリーンショット src 訓練内容③ 比較 img width alt src img width alt src img width alt src インストラクション 例示 比較 結果 gpt 正答率 gpt 正答率 → 約半数の文章は見分けがついていない このようにgpt →gpt 、「自信を持って回答した率」は有意に減少しない → 生成モデルの能力を過小評価している img width alt スクリーンショット src ↑ appendix tab の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧 img width alt src appendix tab の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧 訓練ありのとき、訓練なしのときよりも正答率はわずかに向上。しかし、統計的な有意差はナシ 一応、 、訓練内容②の例示 残された課題・議論 タイトルの通り human is not gold であることが実証されたため、言語モデルの評価方法を再検討すべきと提言。代替の方法として挙げられているのは以下 turingadvice 有用なアドバイスを生成する能力で評価 zellers et al roft 評価者に推測ゲームを通して、人間と機械が生成したテキストの境界を決定させる dugan et al 生成されたテキストを用いて物語を書かせて評価 clark and smith akoury et al 論文内では上記のどれが推奨等は言及なく、その検討自体は残された課題と言えよう どうしても本論文と同じような条件下で人間による評価が必要な場合は、評価者を例題で訓練することを推奨 評価者が暗黙のうちに持っている評価基準が、研究者の意図するものとずれている可能性があるため 本論文内の実験でも、評価者に対して事前の訓練を施すことで(わずかではあるが)評価精度が向上した また、評価者に与えた指示と訓練の詳細を論文に含めることを推奨 関連論文 chmielewski m and kucker s c an mturk crisis shifts in data quality and the impact on study results social psychological and personality science – 本論文の主張からは少し観点がずれるが、そもそも評価者を集めるときに amt を使うこと自体がナンセンスなのでは?と提唱している論文 日本語解説記事 参考情報 slideshare
| 1
|
166,503
| 12,959,551,852
|
IssuesEvent
|
2020-07-20 13:10:23
|
RPTools/maptool
|
https://api.github.com/repos/RPTools/maptool
|
closed
|
add lumens support to Sight (Light already has it)
|
feature tested
|
**Describe the bug**
Lights are currently capable of expressing "negative lumens", causing darkness. There's a note in the Help text of the Campaign Property's **Light** tab that says the dark vision **Sight** is equivalent to 100 lumens. The problem is that I have a creature that can cast darkness at lumens=-100 (per the **Light** tab) but needs to be able to see through it and can't (because the **Sight** tab doesn't have a "lumens" keyword).
**Expected behavior**
Since the **Light** tab allows negative lumens, the **Sight** tab should allow the definition of vision types that can see through it.
**MapTool Info**
- Version: 1.6.1
**Desktop (please complete the following information):**
- OS: macOS
- Version 10.15
**Additional context**
I haven't looked at the code (I have no development environment installed, currently) but this may or may not be a lot of work. Since the **Light** tab says darkvision (which should be any vision type with the **r** modifier) effectively has "lumens=100", I'm hoping there's already a field in the structure and it just needs the proper parsing in the **Sight** tab to be active. 🤞
|
1.0
|
add lumens support to Sight (Light already has it) - **Describe the bug**
Lights are currently capable of expressing "negative lumens", causing darkness. There's a note in the Help text of the Campaign Property's **Light** tab that says the dark vision **Sight** is equivalent to 100 lumens. The problem is that I have a creature that can cast darkness at lumens=-100 (per the **Light** tab) but needs to be able to see through it and can't (because the **Sight** tab doesn't have a "lumens" keyword).
**Expected behavior**
Since the **Light** tab allows negative lumens, the **Sight** tab should allow the definition of vision types that can see through it.
**MapTool Info**
- Version: 1.6.1
**Desktop (please complete the following information):**
- OS: macOS
- Version 10.15
**Additional context**
I haven't looked at the code (I have no development environment installed, currently) but this may or may not be a lot of work. Since the **Light** tab says darkvision (which should be any vision type with the **r** modifier) effectively has "lumens=100", I'm hoping there's already a field in the structure and it just needs the proper parsing in the **Sight** tab to be active. 🤞
|
non_process
|
add lumens support to sight light already has it describe the bug lights are currently capable of expressing negative lumens causing darkness there s a note in the help text of the campaign property s light tab that says the dark vision sight is equivalent to lumens the problem is that i have a creature that can cast darkness at lumens per the light tab but needs to be able to see through it and can t because the sight tab doesn t have a lumens keyword expected behavior since the light tab allows negative lumens the sight tab should allow the definition of vision types that can see through it maptool info version desktop please complete the following information os macos version additional context i haven t looked at the code i have no development environment installed currently but this may or may not be a lot of work since the light tab says darkvision which should be any vision type with the r modifier effectively has lumens i m hoping there s already a field in the structure and it just needs the proper parsing in the sight tab to be active 🤞
| 0
|
711
| 3,203,583,846
|
IssuesEvent
|
2015-10-02 19:47:30
|
johnhbenetech/listTest
|
https://api.github.com/repos/johnhbenetech/listTest
|
opened
|
product.Trachoma_Detection
|
feature.Data_Collection feature.Sensor process.Social_Service_Referral technology.Hardware technology.Narrow_Band_Communication technology.Software user.Poverty_Global_South vertical.Healthcare
|
1 Impact
1 Need
1 Mission fit
1 Legalilty
0 Money: current
1 Money: future
0 Cost
0 Talent
0 Technical Risk
1 Champion
0 Community: users
0 Community: partners
0 Portfolio fit
1 Exit strategies
0 Data/proof
|
1.0
|
product.Trachoma_Detection - 1 Impact
1 Need
1 Mission fit
1 Legalilty
0 Money: current
1 Money: future
0 Cost
0 Talent
0 Technical Risk
1 Champion
0 Community: users
0 Community: partners
0 Portfolio fit
1 Exit strategies
0 Data/proof
|
process
|
product trachoma detection impact need mission fit legalilty money current money future cost talent technical risk champion community users community partners portfolio fit exit strategies data proof
| 1
|
13,395
| 15,867,641,673
|
IssuesEvent
|
2021-04-08 17:09:41
|
NationalSecurityAgency/ghidra
|
https://api.github.com/repos/NationalSecurityAgency/ghidra
|
closed
|
ARM A32/T32 strb.w ldrb.w data acces size of undefined1 somtimes, undefined2 others times
|
Feature: Processor/ARM Type: Question
|
**1. DAT_c28f9080**
strb.w ldrb.w are marking the data as undefined1


**2. DAT_c34eb5aa**
ldrh.w access data as undefined2

any idea why the size mismatch btw the two data ?
|
1.0
|
ARM A32/T32 strb.w ldrb.w data acces size of undefined1 somtimes, undefined2 others times - **1. DAT_c28f9080**
strb.w ldrb.w are marking the data as undefined1


**2. DAT_c34eb5aa**
ldrh.w access data as undefined2

any idea why the size mismatch btw the two data ?
|
process
|
arm strb w ldrb w data acces size of somtimes others times dat strb w ldrb w are marking the data as dat ldrh w access data as any idea why the size mismatch btw the two data
| 1
|
3,982
| 6,911,479,685
|
IssuesEvent
|
2017-11-28 08:31:22
|
pingcap/tikv
|
https://api.github.com/repos/pingcap/tikv
|
closed
|
adjust coprocessor and scheduler slow timer
|
coprocessor enhancement scheduler
|
Now the slow timer for the coprocessor and scheduler operation contains the wait time + execution time. If one operation is too slow, it will cause following operations slow too, and the slow log is very difficult for us to see whether the operation is the real slow operation.
We should use the slow timer for the execution time, use metric for the wait time is enough.
|
1.0
|
adjust coprocessor and scheduler slow timer - Now the slow timer for the coprocessor and scheduler operation contains the wait time + execution time. If one operation is too slow, it will cause following operations slow too, and the slow log is very difficult for us to see whether the operation is the real slow operation.
We should use the slow timer for the execution time, use metric for the wait time is enough.
|
process
|
adjust coprocessor and scheduler slow timer now the slow timer for the coprocessor and scheduler operation contains the wait time execution time if one operation is too slow it will cause following operations slow too and the slow log is very difficult for us to see whether the operation is the real slow operation we should use the slow timer for the execution time use metric for the wait time is enough
| 1
|
696,238
| 23,892,796,936
|
IssuesEvent
|
2022-09-08 12:47:57
|
bedita/manager
|
https://api.github.com/repos/bedita/manager
|
opened
|
Admin - Roles and modules access
|
feature Priority - Normal UI / UX
|
BEM provides modules accesses per role with the configuration `AccessControl`:
An example:
```
'AccessControl' => [
'manager' => [
'hidden' => ['objects'],
'readonly' => ['documents'],
],
'guest' => [
'hidden' => ['objects', 'users'],
'readonly' => ['documents'],
],
],
```
Admin users should be able to change `AccessControl` through an an `Admin - Roles and modules access` page, with an UI like the following:
https://xd.adobe.com/view/48d5c8b0-ea5e-4199-9165-591641d31b3f-1d86/screen/f498cac4-19db-4b0e-82b4-957e5e39eec8/
Note: reuse the pattern of overriding the configuration in `configs` via proper API calls.
|
1.0
|
Admin - Roles and modules access - BEM provides modules accesses per role with the configuration `AccessControl`:
An example:
```
'AccessControl' => [
'manager' => [
'hidden' => ['objects'],
'readonly' => ['documents'],
],
'guest' => [
'hidden' => ['objects', 'users'],
'readonly' => ['documents'],
],
],
```
Admin users should be able to change `AccessControl` through an an `Admin - Roles and modules access` page, with an UI like the following:
https://xd.adobe.com/view/48d5c8b0-ea5e-4199-9165-591641d31b3f-1d86/screen/f498cac4-19db-4b0e-82b4-957e5e39eec8/
Note: reuse the pattern of overriding the configuration in `configs` via proper API calls.
|
non_process
|
admin roles and modules access bem provides modules accesses per role with the configuration accesscontrol an example accesscontrol manager hidden readonly guest hidden readonly admin users should be able to change accesscontrol through an an admin roles and modules access page with an ui like the following note reuse the pattern of overriding the configuration in configs via proper api calls
| 0
|
41,693
| 10,762,733,544
|
IssuesEvent
|
2019-11-01 00:50:24
|
randombit/botan
|
https://api.github.com/repos/randombit/botan
|
closed
|
Building with jom/nmake only works if using MSVC
|
build problem
|
Cannot build the library because there is no libbotan-2.a

|
1.0
|
Building with jom/nmake only works if using MSVC - Cannot build the library because there is no libbotan-2.a

|
non_process
|
building with jom nmake only works if using msvc cannot build the library because there is no libbotan a
| 0
|
109,489
| 9,381,798,949
|
IssuesEvent
|
2019-04-04 20:34:18
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
ccl/logictestccl: TestCCLLogic failed under stress
|
C-test-failure O-robot
|
SHA: https://github.com/cockroachdb/cockroach/commits/c6df752eefe4609b8a5bbada0955f79a2cfb790e
Parameters:
```
TAGS=
GOFLAGS=-parallel=4
```
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=TestCCLLogic PKG=github.com/cockroachdb/cockroach/pkg/ccl/logictestccl TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1217865&tab=buildLog
```
=== RUN TestCCLLogic
--- FAIL: TestCCLLogic (47.49s)
test_log_scope.go:81: test logs captured to: /tmp/logTestCCLLogic865338793
test_log_scope.go:62: use -show-logs to present logs inline
=== RUN TestCCLLogic/5node-dist-opt
--- FAIL: TestCCLLogic/5node-dist-opt (0.00s)
=== RUN TestCCLLogic/5node-dist-opt/zone
=== PAUSE TestCCLLogic/5node-dist-opt/zone
=== CONT TestCCLLogic/5node-dist-opt/zone
--- done: testdata/logic_test/zone with config 5node-dist-opt: 24 tests, 0 failures
--- total: 274 tests, 0 failures
test logs left over in: /tmp/logTestCCLLogic865338793
--- FAIL: TestCCLLogic/5node-dist-opt/zone (45.87s)
logic.go:1536: condition failed to evaluate within 45s:
testdata/logic_test/zone:151: EXPLAIN SELECT * FROM t WHERE k=10
expected:
scan · ·
· table t@tertiary
· spans /10-/11
but found (query options: "retry") :
scan · ·
· table t@primary
· spans /10-/10/#
goroutine 3387 [running]:
runtime/debug.Stack(0xa7a358200, 0xc004197a70, 0x39250e0)
/usr/local/go/src/runtime/debug/stack.go:24 +0xa7
github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon(0x3992e20, 0xc002a16000, 0xc002ec3d60)
/go/src/github.com/cockroachdb/cockroach/pkg/testutils/soon.go:49 +0x103
github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).processSubtest(0xc0010cf680, 0x0, 0x0, 0xc00315d3b0, 0x0, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:1536 +0x1da2
github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).processTestFile(0xc0010cf680, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, 0x0, 0x3219045, 0x2, 0x3219045, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:1186 +0x698
github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).runFile(0xc0010cf680, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, 0x0, 0x3219045, 0x2, 0x3219045, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:2085 +0xe1
github.com/cockroachdb/cockroach/pkg/sql/logictest.RunLogicTest.func1.1(0xc002a16000)
/go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:2196 +0x1cc
testing.tRunner(0xc002a16000, 0xc0029b74a0)
/usr/local/go/src/testing/testing.go:827 +0xbf
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:878 +0x35c
```
|
1.0
|
ccl/logictestccl: TestCCLLogic failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/c6df752eefe4609b8a5bbada0955f79a2cfb790e
Parameters:
```
TAGS=
GOFLAGS=-parallel=4
```
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=TestCCLLogic PKG=github.com/cockroachdb/cockroach/pkg/ccl/logictestccl TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1217865&tab=buildLog
```
=== RUN TestCCLLogic
--- FAIL: TestCCLLogic (47.49s)
test_log_scope.go:81: test logs captured to: /tmp/logTestCCLLogic865338793
test_log_scope.go:62: use -show-logs to present logs inline
=== RUN TestCCLLogic/5node-dist-opt
--- FAIL: TestCCLLogic/5node-dist-opt (0.00s)
=== RUN TestCCLLogic/5node-dist-opt/zone
=== PAUSE TestCCLLogic/5node-dist-opt/zone
=== CONT TestCCLLogic/5node-dist-opt/zone
--- done: testdata/logic_test/zone with config 5node-dist-opt: 24 tests, 0 failures
--- total: 274 tests, 0 failures
test logs left over in: /tmp/logTestCCLLogic865338793
--- FAIL: TestCCLLogic/5node-dist-opt/zone (45.87s)
logic.go:1536: condition failed to evaluate within 45s:
testdata/logic_test/zone:151: EXPLAIN SELECT * FROM t WHERE k=10
expected:
scan · ·
· table t@tertiary
· spans /10-/11
but found (query options: "retry") :
scan · ·
· table t@primary
· spans /10-/10/#
goroutine 3387 [running]:
runtime/debug.Stack(0xa7a358200, 0xc004197a70, 0x39250e0)
/usr/local/go/src/runtime/debug/stack.go:24 +0xa7
github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon(0x3992e20, 0xc002a16000, 0xc002ec3d60)
/go/src/github.com/cockroachdb/cockroach/pkg/testutils/soon.go:49 +0x103
github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).processSubtest(0xc0010cf680, 0x0, 0x0, 0xc00315d3b0, 0x0, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:1536 +0x1da2
github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).processTestFile(0xc0010cf680, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, 0x0, 0x3219045, 0x2, 0x3219045, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:1186 +0x698
github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).runFile(0xc0010cf680, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, 0x0, 0x3219045, 0x2, 0x3219045, ...)
/go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:2085 +0xe1
github.com/cockroachdb/cockroach/pkg/sql/logictest.RunLogicTest.func1.1(0xc002a16000)
/go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:2196 +0x1cc
testing.tRunner(0xc002a16000, 0xc0029b74a0)
/usr/local/go/src/testing/testing.go:827 +0xbf
created by testing.(*T).Run
/usr/local/go/src/testing/testing.go:878 +0x35c
```
|
non_process
|
ccl logictestccl testccllogic failed under stress sha parameters tags goflags parallel to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests testccllogic pkg github com cockroachdb cockroach pkg ccl logictestccl testtimeout stressflags maxtime timeout tee tmp stress log failed test run testccllogic fail testccllogic test log scope go test logs captured to tmp test log scope go use show logs to present logs inline run testccllogic dist opt fail testccllogic dist opt run testccllogic dist opt zone pause testccllogic dist opt zone cont testccllogic dist opt zone done testdata logic test zone with config dist opt tests failures total tests failures test logs left over in tmp fail testccllogic dist opt zone logic go condition failed to evaluate within testdata logic test zone explain select from t where k expected scan · · · table t tertiary · spans but found query options retry scan · · · table t primary · spans goroutine runtime debug stack usr local go src runtime debug stack go github com cockroachdb cockroach pkg testutils succeedssoon go src github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg sql logictest logictest processsubtest go src github com cockroachdb cockroach pkg sql logictest logic go github com cockroachdb cockroach pkg sql logictest logictest processtestfile go src github com cockroachdb cockroach pkg sql logictest logic go github com cockroachdb cockroach pkg sql logictest logictest runfile go src github com cockroachdb cockroach pkg sql logictest logic go github com cockroachdb cockroach pkg sql logictest runlogictest go src github com cockroachdb cockroach pkg sql logictest logic go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go
| 0
|
148,954
| 19,560,751,759
|
IssuesEvent
|
2022-01-03 15:54:23
|
shaimael/Webgoat
|
https://api.github.com/repos/shaimael/Webgoat
|
opened
|
CVE-2021-28170 (Medium) detected in jakarta.el-3.0.3.jar
|
security vulnerability
|
## CVE-2021-28170 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jakarta.el-3.0.3.jar</b></p></summary>
<p>Jakarta Expression Language provides a specification document, API, reference implementation and TCK
that describes an expression language for Java applications.</p>
<p>Library home page: <a href="https://projects.eclipse.org/projects/ee4j.el">https://projects.eclipse.org/projects/ee4j.el</a></p>
<p>Path to dependency file: /webgoat-lessons/vulnerable-components/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-validation-2.4.3.jar (Root Library)
- :x: **jakarta.el-3.0.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/shaimael/Webgoat/commit/06d8d0b5bb8a459ba3d47f61a64fe00c62662d81">06d8d0b5bb8a459ba3d47f61a64fe00c62662d81</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Jakarta Expression Language implementation 3.0.3 and earlier, a bug in the ELParserTokenManager enables invalid EL expressions to be evaluated as if they were valid.
<p>Publish Date: 2021-05-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28170>CVE-2021-28170</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://securitylab.github.com/advisories/GHSL-2020-021-jakarta-el/">https://securitylab.github.com/advisories/GHSL-2020-021-jakarta-el/</a></p>
<p>Release Date: 2021-05-26</p>
<p>Fix Resolution: org.glassfish:jakarta.el:3.0.3.jbossorg-3, 3.0.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.glassfish","packageName":"jakarta.el","packageVersion":"3.0.3","packageFilePaths":["/webgoat-lessons/vulnerable-components/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-validation:2.4.3;org.glassfish:jakarta.el:3.0.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.glassfish:jakarta.el:3.0.3.jbossorg-3, 3.0.4","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-28170","vulnerabilityDetails":"In the Jakarta Expression Language implementation 3.0.3 and earlier, a bug in the ELParserTokenManager enables invalid EL expressions to be evaluated as if they were valid.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28170","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-28170 (Medium) detected in jakarta.el-3.0.3.jar - ## CVE-2021-28170 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jakarta.el-3.0.3.jar</b></p></summary>
<p>Jakarta Expression Language provides a specification document, API, reference implementation and TCK
that describes an expression language for Java applications.</p>
<p>Library home page: <a href="https://projects.eclipse.org/projects/ee4j.el">https://projects.eclipse.org/projects/ee4j.el</a></p>
<p>Path to dependency file: /webgoat-lessons/vulnerable-components/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-validation-2.4.3.jar (Root Library)
- :x: **jakarta.el-3.0.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/shaimael/Webgoat/commit/06d8d0b5bb8a459ba3d47f61a64fe00c62662d81">06d8d0b5bb8a459ba3d47f61a64fe00c62662d81</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In the Jakarta Expression Language implementation 3.0.3 and earlier, a bug in the ELParserTokenManager enables invalid EL expressions to be evaluated as if they were valid.
<p>Publish Date: 2021-05-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28170>CVE-2021-28170</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://securitylab.github.com/advisories/GHSL-2020-021-jakarta-el/">https://securitylab.github.com/advisories/GHSL-2020-021-jakarta-el/</a></p>
<p>Release Date: 2021-05-26</p>
<p>Fix Resolution: org.glassfish:jakarta.el:3.0.3.jbossorg-3, 3.0.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.glassfish","packageName":"jakarta.el","packageVersion":"3.0.3","packageFilePaths":["/webgoat-lessons/vulnerable-components/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-validation:2.4.3;org.glassfish:jakarta.el:3.0.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.glassfish:jakarta.el:3.0.3.jbossorg-3, 3.0.4","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-28170","vulnerabilityDetails":"In the Jakarta Expression Language implementation 3.0.3 and earlier, a bug in the ELParserTokenManager enables invalid EL expressions to be evaluated as if they were valid.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28170","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve medium detected in jakarta el jar cve medium severity vulnerability vulnerable library jakarta el jar jakarta expression language provides a specification document api reference implementation and tck that describes an expression language for java applications library home page a href path to dependency file webgoat lessons vulnerable components pom xml path to vulnerable library home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar dependency hierarchy spring boot starter validation jar root library x jakarta el jar vulnerable library found in head commit a href found in base branch main vulnerability details in the jakarta expression language implementation and earlier a bug in the elparsertokenmanager enables invalid el expressions to be evaluated as if they were valid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org glassfish jakarta el jbossorg isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter validation org glassfish jakarta el isminimumfixversionavailable true minimumfixversion org glassfish jakarta el jbossorg isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in the jakarta expression language implementation and earlier a bug in the elparsertokenmanager enables invalid el expressions to be evaluated as if they were valid vulnerabilityurl
| 0
|
262,688
| 8,272,348,338
|
IssuesEvent
|
2018-09-16 19:12:16
|
letelete/Sleep-Cycle-Alarm
|
https://api.github.com/repos/letelete/Sleep-Cycle-Alarm
|
closed
|
Save wake up at list content after setting up one
|
bug high priority medium
|
List content gets reset when user leaves wake up at tab
|
1.0
|
Save wake up at list content after setting up one - List content gets reset when user leaves wake up at tab
|
non_process
|
save wake up at list content after setting up one list content gets reset when user leaves wake up at tab
| 0
|
15,592
| 19,718,105,577
|
IssuesEvent
|
2022-01-13 13:04:35
|
Altinn/altinn-studio
|
https://api.github.com/repos/Altinn/altinn-studio
|
opened
|
Add support in Application frontend for BPMN gateways
|
area/process kind/user-story
|
<!-- Internal template. If you're external and want to suggest a user story, please use the Feature request template -->
## Description
After backend now supports gateways frontend need to be updated to do the same
## Considerations
- We need to update App API so that app frontend knows that there are multiple flows out of the current task. Example. Go back to data from confirmation and confirm
- We need to design
-
### Ops requirements
<!-- Are there any requirements for monitoring? What is being built and what could go wrong?
Are there any requirements related to backup? -->
## Acceptance criteria
<!-- Describe criteria here (i.e. What is allowed/not allowed (negative tesing), validations, error messages and warnings etc.) -->
## Specification tasks
- [ ] Design GUI for process navigation
- [ ]
- [ ] Development tasks are defined
<!-- Add other spec tasks here using checkboxes -->
## Development tasks
<!-- Add dev tasks here using checkboxes -->
## Test
<!-- Add test tasks here using checkboxes -->
## Definition of done
<!-- This chekclist is to verify that this issue meets [DoD](https://digdir-digitalisering.atlassian.net/wiki/spaces/T3KP/pages/5049246/Definition+of+Done) (Only for project members) before closing. -->
- [ ] Documentation (docs.altinn.studio) is updated (if relevant)
- [ ] Technical documentation
- [ ] User documentation
- [ ] QA
- [ ] Manual test is complete (if relevant)
- [ ] Automated test is implemented (if relevant)
- [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
|
1.0
|
Add support in Application frontend for BPMN gateways - <!-- Internal template. If you're external and want to suggest a user story, please use the Feature request template -->
## Description
After backend now supports gateways frontend need to be updated to do the same
## Considerations
- We need to update App API so that app frontend knows that there are multiple flows out of the current task. Example. Go back to data from confirmation and confirm
- We need to design
-
### Ops requirements
<!-- Are there any requirements for monitoring? What is being built and what could go wrong?
Are there any requirements related to backup? -->
## Acceptance criteria
<!-- Describe criteria here (i.e. What is allowed/not allowed (negative tesing), validations, error messages and warnings etc.) -->
## Specification tasks
- [ ] Design GUI for process navigation
- [ ]
- [ ] Development tasks are defined
<!-- Add other spec tasks here using checkboxes -->
## Development tasks
<!-- Add dev tasks here using checkboxes -->
## Test
<!-- Add test tasks here using checkboxes -->
## Definition of done
<!-- This chekclist is to verify that this issue meets [DoD](https://digdir-digitalisering.atlassian.net/wiki/spaces/T3KP/pages/5049246/Definition+of+Done) (Only for project members) before closing. -->
- [ ] Documentation (docs.altinn.studio) is updated (if relevant)
- [ ] Technical documentation
- [ ] User documentation
- [ ] QA
- [ ] Manual test is complete (if relevant)
- [ ] Automated test is implemented (if relevant)
- [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
|
process
|
add support in application frontend for bpmn gateways description after backend now supports gateways frontend need to be updated to do the same considerations we need to update app api so that app frontend knows that there are multiple flows out of the current task example go back to data from confirmation and confirm we need to design ops requirements are there any requirements for monitoring what is being built and what could go wrong are there any requirements related to backup acceptance criteria specification tasks design gui for process navigation development tasks are defined development tasks test definition of done documentation docs altinn studio is updated if relevant technical documentation user documentation qa manual test is complete if relevant automated test is implemented if relevant all tasks in this userstory are closed i e remaining tasks are moved to other user stories or marked obsolete
| 1
|
120,823
| 4,794,840,303
|
IssuesEvent
|
2016-10-31 22:20:09
|
cyberpwnn/GlacialRealms
|
https://api.github.com/repos/cyberpwnn/GlacialRealms
|
closed
|
Mod+ could kill players on Islands
|
20 Minute ETA bug general high priority
|
Anyone who bypassed the Island Protection rules could attack others and become combat tagged and would end up killing both the player they accidentally hit and themselves.
|
1.0
|
Mod+ could kill players on Islands - Anyone who bypassed the Island Protection rules could attack others and become combat tagged and would end up killing both the player they accidentally hit and themselves.
|
non_process
|
mod could kill players on islands anyone who bypassed the island protection rules could attack others and become combat tagged and would end up killing both the player they accidentally hit and themselves
| 0
|
97,265
| 3,987,907,469
|
IssuesEvent
|
2016-05-09 07:23:53
|
iacopodeenosee/EasyScreenCast
|
https://api.github.com/repos/iacopodeenosee/EasyScreenCast
|
closed
|
Error: No property show_all on this GObject GstDeviceMonitor
|
PRIORITY: normal STATUS: confirmed TYPE: duplicate TYPE: support
|
In Firefox 45, Gnome 3.16.4, I hit the button to install EasyScreenCast and affirmed a subsequent Firefox inquiry, and a red Error mark appeared in place of the install button. Then I followed instructions and pressed Alt+F2, issued the r command, reloaded the installation page, and once more chose to install it. This time the red Error mark reappeared, with a tools button to the left. I pressed the tools button and got the following note:
```
Error: No property show_all on this GObject GstDeviceMonitor
Stack trace:
HelperWebcam<._init@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/utilwebcam.js:38
wrapper@resource:///org/gnome/gjs/modules/lang.js:169
_Base.prototype._construct@resource:///org/gnome/gjs/modules/lang.js:110
Class.prototype._construct/newClass@resource:///org/gnome/gjs/modules/lang.js:204
EasyScreenCastSettingsWidget<._init@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/prefs.js:90
wrapper@resource:///org/gnome/gjs/modules/lang.js:169
buildPrefsWidget@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/prefs.js:730
Application<._selectExtension@resource:///org/gnome/shell/extensionPrefs/main.js:95
wrapper@resource:///org/gnome/gjs/modules/lang.js:169
Application<._onCommandLine@resource:///org/gnome/shell/extensionPrefs/main.js:259
wrapper@resource:///org/gnome/gjs/modules/lang.js:169
main@resource:///org/gnome/shell/extensionPrefs/main.js:397
@<main>:1
```
|
1.0
|
Error: No property show_all on this GObject GstDeviceMonitor - In Firefox 45, Gnome 3.16.4, I hit the button to install EasyScreenCast and affirmed a subsequent Firefox inquiry, and a red Error mark appeared in place of the install button. Then I followed instructions and pressed Alt+F2, issued the r command, reloaded the installation page, and once more chose to install it. This time the red Error mark reappeared, with a tools button to the left. I pressed the tools button and got the following note:
```
Error: No property show_all on this GObject GstDeviceMonitor
Stack trace:
HelperWebcam<._init@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/utilwebcam.js:38
wrapper@resource:///org/gnome/gjs/modules/lang.js:169
_Base.prototype._construct@resource:///org/gnome/gjs/modules/lang.js:110
Class.prototype._construct/newClass@resource:///org/gnome/gjs/modules/lang.js:204
EasyScreenCastSettingsWidget<._init@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/prefs.js:90
wrapper@resource:///org/gnome/gjs/modules/lang.js:169
buildPrefsWidget@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/prefs.js:730
Application<._selectExtension@resource:///org/gnome/shell/extensionPrefs/main.js:95
wrapper@resource:///org/gnome/gjs/modules/lang.js:169
Application<._onCommandLine@resource:///org/gnome/shell/extensionPrefs/main.js:259
wrapper@resource:///org/gnome/gjs/modules/lang.js:169
main@resource:///org/gnome/shell/extensionPrefs/main.js:397
@<main>:1
```
|
non_process
|
error no property show all on this gobject gstdevicemonitor in firefox gnome i hit the button to install easyscreencast and affirmed a subsequent firefox inquiry and a red error mark appeared in place of the install button then i followed instructions and pressed alt issued the r command reloaded the installation page and once more chose to install it this time the red error mark reappeared with a tools button to the left i pressed the tools button and got the following note error no property show all on this gobject gstdevicemonitor stack trace helperwebcam init home pbwolf local share gnome shell extensions easyscreencast iacopodeenosee gmail com utilwebcam js wrapper resource org gnome gjs modules lang js base prototype construct resource org gnome gjs modules lang js class prototype construct newclass resource org gnome gjs modules lang js easyscreencastsettingswidget init home pbwolf local share gnome shell extensions easyscreencast iacopodeenosee gmail com prefs js wrapper resource org gnome gjs modules lang js buildprefswidget home pbwolf local share gnome shell extensions easyscreencast iacopodeenosee gmail com prefs js application selectextension resource org gnome shell extensionprefs main js wrapper resource org gnome gjs modules lang js application oncommandline resource org gnome shell extensionprefs main js wrapper resource org gnome gjs modules lang js main resource org gnome shell extensionprefs main js
| 0
|
110,070
| 13,896,450,391
|
IssuesEvent
|
2020-10-19 17:12:57
|
JeschkeLab/DeerLab
|
https://api.github.com/repos/JeschkeLab/DeerLab
|
opened
|
Simplify parameter boundaries interface for fisignal
|
design
|
Another remnant of the old MATLAB interface. In `fitsignal` the lower/upper boundaries of the different parameter subsets are specified as a single keyword
````python
fitsignal(__, lb = [lb_dd,lb_bg,lb_ex], ub = [ub_dd,ub_bg,ub_ex])
````
Since one seldom needs to re-define all boundaries for all subsets, it would make sense to simplify this to multiple keywords
````python
fitsignal(__, lb_dd = lb_dd, lb_bg = lb_bg lb_ex = lb_ex)
````
this way a quick assignment can be done without needing to worry about the order of the parameter subsets.
|
1.0
|
Simplify parameter boundaries interface for fisignal - Another remnant of the old MATLAB interface. In `fitsignal` the lower/upper boundaries of the different parameter subsets are specified as a single keyword
````python
fitsignal(__, lb = [lb_dd,lb_bg,lb_ex], ub = [ub_dd,ub_bg,ub_ex])
````
Since one seldom needs to re-define all boundaries for all subsets, it would make sense to simplify this to multiple keywords
````python
fitsignal(__, lb_dd = lb_dd, lb_bg = lb_bg lb_ex = lb_ex)
````
this way a quick assignment can be done without needing to worry about the order of the parameter subsets.
|
non_process
|
simplify parameter boundaries interface for fisignal another remnant of the old matlab interface in fitsignal the lower upper boundaries of the different parameter subsets are specified as a single keyword python fitsignal lb ub since one seldom needs to re define all boundaries for all subsets it would make sense to simplify this to multiple keywords python fitsignal lb dd lb dd lb bg lb bg lb ex lb ex this way a quick assignment can be done without needing to worry about the order of the parameter subsets
| 0
|
19,540
| 25,858,235,558
|
IssuesEvent
|
2022-12-13 15:11:16
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
Move samples to client library repositories
|
type: process
|
This is a list of all the samples projects that should likely be moved to client library repositories:
- [x] [asset](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/asset/cloud-client)
- [x] [automl](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/automl)
- [x] [bigquery](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery)
- [x] [bigquery_datatransfer](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery/transfer)
- [x] [bigquery_storage](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery_storage/to_dataframe)
- [x] [bigtable](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigtable)
- https://github.com/googleapis/python-bigtable/pull/38
- [ ] [cloud-sql](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/cloud-sql)
- [ ] [compute](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/compute)
- [x] [container_registry](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/container_registry/container_analysise)
- [x] [datacatalog](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datacatalog/cloud-client)
- [x] [datalabeling](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datalabeling)
- [x] [dataproc](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dataproc)
- [ ] [datastore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datastore)
- [x] [dialogflow](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dialogflow/cloud-client)
- [x] [dlp](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dlp)
- [ ] [dns](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dns/api)
- [x] [document](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/document/cloud-client)
- [x] [error_reporting](error_reporting)
- [ ] [firestore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/firestore/cloud-client)
- [ ] [iot](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/iot/api-client)
- [ ] [jobs](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/jobs)
- [x] [kms](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/kms)
- https://github.com/googleapis/python-kms/pull/14
- [x] [language](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/language)
- [x] [logging](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/logging/cloud-client)
- [x] [media-translation](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/media-translation/cloud-client)
- [ ] [memorystore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/memorystore/redis)
- [ ] [monitoring](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/monitoring/api/v3)
- [x] [pubsub](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/pubsub)
- [x] [scheduler](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/scheduler)
- [x] [secretmanager](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/secretmanager/api-client)
- [x] [spanner](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/spanner/cloud-client)
- [x] [speech](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/speech)
- [ ] [storage](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/storage)
- [x] [tables](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/tables/automl) (technically automl, unclear if these should be merged)
- [x] [tasks](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/tasks)
- [x] [texttospeech](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/texttospeech/cloud-client)
- #3845
- [ ] [trace](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/trace)
- [x] [translate](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/translate)
- [x] [video](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/video/cloud-client)
- [x] [vision](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/vision)
|
1.0
|
Move samples to client library repositories - This is a list of all the samples projects that should likely be moved to client library repositories:
- [x] [asset](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/asset/cloud-client)
- [x] [automl](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/automl)
- [x] [bigquery](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery)
- [x] [bigquery_datatransfer](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery/transfer)
- [x] [bigquery_storage](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery_storage/to_dataframe)
- [x] [bigtable](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigtable)
- https://github.com/googleapis/python-bigtable/pull/38
- [ ] [cloud-sql](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/cloud-sql)
- [ ] [compute](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/compute)
- [x] [container_registry](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/container_registry/container_analysise)
- [x] [datacatalog](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datacatalog/cloud-client)
- [x] [datalabeling](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datalabeling)
- [x] [dataproc](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dataproc)
- [ ] [datastore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datastore)
- [x] [dialogflow](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dialogflow/cloud-client)
- [x] [dlp](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dlp)
- [ ] [dns](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dns/api)
- [x] [document](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/document/cloud-client)
- [x] [error_reporting](error_reporting)
- [ ] [firestore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/firestore/cloud-client)
- [ ] [iot](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/iot/api-client)
- [ ] [jobs](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/jobs)
- [x] [kms](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/kms)
- https://github.com/googleapis/python-kms/pull/14
- [x] [language](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/language)
- [x] [logging](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/logging/cloud-client)
- [x] [media-translation](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/media-translation/cloud-client)
- [ ] [memorystore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/memorystore/redis)
- [ ] [monitoring](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/monitoring/api/v3)
- [x] [pubsub](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/pubsub)
- [x] [scheduler](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/scheduler)
- [x] [secretmanager](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/secretmanager/api-client)
- [x] [spanner](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/spanner/cloud-client)
- [x] [speech](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/speech)
- [ ] [storage](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/storage)
- [x] [tables](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/tables/automl) (technically automl, unclear if these should be merged)
- [x] [tasks](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/tasks)
- [x] [texttospeech](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/texttospeech/cloud-client)
- #3845
- [ ] [trace](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/trace)
- [x] [translate](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/translate)
- [x] [video](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/video/cloud-client)
- [x] [vision](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/vision)
|
process
|
move samples to client library repositories this is a list of all the samples projects that should likely be moved to client library repositories error reporting technically automl unclear if these should be merged
| 1
|
120,755
| 4,793,622,089
|
IssuesEvent
|
2016-10-31 18:40:54
|
CogRob/workspace
|
https://api.github.com/repos/CogRob/workspace
|
opened
|
Continuous Integration
|
P2 - NORMAL PRIORITY
|
We should use CI to ensure all the test passes before merging to master.
|
1.0
|
Continuous Integration - We should use CI to ensure all the test passes before merging to master.
|
non_process
|
continuous integration we should use ci to ensure all the test passes before merging to master
| 0
|
219,585
| 24,501,586,532
|
IssuesEvent
|
2022-10-10 13:13:08
|
nidhi7598/linux-3.0.35
|
https://api.github.com/repos/nidhi7598/linux-3.0.35
|
opened
|
CVE-2019-13631 (Medium) detected in linux-stable-rtv3.8.6
|
security vulnerability
|
## CVE-2019-13631 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In parse_hid_report_descriptor in drivers/input/tablet/gtco.c in the Linux kernel through 5.2.1, a malicious USB device can send an HID report that triggers an out-of-bounds write during generation of debugging messages.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13631>CVE-2019-13631</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-13631">https://nvd.nist.gov/vuln/detail/CVE-2019-13631</a></p>
<p>Release Date: 2019-07-26</p>
<p>Fix Resolution: linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-13631 (Medium) detected in linux-stable-rtv3.8.6 - ## CVE-2019-13631 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In parse_hid_report_descriptor in drivers/input/tablet/gtco.c in the Linux kernel through 5.2.1, a malicious USB device can send an HID report that triggers an out-of-bounds write during generation of debugging messages.
<p>Publish Date: 2019-07-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13631>CVE-2019-13631</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-13631">https://nvd.nist.gov/vuln/detail/CVE-2019-13631</a></p>
<p>Release Date: 2019-07-26</p>
<p>Fix Resolution: linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers input tablet gtco c drivers input tablet gtco c drivers input tablet gtco c vulnerability details in parse hid report descriptor in drivers input tablet gtco c in the linux kernel through a malicious usb device can send an hid report that triggers an out of bounds write during generation of debugging messages publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux yocto gitautoinc gitautoinc step up your open source security game with mend
| 0
|
5,850
| 8,676,158,610
|
IssuesEvent
|
2018-11-30 13:17:21
|
zeebe-io/zeebe
|
https://api.github.com/repos/zeebe-io/zeebe
|
opened
|
Opening RocksDB on partition install
|
broker enhancement logstreams stream processor
|
I think it would be nice if we could create an RocksDB instance per Partition, this would mean that we open the RocksDB on Partition install and give the LogBlockIndex (if the LogBlockIndex is allready migrated to RocksDB) and StreamProcessor access to the database, maybe via dependency.
There could exist an Service/Actor, which creates periodically checkpoints from the database. But we have to
make sure that this causes no inconsistencies if an actor writes to RocksDB and an checkpoint is concurrently created. Possible solutions for this would be to suspend the other actors for this time of checkpoint creation OR they run all in the same actor OR maybe this is even no problem we have to find this out.
This solution would give us a separation of the RocksDB creation, data snapshoting and Event processing, because the StreamProcessorController is only interested in reading and writing events. We could deleted then the whole Snapshot stuff.
|
1.0
|
Opening RocksDB on partition install - I think it would be nice if we could create an RocksDB instance per Partition, this would mean that we open the RocksDB on Partition install and give the LogBlockIndex (if the LogBlockIndex is allready migrated to RocksDB) and StreamProcessor access to the database, maybe via dependency.
There could exist an Service/Actor, which creates periodically checkpoints from the database. But we have to
make sure that this causes no inconsistencies if an actor writes to RocksDB and an checkpoint is concurrently created. Possible solutions for this would be to suspend the other actors for this time of checkpoint creation OR they run all in the same actor OR maybe this is even no problem we have to find this out.
This solution would give us a separation of the RocksDB creation, data snapshoting and Event processing, because the StreamProcessorController is only interested in reading and writing events. We could deleted then the whole Snapshot stuff.
|
process
|
opening rocksdb on partition install i think it would be nice if we could create an rocksdb instance per partition this would mean that we open the rocksdb on partition install and give the logblockindex if the logblockindex is allready migrated to rocksdb and streamprocessor access to the database maybe via dependency there could exist an service actor which creates periodically checkpoints from the database but we have to make sure that this causes no inconsistencies if an actor writes to rocksdb and an checkpoint is concurrently created possible solutions for this would be to suspend the other actors for this time of checkpoint creation or they run all in the same actor or maybe this is even no problem we have to find this out this solution would give us a separation of the rocksdb creation data snapshoting and event processing because the streamprocessorcontroller is only interested in reading and writing events we could deleted then the whole snapshot stuff
| 1
|
255,633
| 19,318,921,753
|
IssuesEvent
|
2021-12-14 01:39:42
|
sillsdev/ptx2pdf
|
https://api.github.com/repos/sillsdev/ptx2pdf
|
closed
|
Add an option to have the header in a second language/script
|
priority enhancement bug Fixed? Documentation
|
Some regions of the world (Biblica South Asia, + BSI India) use 2 different languages or scripts for the header. Currently this is not possible, but perhaps it could be made possible. Consideration also has to be given to digit styles, styling of text, size and fonts etc.
|
1.0
|
Add an option to have the header in a second language/script - Some regions of the world (Biblica South Asia, + BSI India) use 2 different languages or scripts for the header. Currently this is not possible, but perhaps it could be made possible. Consideration also has to be given to digit styles, styling of text, size and fonts etc.
|
non_process
|
add an option to have the header in a second language script some regions of the world biblica south asia bsi india use different languages or scripts for the header currently this is not possible but perhaps it could be made possible consideration also has to be given to digit styles styling of text size and fonts etc
| 0
|
9,986
| 13,034,505,825
|
IssuesEvent
|
2020-07-28 08:49:18
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
closed
|
stellar client activate_through_friendbot no longer works
|
priority_critical process_wontfix type_bug
|
It gives a bad request and the documentation does not state a traling slash after the absolute url
The trailing slash can not be removed when using requests: https://stackoverflow.com/questions/27248521/python-http-request-url-end-slash
also, when using curl:
`curl "https://friendbot.stellar.org?addr=GA47YZA3PKFUZMPLQ3B5F2E3CJIB57TGGU7SPCQT2WAEYKN766PWIMB3"`
curl: (60) SSL certificate problem: certificate has expired
More details here: https://curl.haxx.se/docs/sslcerts.html
curl failed to verify the legitimacy of the server and therefore could not
establish a secure connection to it. To learn more about this situation and
how to fix it, please visit the web page mentioned above.
|
1.0
|
stellar client activate_through_friendbot no longer works - It gives a bad request and the documentation does not state a traling slash after the absolute url
The trailing slash can not be removed when using requests: https://stackoverflow.com/questions/27248521/python-http-request-url-end-slash
also, when using curl:
`curl "https://friendbot.stellar.org?addr=GA47YZA3PKFUZMPLQ3B5F2E3CJIB57TGGU7SPCQT2WAEYKN766PWIMB3"`
curl: (60) SSL certificate problem: certificate has expired
More details here: https://curl.haxx.se/docs/sslcerts.html
curl failed to verify the legitimacy of the server and therefore could not
establish a secure connection to it. To learn more about this situation and
how to fix it, please visit the web page mentioned above.
|
process
|
stellar client activate through friendbot no longer works it gives a bad request and the documentation does not state a traling slash after the absolute url the trailing slash can not be removed when using requests also when using curl curl curl ssl certificate problem certificate has expired more details here curl failed to verify the legitimacy of the server and therefore could not establish a secure connection to it to learn more about this situation and how to fix it please visit the web page mentioned above
| 1
|
324,962
| 24,027,281,428
|
IssuesEvent
|
2022-09-15 12:34:39
|
vuetifyjs/vuetify
|
https://api.github.com/repos/vuetifyjs/vuetify
|
closed
|
[Bug Report][2.6.10] docs(VDataIterator): wrong slot type
|
T: documentation
|
### Environment
**Vuetify Version:** 2.6.10
**Vue Version:** 2.7.10
**Browsers:** Edge 105.0.1343.27
**OS:** Windows 10
### Steps to reproduce

### Expected Behavior
`#default` is a scoped slot
### Actual Behavior
Both slot `#default` and scoped slot `#default` exist in docs.
### Reproduction Link
https://vuetifyjs.com/en/api/v-data-iterator/#slots
<!-- generated by vuetify-issue-helper. DO NOT REMOVE -->
|
1.0
|
[Bug Report][2.6.10] docs(VDataIterator): wrong slot type - ### Environment
**Vuetify Version:** 2.6.10
**Vue Version:** 2.7.10
**Browsers:** Edge 105.0.1343.27
**OS:** Windows 10
### Steps to reproduce

### Expected Behavior
`#default` is a scoped slot
### Actual Behavior
Both slot `#default` and scoped slot `#default` exist in docs.
### Reproduction Link
https://vuetifyjs.com/en/api/v-data-iterator/#slots
<!-- generated by vuetify-issue-helper. DO NOT REMOVE -->
|
non_process
|
docs vdataiterator wrong slot type environment vuetify version vue version browsers edge os windows steps to reproduce expected behavior default is a scoped slot actual behavior both slot default and scoped slot default exist in docs reproduction link
| 0
|
19,067
| 25,088,875,527
|
IssuesEvent
|
2022-11-08 03:28:24
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
[Mirror] make 4.4
|
P2 type: process team-OSS mirror request
|
### Please list the URLs of the archives you'd like to mirror:
Please mirror http://ftpmirror.gnu.org/gnu/make/make-4.4.tar.gz for `rules_foreign_cc`
|
1.0
|
[Mirror] make 4.4 - ### Please list the URLs of the archives you'd like to mirror:
Please mirror http://ftpmirror.gnu.org/gnu/make/make-4.4.tar.gz for `rules_foreign_cc`
|
process
|
make please list the urls of the archives you d like to mirror please mirror for rules foreign cc
| 1
|
16,546
| 21,568,598,986
|
IssuesEvent
|
2022-05-02 04:17:56
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Crime Lab S.F.
|
suggested title in process
|
Title: Crime Lab S.F.
Type (film/tv show): TV show
Film or show in which it appears: Monk
Is the parent film/show streaming anywhere? Peacock, Amazon Prime
About when in the parent film/show does it appear? Season 2, Episode 12, "Mr. Monk and the TV Star"
Actual footage of the film/show can be seen (yes/no)? Yes
Cast:
Brad Terry
Link to Monk wiki: https://monk.fandom.com/wiki/Mr._Monk_and_the_TV_Star
|
1.0
|
Add Crime Lab S.F. - Title: Crime Lab S.F.
Type (film/tv show): TV show
Film or show in which it appears: Monk
Is the parent film/show streaming anywhere? Peacock, Amazon Prime
About when in the parent film/show does it appear? Season 2, Episode 12, "Mr. Monk and the TV Star"
Actual footage of the film/show can be seen (yes/no)? Yes
Cast:
Brad Terry
Link to Monk wiki: https://monk.fandom.com/wiki/Mr._Monk_and_the_TV_Star
|
process
|
add crime lab s f title crime lab s f type film tv show tv show film or show in which it appears monk is the parent film show streaming anywhere peacock amazon prime about when in the parent film show does it appear season episode mr monk and the tv star actual footage of the film show can be seen yes no yes cast brad terry link to monk wiki
| 1
|
19,840
| 26,242,571,521
|
IssuesEvent
|
2023-01-05 12:47:38
|
vesoft-inc/nebula
|
https://api.github.com/repos/vesoft-inc/nebula
|
closed
|
with USER role can do balance data
|
type/bug severity/major auto-sync affects/master process/done
|
as title, however in doc, balance data only god can do


while --enable_authorize=true

<!-- Please check the [FAQ](https://docs.nebula-graph.com.cn/master/20.appendix/0.FAQ/) documentation and old issues before raising an issue in case someone has asked the same question that you are asking. -->
balace leader also can do with user role
**Describe the bug (__required__)**
<!-- A clear and concise description of what the bug is. -->
**Your Environments (__required__)**

nebula-ent-nightly, commit id : f1e5ec3
**How To Reproduce(__required__)**
Steps to reproduce the behavior:
```
1. CREATE USER user1 WITH PASSWORD 'nebula';
2. grant role User on yx_test to user1
3. login in with user1;
4. do: balance data;
```
**Expected behavior**
report error ,only god role can do balance data
|
1.0
|
with USER role can do balance data - as title, however in doc, balance data only god can do


while --enable_authorize=true

<!-- Please check the [FAQ](https://docs.nebula-graph.com.cn/master/20.appendix/0.FAQ/) documentation and old issues before raising an issue in case someone has asked the same question that you are asking. -->
balace leader also can do with user role
**Describe the bug (__required__)**
<!-- A clear and concise description of what the bug is. -->
**Your Environments (__required__)**

nebula-ent-nightly, commit id : f1e5ec3
**How To Reproduce(__required__)**
Steps to reproduce the behavior:
```
1. CREATE USER user1 WITH PASSWORD 'nebula';
2. grant role User on yx_test to user1
3. login in with user1;
4. do: balance data;
```
**Expected behavior**
report error ,only god role can do balance data
|
process
|
with user role can do balance data as title however in doc, balance data only god can do while enable authorize true balace leader also can do with user role describe the bug required your environments required nebula ent nightly commit id how to reproduce required steps to reproduce the behavior create user with password nebula grant role user on yx test to login in with do balance data expected behavior report error only god role can do balance data
| 1
|
11,298
| 14,103,847,985
|
IssuesEvent
|
2020-11-06 10:54:27
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
GO:0097754 Name clathrin-mediated membrane bending
|
PomBase cellular processes low priority
|
missing parent
GO:0048268 clathrin coat assembly
>
or at least there is no link between the membrane bending part and the broader pathway
|
1.0
|
GO:0097754 Name clathrin-mediated membrane bending - missing parent
GO:0048268 clathrin coat assembly
>
or at least there is no link between the membrane bending part and the broader pathway
|
process
|
go name clathrin mediated membrane bending missing parent go clathrin coat assembly or at least there is no link between the membrane bending part and the broader pathway
| 1
|
642,637
| 20,908,855,414
|
IssuesEvent
|
2022-03-24 07:05:53
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
esp32 build error while applicaton in T2 topology
|
bug priority: low platform: ESP32
|
`**Describe the bug**
build error when create an application in T2 topology based on esp32
Please also mention any information which could help others to understand
the problem you're facing:
- What target platform are you using?
usually create an application using T2 topology described in https://docs.zephyrproject.org/latest/guides/west/workspaces.html
- What have you tried to diagnose or workaround this issue?
- no idea on this
**To Reproduce**
Steps to reproduce the behavior:
1. create an application using T2 topology
you could follow the instructions bellow:
```
$ git clone https://gitee.com/fwatch/twatch
$ cd twatch
$ west init -l app/
$ west update # update modules
$ west espressif update # update submodules
```
2. build samples/hell_word
```
west build -p auto -b esp32 -s app
```
3. See error
```
[84/85] Linking C executable bootloader.elf
[85/85] Generating binary image from built executable
esptool.py v3.1-dev
Merged 1 ELF section
Generated D:/workspace/twatch/build/esp-idf/build/bootloader/bootloader.bin
[159/169] Linking C executable zephyr\zephyr_pre0.elf
FAILED: zephyr/zephyr_pre0.elf zephyr/zephyr_pre0.map D:/workspace/twatch/build/zephyr/zephyr_pre0.map
cmd.exe /C "cd . && C:\Users\gouqs\.espressif\tools\zephyr\xtensa-esp32-elf\bin\xtensa-esp32-elf-gcc.exe zephyr/CMakeFiles/zephyr_pre0.dir/misc/empty_file.c.obj -o zephyr\zephyr_pre0.elf zephyr/CMakeFiles/offsets.dir/./arch/xtensa/core/offsets/offsets.c.obj -Wl,-T zephyr/linker_zephyr_pre0.cmd -Wl,-Map=D:/workspace/twatch/build/zephyr/zephyr_pre0.map -Wl,--whole-archive app/libapp.a zephyr/libzephyr.a zephyr/arch/common/libarch__common.a zephyr/arch/arch/xtensa/core/libarch__xtensa__core.a zephyr/lib/libc/minimal/liblib__libc__minimal.a zephyr/lib/posix/liblib__posix.a zephyr/drivers/interrupt_controller/libdrivers__interrupt_controller.a zephyr/drivers/clock_control/libdrivers__clock_control.a zephyr/drivers/console/libdrivers__console.a zephyr/drivers/gpio/libdrivers__gpio.a zephyr/drivers/i2c/libdrivers__i2c.a zephyr/drivers/serial/libdrivers__serial.a zephyr/drivers/timer/libdrivers__timer.a -Wl,--no-whole-archive zephyr/kernel/libkernel.a -L"c:/users/gouqs/.espressif/tools/zephyr/xtensa-esp32-elf/bin/../lib/gcc/xtensa-esp32-elf/8.4.0" -LD:/workspace/twatch/build/zephyr -lgcc zephyr/arch/common/libisr_tables.a -lgcc -no-pie -Wl,--gc-sections -Wl,--build-id=none -Wl,--sort-common=descending -Wl,--sort-section=alignment -Wl,-u,_OffsetAbsSyms -Wl,-u,_ConfigAbsSyms -nostdlib -static -Wl,-X -Wl,-N -Wl,--orphan-handling=warn -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/src/linker/esp32.rom.alias.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.api.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.libgcc.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.redefined.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp32/ld/esp32.peripherals.ld && cmd.exe /C "cd /D D:\workspace\twatch\build\zephyr && "C:\Program Files\CMake\bin\cmake.exe" -E echo ""
c:/users/gouqs/.espressif/tools/zephyr/xtensa-esp32-elf/bin/../lib/gcc/xtensa-esp32-elf/8.4.0/../../../../xtensa-esp32-elf/bin/ld.exe: zephyr/drivers/serial/libdrivers__serial.a(uart_esp32.c.obj):(.literal.uart_esp32_configure+0x0): undefined reference to `uart_periph_signal'
collect2.exe: error: ld returned 1 exit status
ninja: build stopped: subcommand failed.
FATAL ERROR: command exited with status 1: 'C:\Program Files\CMake\bin\cmake.EXE' --build 'D:\workspace\twatch\build'
```
**Expected behavior**
A clear and concise description of what you expected to happen.
there is no error when using T2 topology
**Impact**
**Logs and console output**
**Environment (please complete the following information):**
- OS: (Windows 11 build successful with T1 topology )
**Additional context**
|
1.0
|
esp32 build error while applicaton in T2 topology - `**Describe the bug**
build error when create an application in T2 topology based on esp32
Please also mention any information which could help others to understand
the problem you're facing:
- What target platform are you using?
usually create an application using T2 topology described in https://docs.zephyrproject.org/latest/guides/west/workspaces.html
- What have you tried to diagnose or workaround this issue?
- no idea on this
**To Reproduce**
Steps to reproduce the behavior:
1. create an application using T2 topology
you could follow the instructions bellow:
```
$ git clone https://gitee.com/fwatch/twatch
$ cd twatch
$ west init -l app/
$ west update # update modules
$ west espressif update # update submodules
```
2. build samples/hell_word
```
west build -p auto -b esp32 -s app
```
3. See error
```
[84/85] Linking C executable bootloader.elf
[85/85] Generating binary image from built executable
esptool.py v3.1-dev
Merged 1 ELF section
Generated D:/workspace/twatch/build/esp-idf/build/bootloader/bootloader.bin
[159/169] Linking C executable zephyr\zephyr_pre0.elf
FAILED: zephyr/zephyr_pre0.elf zephyr/zephyr_pre0.map D:/workspace/twatch/build/zephyr/zephyr_pre0.map
cmd.exe /C "cd . && C:\Users\gouqs\.espressif\tools\zephyr\xtensa-esp32-elf\bin\xtensa-esp32-elf-gcc.exe zephyr/CMakeFiles/zephyr_pre0.dir/misc/empty_file.c.obj -o zephyr\zephyr_pre0.elf zephyr/CMakeFiles/offsets.dir/./arch/xtensa/core/offsets/offsets.c.obj -Wl,-T zephyr/linker_zephyr_pre0.cmd -Wl,-Map=D:/workspace/twatch/build/zephyr/zephyr_pre0.map -Wl,--whole-archive app/libapp.a zephyr/libzephyr.a zephyr/arch/common/libarch__common.a zephyr/arch/arch/xtensa/core/libarch__xtensa__core.a zephyr/lib/libc/minimal/liblib__libc__minimal.a zephyr/lib/posix/liblib__posix.a zephyr/drivers/interrupt_controller/libdrivers__interrupt_controller.a zephyr/drivers/clock_control/libdrivers__clock_control.a zephyr/drivers/console/libdrivers__console.a zephyr/drivers/gpio/libdrivers__gpio.a zephyr/drivers/i2c/libdrivers__i2c.a zephyr/drivers/serial/libdrivers__serial.a zephyr/drivers/timer/libdrivers__timer.a -Wl,--no-whole-archive zephyr/kernel/libkernel.a -L"c:/users/gouqs/.espressif/tools/zephyr/xtensa-esp32-elf/bin/../lib/gcc/xtensa-esp32-elf/8.4.0" -LD:/workspace/twatch/build/zephyr -lgcc zephyr/arch/common/libisr_tables.a -lgcc -no-pie -Wl,--gc-sections -Wl,--build-id=none -Wl,--sort-common=descending -Wl,--sort-section=alignment -Wl,-u,_OffsetAbsSyms -Wl,-u,_ConfigAbsSyms -nostdlib -static -Wl,-X -Wl,-N -Wl,--orphan-handling=warn -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/src/linker/esp32.rom.alias.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.api.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.libgcc.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.redefined.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp32/ld/esp32.peripherals.ld && cmd.exe /C "cd /D D:\workspace\twatch\build\zephyr && "C:\Program Files\CMake\bin\cmake.exe" -E echo ""
c:/users/gouqs/.espressif/tools/zephyr/xtensa-esp32-elf/bin/../lib/gcc/xtensa-esp32-elf/8.4.0/../../../../xtensa-esp32-elf/bin/ld.exe: zephyr/drivers/serial/libdrivers__serial.a(uart_esp32.c.obj):(.literal.uart_esp32_configure+0x0): undefined reference to `uart_periph_signal'
collect2.exe: error: ld returned 1 exit status
ninja: build stopped: subcommand failed.
FATAL ERROR: command exited with status 1: 'C:\Program Files\CMake\bin\cmake.EXE' --build 'D:\workspace\twatch\build'
```
**Expected behavior**
A clear and concise description of what you expected to happen.
there is no error when using T2 topology
**Impact**
**Logs and console output**
**Environment (please complete the following information):**
- OS: (Windows 11 build successful with T1 topology )
**Additional context**
|
non_process
|
build error while applicaton in topology describe the bug build error when create an application in topology based on please also mention any information which could help others to understand the problem you re facing what target platform are you using usually create an application using topology described in what have you tried to diagnose or workaround this issue no idea on this to reproduce steps to reproduce the behavior create an application using topology you could follow the instructions bellow git clone cd twatch west init l app west update update modules west espressif update update submodules build samples hell word west build p auto b s app see error linking c executable bootloader elf generating binary image from built executable esptool py dev merged elf section generated d workspace twatch build esp idf build bootloader bootloader bin linking c executable zephyr zephyr elf failed zephyr zephyr elf zephyr zephyr map d workspace twatch build zephyr zephyr map cmd exe c cd c users gouqs espressif tools zephyr xtensa elf bin xtensa elf gcc exe zephyr cmakefiles zephyr dir misc empty file c obj o zephyr zephyr elf zephyr cmakefiles offsets dir arch xtensa core offsets offsets c obj wl t zephyr linker zephyr cmd wl map d workspace twatch build zephyr zephyr map wl whole archive app libapp a zephyr libzephyr a zephyr arch common libarch common a zephyr arch arch xtensa core libarch xtensa core a zephyr lib libc minimal liblib libc minimal a zephyr lib posix liblib posix a zephyr drivers interrupt controller libdrivers interrupt controller a zephyr drivers clock control libdrivers clock control a zephyr drivers console libdrivers console a zephyr drivers gpio libdrivers gpio a zephyr drivers libdrivers a zephyr drivers serial libdrivers serial a zephyr drivers timer libdrivers timer a wl no whole archive zephyr kernel libkernel a l c users gouqs espressif tools zephyr xtensa elf bin lib gcc xtensa elf ld workspace twatch build zephyr lgcc zephyr arch common libisr tables a lgcc no pie wl gc sections wl build id none wl sort common descending wl sort section alignment wl u offsetabssyms wl u configabssyms nostdlib static wl x wl n wl orphan handling warn td workspace twatch modules hal espressif zephyr src linker rom alias ld td workspace twatch modules hal espressif zephyr components esp rom ld rom ld td workspace twatch modules hal espressif zephyr components esp rom ld rom api ld td workspace twatch modules hal espressif zephyr components esp rom ld rom libgcc ld td workspace twatch modules hal espressif zephyr components esp rom ld rom redefined ld td workspace twatch modules hal espressif zephyr components ld peripherals ld cmd exe c cd d d workspace twatch build zephyr c program files cmake bin cmake exe e echo c users gouqs espressif tools zephyr xtensa elf bin lib gcc xtensa elf xtensa elf bin ld exe zephyr drivers serial libdrivers serial a uart c obj literal uart configure undefined reference to uart periph signal exe error ld returned exit status ninja build stopped subcommand failed fatal error command exited with status c program files cmake bin cmake exe build d workspace twatch build expected behavior a clear and concise description of what you expected to happen there is no error when using topology impact logs and console output environment please complete the following information os windows build successful with topology additional context
| 0
|
16,610
| 21,671,089,029
|
IssuesEvent
|
2022-05-08 00:39:42
|
veritablequandary/senatebot
|
https://api.github.com/repos/veritablequandary/senatebot
|
closed
|
[FR] Welcome Menu - Specific Guild
|
feature inprocess
|
**Description**
When users join the server and are sent the welcome menu, they should be given the option to contact the recruiters for a specific guild. That selection will notify that guild's recruiters in the recruitment channel.
**Suggested By**
Brax Nightfighter
|
1.0
|
[FR] Welcome Menu - Specific Guild - **Description**
When users join the server and are sent the welcome menu, they should be given the option to contact the recruiters for a specific guild. That selection will notify that guild's recruiters in the recruitment channel.
**Suggested By**
Brax Nightfighter
|
process
|
welcome menu specific guild description when users join the server and are sent the welcome menu they should be given the option to contact the recruiters for a specific guild that selection will notify that guild s recruiters in the recruitment channel suggested by brax nightfighter
| 1
|
361,733
| 25,350,082,628
|
IssuesEvent
|
2022-11-19 17:09:22
|
ahmed-nazir/Capstone
|
https://api.github.com/repos/ahmed-nazir/Capstone
|
closed
|
[DOC] - VnV Plan Section 6.1, 6.3
|
documentation
|
**What document(s) are being updated?**
VnVPlan.tex
**What is being added/updated? (optional)**
Brief description of documentation change
|
1.0
|
[DOC] - VnV Plan Section 6.1, 6.3 - **What document(s) are being updated?**
VnVPlan.tex
**What is being added/updated? (optional)**
Brief description of documentation change
|
non_process
|
vnv plan section what document s are being updated vnvplan tex what is being added updated optional brief description of documentation change
| 0
|
287,424
| 24,829,141,599
|
IssuesEvent
|
2022-10-26 00:37:51
|
jessicalewinter/Utilities
|
https://api.github.com/repos/jessicalewinter/Utilities
|
opened
|
Migrar módulos do +Sources/Utilities/UIKit/Extensions/UITableView+Dequeuable.swift
|
testing-modules
|
Esse módulo possui TestDoubles em sua interface e deve ser migrado para um novo módulo de Testing
|
1.0
|
Migrar módulos do +Sources/Utilities/UIKit/Extensions/UITableView+Dequeuable.swift - Esse módulo possui TestDoubles em sua interface e deve ser migrado para um novo módulo de Testing
|
non_process
|
migrar módulos do sources utilities uikit extensions uitableview dequeuable swift esse módulo possui testdoubles em sua interface e deve ser migrado para um novo módulo de testing
| 0
|
7,705
| 10,799,971,794
|
IssuesEvent
|
2019-11-06 13:21:22
|
prisma-labs/issues
|
https://api.github.com/repos/prisma-labs/issues
|
opened
|
Create a Project Statuses Board
|
type/process
|
We need a table that lists the primary projects we are or have worked on and what there status is. We sometimes get asked questions about why some project is is inactive or what its plans are.
- Make it a new notion table in the root `Labs` page
- Put it below the roadmap
- Entries should include `nexus` `nexus-prisma` `graphql-playground` `pumpkins`
- Columns should include `name: string` `status: Status` `repo: Link` `reason: string`
- `Status` can be single select of `active this quarter` | `inactive this quarter` | `indefinitely inactive`
- `Reason` should provide an explanation for the current status (if needed)
- Idea: Consider using emojis or emoji+text for each `status`. For example: 🛫|🛬|✈️ or maybe weather themed ☀️|⛈|🌤|☁️
|
1.0
|
Create a Project Statuses Board - We need a table that lists the primary projects we are or have worked on and what there status is. We sometimes get asked questions about why some project is is inactive or what its plans are.
- Make it a new notion table in the root `Labs` page
- Put it below the roadmap
- Entries should include `nexus` `nexus-prisma` `graphql-playground` `pumpkins`
- Columns should include `name: string` `status: Status` `repo: Link` `reason: string`
- `Status` can be single select of `active this quarter` | `inactive this quarter` | `indefinitely inactive`
- `Reason` should provide an explanation for the current status (if needed)
- Idea: Consider using emojis or emoji+text for each `status`. For example: 🛫|🛬|✈️ or maybe weather themed ☀️|⛈|🌤|☁️
|
process
|
create a project statuses board we need a table that lists the primary projects we are or have worked on and what there status is we sometimes get asked questions about why some project is is inactive or what its plans are make it a new notion table in the root labs page put it below the roadmap entries should include nexus nexus prisma graphql playground pumpkins columns should include name string status status repo link reason string status can be single select of active this quarter inactive this quarter indefinitely inactive reason should provide an explanation for the current status if needed idea consider using emojis or emoji text for each status for example 🛫 🛬 ✈️ or maybe weather themed ☀️ ⛈ 🌤 ☁️
| 1
|
10,643
| 13,446,196,249
|
IssuesEvent
|
2020-09-08 12:37:02
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
PARs - Allow deleting already added products & substances
|
EPIC - PARs process
|
When the user enters multiple products and/or substances they should be allowed to delete them.
Designs:
- https://app.zeplin.io/project/5dd51ae21205c944f8c1d35b/screen/5ebbfcb92c78c074bef52070
- https://app.zeplin.io/project/5dd51ae21205c944f8c1d35b/screen/5ebe9acb07baf73069b9be30
|
1.0
|
PARs - Allow deleting already added products & substances - When the user enters multiple products and/or substances they should be allowed to delete them.
Designs:
- https://app.zeplin.io/project/5dd51ae21205c944f8c1d35b/screen/5ebbfcb92c78c074bef52070
- https://app.zeplin.io/project/5dd51ae21205c944f8c1d35b/screen/5ebe9acb07baf73069b9be30
|
process
|
pars allow deleting already added products substances when the user enters multiple products and or substances they should be allowed to delete them designs
| 1
|
6,107
| 8,966,962,329
|
IssuesEvent
|
2019-01-29 01:09:38
|
knative/serving
|
https://api.github.com/repos/knative/serving
|
closed
|
Simplify the traffic.go
|
area/API kind/cleanup kind/process
|
<!--
Pro-tip: You can leave this block commented, and it still works!
Select the appropriate areas for your issue:
/area API
Classify what kind of issue this is:
/kind cleanup
/kind process
-->
## Proposal
Currently the route/traffic/traffic.go is quite complex, but it doesn't have to be so, since we only permit a single revision for a target, so the whole flattening process can be rewritten with fewer lines.
- simplify the data structures
- simplify the algorithms
- remove the unnecessary code
- fix the linter errors with stutter named structures
|
1.0
|
Simplify the traffic.go - <!--
Pro-tip: You can leave this block commented, and it still works!
Select the appropriate areas for your issue:
/area API
Classify what kind of issue this is:
/kind cleanup
/kind process
-->
## Proposal
Currently the route/traffic/traffic.go is quite complex, but it doesn't have to be so, since we only permit a single revision for a target, so the whole flattening process can be rewritten with fewer lines.
- simplify the data structures
- simplify the algorithms
- remove the unnecessary code
- fix the linter errors with stutter named structures
|
process
|
simplify the traffic go pro tip you can leave this block commented and it still works select the appropriate areas for your issue area api classify what kind of issue this is kind cleanup kind process proposal currently the route traffic traffic go is quite complex but it doesn t have to be so since we only permit a single revision for a target so the whole flattening process can be rewritten with fewer lines simplify the data structures simplify the algorithms remove the unnecessary code fix the linter errors with stutter named structures
| 1
|
7,569
| 10,684,225,854
|
IssuesEvent
|
2019-10-22 09:59:58
|
CGAL/cgal
|
https://api.github.com/repos/CGAL/cgal
|
closed
|
Bad results after MST normal orientation and Poisson reconstruction
|
Pkg::Point_set_processing_3 question
|
[Keypoints.txt](https://github.com/CGAL/cgal/files/1888743/Keypoints.txt)
_Please use the following template to help us solving your issue._
## Issue Details
I want to estimate normals of point cloud using pca_estimate_normals, but i found it generates very bad result compare with pcl using same parameters(k neighbor )
test file is attached in the attachment:please change the file extension to ply
## Source Code
CGAL::pca_estimate_normals<CGAL::Sequential_tag>(points.begin(), points.end(),
CGAL::First_of_pair_property_map<PointVectorPair>(),
CGAL::Second_of_pair_property_map<PointVectorPair>(),
nb_neighbors_pca_normals);
## Environment
* Operating system (Windows/Mac/Linux, 32/64 bits):windows 10
* Compiler: visual studio 2017
* Release or debug mode: release
* Specific flags used (if any):
* CGAL version:4.11.1
* Boost version:1.66
* Other libraries versions if used (Eigen, TBB, etc.):eigen tbb qt
|
1.0
|
Bad results after MST normal orientation and Poisson reconstruction -
[Keypoints.txt](https://github.com/CGAL/cgal/files/1888743/Keypoints.txt)
_Please use the following template to help us solving your issue._
## Issue Details
I want to estimate normals of point cloud using pca_estimate_normals, but i found it generates very bad result compare with pcl using same parameters(k neighbor )
test file is attached in the attachment:please change the file extension to ply
## Source Code
CGAL::pca_estimate_normals<CGAL::Sequential_tag>(points.begin(), points.end(),
CGAL::First_of_pair_property_map<PointVectorPair>(),
CGAL::Second_of_pair_property_map<PointVectorPair>(),
nb_neighbors_pca_normals);
## Environment
* Operating system (Windows/Mac/Linux, 32/64 bits):windows 10
* Compiler: visual studio 2017
* Release or debug mode: release
* Specific flags used (if any):
* CGAL version:4.11.1
* Boost version:1.66
* Other libraries versions if used (Eigen, TBB, etc.):eigen tbb qt
|
process
|
bad results after mst normal orientation and poisson reconstruction please use the following template to help us solving your issue issue details i want to estimate normals of point cloud using pca estimate normals but i found it generates very bad result compare with pcl using same parameters k neighbor test file is attached in the attachment please change the file extension to ply source code cgal pca estimate normals points begin points end cgal first of pair property map cgal second of pair property map nb neighbors pca normals environment operating system windows mac linux bits windows compiler visual studio release or debug mode release specific flags used if any cgal version boost version other libraries versions if used eigen tbb etc eigen tbb qt
| 1
|
9,927
| 12,964,607,225
|
IssuesEvent
|
2020-07-20 20:47:14
|
crim-ca/weaver
|
https://api.github.com/repos/crim-ca/weaver
|
opened
|
[BUG] wrong workflow_input_value usage
|
CWL OGC job process/workflow process/wps3 triage/bug
|
**Describe the bug**
`workflow_input_value` can have multiple types (eg: href, float, int) but `wps3_process.execute()` expects it to be a href.
When it isn't and href, `wps3_process.execute()` fails.
Either when trying `workflow_input_value["location"]` when `workflow_input_value` is a float (which isn't subscriptable), or later on, during the `.startswith()` validation, which fails since type mismatch.
**To Reproduce**
1. Run a workflow, for eg: `pytest tests/functional/test_ems_end2end.py -k "test_workflow_subset_picker"`
2. Notice the TypeError message in the logs (see Screenshots section)
**Expected behavior**
Having data other than `href` parsed as `data`, to avoid type mismatch.
**Screenshots**
```
weaver_1 | File "/opt/local/src/weaver/weaver/processes/wps_workflow.py", line 455, in execute
weaver_1 | self.results = self.wps_process.execute(self.builder.job, self.outdir, self.expected_outputs)
weaver_1 | File "/opt/local/src/weaver/weaver/processes/wps3_process.py", line 256, in execute
weaver_1 | execute_req_input_val: workflow_input_value["location"]})
weaver_1 | TypeError: 'float' object is not subscriptable
```
**Context (please complete the following information):**
- OS: linux
- Instance: local EMS
- Version 1.13.1
Note that although `test_workflow_subset_picker` is not currently working, in the past this test passed successfully on a local EMS.
|
2.0
|
[BUG] wrong workflow_input_value usage - **Describe the bug**
`workflow_input_value` can have multiple types (eg: href, float, int) but `wps3_process.execute()` expects it to be a href.
When it isn't and href, `wps3_process.execute()` fails.
Either when trying `workflow_input_value["location"]` when `workflow_input_value` is a float (which isn't subscriptable), or later on, during the `.startswith()` validation, which fails since type mismatch.
**To Reproduce**
1. Run a workflow, for eg: `pytest tests/functional/test_ems_end2end.py -k "test_workflow_subset_picker"`
2. Notice the TypeError message in the logs (see Screenshots section)
**Expected behavior**
Having data other than `href` parsed as `data`, to avoid type mismatch.
**Screenshots**
```
weaver_1 | File "/opt/local/src/weaver/weaver/processes/wps_workflow.py", line 455, in execute
weaver_1 | self.results = self.wps_process.execute(self.builder.job, self.outdir, self.expected_outputs)
weaver_1 | File "/opt/local/src/weaver/weaver/processes/wps3_process.py", line 256, in execute
weaver_1 | execute_req_input_val: workflow_input_value["location"]})
weaver_1 | TypeError: 'float' object is not subscriptable
```
**Context (please complete the following information):**
- OS: linux
- Instance: local EMS
- Version 1.13.1
Note that although `test_workflow_subset_picker` is not currently working, in the past this test passed successfully on a local EMS.
|
process
|
wrong workflow input value usage describe the bug workflow input value can have multiple types eg href float int but process execute expects it to be a href when it isn t and href process execute fails either when trying workflow input value when workflow input value is a float which isn t subscriptable or later on during the startswith validation which fails since type mismatch to reproduce run a workflow for eg pytest tests functional test ems py k test workflow subset picker notice the typeerror message in the logs see screenshots section expected behavior having data other than href parsed as data to avoid type mismatch screenshots weaver file opt local src weaver weaver processes wps workflow py line in execute weaver self results self wps process execute self builder job self outdir self expected outputs weaver file opt local src weaver weaver processes process py line in execute weaver execute req input val workflow input value weaver typeerror float object is not subscriptable context please complete the following information os linux instance local ems version note that although test workflow subset picker is not currently working in the past this test passed successfully on a local ems
| 1
|
385,355
| 26,633,104,122
|
IssuesEvent
|
2023-01-24 19:27:48
|
0x192/universal-android-debloater
|
https://api.github.com/repos/0x192/universal-android-debloater
|
opened
|
Samsung App Separation official docs
|
package::documentation
|
**Your phone**: \<phone model\>
**Packages documentation to update:**
```
com.samsung.android.appseparation
...
```
## Documentation change
### com.samsung.android.appseparation
**List**: Pending :arrow_right: OEM
**Removal**: Expert (no diff)
**Current description**
>
(none)
**Proposed description**
> https://docs.samsungknox.com/admin/knox-platform-for-enterprise/separated-apps.htm
I know, it's not very helpful, but it's better than nothing
|
1.0
|
Samsung App Separation official docs - **Your phone**: \<phone model\>
**Packages documentation to update:**
```
com.samsung.android.appseparation
...
```
## Documentation change
### com.samsung.android.appseparation
**List**: Pending :arrow_right: OEM
**Removal**: Expert (no diff)
**Current description**
>
(none)
**Proposed description**
> https://docs.samsungknox.com/admin/knox-platform-for-enterprise/separated-apps.htm
I know, it's not very helpful, but it's better than nothing
|
non_process
|
samsung app separation official docs your phone packages documentation to update com samsung android appseparation documentation change com samsung android appseparation list pending arrow right oem removal expert no diff current description none proposed description i know it s not very helpful but it s better than nothing
| 0
|
26,327
| 5,243,702,331
|
IssuesEvent
|
2017-01-31 21:25:53
|
symfony/symfony-docs
|
https://api.github.com/repos/symfony/symfony-docs
|
closed
|
Document the deprecation of case-insensitive service ids
|
DependencyInjection hasPR Missing Documentation
|
Deprecated in 3.3 and remove in 4.0.
See https://github.com/symfony/symfony/pull/21223 and http://symfony.com/blog/new-in-symfony-3-3-dependency-injection-deprecations
|
1.0
|
Document the deprecation of case-insensitive service ids - Deprecated in 3.3 and remove in 4.0.
See https://github.com/symfony/symfony/pull/21223 and http://symfony.com/blog/new-in-symfony-3-3-dependency-injection-deprecations
|
non_process
|
document the deprecation of case insensitive service ids deprecated in and remove in see and
| 0
|
15,326
| 19,433,139,326
|
IssuesEvent
|
2021-12-21 14:16:00
|
threefoldtech/tfchain
|
https://api.github.com/repos/threefoldtech/tfchain
|
closed
|
change uncle filter in pallet authorship config for NPOS chain
|
process_wontfix
|
FitlerUncle is currently implemented as an empty tuple `()`, which is a dummy handler. As per the pallet_authorship code docs, we should probably use the also provided `OnePerAuthorPerHeight`.
Link: https://github.com/paritytech/substrate/blob/49a4103f4bfef55be20a5c6d26e18ff3003c3353/frame/authorship/src/lib.rs#L55-L57
|
1.0
|
change uncle filter in pallet authorship config for NPOS chain - FitlerUncle is currently implemented as an empty tuple `()`, which is a dummy handler. As per the pallet_authorship code docs, we should probably use the also provided `OnePerAuthorPerHeight`.
Link: https://github.com/paritytech/substrate/blob/49a4103f4bfef55be20a5c6d26e18ff3003c3353/frame/authorship/src/lib.rs#L55-L57
|
process
|
change uncle filter in pallet authorship config for npos chain fitleruncle is currently implemented as an empty tuple which is a dummy handler as per the pallet authorship code docs we should probably use the also provided oneperauthorperheight link
| 1
|
29,234
| 23,837,141,723
|
IssuesEvent
|
2022-09-06 07:11:01
|
woocommerce/woocommerce
|
https://api.github.com/repos/woocommerce/woocommerce
|
opened
|
RFC: noUnusedParameters yes/no?
|
needs: discussion tool: monorepo infrastructure
|
## Background
In the root directory's `tsconfig.base.json` we have the [noUnusedParameters](https://www.typescriptlang.org/tsconfig#noUnusedParameters) turned on - however as of present moment this rule is not applied to woocommerce-admin or the packages since they have their own `tsconfig.json` which do not extend from the root `tsconfig.base.json`
When this rule is applied, there were some violations of this rule, which look like:

In most instances, it can be inferred that the unused parameter serves as documentation for potential future usage.
Is there a better solution aside from
1. Disabling the rule entirely
2. Overriding each instance with a @ts-ignore
|
1.0
|
RFC: noUnusedParameters yes/no? - ## Background
In the root directory's `tsconfig.base.json` we have the [noUnusedParameters](https://www.typescriptlang.org/tsconfig#noUnusedParameters) turned on - however as of present moment this rule is not applied to woocommerce-admin or the packages since they have their own `tsconfig.json` which do not extend from the root `tsconfig.base.json`
When this rule is applied, there were some violations of this rule, which look like:

In most instances, it can be inferred that the unused parameter serves as documentation for potential future usage.
Is there a better solution aside from
1. Disabling the rule entirely
2. Overriding each instance with a @ts-ignore
|
non_process
|
rfc nounusedparameters yes no background in the root directory s tsconfig base json we have the turned on however as of present moment this rule is not applied to woocommerce admin or the packages since they have their own tsconfig json which do not extend from the root tsconfig base json when this rule is applied there were some violations of this rule which look like in most instances it can be inferred that the unused parameter serves as documentation for potential future usage is there a better solution aside from disabling the rule entirely overriding each instance with a ts ignore
| 0
|
12,778
| 15,163,015,843
|
IssuesEvent
|
2021-02-12 11:30:55
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
closed
|
Create interface(s) for Flue Gas property calculation methods
|
Calculator Process Heating
|
Calculations methods for O2 and Excess Air take an untyped/any input object. Needs a joint interface or separate for each (Mass/Volume) input. Methods also exist in Air-heating calc (see for example).

|
1.0
|
Create interface(s) for Flue Gas property calculation methods - Calculations methods for O2 and Excess Air take an untyped/any input object. Needs a joint interface or separate for each (Mass/Volume) input. Methods also exist in Air-heating calc (see for example).

|
process
|
create interface s for flue gas property calculation methods calculations methods for and excess air take an untyped any input object needs a joint interface or separate for each mass volume input methods also exist in air heating calc see for example
| 1
|
757,780
| 26,528,071,769
|
IssuesEvent
|
2023-01-19 10:17:56
|
consta-design-system/uikit
|
https://api.github.com/repos/consta-design-system/uikit
|
opened
|
Table: Вертикальные связи между уровнями
|
feature 🔥🔥 priority
|
У нас возникла следующая потребность: в иерархических таблицах нам очень нужны вертикальные связи между уровнями таблиц. Желаемый вид отобразили на макете, скриншот во вложении.

|
1.0
|
Table: Вертикальные связи между уровнями - У нас возникла следующая потребность: в иерархических таблицах нам очень нужны вертикальные связи между уровнями таблиц. Желаемый вид отобразили на макете, скриншот во вложении.

|
non_process
|
table вертикальные связи между уровнями у нас возникла следующая потребность в иерархических таблицах нам очень нужны вертикальные связи между уровнями таблиц желаемый вид отобразили на макете скриншот во вложении
| 0
|
21,297
| 28,495,290,970
|
IssuesEvent
|
2023-04-18 13:50:40
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
System hangs when a php interpreter is executed via Process
|
Bug Process Status: Needs Review
|
### Symfony version(s) affected
6.2.6
### Description
Server setup:
PHP: 8.2.2
Server API: CGI/FastCGI
When I run a PHP script via Process, Symfony creates lots of `/usr/lib/cgi-bin/php8.2` processes, which consume all RAM, server hangs and the OOM killer terminates all Apache instances.
### How to reproduce
1. Create a test project:
```
composer create-project symfony/skeleton:"6.2.*" testing
cd testing
composer require symfony/apache-pack
composer require symfony/process
```
2. `config/routes.yaml` :
```
controllers:
resource:
path: ../src/Controller/
namespace: App\Controller
type: attribute
test:
path: /test
controller: App\Controller\Test::test
```
3. `src/Controller/Test.php`:
```
<?php
namespace App\Controller;
use Symfony\Component\HttpFoundation\Response;
use Symfony\Component\Process\PhpExecutableFinder;
use Symfony\Component\Process\Exception\ProcessFailedException;
use Symfony\Component\Process\Process;
class Test
{
public function test(): Response
{
$phpBinaryPath = (new PhpExecutableFinder())->find();
$process = new Process([$phpBinaryPath, './test.php']);
try {
$process->run();
if (!$process->isSuccessful()) {
throw new ProcessFailedException($process);
}
} catch (\Exception $e) {
throw new ProcessFailedException($process);
}
return new Response(
'result: ' . $process->getOutput()
);
}
}
```
4. `public/test.php`:
```
<?php echo 'OK';
```
Now open the path `/test` in your web browser.
Symfony starts lots of `/usr/lib/cgi-bin/php8.2` processes, which consume all memory and CPU.
I supposed that the issue might be related to HTTP headers, generated by the CGI version.
However, this does not help - `new Process([$phpBinaryPath, '-q', './test.php'])`.
If I run either command directly on the server, all is good - `/usr/lib/cgi-bin/php8.2 ./test.php` / `/usr/lib/cgi-bin/php8.2 -q ./test.php`.
If I hardcode the path to CLI PHP (`$phpBinaryPath = '/bin/php';`) that test script works correctly.
### Possible Solution
`PhpExecutableFinder` should probably return the CLI version instead of CGI.
### Additional Context
_No response_
|
1.0
|
System hangs when a php interpreter is executed via Process - ### Symfony version(s) affected
6.2.6
### Description
Server setup:
PHP: 8.2.2
Server API: CGI/FastCGI
When I run a PHP script via Process, Symfony creates lots of `/usr/lib/cgi-bin/php8.2` processes, which consume all RAM, server hangs and the OOM killer terminates all Apache instances.
### How to reproduce
1. Create a test project:
```
composer create-project symfony/skeleton:"6.2.*" testing
cd testing
composer require symfony/apache-pack
composer require symfony/process
```
2. `config/routes.yaml` :
```
controllers:
resource:
path: ../src/Controller/
namespace: App\Controller
type: attribute
test:
path: /test
controller: App\Controller\Test::test
```
3. `src/Controller/Test.php`:
```
<?php
namespace App\Controller;
use Symfony\Component\HttpFoundation\Response;
use Symfony\Component\Process\PhpExecutableFinder;
use Symfony\Component\Process\Exception\ProcessFailedException;
use Symfony\Component\Process\Process;
class Test
{
public function test(): Response
{
$phpBinaryPath = (new PhpExecutableFinder())->find();
$process = new Process([$phpBinaryPath, './test.php']);
try {
$process->run();
if (!$process->isSuccessful()) {
throw new ProcessFailedException($process);
}
} catch (\Exception $e) {
throw new ProcessFailedException($process);
}
return new Response(
'result: ' . $process->getOutput()
);
}
}
```
4. `public/test.php`:
```
<?php echo 'OK';
```
Now open the path `/test` in your web browser.
Symfony starts lots of `/usr/lib/cgi-bin/php8.2` processes, which consume all memory and CPU.
I supposed that the issue might be related to HTTP headers, generated by the CGI version.
However, this does not help - `new Process([$phpBinaryPath, '-q', './test.php'])`.
If I run either command directly on the server, all is good - `/usr/lib/cgi-bin/php8.2 ./test.php` / `/usr/lib/cgi-bin/php8.2 -q ./test.php`.
If I hardcode the path to CLI PHP (`$phpBinaryPath = '/bin/php';`) that test script works correctly.
### Possible Solution
`PhpExecutableFinder` should probably return the CLI version instead of CGI.
### Additional Context
_No response_
|
process
|
system hangs when a php interpreter is executed via process symfony version s affected description server setup php server api cgi fastcgi when i run a php script via process symfony creates lots of usr lib cgi bin processes which consume all ram server hangs and the oom killer terminates all apache instances how to reproduce create a test project composer create project symfony skeleton testing cd testing composer require symfony apache pack composer require symfony process config routes yaml controllers resource path src controller namespace app controller type attribute test path test controller app controller test test src controller test php php namespace app controller use symfony component httpfoundation response use symfony component process phpexecutablefinder use symfony component process exception processfailedexception use symfony component process process class test public function test response phpbinarypath new phpexecutablefinder find process new process try process run if process issuccessful throw new processfailedexception process catch exception e throw new processfailedexception process return new response result process getoutput public test php php echo ok now open the path test in your web browser symfony starts lots of usr lib cgi bin processes which consume all memory and cpu i supposed that the issue might be related to http headers generated by the cgi version however this does not help new process if i run either command directly on the server all is good usr lib cgi bin test php usr lib cgi bin q test php if i hardcode the path to cli php phpbinarypath bin php that test script works correctly possible solution phpexecutablefinder should probably return the cli version instead of cgi additional context no response
| 1
|
17,184
| 22,766,639,236
|
IssuesEvent
|
2022-07-08 05:31:24
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
opened
|
Obsoletion notice: GO:0039611 suppression by virus of host translation initiation factor activity
|
obsoletion multi-species process
|
Dear all,
The proposal has been made to obsolete GO:0039611 suppression by virus of host translation initiation factor activity. The reason for obsoletion is that this a a molecular function represented as a biological process.
The term will be replaced by 'GO:0039606 suppression by virus of host translation initiation' . There were 5 experimental annotations (updated). There is one keyword mapping that can also be moved to 'GO:0039606 suppression by virus of host translation initiation', KW-1075 Inhibition of host translation factors by virus. This term is not in any subsets.
You can comment on the ticket:
Thanks, Pascale
|
1.0
|
Obsoletion notice: GO:0039611 suppression by virus of host translation initiation factor activity - Dear all,
The proposal has been made to obsolete GO:0039611 suppression by virus of host translation initiation factor activity. The reason for obsoletion is that this a a molecular function represented as a biological process.
The term will be replaced by 'GO:0039606 suppression by virus of host translation initiation' . There were 5 experimental annotations (updated). There is one keyword mapping that can also be moved to 'GO:0039606 suppression by virus of host translation initiation', KW-1075 Inhibition of host translation factors by virus. This term is not in any subsets.
You can comment on the ticket:
Thanks, Pascale
|
process
|
obsoletion notice go suppression by virus of host translation initiation factor activity dear all the proposal has been made to obsolete go suppression by virus of host translation initiation factor activity the reason for obsoletion is that this a a molecular function represented as a biological process the term will be replaced by go suppression by virus of host translation initiation there were experimental annotations updated there is one keyword mapping that can also be moved to go suppression by virus of host translation initiation kw inhibition of host translation factors by virus this term is not in any subsets you can comment on the ticket thanks pascale
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.