Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
343,050
30,650,980,658
IssuesEvent
2023-07-25 08:53:33
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix device.test_gpu_is_available
Sub Task Failing Test
| | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix device.test_gpu_is_available - | | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5654611290/job/15318026774"><img src=https://img.shields.io/badge/-success-success></a>
non_process
fix device test gpu is available jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
0
586
3,062,240,628
IssuesEvent
2015-08-16 11:34:49
yeoman/yeoman-app
https://api.github.com/repos/yeoman/yeoman-app
closed
Radio buttons not working for me
bug renderer process
All of the radio buttons generated from prompts `type: list` are not working with the following JS error in the console: ``` [6016:0726/125628:INFO:CONSOLE(28)] "Uncaught TypeError: Cannot read property 'value' of null", source: .../yeoman-app/src/renderer/components/prompts/list.jsx (28) ```
1.0
Radio buttons not working for me - All of the radio buttons generated from prompts `type: list` are not working with the following JS error in the console: ``` [6016:0726/125628:INFO:CONSOLE(28)] "Uncaught TypeError: Cannot read property 'value' of null", source: .../yeoman-app/src/renderer/components/prompts/list.jsx (28) ```
process
radio buttons not working for me all of the radio buttons generated from prompts type list are not working with the following js error in the console uncaught typeerror cannot read property value of null source yeoman app src renderer components prompts list jsx
1
107
2,545,536,557
IssuesEvent
2015-01-29 17:44:28
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
opened
Remove Neo4jGraphTraversal.
enhancement neo4j process
The only reason we need Neo4jGraphTraversal is cause of mid-traversal `cypher()`. Because of this, we have lots of overhead -- source code generation, some tests can't run because it method counts are off, etc. Also, lots of Neo4j code will go away .... e.g. Neo4jElementTraversal, Neo4jVertexTraversal, Neo4jGraphTraversals, Neo4jDefualGraphTraversal, Neo4jEdgeTravaersal, Neo4jVertexPropertyTraversal.... I'm wondering if we get rid of `g.V().out().id().cypher('MATCH...').select('a').outE()` in favor of JUST g.cypher(MATCH).select('a'). In short, cypher() is a method off `Neo4jGraph` and yields a `GraphTraversal`. Thoughts?
1.0
Remove Neo4jGraphTraversal. - The only reason we need Neo4jGraphTraversal is cause of mid-traversal `cypher()`. Because of this, we have lots of overhead -- source code generation, some tests can't run because it method counts are off, etc. Also, lots of Neo4j code will go away .... e.g. Neo4jElementTraversal, Neo4jVertexTraversal, Neo4jGraphTraversals, Neo4jDefualGraphTraversal, Neo4jEdgeTravaersal, Neo4jVertexPropertyTraversal.... I'm wondering if we get rid of `g.V().out().id().cypher('MATCH...').select('a').outE()` in favor of JUST g.cypher(MATCH).select('a'). In short, cypher() is a method off `Neo4jGraph` and yields a `GraphTraversal`. Thoughts?
process
remove the only reason we need is cause of mid traversal cypher because of this we have lots of overhead source code generation some tests can t run because it method counts are off etc also lots of code will go away e g i m wondering if we get rid of g v out id cypher match select a oute in favor of just g cypher match select a in short cypher is a method off and yields a graphtraversal thoughts
1
19,618
25,970,985,497
IssuesEvent
2022-12-19 11:11:55
googleapis/gapic-generator-csharp
https://api.github.com/repos/googleapis/gapic-generator-csharp
closed
Security Policy violation Binary Artifacts
type: process allstar
_This issue was automatically created by [Allstar](https://github.com/ossf/allstar/)._ **Security Policy Violation** Project is out of compliance with Binary Artifacts policy: binaries present in source code **Rule Description** Binary Artifacts are an increased security risk in your repository. Binary artifacts cannot be reviewed, allowing the introduction of possibly obsolete or maliciously subverted executables. For more information see the [Security Scorecards Documentation](https://github.com/ossf/scorecard/blob/main/docs/checks.md#binary-artifacts) for Binary Artifacts. **Remediation Steps** To remediate, remove the generated executable artifacts from the repository. **Artifacts Found** - tools/grpc_csharp_plugin - tools/grpc_csharp_plugin.exe - tools/protoc - tools/protoc.exe **Additional Information** This policy is drawn from [Security Scorecards](https://github.com/ossf/scorecard/), which is a tool that scores a project's adherence to security best practices. You may wish to run a Scorecards scan directly on this repository for more details. --- <!-- Edit section #updates --><!-- Current result text hash: 1fb9f938741aaa770502d7faa1836bd83fd751f88f55304b5deaf90bc8fae266 --><!-- Edit section #updates --> Allstar has been installed on all Google managed GitHub orgs. Policies are gradually being rolled out and enforced by the GOSST and OSPO teams. Learn more at http://go/allstar This issue will auto resolve when the policy is in compliance. Issue created by Allstar. See https://github.com/ossf/allstar/ for more information. For questions specific to the repository, please contact the owner or maintainer.
1.0
Security Policy violation Binary Artifacts - _This issue was automatically created by [Allstar](https://github.com/ossf/allstar/)._ **Security Policy Violation** Project is out of compliance with Binary Artifacts policy: binaries present in source code **Rule Description** Binary Artifacts are an increased security risk in your repository. Binary artifacts cannot be reviewed, allowing the introduction of possibly obsolete or maliciously subverted executables. For more information see the [Security Scorecards Documentation](https://github.com/ossf/scorecard/blob/main/docs/checks.md#binary-artifacts) for Binary Artifacts. **Remediation Steps** To remediate, remove the generated executable artifacts from the repository. **Artifacts Found** - tools/grpc_csharp_plugin - tools/grpc_csharp_plugin.exe - tools/protoc - tools/protoc.exe **Additional Information** This policy is drawn from [Security Scorecards](https://github.com/ossf/scorecard/), which is a tool that scores a project's adherence to security best practices. You may wish to run a Scorecards scan directly on this repository for more details. --- <!-- Edit section #updates --><!-- Current result text hash: 1fb9f938741aaa770502d7faa1836bd83fd751f88f55304b5deaf90bc8fae266 --><!-- Edit section #updates --> Allstar has been installed on all Google managed GitHub orgs. Policies are gradually being rolled out and enforced by the GOSST and OSPO teams. Learn more at http://go/allstar This issue will auto resolve when the policy is in compliance. Issue created by Allstar. See https://github.com/ossf/allstar/ for more information. For questions specific to the repository, please contact the owner or maintainer.
process
security policy violation binary artifacts this issue was automatically created by security policy violation project is out of compliance with binary artifacts policy binaries present in source code rule description binary artifacts are an increased security risk in your repository binary artifacts cannot be reviewed allowing the introduction of possibly obsolete or maliciously subverted executables for more information see the for binary artifacts remediation steps to remediate remove the generated executable artifacts from the repository artifacts found tools grpc csharp plugin tools grpc csharp plugin exe tools protoc tools protoc exe additional information this policy is drawn from which is a tool that scores a project s adherence to security best practices you may wish to run a scorecards scan directly on this repository for more details allstar has been installed on all google managed github orgs policies are gradually being rolled out and enforced by the gosst and ospo teams learn more at this issue will auto resolve when the policy is in compliance issue created by allstar see for more information for questions specific to the repository please contact the owner or maintainer
1
18,704
24,598,953,416
IssuesEvent
2022-10-14 10:44:30
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
[postgresqlExtensions] consider a destructive change check for dropped-and-recreated extensions in Migrate
process/candidate kind/improvement topic: migrate tech/engines/migration engine team/schema topic: postgresqlExtensions
Extensions create schema items. Dropping an extension can cause data loss. We drop extensions when we want to relocate non-relocatable extensions. However, this is safe for _most_ extensions, that do not create tables where users are supposed to insert data.
1.0
[postgresqlExtensions] consider a destructive change check for dropped-and-recreated extensions in Migrate - Extensions create schema items. Dropping an extension can cause data loss. We drop extensions when we want to relocate non-relocatable extensions. However, this is safe for _most_ extensions, that do not create tables where users are supposed to insert data.
process
consider a destructive change check for dropped and recreated extensions in migrate extensions create schema items dropping an extension can cause data loss we drop extensions when we want to relocate non relocatable extensions however this is safe for most extensions that do not create tables where users are supposed to insert data
1
4,983
7,816,432,955
IssuesEvent
2018-06-13 04:39:59
h-a-m-r-time/grace-shopper
https://api.github.com/repos/h-a-m-r-time/grace-shopper
closed
Deploy Research/Discussion
process
Discuss who's responsiblity to keep deploys going forward, maybe setting up a rolling automated overnight deploy plus allowing us to kick off a deploy whenever necessary
1.0
Deploy Research/Discussion - Discuss who's responsiblity to keep deploys going forward, maybe setting up a rolling automated overnight deploy plus allowing us to kick off a deploy whenever necessary
process
deploy research discussion discuss who s responsiblity to keep deploys going forward maybe setting up a rolling automated overnight deploy plus allowing us to kick off a deploy whenever necessary
1
48,667
25,750,301,468
IssuesEvent
2022-12-08 12:46:57
AutomatedProcessImprovement/Simod
https://api.github.com/repos/AutomatedProcessImprovement/Simod
closed
Remove time.sleep from parallel processing in the utilities module
performance experiment
This line, https://github.com/AutomatedProcessImprovement/Simod/blob/master/src/simod/utilities.py#L143, contributes to about 15% of running time of Simod.
True
Remove time.sleep from parallel processing in the utilities module - This line, https://github.com/AutomatedProcessImprovement/Simod/blob/master/src/simod/utilities.py#L143, contributes to about 15% of running time of Simod.
non_process
remove time sleep from parallel processing in the utilities module this line contributes to about of running time of simod
0
17,978
23,995,303,975
IssuesEvent
2022-09-14 07:00:46
ankidroid/Anki-Android
https://api.github.com/repos/ankidroid/Anki-Android
closed
Tests: Use hamcrest.io matchers for files
Priority-Low Good First Issue! Stale Test process
http://hamcrest.org/JavaHamcrest/javadoc/2.2/org/hamcrest/io/FileMatchers.html is now usable in the codebase. Let's use it to improve tests. <details><summary>old</summary> I'd really like to see more custom Hamcrest matchers in the codebase For example: "file exists" would be useful, so we can reduce the verbosity of error messages. And something along the lines of `hasSingleElementEqualTo` _Originally posted by @david-allison in https://github.com/ankidroid/Anki-Android/issues/10385#issuecomment-1049199502_ </details>
1.0
Tests: Use hamcrest.io matchers for files - http://hamcrest.org/JavaHamcrest/javadoc/2.2/org/hamcrest/io/FileMatchers.html is now usable in the codebase. Let's use it to improve tests. <details><summary>old</summary> I'd really like to see more custom Hamcrest matchers in the codebase For example: "file exists" would be useful, so we can reduce the verbosity of error messages. And something along the lines of `hasSingleElementEqualTo` _Originally posted by @david-allison in https://github.com/ankidroid/Anki-Android/issues/10385#issuecomment-1049199502_ </details>
process
tests use hamcrest io matchers for files is now usable in the codebase let s use it to improve tests old i d really like to see more custom hamcrest matchers in the codebase for example file exists would be useful so we can reduce the verbosity of error messages and something along the lines of hassingleelementequalto originally posted by david allison in
1
18,122
24,158,121,879
IssuesEvent
2022-09-22 09:22:07
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Processing algorithm "Split with lines" does not split polygon
Feedback Processing Bug
### What is the bug or the crash? When using the processing algorithm "Split with lines" to split poylgon with certain line data, it does not split polygon at all. Here is a simple (topologically correct) line layer to test the bug with: [lines_simplified.zip](https://github.com/qgis/QGIS/files/9581708/lines_simplified.zip) ### Steps to reproduce the issue 1. Add lines_simplified.json to the project ![image](https://user-images.githubusercontent.com/33314057/190583751-450e71d4-f6a7-4a90-b9ab-fbfccebb0269.png) 2. Create polygon from layer's extent (with alg "Polygon from layer extent") ![image](https://user-images.githubusercontent.com/33314057/190583872-abb1be63-95db-4416-befa-bc21df36f129.png) 3. Run algorithm "Split with lines" using the extent layer and the line layer and get the following output ![image](https://user-images.githubusercontent.com/33314057/190584560-652cc702-f16b-4c30-a76e-3c7dec85684b.png) The output has actually two polyons. One large one, and one tiny one of the upper right corner ![image](https://user-images.githubusercontent.com/33314057/190584795-151fd15f-5785-44fd-86bf-05d77075616c.png) ### Versions QGISin versio 3.22.9-Białowieża QGIS-koodin versio a8e9e6fa Qt:n versio 5.15.3 Python-versio 3.9.5 GDAL/OGR:n versio 3.5.1 PROJ-versio 9.0.1 EPSG-rekisteritietokannan versio v10.064 (2022-05-19) GEOS-versio 3.10.3-CAPI-1.16.1 SQLite-versio 3.38.1 PDAL-versio 2.3.0 PostgreSQL:n asiakkaan versio 14.3 SpatiaLite-versio 5.0.1 QWT-versio 6.1.6 QScintilla2-versio 2.13.1 Käyttöjärjestelmän versio Windows 10 Version 2009 Aktiiviset python-lisäosat db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 plugin_reloader 0.9.1 Ilre 0.1.0 Also tested with QGIS 3.26.2 on Arch Linux ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context The problem might lie with the dataset used. The same dataset causes the problems in #50226
1.0
Processing algorithm "Split with lines" does not split polygon - ### What is the bug or the crash? When using the processing algorithm "Split with lines" to split poylgon with certain line data, it does not split polygon at all. Here is a simple (topologically correct) line layer to test the bug with: [lines_simplified.zip](https://github.com/qgis/QGIS/files/9581708/lines_simplified.zip) ### Steps to reproduce the issue 1. Add lines_simplified.json to the project ![image](https://user-images.githubusercontent.com/33314057/190583751-450e71d4-f6a7-4a90-b9ab-fbfccebb0269.png) 2. Create polygon from layer's extent (with alg "Polygon from layer extent") ![image](https://user-images.githubusercontent.com/33314057/190583872-abb1be63-95db-4416-befa-bc21df36f129.png) 3. Run algorithm "Split with lines" using the extent layer and the line layer and get the following output ![image](https://user-images.githubusercontent.com/33314057/190584560-652cc702-f16b-4c30-a76e-3c7dec85684b.png) The output has actually two polyons. One large one, and one tiny one of the upper right corner ![image](https://user-images.githubusercontent.com/33314057/190584795-151fd15f-5785-44fd-86bf-05d77075616c.png) ### Versions QGISin versio 3.22.9-Białowieża QGIS-koodin versio a8e9e6fa Qt:n versio 5.15.3 Python-versio 3.9.5 GDAL/OGR:n versio 3.5.1 PROJ-versio 9.0.1 EPSG-rekisteritietokannan versio v10.064 (2022-05-19) GEOS-versio 3.10.3-CAPI-1.16.1 SQLite-versio 3.38.1 PDAL-versio 2.3.0 PostgreSQL:n asiakkaan versio 14.3 SpatiaLite-versio 5.0.1 QWT-versio 6.1.6 QScintilla2-versio 2.13.1 Käyttöjärjestelmän versio Windows 10 Version 2009 Aktiiviset python-lisäosat db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 plugin_reloader 0.9.1 Ilre 0.1.0 Also tested with QGIS 3.26.2 on Arch Linux ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context The problem might lie with the dataset used. The same dataset causes the problems in #50226
process
processing algorithm split with lines does not split polygon what is the bug or the crash when using the processing algorithm split with lines to split poylgon with certain line data it does not split polygon at all here is a simple topologically correct line layer to test the bug with steps to reproduce the issue add lines simplified json to the project create polygon from layer s extent with alg polygon from layer extent run algorithm split with lines using the extent layer and the line layer and get the following output the output has actually two polyons one large one and one tiny one of the upper right corner versions qgisin versio białowieża qgis koodin versio qt n versio python versio gdal ogr n versio proj versio epsg rekisteritietokannan versio geos versio capi sqlite versio pdal versio postgresql n asiakkaan versio spatialite versio qwt versio versio käyttöjärjestelmän versio windows version aktiiviset python lisäosat db manager grassprovider metasearch processing sagaprovider plugin reloader ilre also tested with qgis on arch linux supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context the problem might lie with the dataset used the same dataset causes the problems in
1
12,133
14,740,951,908
IssuesEvent
2021-01-07 09:52:32
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Customer / Client Portal Management
anc-process anp-not prioritized ant-child/secondary ant-enhancement customer portal
In GitLab by @kdjstudios on Dec 14, 2018, 08:43 **Submitted by:** Kyle **Helpdesk:** NA **Server:** ALL **Client/Site:** All **Account:** NA **Issue:** We need a better way to manage and maintain client portal logins. - Currently if a user needs to switch from managing one customer to another there is no way for operations to remove a login from the UI. - Also merging multiple customer logins into one user portal is difficult too and needs to be improved. #576 #1065 (Mock up #597) - There is no easy way for a customer/client to change their password if they forgot it. (Currently a new invite has to be sent) #1100
1.0
Customer / Client Portal Management - In GitLab by @kdjstudios on Dec 14, 2018, 08:43 **Submitted by:** Kyle **Helpdesk:** NA **Server:** ALL **Client/Site:** All **Account:** NA **Issue:** We need a better way to manage and maintain client portal logins. - Currently if a user needs to switch from managing one customer to another there is no way for operations to remove a login from the UI. - Also merging multiple customer logins into one user portal is difficult too and needs to be improved. #576 #1065 (Mock up #597) - There is no easy way for a customer/client to change their password if they forgot it. (Currently a new invite has to be sent) #1100
process
customer client portal management in gitlab by kdjstudios on dec submitted by kyle helpdesk na server all client site all account na issue we need a better way to manage and maintain client portal logins currently if a user needs to switch from managing one customer to another there is no way for operations to remove a login from the ui also merging multiple customer logins into one user portal is difficult too and needs to be improved mock up there is no easy way for a customer client to change their password if they forgot it currently a new invite has to be sent
1
249,025
26,875,573,761
IssuesEvent
2023-02-05 01:05:28
gabriel-milan/nginx-ingress-django-auth
https://api.github.com/repos/gabriel-milan/nginx-ingress-django-auth
opened
CVE-2023-23969 (Medium) detected in Django-3.2.13-py3-none-any.whl
security vulnerability
## CVE-2023-23969 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-3.2.13-py3-none-any.whl</b></p></summary> <p>A high-level Python web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl">https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl</a></p> <p> Dependency Hierarchy: - :x: **Django-3.2.13-py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Django 3.2 before 3.2.17, 4.0 before 4.0.9, and 4.1 before 4.1.6, the parsed values of Accept-Language headers are cached in order to avoid repetitive parsing. This leads to a potential denial-of-service vector via excessive memory usage if the raw value of Accept-Language headers is very large. <p>Publish Date: 2023-02-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23969>CVE-2023-23969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.djangoproject.com/weblog/2023/feb/01/security-releases/">https://www.djangoproject.com/weblog/2023/feb/01/security-releases/</a></p> <p>Release Date: 2023-02-01</p> <p>Fix Resolution: django - 3.2.17, 4.0.9, 4.1.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-23969 (Medium) detected in Django-3.2.13-py3-none-any.whl - ## CVE-2023-23969 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-3.2.13-py3-none-any.whl</b></p></summary> <p>A high-level Python web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl">https://files.pythonhosted.org/packages/c3/68/b63abc009adee25c1bd266b3740e173eab656608cf21641594a37e02cf57/Django-3.2.13-py3-none-any.whl</a></p> <p> Dependency Hierarchy: - :x: **Django-3.2.13-py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Django 3.2 before 3.2.17, 4.0 before 4.0.9, and 4.1 before 4.1.6, the parsed values of Accept-Language headers are cached in order to avoid repetitive parsing. This leads to a potential denial-of-service vector via excessive memory usage if the raw value of Accept-Language headers is very large. <p>Publish Date: 2023-02-01 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-23969>CVE-2023-23969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.djangoproject.com/weblog/2023/feb/01/security-releases/">https://www.djangoproject.com/weblog/2023/feb/01/security-releases/</a></p> <p>Release Date: 2023-02-01</p> <p>Fix Resolution: django - 3.2.17, 4.0.9, 4.1.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in django none any whl cve medium severity vulnerability vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href dependency hierarchy x django none any whl vulnerable library found in base branch master vulnerability details in django before before and before the parsed values of accept language headers are cached in order to avoid repetitive parsing this leads to a potential denial of service vector via excessive memory usage if the raw value of accept language headers is very large publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution django step up your open source security game with mend
0
18,313
2,615,168,421
IssuesEvent
2015-03-01 06:49:55
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
Cutting v7 for www
auto-migrated Milestone-7 Priority-P0 Type-Task
``` Revision was: http://code.google.com/p/html5rocks/source/detail?r=098033a3cfc174cc447581e16256 874afd91570a&name=v4 Only www was pushed. ``` Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 5 Aug 2011 at 7:01
1.0
Cutting v7 for www - ``` Revision was: http://code.google.com/p/html5rocks/source/detail?r=098033a3cfc174cc447581e16256 874afd91570a&name=v4 Only www was pushed. ``` Original issue reported on code.google.com by `ericbide...@html5rocks.com` on 5 Aug 2011 at 7:01
non_process
cutting for www revision was name only www was pushed original issue reported on code google com by ericbide com on aug at
0
93,575
26,988,813,137
IssuesEvent
2023-02-09 18:07:47
bitcoin/bitcoin
https://api.github.com/repos/bitcoin/bitcoin
closed
macos depends build does not cache in the CI
Bug Build system
At least in the CI, the depends build does not cache. E.g. https://cirrus-ci.com/task/5094611354386432?logs=ci#L789 This causes the build to take an hour when it should be done in less than 10 minutes witch a ccache.
1.0
macos depends build does not cache in the CI - At least in the CI, the depends build does not cache. E.g. https://cirrus-ci.com/task/5094611354386432?logs=ci#L789 This causes the build to take an hour when it should be done in less than 10 minutes witch a ccache.
non_process
macos depends build does not cache in the ci at least in the ci the depends build does not cache e g this causes the build to take an hour when it should be done in less than minutes witch a ccache
0
181,869
14,891,370,600
IssuesEvent
2021-01-21 00:42:24
lightningnetwork/lnd
https://api.github.com/repos/lightningnetwork/lnd
closed
Docker image not exists
docker documentation
The [DOCKER Readme](https://github.com/lightningnetwork/lnd/blob/master/docs/DOCKER.md) specifies to use the [lightninglabs/lnd](https://hub.docker.com/r/lightninglabs/lnd/) image, but it currently does not exist on dockerhub.
1.0
Docker image not exists - The [DOCKER Readme](https://github.com/lightningnetwork/lnd/blob/master/docs/DOCKER.md) specifies to use the [lightninglabs/lnd](https://hub.docker.com/r/lightninglabs/lnd/) image, but it currently does not exist on dockerhub.
non_process
docker image not exists the specifies to use the image but it currently does not exist on dockerhub
0
299,899
25,936,193,734
IssuesEvent
2022-12-16 14:22:14
DucTrann1310/FeedbackOnline
https://api.github.com/repos/DucTrann1310/FeedbackOnline
opened
[BugID_34]_FUNC_Quản lý topic_Tạo mới topic_Tạo mới topic thành công khi tên topic đã tồn tại trong hệ thống
bug Open Integration Test Fun_Wrong Business logic Priority_Medium Severity_Medium
Precondition : - Admin đang ở màn hình tạo mới Topic - Tồn tại một record với TenTopic = 'Hoangnhi' trong DB table Topic Step: 1. Nhập "Hoangnhi" vào [Tên Topic] textbox 2. Click [Hoàn tất] button Actual output: 1. Quay về màn hình Quản lý topic 2. Topic vừa tạo mới lưu ở DB Topic và refresh grid Expected output: Hiển thị inline message màu đỏ"Tên Topic đã tồn tại!" phía dưới trường Tên Topic và set focus tại đó
1.0
[BugID_34]_FUNC_Quản lý topic_Tạo mới topic_Tạo mới topic thành công khi tên topic đã tồn tại trong hệ thống - Precondition : - Admin đang ở màn hình tạo mới Topic - Tồn tại một record với TenTopic = 'Hoangnhi' trong DB table Topic Step: 1. Nhập "Hoangnhi" vào [Tên Topic] textbox 2. Click [Hoàn tất] button Actual output: 1. Quay về màn hình Quản lý topic 2. Topic vừa tạo mới lưu ở DB Topic và refresh grid Expected output: Hiển thị inline message màu đỏ"Tên Topic đã tồn tại!" phía dưới trường Tên Topic và set focus tại đó
non_process
func quản lý topic tạo mới topic tạo mới topic thành công khi tên topic đã tồn tại trong hệ thống precondition admin đang ở màn hình tạo mới topic tồn tại một record với tentopic hoangnhi trong db table topic step nhập hoangnhi vào textbox click button actual output quay về màn hình quản lý topic topic vừa tạo mới lưu ở db topic và refresh grid expected output hiển thị inline message màu đỏ tên topic đã tồn tại phía dưới trường tên topic và set focus tại đó
0
18,668
24,584,819,983
IssuesEvent
2022-10-13 18:42:51
googleapis/google-cloud-ruby
https://api.github.com/repos/googleapis/google-cloud-ruby
closed
Example Rails app
type: process
It would be useful to have a sample rails app that configures and uses many of the most popular gems, including all the stackdriver gems. The app should be configured to run using forked processes on Puma, and demonstrate how to work around the restrictions on delaying GRPC loading until after the fork.
1.0
Example Rails app - It would be useful to have a sample rails app that configures and uses many of the most popular gems, including all the stackdriver gems. The app should be configured to run using forked processes on Puma, and demonstrate how to work around the restrictions on delaying GRPC loading until after the fork.
process
example rails app it would be useful to have a sample rails app that configures and uses many of the most popular gems including all the stackdriver gems the app should be configured to run using forked processes on puma and demonstrate how to work around the restrictions on delaying grpc loading until after the fork
1
4,375
7,260,516,209
IssuesEvent
2018-02-18 10:54:51
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE] Extra control when smoothing geometries
Automatic new feature Processing
Original commit: https://github.com/qgis/QGIS/commit/4d60d0cdb54cc75f15c4e841aa87a3fd6994c918 by nyalldawson Add option to QgsGeometry::smooth to not smooth segments shorter than a certain threshold or sharp corners with an angle exceeding a threshold Expose the angle threshold to processing smooth algorithm Also: - optimise QgsGeometry::smooth for new geometry classes - Fix smooth does not work with geometries containing Z/M
1.0
[FEATURE] Extra control when smoothing geometries - Original commit: https://github.com/qgis/QGIS/commit/4d60d0cdb54cc75f15c4e841aa87a3fd6994c918 by nyalldawson Add option to QgsGeometry::smooth to not smooth segments shorter than a certain threshold or sharp corners with an angle exceeding a threshold Expose the angle threshold to processing smooth algorithm Also: - optimise QgsGeometry::smooth for new geometry classes - Fix smooth does not work with geometries containing Z/M
process
extra control when smoothing geometries original commit by nyalldawson add option to qgsgeometry smooth to not smooth segments shorter than a certain threshold or sharp corners with an angle exceeding a threshold expose the angle threshold to processing smooth algorithm also optimise qgsgeometry smooth for new geometry classes fix smooth does not work with geometries containing z m
1
18,386
24,515,230,732
IssuesEvent
2022-10-11 03:59:58
f5devcentral/container-egress-service
https://api.github.com/repos/f5devcentral/container-egress-service
closed
CVE-2021-44716
fixed processing
[CVE-2021-44716](https://nvd.nist.gov/vuln/detail/CVE-2021-44716) Published: January 01, 2022; 12:15:08 AM -0500 V3.1: 7.5 **HIGH** V2.0: 5.0 MEDIUM > net/http in Go before 1.16.12 and 1.17.x before 1.17.5 allows uncontrolled memory consumption in the header canonicalization cache via HTTP/2 requests. **To Reproduce** https://github.com/f5devcentral/container-egress-service/blob/3e8f64bb9249ae60325fa7cd71e77b078abcfef2/go.sum#L428 ``` golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180906233101-161cd47e91fd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20181023162649-9b4f9f5ad519/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20181201002055-351d144fa1fc/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190213061140-3a22650c66bd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190311183353-d8887717615a/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190501004415-9ce7a6920f09/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190503192946-f4e77d36d62c/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190603091049-60506f45cf65/go.mod h1:HSz+uSET+XFnRR8LxR5pz3Of3rY3CfYBVs4xY44aLks= golang.org/x/net v0.0.0-20190613194153-d28f0bde5980/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190628185345-da137c7871d7/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190724013045-ca1201d0de80/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190827160401-ba9fcec4b297/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190923162816-aa69164e4478/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200114155413-6afb5195e5aa/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200202094626-16171245cfb2/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200222125558-5a598a2470a0/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200301022130-244492dfa37a/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200324143707-d3edc9973b7e/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200501053045-e0ff5e5a1de5/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200506145744-7e3656a0809f/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200513185701-a91f0712d120/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200520182314-0ba52f642ac2/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200625001655-4c5254603344/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/net v0.0.0-20200707034311-ab3426394381/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/net v0.0.0-20200822124328-c89045814202/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.0.0-20201031054903-ff519b6c9102/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.0.0-20201110031124-69a78807bb2b/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.0.0-20201209123823-ac852fbbde11/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= golang.org/x/net v0.0.0-20210119194325-5f4716e94777/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= golang.org/x/net v0.0.0-20210226172049-e18ecbb05110/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= golang.org/x/net v0.0.0-20210316092652-d523dce5a7f4/go.mod h1:RBQZq4jEuRlivfhVLdyRGr576XBO4/greRjx4P4O3yc= golang.org/x/net v0.0.0-20210405180319-a5a99cb37ef4/go.mod h1:p54w0d4576C0XHj96bSt6lcn1PtDYWL6XObtHCRCNQM= golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2 h1:CIJ76btIcR3eFI5EgSo6k1qKw9KJexJuRLI9G7Hp5wE= golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y= ``` **Expected behavior** golang.org/x/net 0.0.0-20211209124913-491a49abca63 or higher
1.0
CVE-2021-44716 - [CVE-2021-44716](https://nvd.nist.gov/vuln/detail/CVE-2021-44716) Published: January 01, 2022; 12:15:08 AM -0500 V3.1: 7.5 **HIGH** V2.0: 5.0 MEDIUM > net/http in Go before 1.16.12 and 1.17.x before 1.17.5 allows uncontrolled memory consumption in the header canonicalization cache via HTTP/2 requests. **To Reproduce** https://github.com/f5devcentral/container-egress-service/blob/3e8f64bb9249ae60325fa7cd71e77b078abcfef2/go.sum#L428 ``` golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20180906233101-161cd47e91fd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20181023162649-9b4f9f5ad519/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20181201002055-351d144fa1fc/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190108225652-1e06a53dbb7e/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190213061140-3a22650c66bd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4= golang.org/x/net v0.0.0-20190311183353-d8887717615a/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190404232315-eb5bcb51f2a3/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190501004415-9ce7a6920f09/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190503192946-f4e77d36d62c/go.mod h1:t9HGtf8HONx5eT2rtn7q6eTqICYqUVnKs3thJo3Qplg= golang.org/x/net v0.0.0-20190603091049-60506f45cf65/go.mod h1:HSz+uSET+XFnRR8LxR5pz3Of3rY3CfYBVs4xY44aLks= golang.org/x/net v0.0.0-20190613194153-d28f0bde5980/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190620200207-3b0461eec859/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190628185345-da137c7871d7/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190724013045-ca1201d0de80/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190827160401-ba9fcec4b297/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20190923162816-aa69164e4478/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20191209160850-c0dbc17a3553/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200114155413-6afb5195e5aa/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200202094626-16171245cfb2/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200222125558-5a598a2470a0/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200226121028-0de0cce0169b/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200301022130-244492dfa37a/go.mod h1:z5CRVTTTmAJ677TzLLGU+0bjPO0LkuOLi4/5GtJWs/s= golang.org/x/net v0.0.0-20200324143707-d3edc9973b7e/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200501053045-e0ff5e5a1de5/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200506145744-7e3656a0809f/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200513185701-a91f0712d120/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200520182314-0ba52f642ac2/go.mod h1:qpuaurCH72eLCgpAm/N6yyVIVM9cpaDIP3A8BGJEC5A= golang.org/x/net v0.0.0-20200625001655-4c5254603344/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/net v0.0.0-20200707034311-ab3426394381/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/net v0.0.0-20200822124328-c89045814202/go.mod h1:/O7V0waA8r7cgGh81Ro3o1hOxt32SMVPicZroKQ2sZA= golang.org/x/net v0.0.0-20201021035429-f5854403a974/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.0.0-20201031054903-ff519b6c9102/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.0.0-20201110031124-69a78807bb2b/go.mod h1:sp8m0HH+o8qH0wwXwYZr8TS3Oi6o0r6Gce1SSxlDquU= golang.org/x/net v0.0.0-20201209123823-ac852fbbde11/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= golang.org/x/net v0.0.0-20210119194325-5f4716e94777/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= golang.org/x/net v0.0.0-20210226172049-e18ecbb05110/go.mod h1:m0MpNAwzfU5UDzcl9v0D8zg8gWTRqZa9RBIspLL5mdg= golang.org/x/net v0.0.0-20210316092652-d523dce5a7f4/go.mod h1:RBQZq4jEuRlivfhVLdyRGr576XBO4/greRjx4P4O3yc= golang.org/x/net v0.0.0-20210405180319-a5a99cb37ef4/go.mod h1:p54w0d4576C0XHj96bSt6lcn1PtDYWL6XObtHCRCNQM= golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2 h1:CIJ76btIcR3eFI5EgSo6k1qKw9KJexJuRLI9G7Hp5wE= golang.org/x/net v0.0.0-20211112202133-69e39bad7dc2/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y= ``` **Expected behavior** golang.org/x/net 0.0.0-20211209124913-491a49abca63 or higher
process
cve published  january am   high   medium net http in go before and x before allows uncontrolled memory consumption in the header canonicalization cache via http requests to reproduce golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod hsz uset golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod s golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net go mod golang org x net golang org x net go mod expected behavior golang org x net or higher
1
297,743
25,759,860,493
IssuesEvent
2022-12-08 19:30:05
rackslab/racksdb
https://api.github.com/repos/rackslab/racksdb
opened
Define unit tests for core generic components
tests
Unit tests must be defined with a good coverage of core generic components. It should at least cover the following points: - [ ] Back references in Schema - [ ] Default properties in Schema - [ ] Boolean properties in Schema - [ ] Circular dependencies in schemas - [ ] Generic DB loads
1.0
Define unit tests for core generic components - Unit tests must be defined with a good coverage of core generic components. It should at least cover the following points: - [ ] Back references in Schema - [ ] Default properties in Schema - [ ] Boolean properties in Schema - [ ] Circular dependencies in schemas - [ ] Generic DB loads
non_process
define unit tests for core generic components unit tests must be defined with a good coverage of core generic components it should at least cover the following points back references in schema default properties in schema boolean properties in schema circular dependencies in schemas generic db loads
0
765
3,250,945,546
IssuesEvent
2015-10-19 06:29:26
e-government-ua/i
https://api.github.com/repos/e-government-ua/i
closed
На главном портале добавить в объект markers новый валидатор по КВЕД
active bug In process of testing test _central-js
(в под-объект validate) id = KVED //вид экономической деятельности по КВЕД. Логика: две цифры точка две цифры (первые две цифры не могут быть 04, 34, 40, 44, 48, 54, 57, 67, 76, 83, 89) Сообщение: Такого КВЕД не існує - (ви не можете вписувати літери) Поля: kved
1.0
На главном портале добавить в объект markers новый валидатор по КВЕД - (в под-объект validate) id = KVED //вид экономической деятельности по КВЕД. Логика: две цифры точка две цифры (первые две цифры не могут быть 04, 34, 40, 44, 48, 54, 57, 67, 76, 83, 89) Сообщение: Такого КВЕД не існує - (ви не можете вписувати літери) Поля: kved
process
на главном портале добавить в объект markers новый валидатор по квед в под объект validate id kved вид экономической деятельности по квед логика две цифры точка две цифры первые две цифры не могут быть сообщение такого квед не існує ви не можете вписувати літери поля kved
1
17,589
23,408,195,825
IssuesEvent
2022-08-12 14:46:11
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Add htsPrecompile acceptance tests support for nonFungible token allowance/approval methods
enhancement limechain P2 process
### Problem No htsPrecompile acceptance tests support for non fungible token allowance/approval methods exists ### Solution Add support for the non fungible token allowance/approval verifications: - getApproved(address token, int64 serialNumber) external returns (int64 responseCode, address approved) - setApprovalForAll(address token, address operator, bool approved) external returns (int64 responseCode) - isApprovedForAll(address token, address owner, address operator) external returns (int64 responseCode, bool approved) ### Alternatives _No response_
1.0
Add htsPrecompile acceptance tests support for nonFungible token allowance/approval methods - ### Problem No htsPrecompile acceptance tests support for non fungible token allowance/approval methods exists ### Solution Add support for the non fungible token allowance/approval verifications: - getApproved(address token, int64 serialNumber) external returns (int64 responseCode, address approved) - setApprovalForAll(address token, address operator, bool approved) external returns (int64 responseCode) - isApprovedForAll(address token, address owner, address operator) external returns (int64 responseCode, bool approved) ### Alternatives _No response_
process
add htsprecompile acceptance tests support for nonfungible token allowance approval methods problem no htsprecompile acceptance tests support for non fungible token allowance approval methods exists solution add support for the non fungible token allowance approval verifications getapproved address token serialnumber external returns responsecode address approved setapprovalforall address token address operator bool approved external returns responsecode isapprovedforall address token address owner address operator external returns responsecode bool approved alternatives no response
1
771,644
27,088,149,438
IssuesEvent
2023-02-14 18:37:38
googleapis/google-cloud-ruby
https://api.github.com/repos/googleapis/google-cloud-ruby
closed
[Nightly CI Failures] Failures detected for google-cloud-talent-v4
type: bug priority: p1 nightly failure
At 2023-02-09 08:59:34 UTC, detected failures in google-cloud-talent-v4 for: rubocop report_key_0af3fa4a3a8bd1dc69005961b063920e
1.0
[Nightly CI Failures] Failures detected for google-cloud-talent-v4 - At 2023-02-09 08:59:34 UTC, detected failures in google-cloud-talent-v4 for: rubocop report_key_0af3fa4a3a8bd1dc69005961b063920e
non_process
failures detected for google cloud talent at utc detected failures in google cloud talent for rubocop report key
0
6,275
9,231,176,627
IssuesEvent
2019-03-13 01:09:14
EthVM/EthVM
https://api.github.com/repos/EthVM/EthVM
closed
Usage of Buffer correctly
bug milestone:2 priority:medium project:api project:explorer project:processing
Currently in backend we're storing BigIntegers as Decimal128 which is an special type supported by Mongo. Whenever we try to send the value to the front-end as Buffer, I've found issues to convert those value back to BN (js). So we opted to send the values as raw strings. This task should take care of researching and possible fixing this little issue.
1.0
Usage of Buffer correctly - Currently in backend we're storing BigIntegers as Decimal128 which is an special type supported by Mongo. Whenever we try to send the value to the front-end as Buffer, I've found issues to convert those value back to BN (js). So we opted to send the values as raw strings. This task should take care of researching and possible fixing this little issue.
process
usage of buffer correctly currently in backend we re storing bigintegers as which is an special type supported by mongo whenever we try to send the value to the front end as buffer i ve found issues to convert those value back to bn js so we opted to send the values as raw strings this task should take care of researching and possible fixing this little issue
1
175,851
21,336,915,248
IssuesEvent
2022-04-18 15:36:27
ShaikUsaf/frameworks_native_AOSP10_r33_ShaikUsaf-frameworks_native_AOSP10_r33_CVE-2020-0226
https://api.github.com/repos/ShaikUsaf/frameworks_native_AOSP10_r33_ShaikUsaf-frameworks_native_AOSP10_r33_CVE-2020-0226
opened
CVE-2020-0392 (High) detected in multiple libraries
security vulnerability
## CVE-2020-0392 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r45</b>, <b>nativeandroid-10.0.0_r45</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In getLayerDebugInfo of SurfaceFlinger.cpp, there is a possible code execution due to a double free. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-9 Android-10 Android-11Android ID: A-150226608 <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0392>CVE-2020-0392</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: android-9.0.0_r60,android-10.0.0_r46</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-0392 (High) detected in multiple libraries - ## CVE-2020-0392 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r47</b>, <b>nativeandroid-10.0.0_r45</b>, <b>nativeandroid-10.0.0_r45</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In getLayerDebugInfo of SurfaceFlinger.cpp, there is a possible code execution due to a double free. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-9 Android-10 Android-11Android ID: A-150226608 <p>Publish Date: 2020-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-0392>CVE-2020-0392</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://source.android.com/security/bulletin/2020-09-01">https://source.android.com/security/bulletin/2020-09-01</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: android-9.0.0_r60,android-10.0.0_r46</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries nativeandroid nativeandroid nativeandroid nativeandroid nativeandroid vulnerability details in getlayerdebuginfo of surfaceflinger cpp there is a possible code execution due to a double free this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android android android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution android android step up your open source security game with whitesource
0
365,395
25,533,630,341
IssuesEvent
2022-11-29 10:17:58
SovereignCloudStack/Docs
https://api.github.com/repos/SovereignCloudStack/Docs
closed
Release Notes for R3
documentation IaaS Container Ops release3 IAM
We have started to collect the central pieces of the release notes. Please have a look at https://github.com/SovereignCloudStack/Docs/blob/feat/RelNotes3/Release-Notes/Release3.md in branch feat/RelNotes3 of this Docs repository. Pull requests and comments are more than welcome!
1.0
Release Notes for R3 - We have started to collect the central pieces of the release notes. Please have a look at https://github.com/SovereignCloudStack/Docs/blob/feat/RelNotes3/Release-Notes/Release3.md in branch feat/RelNotes3 of this Docs repository. Pull requests and comments are more than welcome!
non_process
release notes for we have started to collect the central pieces of the release notes please have a look at in branch feat of this docs repository pull requests and comments are more than welcome
0
764,307
26,794,171,007
IssuesEvent
2023-02-01 10:38:23
unitaryfund/mitiq
https://api.github.com/repos/unitaryfund/mitiq
closed
Improve the handling of negative probabilities in REM
priority/mid feature-request rem
## Issue Description At the moment we clip negative probabilities when running REM. An alternative solution, typically used in the literature, is to return the closest positive probability distribution. I.e., after getting the adjusted quasi-prob distribution $p' = C^{-1} p$, where $C$ is the confusion matrix, instead of returning $p_{\rm positive} = clip(p) / |clip(p)|_1$ it is probably better to return something the optimal solution of: $$ \min_{p_{\rm positive}} \||p' - p_{\rm positive}\||_1$$ See e.g. https://arxiv.org/abs/2006.14044 ## Additional comment In principle, if we are just interested in computing expectation values, one can also work with the non-positive distribution $p' = C^{-1} p$. Actually this would be the optimal approach since if we avoid any post-processing we get an unbiased estimate of the expectation value, similar to PEC. See e.g. https://arxiv.org/abs/2006.14044 . Not sure however how easy is the implementation of this approach in Mitiq.
1.0
Improve the handling of negative probabilities in REM - ## Issue Description At the moment we clip negative probabilities when running REM. An alternative solution, typically used in the literature, is to return the closest positive probability distribution. I.e., after getting the adjusted quasi-prob distribution $p' = C^{-1} p$, where $C$ is the confusion matrix, instead of returning $p_{\rm positive} = clip(p) / |clip(p)|_1$ it is probably better to return something the optimal solution of: $$ \min_{p_{\rm positive}} \||p' - p_{\rm positive}\||_1$$ See e.g. https://arxiv.org/abs/2006.14044 ## Additional comment In principle, if we are just interested in computing expectation values, one can also work with the non-positive distribution $p' = C^{-1} p$. Actually this would be the optimal approach since if we avoid any post-processing we get an unbiased estimate of the expectation value, similar to PEC. See e.g. https://arxiv.org/abs/2006.14044 . Not sure however how easy is the implementation of this approach in Mitiq.
non_process
improve the handling of negative probabilities in rem issue description at the moment we clip negative probabilities when running rem an alternative solution typically used in the literature is to return the closest positive probability distribution i e after getting the adjusted quasi prob distribution p c p where c is the confusion matrix instead of returning p rm positive clip p clip p it is probably better to return something the optimal solution of min p rm positive p p rm positive see e g additional comment in principle if we are just interested in computing expectation values one can also work with the non positive distribution p c p actually this would be the optimal approach since if we avoid any post processing we get an unbiased estimate of the expectation value similar to pec see e g not sure however how easy is the implementation of this approach in mitiq
0
21,321
28,931,475,805
IssuesEvent
2023-05-09 00:00:12
devinit/DIwebsite-redesign
https://api.github.com/repos/devinit/DIwebsite-redesign
closed
Enhance accessibility on the Short Publication Page
Enhancement wontfix Feedback to Process
Perform an accessibility audit on the Short Publications Page and fix any flagged accessibility issues. Resources: [WAVE](https://chrome.google.com/webstore/detail/wave-evaluation-tool/jbbplnpkjmmeebjpijfedlgcdilocofh) For full coverage, you'll need to create a test page that makes use of every possible StreamField item.
1.0
Enhance accessibility on the Short Publication Page - Perform an accessibility audit on the Short Publications Page and fix any flagged accessibility issues. Resources: [WAVE](https://chrome.google.com/webstore/detail/wave-evaluation-tool/jbbplnpkjmmeebjpijfedlgcdilocofh) For full coverage, you'll need to create a test page that makes use of every possible StreamField item.
process
enhance accessibility on the short publication page perform an accessibility audit on the short publications page and fix any flagged accessibility issues resources for full coverage you ll need to create a test page that makes use of every possible streamfield item
1
410,317
11,986,254,271
IssuesEvent
2020-04-07 18:58:16
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
coronavirus.saude.gov.br - Desktop site instead of mobile site on "Release Notes" page
browser-firefox engine-gecko priority-normal severity-minor type-covid19
<!-- @browser: Firefox Preview Nightly 200405 (🦎: 76.0a1-20200403063228) --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:77.0) Gecko/20100101 Firefox/77.0 --> <!-- @reported_with: --> **URL**: https://coronavirus.saude.gov.br/templates/padraogoverno01/releaseNotes.html **Browser / Version**: Firefox Preview Nightly 200405 (🦎: 76.0a1-20200403063228) **Operating System**: Samsung Galaxy S6 Edge (Android 7.0) - 1440 x 2560 pixels (~577 ppi pixel density) **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: 1. Navigate to https://coronavirus.saude.gov.br/templates/padraogoverno01/releaseNotes.html 2. Observe the page. **Expected Behavior:** Mobile layout is displayed. **Actual Behavior:** Desktop layout is displayed. **Notes:** 1. Screenshot attached. 2. The issue is not reproducible on Chrome 80.0.3987.162. Watchers: @softvision-oana-arbuzov @softvision-sergiulogigan @cipriansv <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/b87f0add-10fc-4207-8294-c0ce33731ed4.jpg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
coronavirus.saude.gov.br - Desktop site instead of mobile site on "Release Notes" page - <!-- @browser: Firefox Preview Nightly 200405 (🦎: 76.0a1-20200403063228) --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:77.0) Gecko/20100101 Firefox/77.0 --> <!-- @reported_with: --> **URL**: https://coronavirus.saude.gov.br/templates/padraogoverno01/releaseNotes.html **Browser / Version**: Firefox Preview Nightly 200405 (🦎: 76.0a1-20200403063228) **Operating System**: Samsung Galaxy S6 Edge (Android 7.0) - 1440 x 2560 pixels (~577 ppi pixel density) **Tested Another Browser**: Yes Chrome **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: 1. Navigate to https://coronavirus.saude.gov.br/templates/padraogoverno01/releaseNotes.html 2. Observe the page. **Expected Behavior:** Mobile layout is displayed. **Actual Behavior:** Desktop layout is displayed. **Notes:** 1. Screenshot attached. 2. The issue is not reproducible on Chrome 80.0.3987.162. Watchers: @softvision-oana-arbuzov @softvision-sergiulogigan @cipriansv <details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/b87f0add-10fc-4207-8294-c0ce33731ed4.jpg'></details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
coronavirus saude gov br desktop site instead of mobile site on release notes page url browser version firefox preview nightly 🦎 operating system samsung galaxy edge android x pixels ppi pixel density tested another browser yes chrome problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce navigate to observe the page expected behavior mobile layout is displayed actual behavior desktop layout is displayed notes screenshot attached the issue is not reproducible on chrome watchers softvision oana arbuzov softvision sergiulogigan cipriansv view the screenshot img alt screenshot src browser configuration none from with ❤️
0
18,301
24,414,659,728
IssuesEvent
2022-10-05 14:57:39
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Prevent new uses of old ztest API
Process
Discussed in the TSC meeting, we need a way to prevent new uses of the ztest API. I believe that the best way to do this will be to add a new check for PRs similar to a language filter that checks for inappropriate language. To do this we'd need: 1. Create a new repo something like `zephyrproject-rtos/word-blocklist` 2. Provide an implementation of the GitHub check that scans for new test suites: a. `ztest_register_test_suite` b. `ztest_test_suite` I don't believe that we should block adding more tests to existing suites. That will automatically happen as people migrate to the V2 API. Once we're done, `zephyrproject-rtos/word-blocklist` can remain and we can use this for other migrations. It's also possible that we can find an existing GitHub CI for profanity filter and simply fork it. I don't think it'll be a bad idea anyway to have a profanity filter.
1.0
Prevent new uses of old ztest API - Discussed in the TSC meeting, we need a way to prevent new uses of the ztest API. I believe that the best way to do this will be to add a new check for PRs similar to a language filter that checks for inappropriate language. To do this we'd need: 1. Create a new repo something like `zephyrproject-rtos/word-blocklist` 2. Provide an implementation of the GitHub check that scans for new test suites: a. `ztest_register_test_suite` b. `ztest_test_suite` I don't believe that we should block adding more tests to existing suites. That will automatically happen as people migrate to the V2 API. Once we're done, `zephyrproject-rtos/word-blocklist` can remain and we can use this for other migrations. It's also possible that we can find an existing GitHub CI for profanity filter and simply fork it. I don't think it'll be a bad idea anyway to have a profanity filter.
process
prevent new uses of old ztest api discussed in the tsc meeting we need a way to prevent new uses of the ztest api i believe that the best way to do this will be to add a new check for prs similar to a language filter that checks for inappropriate language to do this we d need create a new repo something like zephyrproject rtos word blocklist provide an implementation of the github check that scans for new test suites a ztest register test suite b ztest test suite i don t believe that we should block adding more tests to existing suites that will automatically happen as people migrate to the api once we re done zephyrproject rtos word blocklist can remain and we can use this for other migrations it s also possible that we can find an existing github ci for profanity filter and simply fork it i don t think it ll be a bad idea anyway to have a profanity filter
1
11,098
13,940,133,617
IssuesEvent
2020-10-22 17:27:31
googleapis/java-webrisk
https://api.github.com/repos/googleapis/java-webrisk
closed
Dependency Dashboard
api: webrisk type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-webrisk-1.x -->[chore(deps): update dependency com.google.cloud:google-cloud-webrisk to v1.1.1](../pull/218) - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->[chore(deps): update dependency com.google.cloud:libraries-bom to v10](../pull/229) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-webrisk-1.x -->[chore(deps): update dependency com.google.cloud:google-cloud-webrisk to v1.1.1](../pull/218) - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->[chore(deps): update dependency com.google.cloud:libraries-bom to v10](../pull/229) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any pull pull check this box to trigger a request for renovate to run again on this repository
1
53,936
13,230,144,888
IssuesEvent
2020-08-18 09:20:33
feelpp/feelpp
https://api.github.com/repos/feelpp/feelpp
opened
Re-enable doxygen
module:build module:doc type:feature
As discussed with @vincentchabannes we should re-enable doxygen to build the API documentation This shall work as follows - [ ] add github action to build the api docs and push them to gh-pages - [ ] update [Doxyfile.in](https://github.com/feelpp/feelpp/blob/develop/doc/api/Doxyfile.in) - [ ] use grouping to add classes to modules (here module means a subdirectory) We have to sets of modules - feelpp modules, the name of the module should be `feelpp/<module name>` - feelpp-toolboxes modules, the name of the module should be `feelpp-toolboxes/<module name>`
1.0
Re-enable doxygen - As discussed with @vincentchabannes we should re-enable doxygen to build the API documentation This shall work as follows - [ ] add github action to build the api docs and push them to gh-pages - [ ] update [Doxyfile.in](https://github.com/feelpp/feelpp/blob/develop/doc/api/Doxyfile.in) - [ ] use grouping to add classes to modules (here module means a subdirectory) We have to sets of modules - feelpp modules, the name of the module should be `feelpp/<module name>` - feelpp-toolboxes modules, the name of the module should be `feelpp-toolboxes/<module name>`
non_process
re enable doxygen as discussed with vincentchabannes we should re enable doxygen to build the api documentation this shall work as follows add github action to build the api docs and push them to gh pages update use grouping to add classes to modules here module means a subdirectory we have to sets of modules feelpp modules the name of the module should be feelpp feelpp toolboxes modules the name of the module should be feelpp toolboxes
0
4,474
7,341,343,093
IssuesEvent
2018-03-07 01:34:00
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
How to enable these Preview features in Azure SQL?
cxp in-process product-question sql-database triaged
How to enable these Preview features in Azure SQL? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9ca439af-510f-303c-e72c-424b78ac0ac7 * Version Independent ID: 616d8cf5-1631-1179-0c59-f8b4f6c1f7bb * [Content](https://docs.microsoft.com/en-us/azure/sql-database/sql-database-data-discovery-and-classification) * [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/sql-database/sql-database-data-discovery-and-classification.md) * Service: sql-database
1.0
How to enable these Preview features in Azure SQL? - How to enable these Preview features in Azure SQL? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9ca439af-510f-303c-e72c-424b78ac0ac7 * Version Independent ID: 616d8cf5-1631-1179-0c59-f8b4f6c1f7bb * [Content](https://docs.microsoft.com/en-us/azure/sql-database/sql-database-data-discovery-and-classification) * [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/sql-database/sql-database-data-discovery-and-classification.md) * Service: sql-database
process
how to enable these preview features in azure sql how to enable these preview features in azure sql document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id service sql database
1
81,794
15,630,140,519
IssuesEvent
2021-03-22 01:28:53
TrapTeamCCNZ/webui
https://api.github.com/repos/TrapTeamCCNZ/webui
closed
CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - autoclosed
security vulnerability
## CVE-2020-7733 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary> <p>Lightweight JavaScript-based user-agent string parser</p> <p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p> <p>Path to dependency file: webui/AdminLTE-3.0.0-rc.6/AdminLTE-3.0.0-rc.6/package.json</p> <p>Path to vulnerable library: webui/AdminLTE-3.0.0-rc.6/AdminLTE-3.0.0-rc.6/node_modules/ua-parser-js/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.7.tgz (Root Library) - :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TrapTeamCCNZ/webui/commit/914b985d163674d879ae2e213efd026abf814948">914b985d163674d879ae2e213efd026abf814948</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA. <p>Publish Date: 2020-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7733>CVE-2020-7733</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 0.7.22</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7733 (High) detected in ua-parser-js-0.7.17.tgz - autoclosed - ## CVE-2020-7733 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ua-parser-js-0.7.17.tgz</b></p></summary> <p>Lightweight JavaScript-based user-agent string parser</p> <p>Library home page: <a href="https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz">https://registry.npmjs.org/ua-parser-js/-/ua-parser-js-0.7.17.tgz</a></p> <p>Path to dependency file: webui/AdminLTE-3.0.0-rc.6/AdminLTE-3.0.0-rc.6/package.json</p> <p>Path to vulnerable library: webui/AdminLTE-3.0.0-rc.6/AdminLTE-3.0.0-rc.6/node_modules/ua-parser-js/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.7.tgz (Root Library) - :x: **ua-parser-js-0.7.17.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TrapTeamCCNZ/webui/commit/914b985d163674d879ae2e213efd026abf814948">914b985d163674d879ae2e213efd026abf814948</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package ua-parser-js before 0.7.22 are vulnerable to Regular Expression Denial of Service (ReDoS) via the regex for Redmi Phones and Mi Pad Tablets UA. <p>Publish Date: 2020-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7733>CVE-2020-7733</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7733</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 0.7.22</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in ua parser js tgz autoclosed cve high severity vulnerability vulnerable library ua parser js tgz lightweight javascript based user agent string parser library home page a href path to dependency file webui adminlte rc adminlte rc package json path to vulnerable library webui adminlte rc adminlte rc node modules ua parser js package json dependency hierarchy browser sync tgz root library x ua parser js tgz vulnerable library found in head commit a href vulnerability details the package ua parser js before are vulnerable to regular expression denial of service redos via the regex for redmi phones and mi pad tablets ua publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
260,093
19,657,469,531
IssuesEvent
2022-01-10 13:58:54
felixstrobel/mailtm
https://api.github.com/repos/felixstrobel/mailtm
opened
Add in-code documentation
documentation enhancement
We should probably write an in-code documentation in [GoDoc](https://go.dev/blog/godoc) to improve the development speed.
1.0
Add in-code documentation - We should probably write an in-code documentation in [GoDoc](https://go.dev/blog/godoc) to improve the development speed.
non_process
add in code documentation we should probably write an in code documentation in to improve the development speed
0
14,865
18,273,498,331
IssuesEvent
2021-10-04 16:03:25
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Re-introspection support using `db pull` with MongoDB
kind/feature process/candidate topic: re-introspection tech/engines tech/typescript team/migrations topic: mongodb
Introspection support for MongoDB (Preview) was shipped with Prisma version 3.2.0 without re-introspection. Meaning that during re-introspection some information will be lost (relations, comments, mapped fields...)
1.0
Re-introspection support using `db pull` with MongoDB - Introspection support for MongoDB (Preview) was shipped with Prisma version 3.2.0 without re-introspection. Meaning that during re-introspection some information will be lost (relations, comments, mapped fields...)
process
re introspection support using db pull with mongodb introspection support for mongodb preview was shipped with prisma version without re introspection meaning that during re introspection some information will be lost relations comments mapped fields
1
140,605
12,943,159,681
IssuesEvent
2020-07-18 05:29:15
Naeemkh/tsprocess
https://api.github.com/repos/Naeemkh/tsprocess
closed
doc: Add doctest
documentation
Add doctest to make sure that documents and the codebase are getting updated together. Read more: https://docs.python.org/3/library/doctest.html
1.0
doc: Add doctest - Add doctest to make sure that documents and the codebase are getting updated together. Read more: https://docs.python.org/3/library/doctest.html
non_process
doc add doctest add doctest to make sure that documents and the codebase are getting updated together read more
0
10,429
13,218,985,374
IssuesEvent
2020-08-17 09:40:52
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
Standardized anomalies should be unitless.
bug preprocessor
**Describe the bug** The title says it all. Found by @hb326 while testing https://github.com/ESMValGroup/ESMValTool/issues/1687.
1.0
Standardized anomalies should be unitless. - **Describe the bug** The title says it all. Found by @hb326 while testing https://github.com/ESMValGroup/ESMValTool/issues/1687.
process
standardized anomalies should be unitless describe the bug the title says it all found by while testing
1
19,680
26,031,908,370
IssuesEvent
2022-12-21 22:20:22
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
What, exactly, is the set of invalid characters for a build number?
doc-enhancement devops/prod Pri1 devops-cicd-process/tech
In this documentation, with respect to the token `$(Build.DefinitionName)`, the following note is provided in the table: > Note: The pipeline name must not contain invalid or whitespace characters. But this article fails to define what are invalid characters. For example, I want to set the date _and time_ in my build number. Are colons valid characters? --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=classic) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
What, exactly, is the set of invalid characters for a build number? - In this documentation, with respect to the token `$(Build.DefinitionName)`, the following note is provided in the table: > Note: The pipeline name must not contain invalid or whitespace characters. But this article fails to define what are invalid characters. For example, I want to set the date _and time_ in my build number. Are colons valid characters? --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=classic) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
what exactly is the set of invalid characters for a build number in this documentation with respect to the token build definitionname the following note is provided in the table note the pipeline name must not contain invalid or whitespace characters but this article fails to define what are invalid characters for example i want to set the date and time in my build number are colons valid characters document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
14,924
18,359,529,127
IssuesEvent
2021-10-09 01:45:46
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
Testcafe version of browser.waitUntil and browser.execute
TYPE: bug AREA: client FREQUENCY: level 1 SYSTEM: client side processing STATE: Stale
<!-- If you have all reproduction steps with a complete sample app, please share as many details as possible in the sections below. Make sure that you tried using the latest TestCafe version (https://github.com/DevExpress/testcafe/releases), where this behavior might have been already addressed. Before submitting an issue, please check CONTRIBUTING.md and existing issues in this repository (https://github.com/DevExpress/testcafe/issues) in case a similar issue exists or was already addressed. This may save your time (and ours). --> ### What is your Test Scenario? Replicate the following code written in Selenium to TestCafe framework. ``` browser.waitUntil( () => { return browser.execute(() => { return ( navigator && navigator.serviceWorker && navigator.serviceWorker.controller && navigator.serviceWorker.controller.state && // We want to ensure the Service Worker is fully activated, // other states don't mean success. navigator.serviceWorker.controller.state === 'activated' ); }).value; }, 10000, 'Service Worker registration failed!' ); ``` ### What is the Current behavior? Can't figure out how to write the above code using ClientFunctions ### What is the Expected behavior? Use ClientFunctions to test for Service Workers ### Your Environment details: * testcafe version: 1.2.1 * node.js version: v10.15.3 * command-line arguments: testcafe chrome -d --skip-js-errors * browser name and version: Chrome * platform and version: macOS 10.14
1.0
Testcafe version of browser.waitUntil and browser.execute - <!-- If you have all reproduction steps with a complete sample app, please share as many details as possible in the sections below. Make sure that you tried using the latest TestCafe version (https://github.com/DevExpress/testcafe/releases), where this behavior might have been already addressed. Before submitting an issue, please check CONTRIBUTING.md and existing issues in this repository (https://github.com/DevExpress/testcafe/issues) in case a similar issue exists or was already addressed. This may save your time (and ours). --> ### What is your Test Scenario? Replicate the following code written in Selenium to TestCafe framework. ``` browser.waitUntil( () => { return browser.execute(() => { return ( navigator && navigator.serviceWorker && navigator.serviceWorker.controller && navigator.serviceWorker.controller.state && // We want to ensure the Service Worker is fully activated, // other states don't mean success. navigator.serviceWorker.controller.state === 'activated' ); }).value; }, 10000, 'Service Worker registration failed!' ); ``` ### What is the Current behavior? Can't figure out how to write the above code using ClientFunctions ### What is the Expected behavior? Use ClientFunctions to test for Service Workers ### Your Environment details: * testcafe version: 1.2.1 * node.js version: v10.15.3 * command-line arguments: testcafe chrome -d --skip-js-errors * browser name and version: Chrome * platform and version: macOS 10.14
process
testcafe version of browser waituntil and browser execute if you have all reproduction steps with a complete sample app please share as many details as possible in the sections below make sure that you tried using the latest testcafe version where this behavior might have been already addressed before submitting an issue please check contributing md and existing issues in this repository  in case a similar issue exists or was already addressed this may save your time and ours what is your test scenario replicate the following code written in selenium to testcafe framework browser waituntil return browser execute return navigator navigator serviceworker navigator serviceworker controller navigator serviceworker controller state we want to ensure the service worker is fully activated other states don t mean success navigator serviceworker controller state activated value service worker registration failed what is the current behavior can t figure out how to write the above code using clientfunctions what is the expected behavior use clientfunctions to test for service workers your environment details testcafe version node js version command line arguments testcafe chrome d skip js errors browser name and version chrome platform and version macos
1
303,875
23,043,226,940
IssuesEvent
2022-07-23 13:35:00
numpy/numpy
https://api.github.com/repos/numpy/numpy
opened
DOC: einsum examples of diagonalising and writeable view
04 - Documentation
### Issue with current documentation: There are two issues I think worth expanding the docs with. ### 1) Diagonalising vectors in tensors One can diagonalise a simple vector in variety of ways; ```python >>> a = np.array([1, 2, 3]) >>> a * np.eye(3). # broadcasting >>> np.multiply(a, np.eye(3)). # method broadcasting >>> np.diag(a). # targeted method >>> np.einsum('i, ij -> ij', a, np.eye(3)). # explicit einsum ``` I think it is useful to document `einsum` can do this since it can then be used for larger tensors where broadcasting does not. My use case was to convert a set of 1d samples to a set of 2d-diagonal samples, ```python >>> 1d_samples = np.array([1, 2, 3], [10, 20, 30]]) >>> 2d_samples = np.einsum("ki, ij -> kij", 1d_samples, np.eye(3)) >>> 2d_samples[1, :, :] array([[[ 1., 0., 0.], [ 0., 2., 0.], [ 0., 0., 3.]], [[10., 0., 0.], [ 0., 20., 0.], [ 0., 0., 30.]]]) ``` ### 2) Writable views In the docs it is written that, for example as of NumPy 1.10 > np.einsum('ii->i', a) will return a writeable view of the diagonal of a 2D array. I tried to use this the other and couldn't see how it was writeable to the original array. And thus, if the `return` is just a new ndarray isn't it expected that that array should be writeable as per any other array? I find this a confusing part of the docs if the meaning is just to say an array is returned. ### Idea or request for content: as above
1.0
DOC: einsum examples of diagonalising and writeable view - ### Issue with current documentation: There are two issues I think worth expanding the docs with. ### 1) Diagonalising vectors in tensors One can diagonalise a simple vector in variety of ways; ```python >>> a = np.array([1, 2, 3]) >>> a * np.eye(3). # broadcasting >>> np.multiply(a, np.eye(3)). # method broadcasting >>> np.diag(a). # targeted method >>> np.einsum('i, ij -> ij', a, np.eye(3)). # explicit einsum ``` I think it is useful to document `einsum` can do this since it can then be used for larger tensors where broadcasting does not. My use case was to convert a set of 1d samples to a set of 2d-diagonal samples, ```python >>> 1d_samples = np.array([1, 2, 3], [10, 20, 30]]) >>> 2d_samples = np.einsum("ki, ij -> kij", 1d_samples, np.eye(3)) >>> 2d_samples[1, :, :] array([[[ 1., 0., 0.], [ 0., 2., 0.], [ 0., 0., 3.]], [[10., 0., 0.], [ 0., 20., 0.], [ 0., 0., 30.]]]) ``` ### 2) Writable views In the docs it is written that, for example as of NumPy 1.10 > np.einsum('ii->i', a) will return a writeable view of the diagonal of a 2D array. I tried to use this the other and couldn't see how it was writeable to the original array. And thus, if the `return` is just a new ndarray isn't it expected that that array should be writeable as per any other array? I find this a confusing part of the docs if the meaning is just to say an array is returned. ### Idea or request for content: as above
non_process
doc einsum examples of diagonalising and writeable view issue with current documentation there are two issues i think worth expanding the docs with diagonalising vectors in tensors one can diagonalise a simple vector in variety of ways python a np array a np eye broadcasting np multiply a np eye method broadcasting np diag a targeted method np einsum i ij ij a np eye explicit einsum i think it is useful to document einsum can do this since it can then be used for larger tensors where broadcasting does not my use case was to convert a set of samples to a set of diagonal samples python samples np array samples np einsum ki ij kij samples np eye samples array writable views in the docs it is written that for example as of numpy np einsum ii i a will return a writeable view of the diagonal of a array i tried to use this the other and couldn t see how it was writeable to the original array and thus if the return is just a new ndarray isn t it expected that that array should be writeable as per any other array i find this a confusing part of the docs if the meaning is just to say an array is returned idea or request for content as above
0
10,595
13,402,035,804
IssuesEvent
2020-09-03 18:17:51
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
opened
Gluesync tool should search to update partitions at dates before the table creation time
enhancement p0 p1 team:data processing
### Describe the bug > **NOTE**: A quickfix for this should be included in the 1.8 to avoid surprises during upgrades Currently `gluesync` needs a specific date in the past to sync partitions at dates before the Glue table creation time. This is problematic because historical data and events from misconfigured devices can put partitions at arbitrary points in the past. Especially in 1.8 release we added the `source_id` and `source_label` fields to *all tables*, which requires all Glue tables to be updated and all partitions synced. For tables that have data at some point in the past of table creation time this **will break queries** after deployment and will require manual intervention to fix. Specifically the person deploying should find out which is the oldest date they have data for in the problematic tables and run `gluesync` for these with the appropriate start date. ### Expected behavior Ideally the deployment should handle the migration transparently to the user ### Proposed fix After a short discussion with @rleighton we decided on adding a query to the Glue API for table partitions *before table creation time* and also syncing those during a table sync. This will add some delay to the tool but it will ensure that all partitions of the table will be updated to the latest schema.
1.0
Gluesync tool should search to update partitions at dates before the table creation time - ### Describe the bug > **NOTE**: A quickfix for this should be included in the 1.8 to avoid surprises during upgrades Currently `gluesync` needs a specific date in the past to sync partitions at dates before the Glue table creation time. This is problematic because historical data and events from misconfigured devices can put partitions at arbitrary points in the past. Especially in 1.8 release we added the `source_id` and `source_label` fields to *all tables*, which requires all Glue tables to be updated and all partitions synced. For tables that have data at some point in the past of table creation time this **will break queries** after deployment and will require manual intervention to fix. Specifically the person deploying should find out which is the oldest date they have data for in the problematic tables and run `gluesync` for these with the appropriate start date. ### Expected behavior Ideally the deployment should handle the migration transparently to the user ### Proposed fix After a short discussion with @rleighton we decided on adding a query to the Glue API for table partitions *before table creation time* and also syncing those during a table sync. This will add some delay to the tool but it will ensure that all partitions of the table will be updated to the latest schema.
process
gluesync tool should search to update partitions at dates before the table creation time describe the bug note a quickfix for this should be included in the to avoid surprises during upgrades currently gluesync needs a specific date in the past to sync partitions at dates before the glue table creation time this is problematic because historical data and events from misconfigured devices can put partitions at arbitrary points in the past especially in release we added the source id and source label fields to all tables which requires all glue tables to be updated and all partitions synced for tables that have data at some point in the past of table creation time this will break queries after deployment and will require manual intervention to fix specifically the person deploying should find out which is the oldest date they have data for in the problematic tables and run gluesync for these with the appropriate start date expected behavior ideally the deployment should handle the migration transparently to the user proposed fix after a short discussion with rleighton we decided on adding a query to the glue api for table partitions before table creation time and also syncing those during a table sync this will add some delay to the tool but it will ensure that all partitions of the table will be updated to the latest schema
1
18,932
24,886,933,328
IssuesEvent
2022-10-28 08:35:53
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
Error: [libs/dml/src/render/render_configuration.rs:58:54] called `Option::unwrap()` on a `None` value
bug/1-unconfirmed kind/bug process/candidate tech/engines/introspection engine topic: error reporting team/schema
<!-- If required, please update the title to be clear and descriptive --> Command: `prisma db pull` Version: `4.4.0` Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6` Report: https://prisma-errors.netlify.app/report/14391 OS: `x64 darwin 21.6.0` ```
1.0
Error: [libs/dml/src/render/render_configuration.rs:58:54] called `Option::unwrap()` on a `None` value - <!-- If required, please update the title to be clear and descriptive --> Command: `prisma db pull` Version: `4.4.0` Binary Version: `f352a33b70356f46311da8b00d83386dd9f145d6` Report: https://prisma-errors.netlify.app/report/14391 OS: `x64 darwin 21.6.0` ```
process
error called option unwrap on a none value command prisma db pull version binary version report os darwin
1
9,325
12,338,986,459
IssuesEvent
2020-05-14 17:20:55
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
closed
Consider using the RACI framework for responsibility assignment
C-process T-docs
RACI is described here: https://en.wikipedia.org/wiki/Responsibility_assignment_matrix tl;dr: this allows a distinction between people who are *accountable* for the outcome, and people who actually do the work (can be the same as accountable, but not necessary); as well as calling out explicitly the folks who can provide info vs people who are informed only. https://github.com/unicode-org/icu4x/blob/master/docs/triaging.md#assignee talks about a "champion". This would be "accountable". Using prior art in responsibility assingment allows us not to spend time reinventing that.
1.0
Consider using the RACI framework for responsibility assignment - RACI is described here: https://en.wikipedia.org/wiki/Responsibility_assignment_matrix tl;dr: this allows a distinction between people who are *accountable* for the outcome, and people who actually do the work (can be the same as accountable, but not necessary); as well as calling out explicitly the folks who can provide info vs people who are informed only. https://github.com/unicode-org/icu4x/blob/master/docs/triaging.md#assignee talks about a "champion". This would be "accountable". Using prior art in responsibility assingment allows us not to spend time reinventing that.
process
consider using the raci framework for responsibility assignment raci is described here tl dr this allows a distinction between people who are accountable for the outcome and people who actually do the work can be the same as accountable but not necessary as well as calling out explicitly the folks who can provide info vs people who are informed only talks about a champion this would be accountable using prior art in responsibility assingment allows us not to spend time reinventing that
1
16,307
11,907,355,290
IssuesEvent
2020-03-30 22:07:07
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Soil downloader doesn't work in German locale
bug interface/infrastructure
Attempts to use the soil download in a German locale fail because the latitude and longitude values in the request uri get written in a culture-sensitive manner (using a comma as decimal place separator): ``` http://www.asris.csiro.au/ASRISApi/api/APSIM/getClosestApsoil?maxCnt=5&longitude=151,26&latitude=-27,18 ```
1.0
Soil downloader doesn't work in German locale - Attempts to use the soil download in a German locale fail because the latitude and longitude values in the request uri get written in a culture-sensitive manner (using a comma as decimal place separator): ``` http://www.asris.csiro.au/ASRISApi/api/APSIM/getClosestApsoil?maxCnt=5&longitude=151,26&latitude=-27,18 ```
non_process
soil downloader doesn t work in german locale attempts to use the soil download in a german locale fail because the latitude and longitude values in the request uri get written in a culture sensitive manner using a comma as decimal place separator
0
8,002
11,196,615,064
IssuesEvent
2020-01-03 10:37:22
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
`Clip raster by extension` batch mode: no path for output choosable
Bug Feedback High Priority Processing Regression
In the ``Clip raster by extent`` of gdal, in the **batch mode** it is not possible to choose an output for the raster files (and having the nice *Fill automatically with... (parameter, numbers..)* options. It seems to be just an empty string. With the ``Clip raster by mask`` always gdal, all seems working fine: 1. Clip raster by extent (not working) ![immagine](https://user-images.githubusercontent.com/2884884/67203763-40797f80-f40c-11e9-9dad-ce6cca1d8b62.png) 2. Clip raster by mask (working) ![immagine](https://user-images.githubusercontent.com/2884884/67203818-57b86d00-f40c-11e9-8698-54a9863b2641.png) QGIS 3.9 on Linux Debian Sid
1.0
`Clip raster by extension` batch mode: no path for output choosable - In the ``Clip raster by extent`` of gdal, in the **batch mode** it is not possible to choose an output for the raster files (and having the nice *Fill automatically with... (parameter, numbers..)* options. It seems to be just an empty string. With the ``Clip raster by mask`` always gdal, all seems working fine: 1. Clip raster by extent (not working) ![immagine](https://user-images.githubusercontent.com/2884884/67203763-40797f80-f40c-11e9-9dad-ce6cca1d8b62.png) 2. Clip raster by mask (working) ![immagine](https://user-images.githubusercontent.com/2884884/67203818-57b86d00-f40c-11e9-8698-54a9863b2641.png) QGIS 3.9 on Linux Debian Sid
process
clip raster by extension batch mode no path for output choosable in the clip raster by extent of gdal in the batch mode it is not possible to choose an output for the raster files and having the nice fill automatically with parameter numbers options it seems to be just an empty string with the clip raster by mask always gdal all seems working fine clip raster by extent not working clip raster by mask working qgis on linux debian sid
1
143,151
13,055,525,550
IssuesEvent
2020-07-30 01:57:20
miguelangelsoria/Reclamaciones
https://api.github.com/repos/miguelangelsoria/Reclamaciones
reopened
Actualización a base de datos
documentation
SE cambia el nombre de la tabla Trabajadores por el de Registros
1.0
Actualización a base de datos - SE cambia el nombre de la tabla Trabajadores por el de Registros
non_process
actualización a base de datos se cambia el nombre de la tabla trabajadores por el de registros
0
5,735
8,577,758,676
IssuesEvent
2018-11-13 01:33:12
googleapis/cloud-trace-nodejs
https://api.github.com/repos/googleapis/cloud-trace-nodejs
closed
Move samples from docs-samples
type: process
Hi there, our docs-samples repo still have some [trace samples](https://github.com/GoogleCloudPlatform/nodejs-docs-samples/tree/master/trace). For Node.js, all samples live with their client library if they have one. Can you move the samples over please. Thanks! cc @ofrobots
1.0
Move samples from docs-samples - Hi there, our docs-samples repo still have some [trace samples](https://github.com/GoogleCloudPlatform/nodejs-docs-samples/tree/master/trace). For Node.js, all samples live with their client library if they have one. Can you move the samples over please. Thanks! cc @ofrobots
process
move samples from docs samples hi there our docs samples repo still have some for node js all samples live with their client library if they have one can you move the samples over please thanks cc ofrobots
1
36,852
8,167,589,516
IssuesEvent
2018-08-26 00:57:07
idaholab/raven
https://api.github.com/repos/idaholab/raven
opened
codecoverage is months old
defect devel master priority_critical
-------- Issue Description -------- ##### What did you expect to see happen? The code coverage script to be functional. The code coverage is months old since the check_coverage.sh is broken. This is causing the master branch to be red since a while. ##### What did you see instead? The code coverage script is currently broken ##### Do you have a suggested fix for the development team? ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or improvement? - [ ] 2. Is it tagged with a priority: critical, normal or minor? - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
1.0
codecoverage is months old - -------- Issue Description -------- ##### What did you expect to see happen? The code coverage script to be functional. The code coverage is months old since the check_coverage.sh is broken. This is causing the master branch to be red since a while. ##### What did you see instead? The code coverage script is currently broken ##### Do you have a suggested fix for the development team? ##### Please attach the input file(s) that generate this error. The simpler the input, the faster we can find the issue. ---------------- For Change Control Board: Issue Review ---------------- This review should occur before any development is performed as a response to this issue. - [ ] 1. Is it tagged with a type: defect or improvement? - [ ] 2. Is it tagged with a priority: critical, normal or minor? - [ ] 3. If it will impact requirements or requirements tests, is it tagged with requirements? - [ ] 4. If it is a defect, can it cause wrong results for users? If so an email needs to be sent to the users. - [ ] 5. Is a rationale provided? (Such as explaining why the improvement is needed or why current code is wrong.) ------- For Change Control Board: Issue Closure ------- This review should occur when the issue is imminently going to be closed. - [ ] 1. If the issue is a defect, is the defect fixed? - [ ] 2. If the issue is a defect, is the defect tested for in the regression test system? (If not explain why not.) - [ ] 3. If the issue can impact users, has an email to the users group been written (the email should specify if the defect impacts stable or master)? - [ ] 4. If the issue is a defect, does it impact the latest stable branch? If yes, is there any issue tagged with stable (create if needed)? - [ ] 5. If the issue is being closed without a merge request, has an explanation of why it is being closed been provided?
non_process
codecoverage is months old issue description what did you expect to see happen the code coverage script to be functional the code coverage is months old since the check coverage sh is broken this is causing the master branch to be red since a while what did you see instead the code coverage script is currently broken do you have a suggested fix for the development team please attach the input file s that generate this error the simpler the input the faster we can find the issue for change control board issue review this review should occur before any development is performed as a response to this issue is it tagged with a type defect or improvement is it tagged with a priority critical normal or minor if it will impact requirements or requirements tests is it tagged with requirements if it is a defect can it cause wrong results for users if so an email needs to be sent to the users is a rationale provided such as explaining why the improvement is needed or why current code is wrong for change control board issue closure this review should occur when the issue is imminently going to be closed if the issue is a defect is the defect fixed if the issue is a defect is the defect tested for in the regression test system if not explain why not if the issue can impact users has an email to the users group been written the email should specify if the defect impacts stable or master if the issue is a defect does it impact the latest stable branch if yes is there any issue tagged with stable create if needed if the issue is being closed without a merge request has an explanation of why it is being closed been provided
0
63,643
26,478,389,186
IssuesEvent
2023-01-17 12:58:43
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Upgrade Proof URL Migration
Team:AppServicesSv Epic:KBNA-59
Currently Kibana URLs can break when our users migrate across minor versions. This can happen for various reasons, including: (1) because app teams change the URL structure, (2) because the ID of a saved object that was referenced in the URL changed; (3) because data structure, say global filters, has changed its shape. The scope of this project is to identify actual and possible causes why Kibana URLs might break across minor upgrades, followed by implementation of guidelines, best practices and possibly utility libraries which will help app teams ensure their URLs are upgrade proof.
1.0
Upgrade Proof URL Migration - Currently Kibana URLs can break when our users migrate across minor versions. This can happen for various reasons, including: (1) because app teams change the URL structure, (2) because the ID of a saved object that was referenced in the URL changed; (3) because data structure, say global filters, has changed its shape. The scope of this project is to identify actual and possible causes why Kibana URLs might break across minor upgrades, followed by implementation of guidelines, best practices and possibly utility libraries which will help app teams ensure their URLs are upgrade proof.
non_process
upgrade proof url migration currently kibana urls can break when our users migrate across minor versions this can happen for various reasons including because app teams change the url structure because the id of a saved object that was referenced in the url changed because data structure say global filters has changed its shape the scope of this project is to identify actual and possible causes why kibana urls might break across minor upgrades followed by implementation of guidelines best practices and possibly utility libraries which will help app teams ensure their urls are upgrade proof
0
20,988
11,572,818,480
IssuesEvent
2020-02-21 01:19:35
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
Could not enable aks dashboard on aks cluster
AKS Service Attention
## Describe the bug Was trying to enable the kubernetes dashboard on an existing aks cluster deployed through Terraform. **Command Name** `az aks enable-addons` **Errors:** ``` argument of type 'NoneType' is not iterable Traceback (most recent call last): Temp\pip-install-0u_uiwkq\knack\knack\cli.py, ln 206, in invoke azure\cli\core\commands\__init__.py, ln 608, in execute azure\cli\core\commands\__init__.py, ln 666, in _run_jobs_serially azure\cli\core\commands\__init__.py, ln 659, in _run_job Local\Temp\pip-install-0u_uiwkq\six\six.py, ln 696, in reraise azure\cli\core\commands\__init__.py, ln 636, in _run_job azure\cli\core\commands\__init__.py, ln 306, in __call__ azure-cli-core\azure\cli\core\__init__.py, ln 493, in default_command_handler azure\cli\command_modules\acs\custom.py, ln 1844, in aks_enable_addons azure\cli\command_modules\acs\custom.py, ln 2487, in _ensure_container_insights_for_monitoring TypeError: argument of type 'NoneType' is not iterable ``` ## To Reproduce: Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information. - _Put any pre-requisite steps here..._ - `az aks enable-addons --addons {} -n {} -g {}` ## Expected Behavior ## Environment Summary ``` Windows-10-10.0.18362-SP0 Python 3.6.6 Shell: powershell.exe azure-cli 2.0.80 * Extensions: azure-devops 0.17.0 ``` ## Additional Context <!--Please don't remove this:--> <!--auto-generated-->
1.0
Could not enable aks dashboard on aks cluster - ## Describe the bug Was trying to enable the kubernetes dashboard on an existing aks cluster deployed through Terraform. **Command Name** `az aks enable-addons` **Errors:** ``` argument of type 'NoneType' is not iterable Traceback (most recent call last): Temp\pip-install-0u_uiwkq\knack\knack\cli.py, ln 206, in invoke azure\cli\core\commands\__init__.py, ln 608, in execute azure\cli\core\commands\__init__.py, ln 666, in _run_jobs_serially azure\cli\core\commands\__init__.py, ln 659, in _run_job Local\Temp\pip-install-0u_uiwkq\six\six.py, ln 696, in reraise azure\cli\core\commands\__init__.py, ln 636, in _run_job azure\cli\core\commands\__init__.py, ln 306, in __call__ azure-cli-core\azure\cli\core\__init__.py, ln 493, in default_command_handler azure\cli\command_modules\acs\custom.py, ln 1844, in aks_enable_addons azure\cli\command_modules\acs\custom.py, ln 2487, in _ensure_container_insights_for_monitoring TypeError: argument of type 'NoneType' is not iterable ``` ## To Reproduce: Steps to reproduce the behavior. Note that argument values have been redacted, as they may contain sensitive information. - _Put any pre-requisite steps here..._ - `az aks enable-addons --addons {} -n {} -g {}` ## Expected Behavior ## Environment Summary ``` Windows-10-10.0.18362-SP0 Python 3.6.6 Shell: powershell.exe azure-cli 2.0.80 * Extensions: azure-devops 0.17.0 ``` ## Additional Context <!--Please don't remove this:--> <!--auto-generated-->
non_process
could not enable aks dashboard on aks cluster describe the bug was trying to enable the kubernetes dashboard on an existing aks cluster deployed through terraform command name az aks enable addons errors argument of type nonetype is not iterable traceback most recent call last temp pip install uiwkq knack knack cli py ln in invoke azure cli core commands init py ln in execute azure cli core commands init py ln in run jobs serially azure cli core commands init py ln in run job local temp pip install uiwkq six six py ln in reraise azure cli core commands init py ln in run job azure cli core commands init py ln in call azure cli core azure cli core init py ln in default command handler azure cli command modules acs custom py ln in aks enable addons azure cli command modules acs custom py ln in ensure container insights for monitoring typeerror argument of type nonetype is not iterable to reproduce steps to reproduce the behavior note that argument values have been redacted as they may contain sensitive information put any pre requisite steps here az aks enable addons addons n g expected behavior environment summary windows python shell powershell exe azure cli extensions azure devops additional context
0
36,102
7,860,889,754
IssuesEvent
2018-06-21 21:34:40
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
TRAC appears to convert character sets incorrectly (Trac #6)
Migrated from Trac defect senkbeil@uwm.edu utilities
Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not. Attachments: [plot_explicit_ta_configs.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_explicit_ta_configs.maff) [plot_new_pdf_config_1_plot_2.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_config_1_plot_2.maff) [plot_combo_pdf_run_3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_combo_pdf_run_3.maff) [plot_input_fields_rtp3_thlp3_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_input_fields_rtp3_thlp3_1.maff) [plot_new_pdf_20180522_test_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_20180522_test_1.maff) [plot_attempts_8_10.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempts_8_10.maff) [plot_attempt_8_only.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempt_8_only.maff) [plot_beta_1p3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3.maff) [plot_beta_1p3_all.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3_all.maff) Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/6 ```json { "status": "closed", "changetime": "2009-05-13T18:15:45", "description": "Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.", "reporter": "dschanen@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242238545000000", "component": "utilities", "summary": "TRAC appears to convert character sets incorrectly", "priority": "minor", "keywords": "", "time": "2009-05-04T20:44:26", "milestone": "", "owner": "senkbeil@uwm.edu", "type": "defect" } ```
1.0
TRAC appears to convert character sets incorrectly (Trac #6) - Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not. Attachments: [plot_explicit_ta_configs.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_explicit_ta_configs.maff) [plot_new_pdf_config_1_plot_2.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_config_1_plot_2.maff) [plot_combo_pdf_run_3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_combo_pdf_run_3.maff) [plot_input_fields_rtp3_thlp3_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_input_fields_rtp3_thlp3_1.maff) [plot_new_pdf_20180522_test_1.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_new_pdf_20180522_test_1.maff) [plot_attempts_8_10.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempts_8_10.maff) [plot_attempt_8_only.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_attempt_8_only.maff) [plot_beta_1p3.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3.maff) [plot_beta_1p3_all.maff](https://github.com/larson-group/trac_attachment_archive/blob/master/trac_test/822/plot_beta_1p3_all.maff) Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/6 ```json { "status": "closed", "changetime": "2009-05-13T18:15:45", "description": "Some of our source files (e.g. microphys_driver.F90) contain some non-ascii characters. According to the TRAC website it uses Unicode for character encoding, so I think these should display properly, but they do not.", "reporter": "dschanen@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1242238545000000", "component": "utilities", "summary": "TRAC appears to convert character sets incorrectly", "priority": "minor", "keywords": "", "time": "2009-05-04T20:44:26", "milestone": "", "owner": "senkbeil@uwm.edu", "type": "defect" } ```
non_process
trac appears to convert character sets incorrectly trac some of our source files e g microphys driver contain some non ascii characters according to the trac website it uses unicode for character encoding so i think these should display properly but they do not attachments migrated from json status closed changetime description some of our source files e g microphys driver contain some non ascii characters according to the trac website it uses unicode for character encoding so i think these should display properly but they do not reporter dschanen uwm edu cc resolution verified by v larson ts component utilities summary trac appears to convert character sets incorrectly priority minor keywords time milestone owner senkbeil uwm edu type defect
0
21,160
3,466,366,909
IssuesEvent
2015-12-22 02:49:09
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
CakePHP 3.1: cakephp-plugins.php has the wrong path?
Defect
<?php $baseDir = dirname(dirname(__FILE__)); return [ 'plugins' => [ 'ADmad/HybridAuth' => 'C:\xampp\htdocs\mysite/vendor/admad/cakephp-hybridauth/', 'Admin' => 'C:\xampp\htdocs\mysite/plugins/Admin/' ] ] After I changed the name of the project folder, I received a number of errors that occur that the plugin does not exist. The problem is that the paths is written as above, instead: 'ADmad/HybridAuth' => $baseDir . '/vendor/admad/cakephp-hybridauth/', 'Admin' => $baseDir . '/plugins/Admin/',
1.0
CakePHP 3.1: cakephp-plugins.php has the wrong path? - <?php $baseDir = dirname(dirname(__FILE__)); return [ 'plugins' => [ 'ADmad/HybridAuth' => 'C:\xampp\htdocs\mysite/vendor/admad/cakephp-hybridauth/', 'Admin' => 'C:\xampp\htdocs\mysite/plugins/Admin/' ] ] After I changed the name of the project folder, I received a number of errors that occur that the plugin does not exist. The problem is that the paths is written as above, instead: 'ADmad/HybridAuth' => $baseDir . '/vendor/admad/cakephp-hybridauth/', 'Admin' => $baseDir . '/plugins/Admin/',
non_process
cakephp cakephp plugins php has the wrong path php basedir dirname dirname file return plugins admad hybridauth c xampp htdocs mysite vendor admad cakephp hybridauth admin c xampp htdocs mysite plugins admin after i changed the name of the project folder i received a number of errors that occur that the plugin does not exist the problem is that the paths is written as above instead admad hybridauth basedir vendor admad cakephp hybridauth admin basedir plugins admin
0
16,510
21,519,225,157
IssuesEvent
2022-04-28 12:52:15
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Job of cancelled instance can be activated if an error was thrown on it
kind/bug scope/broker severity/high area/reliability team/process-automation
**Describe the Bug** A job can still be activated (that is, Zeebe will return it as part of an ActivateJobsResponse) even after its associated process instance has been cancelled, if an error was previously thrown on that job. **To Reproduce** @korthout's method is easier than mine, especially for Camunda folks: > I was also able to reproduce it through zbctl: > > ``` > zbctl deploy ./one_task.bpmn > sleep 1 > zbctl create instance one_task > zbctl activate jobs do_something > zbctl throwError job 2251799813685256 --errorCode FOO --errorMessage foobar > sleep 1 > zbctl cancel instance 2251799813685251 > sleep 1 > zbctl activate jobs do_something > ``` > > This activated the job and returned: > > ``` > { > "jobs": [ > { > "key": "2251799813685256", > "type": "do_something", > "processInstanceKey": "2251799813685251", > "bpmnProcessId": "one_task", > "processDefinitionVersion": 1, > "processDefinitionKey": "2251799813685249", > "elementId": "do_something", > "elementInstanceKey": "2251799813685255", > "customHeaders": "{\"what_to_do\":\"nothing\"}", > "worker": "zbctl", > "retries": 2, > "deadline": "1642166365427", > "variables": "{}" > } > ] > } > ``` <details> <summary>My original method</summary> ZIP file: [reproduce_bug.zip](https://github.com/camunda-cloud/zeebe/files/7866658/reproduce_bug.zip) 1. Ensure you have ruby installed, as well as the [zeebe_client ruby gem](https://github.com/zeebe-io/zeebe-client-ruby). 2. Download the attached ZIP file and expand it to reveal the `reproduce_bug.rb` and `one_task.bpmn` files, and put them in a directory together. 3. Ensure you have Zeebe running. If you need to change Zeebe's address away from `localhost:26500`, edit the `reproduce_bug.rb` file. 4. Run `ruby reproduce_bug.rb` and observe the output. The script will deploy the included BPMN, create an instance of that process, and activate jobs for that process instance. If the jobs that Zeebe returns are for a running process instance, it will throw an error on them. It will then cancel that process instance and loop around and create another instance. If the jobs that it receives from Zeebe are from a process instance that was already successfully cancelled, it will write a line of output including "[!!! BUG !!!]". The script will exit after 5 occurrences of the bug or 50 loops, whichever happens first. </details> **Expected Behavior** After Zeebe successfully cancels a process instance, no jobs from that instance should ever be returned to a future ActivateJobs request. **Actual Behavior** Jobs which previously had an error thrown on them before their process instance was cancelled are being returned to ActivateJobs requests that are placed after their process instance was cancelled. **Log/Stacktrace** <details> <summary>My original method</summary> Here is the output of a sample run of the `reproduce_bug.rb` script: ``` $ ruby reproduce_bug.rb deploying process definition from /usr/src/app/one_task.bpmn deployed process `one_task` with key 2251799813685249 created instance 2251799813685423 activated job 2251799813685429 for process instance 2251799813685423 threw error on job 2251799813685429 cancelled instance 2251799813685423 created instance 2251799813685432 [!!! BUG !!!] activated job 2251799813685429 for cancelled process instance 2251799813685423 [!!! BUG !!!] cancelled instance 2251799813685432 created instance 2251799813685439 activated job 2251799813685445 for process instance 2251799813685439 threw error on job 2251799813685445 cancelled instance 2251799813685439 created instance 2251799813685448 [!!! BUG !!!] activated job 2251799813685445 for cancelled process instance 2251799813685439 [!!! BUG !!!] cancelled instance 2251799813685448 created instance 2251799813685455 activated job 2251799813685461 for process instance 2251799813685455 threw error on job 2251799813685461 cancelled instance 2251799813685455 created instance 2251799813685464 [!!! BUG !!!] activated job 2251799813685461 for cancelled process instance 2251799813685455 [!!! BUG !!!] cancelled instance 2251799813685464 created instance 2251799813685471 activated job 2251799813685477 for process instance 2251799813685471 threw error on job 2251799813685477 cancelled instance 2251799813685471 created instance 2251799813685480 [!!! BUG !!!] activated job 2251799813685429 for cancelled process instance 2251799813685423 [!!! BUG !!!] [!!! BUG !!!] activated job 2251799813685445 for cancelled process instance 2251799813685439 [!!! BUG !!!] [!!! BUG !!!] activated job 2251799813685477 for cancelled process instance 2251799813685471 [!!! BUG !!!] cancelled instance 2251799813685480 ``` </details> **Environment:** - OS: Debian GNU/Linux 10 in a Docker container - Zeebe Version: 1.2.9 and 1.3.0 (verified against both) - Configuration: zeebe-kafka-exporter v3.0.0
1.0
Job of cancelled instance can be activated if an error was thrown on it - **Describe the Bug** A job can still be activated (that is, Zeebe will return it as part of an ActivateJobsResponse) even after its associated process instance has been cancelled, if an error was previously thrown on that job. **To Reproduce** @korthout's method is easier than mine, especially for Camunda folks: > I was also able to reproduce it through zbctl: > > ``` > zbctl deploy ./one_task.bpmn > sleep 1 > zbctl create instance one_task > zbctl activate jobs do_something > zbctl throwError job 2251799813685256 --errorCode FOO --errorMessage foobar > sleep 1 > zbctl cancel instance 2251799813685251 > sleep 1 > zbctl activate jobs do_something > ``` > > This activated the job and returned: > > ``` > { > "jobs": [ > { > "key": "2251799813685256", > "type": "do_something", > "processInstanceKey": "2251799813685251", > "bpmnProcessId": "one_task", > "processDefinitionVersion": 1, > "processDefinitionKey": "2251799813685249", > "elementId": "do_something", > "elementInstanceKey": "2251799813685255", > "customHeaders": "{\"what_to_do\":\"nothing\"}", > "worker": "zbctl", > "retries": 2, > "deadline": "1642166365427", > "variables": "{}" > } > ] > } > ``` <details> <summary>My original method</summary> ZIP file: [reproduce_bug.zip](https://github.com/camunda-cloud/zeebe/files/7866658/reproduce_bug.zip) 1. Ensure you have ruby installed, as well as the [zeebe_client ruby gem](https://github.com/zeebe-io/zeebe-client-ruby). 2. Download the attached ZIP file and expand it to reveal the `reproduce_bug.rb` and `one_task.bpmn` files, and put them in a directory together. 3. Ensure you have Zeebe running. If you need to change Zeebe's address away from `localhost:26500`, edit the `reproduce_bug.rb` file. 4. Run `ruby reproduce_bug.rb` and observe the output. The script will deploy the included BPMN, create an instance of that process, and activate jobs for that process instance. If the jobs that Zeebe returns are for a running process instance, it will throw an error on them. It will then cancel that process instance and loop around and create another instance. If the jobs that it receives from Zeebe are from a process instance that was already successfully cancelled, it will write a line of output including "[!!! BUG !!!]". The script will exit after 5 occurrences of the bug or 50 loops, whichever happens first. </details> **Expected Behavior** After Zeebe successfully cancels a process instance, no jobs from that instance should ever be returned to a future ActivateJobs request. **Actual Behavior** Jobs which previously had an error thrown on them before their process instance was cancelled are being returned to ActivateJobs requests that are placed after their process instance was cancelled. **Log/Stacktrace** <details> <summary>My original method</summary> Here is the output of a sample run of the `reproduce_bug.rb` script: ``` $ ruby reproduce_bug.rb deploying process definition from /usr/src/app/one_task.bpmn deployed process `one_task` with key 2251799813685249 created instance 2251799813685423 activated job 2251799813685429 for process instance 2251799813685423 threw error on job 2251799813685429 cancelled instance 2251799813685423 created instance 2251799813685432 [!!! BUG !!!] activated job 2251799813685429 for cancelled process instance 2251799813685423 [!!! BUG !!!] cancelled instance 2251799813685432 created instance 2251799813685439 activated job 2251799813685445 for process instance 2251799813685439 threw error on job 2251799813685445 cancelled instance 2251799813685439 created instance 2251799813685448 [!!! BUG !!!] activated job 2251799813685445 for cancelled process instance 2251799813685439 [!!! BUG !!!] cancelled instance 2251799813685448 created instance 2251799813685455 activated job 2251799813685461 for process instance 2251799813685455 threw error on job 2251799813685461 cancelled instance 2251799813685455 created instance 2251799813685464 [!!! BUG !!!] activated job 2251799813685461 for cancelled process instance 2251799813685455 [!!! BUG !!!] cancelled instance 2251799813685464 created instance 2251799813685471 activated job 2251799813685477 for process instance 2251799813685471 threw error on job 2251799813685477 cancelled instance 2251799813685471 created instance 2251799813685480 [!!! BUG !!!] activated job 2251799813685429 for cancelled process instance 2251799813685423 [!!! BUG !!!] [!!! BUG !!!] activated job 2251799813685445 for cancelled process instance 2251799813685439 [!!! BUG !!!] [!!! BUG !!!] activated job 2251799813685477 for cancelled process instance 2251799813685471 [!!! BUG !!!] cancelled instance 2251799813685480 ``` </details> **Environment:** - OS: Debian GNU/Linux 10 in a Docker container - Zeebe Version: 1.2.9 and 1.3.0 (verified against both) - Configuration: zeebe-kafka-exporter v3.0.0
process
job of cancelled instance can be activated if an error was thrown on it describe the bug a job can still be activated that is zeebe will return it as part of an activatejobsresponse even after its associated process instance has been cancelled if an error was previously thrown on that job to reproduce korthout s method is easier than mine especially for camunda folks i was also able to reproduce it through zbctl zbctl deploy one task bpmn sleep zbctl create instance one task zbctl activate jobs do something zbctl throwerror job errorcode foo errormessage foobar sleep zbctl cancel instance sleep zbctl activate jobs do something this activated the job and returned jobs key type do something processinstancekey bpmnprocessid one task processdefinitionversion processdefinitionkey elementid do something elementinstancekey customheaders what to do nothing worker zbctl retries deadline variables my original method zip file ensure you have ruby installed as well as the download the attached zip file and expand it to reveal the reproduce bug rb and one task bpmn files and put them in a directory together ensure you have zeebe running if you need to change zeebe s address away from localhost edit the reproduce bug rb file run ruby reproduce bug rb and observe the output the script will deploy the included bpmn create an instance of that process and activate jobs for that process instance if the jobs that zeebe returns are for a running process instance it will throw an error on them it will then cancel that process instance and loop around and create another instance if the jobs that it receives from zeebe are from a process instance that was already successfully cancelled it will write a line of output including the script will exit after occurrences of the bug or loops whichever happens first expected behavior after zeebe successfully cancels a process instance no jobs from that instance should ever be returned to a future activatejobs request actual behavior jobs which previously had an error thrown on them before their process instance was cancelled are being returned to activatejobs requests that are placed after their process instance was cancelled log stacktrace my original method here is the output of a sample run of the reproduce bug rb script ruby reproduce bug rb deploying process definition from usr src app one task bpmn deployed process one task with key created instance activated job for process instance threw error on job cancelled instance created instance activated job for cancelled process instance cancelled instance created instance activated job for process instance threw error on job cancelled instance created instance activated job for cancelled process instance cancelled instance created instance activated job for process instance threw error on job cancelled instance created instance activated job for cancelled process instance cancelled instance created instance activated job for process instance threw error on job cancelled instance created instance activated job for cancelled process instance activated job for cancelled process instance activated job for cancelled process instance cancelled instance environment os debian gnu linux in a docker container zeebe version and verified against both configuration zeebe kafka exporter
1
19,723
26,073,834,701
IssuesEvent
2022-12-24 07:07:41
pyanodon/pybugreports
https://api.github.com/repos/pyanodon/pybugreports
closed
Incompatable with 248k Mod
mod:pypostprocessing crash compatibility
### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [ ] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [X] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [X] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [X] Pypostprocessing failure - [ ] Other ### What is the problem? [factorio-current (1).log](https://github.com/pyanodon/pybugreports/files/9637768/factorio-current.1.log) ### Steps to reproduce Load PYbeta and 248k mods ### Additional context _No response_ ### Log file _No response_
1.0
Incompatable with 248k Mod - ### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [ ] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [X] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [X] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [ ] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [X] Pypostprocessing failure - [ ] Other ### What is the problem? [factorio-current (1).log](https://github.com/pyanodon/pybugreports/files/9637768/factorio-current.1.log) ### Steps to reproduce Load PYbeta and 248k mods ### Additional context _No response_ ### Log file _No response_
process
incompatable with mod mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem steps to reproduce load pybeta and mods additional context no response log file no response
1
5,897
8,712,368,632
IssuesEvent
2018-12-06 22:00:28
nodejs/node
https://api.github.com/repos/nodejs/node
closed
RFC process.on("shutdown") proposal
feature request process
See this popular question on so https://stackoverflow.com/q/14031763/3748498 By now there is no clean API to do cleanup when node.js exits, in order to catch all situation you need to do following: ```js for (const event of ["exit", "SIGINT", "SIGUSR1", "SIGUSR2", "uncaughtException", "SIGTERM"]) { process.on(event, cleanUp) } ``` This issue is a place for discussion how to fix it. My proposal is `process.on("shutdown", cleanUp)`.
1.0
RFC process.on("shutdown") proposal - See this popular question on so https://stackoverflow.com/q/14031763/3748498 By now there is no clean API to do cleanup when node.js exits, in order to catch all situation you need to do following: ```js for (const event of ["exit", "SIGINT", "SIGUSR1", "SIGUSR2", "uncaughtException", "SIGTERM"]) { process.on(event, cleanUp) } ``` This issue is a place for discussion how to fix it. My proposal is `process.on("shutdown", cleanUp)`.
process
rfc process on shutdown proposal see this popular question on so by now there is no clean api to do cleanup when node js exits in order to catch all situation you need to do following js for const event of process on event cleanup this issue is a place for discussion how to fix it my proposal is process on shutdown cleanup
1
117,455
11,947,011,036
IssuesEvent
2020-04-03 09:10:18
marmelab/react-admin
https://api.github.com/repos/marmelab/react-admin
closed
Document how to override notifications messages
documentation enhancement needs more info old
This has already been asked before: #180 but the answer provided was to basically not use the Admin component. I believe this could be a useful feature. Can we discuss some alternative approaches to provide an easier to use and user friendly server messaging? I propose to add support for a mapping file, similar to the way Translation messages are implemented, such that every error (or even every response) received from the server will be looked up in this server messages mapping file by the Admin on rest saga.
1.0
Document how to override notifications messages - This has already been asked before: #180 but the answer provided was to basically not use the Admin component. I believe this could be a useful feature. Can we discuss some alternative approaches to provide an easier to use and user friendly server messaging? I propose to add support for a mapping file, similar to the way Translation messages are implemented, such that every error (or even every response) received from the server will be looked up in this server messages mapping file by the Admin on rest saga.
non_process
document how to override notifications messages this has already been asked before but the answer provided was to basically not use the admin component i believe this could be a useful feature can we discuss some alternative approaches to provide an easier to use and user friendly server messaging i propose to add support for a mapping file similar to the way translation messages are implemented such that every error or even every response received from the server will be looked up in this server messages mapping file by the admin on rest saga
0
116,612
17,380,518,793
IssuesEvent
2021-07-31 16:03:11
AlexRogalskiy/charts
https://api.github.com/repos/AlexRogalskiy/charts
opened
CVE-2020-7656 (Medium) detected in jquery-1.8.1.min.js
security vulnerability
## CVE-2020-7656 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: charts/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: charts/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/charts/commit/6ec1d9fe297e071640b6fc8bc19516a88e99467a">6ec1d9fe297e071640b6fc8bc19516a88e99467a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p> <p>Release Date: 2020-05-28</p> <p>Fix Resolution: jquery - 1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7656 (Medium) detected in jquery-1.8.1.min.js - ## CVE-2020-7656 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: charts/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: charts/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/charts/commit/6ec1d9fe297e071640b6fc8bc19516a88e99467a">6ec1d9fe297e071640b6fc8bc19516a88e99467a</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-q4m3-2j7h-f7xw">https://github.com/advisories/GHSA-q4m3-2j7h-f7xw</a></p> <p>Release Date: 2020-05-28</p> <p>Fix Resolution: jquery - 1.9.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file charts node modules redeyed examples browser index html path to vulnerable library charts node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
36,426
5,051,575,899
IssuesEvent
2016-12-20 22:19:53
DirtyHairy/stella
https://api.github.com/repos/DirtyHairy/stella
closed
Tapper (from AtariAge)
regression test
Improvement in the new core: Old stella versions had a small graphical glitch in "Tapper" (circled in yellow in the picture) Now it's fine with the new core (and in Stellerator too) ![tapper](https://cloud.githubusercontent.com/assets/16061066/21370164/0e7b8ab6-c6e5-11e6-9d7e-6798cca420f1.png)
1.0
Tapper (from AtariAge) - Improvement in the new core: Old stella versions had a small graphical glitch in "Tapper" (circled in yellow in the picture) Now it's fine with the new core (and in Stellerator too) ![tapper](https://cloud.githubusercontent.com/assets/16061066/21370164/0e7b8ab6-c6e5-11e6-9d7e-6798cca420f1.png)
non_process
tapper from atariage improvement in the new core old stella versions had a small graphical glitch in tapper circled in yellow in the picture now it s fine with the new core and in stellerator too
0
21,049
27,993,405,596
IssuesEvent
2023-03-27 06:36:47
TeamAidemy/ds-paper-summaries
https://api.github.com/repos/TeamAidemy/ds-paper-summaries
opened
All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text
Natural language processing Evaluation
Elizabeth Clark, Tal August, Sofia Serrano, Nikita Haduong, Suchin Gururangan, and Noah A. Smith. 2022. “All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text.” arXiv:2107.00061 [cs.CL]. https://arxiv.org/abs/2107.00061. - 言語モデルが生成したテキストの評価を人間が行っているが、そもそもそれ('Human' Is Gold)でよいのか?と警鐘を鳴らす論文 - 人間が書いた文章と機械(GPT-2, 3)が書いた文章の識別タスクを評価者に実施してもらった結果、約半数は見分けがつかないことを実証 - 評価者を訓練すれば区別できるようになるのでは? → 訓練しても正答率は微増に留まり、統計的な有意差は確認されなかった - だから、そもそも評価方法を再考しようねと提言 ## Abstract > Human evaluations are typically considered the gold standard in natural language generation, but as models' fluency improves, how well can evaluators detect and judge machine-generated text? We run a study assessing non-experts' ability to distinguish between human- and machine-authored text (GPT2 and GPT3) in three domains (stories, news articles, and recipes). We find that, without training, evaluators distinguished between GPT3- and human-authored text at random chance level. We explore three approaches for quickly training evaluators to better identify GPT3-authored text (detailed instructions, annotated examples, and paired examples) and find that while evaluators' accuracy improved up to 55%, it did not significantly improve across the three domains. Given the inconsistent results across text domains and the often contradictory reasons evaluators gave for their judgments, we examine the role untrained human evaluations play in NLG evaluation and provide recommendations to NLG researchers for improving human evaluations of text generated from state-of-the-art models. (DeepL翻訳) 自然言語生成において、人間の評価は一般的にゴールドスタンダードと考えられているが、モデルの流暢性が向上するにつれ、評価者は機械が生成したテキストをどれだけ検出し判断できるのだろうか?我々は、3つのドメイン(ストーリー、ニュース記事、レシピ)において、非専門家が人間が作成したテキストと機械が作成したテキスト(GPT2とGPT3)を区別する能力を評価する研究を実施した。その結果、訓練なしでも、評価者はGPT3と人間作成のテキストをランダムな確率レベルで区別することがわかった。また、GPT3が作成したテキストをより適切に識別するために、3つのアプローチ(詳細な説明、注釈付き例、ペア例)を検討し、評価者の精度が最大55%向上するものの、3つのドメインで有意な向上が見られないことを発見しました。テキストドメイン間で一貫性のない結果と、評価者が判断した理由がしばしば矛盾していたことから、訓練されていない人間の評価がNLG評価において果たす役割を検証し、NLG研究者に最先端のモデルから生成されたテキストの人間評価を改善するための提言を行う。 ## コード まとめ作成時点では無し ## 解決した課題/先行研究との比較 GPT-2, 3 と言語モデルの発展につれて、どんどん流暢な文章が生成可能になってきたが、果たして人間は言語モデルが生成した文章と人間が書いた文章を見分けられるだろうか?という疑問からスタート <!-- <img width="411" alt="スクリーンショット 2023-03-14 10 40 38" src="https://user-images.githubusercontent.com/68265677/224870206-e1e00ba1-beac-4949-b7f2-15e3ce971d40.png"> *あなた自身は、これを人間が書いた文章なのかAIが書いた文章なのか見抜けますか?(正解は本論文のFig.1参照)* --> | <img width="500" alt="スクリーンショット 2023-03-14 10 40 38" src="https://user-images.githubusercontent.com/68265677/224870206-e1e00ba1-beac-4949-b7f2-15e3ce971d40.png"> | |:--:| | *あなた自身は、これを人間が書いた文章なのかAIが書いた文章なのか見抜けますか?(正解は本論文のFig.1参照)* | - 人間の言語モデル識別能力に関して調査した以下のような先行研究が存在 - GPT2-L が生成したテキストを、訓練された評価者でも、約30%は「これは人間が書いた」と識別してしまう (Ippolito et al., 2020) - GPT3-davinci が生成したニュース記事を、評価者は約50%の確率で「これは人間が書いた」と識別してしまう (Brown et al., 2020) - 本論文は「評価の設定・データ・参加者を変えて調査してみた」という、上記の派生系の1種としての位置づけ。特にいずれかを否定する趣旨ではない **本論文内で直接解決した課題はなく、今後はこうすべきという提言に留まる(詳細は後述)** ## 評価指標 - 人間が書いた文章と機械(GPT-2, GPT-3)が書いた文章が(当然、各文章を人間と機械のいずれが書いたかは非公開の状態で)計5つ提示されるので、評価者はそれぞれを4段階で評価 + なぜそう思ったかのコメントを残す - Definitely human-written - Possibly human-written - Possibly machine-generated - Definitely machine-generated - 評価者は[Amazon Mechanical Turk (AMT)](https://www.mturk.com/)で収集 - 言語面でのアンバランスが生じないように→米国在住の方に限定 - 雑な評価をしないように→1,000以上のAMTタスクを完了&95%以上の合格率を持つ方に限定 - 2つの条件下でそれぞれ実験 - いずれの条件の場合も評価者一人あたりは計30文章を評価する - 3ドメイン(①物語、②ニュース記事、③料理のレシピ) - 2モデル(GPT-2, GPT-3) - 5つの文章が提示される - ∴ 3 x 2 x 5 = 30 - 評価前の訓練**なし**: 130名 - 評価前の訓練**あり**: 195名(上記130名とは完全に別) <!-- - 訓練内容① インストラクション <img width="418" alt="スクリーンショット 2023-02-13 17 11 51" src="https://user-images.githubusercontent.com/68265677/218404698-d315b143-d47a-4d30-89a1-c9fe289ce4e0.png"> - 訓練内容② 例示 <img width="209" alt="スクリーンショット 2023-02-13 17 11 59" src="https://user-images.githubusercontent.com/68265677/218404669-6dc63ba3-207a-40f9-8d30-3f1d29677cd7.png"> - 訓練内容③ 比較 <img width="324" alt="スクリーンショット 2023-02-13 17 12 15" src="https://user-images.githubusercontent.com/68265677/218404645-495998d3-6fa7-4870-9438-d6b46c80a980.png"> --> | <img width="350" alt="training1" src="https://user-images.githubusercontent.com/68265677/218404698-d315b143-d47a-4d30-89a1-c9fe289ce4e0.png"> | <img width="350" alt="training2" src="https://user-images.githubusercontent.com/68265677/218404669-6dc63ba3-207a-40f9-8d30-3f1d29677cd7.png"> | <img width="350" alt="training3" src="https://user-images.githubusercontent.com/68265677/218404645-495998d3-6fa7-4870-9438-d6b46c80a980.png"> | | :--: | :--: | :--: | | *訓練1 インストラクション* | *訓練2 例示* | *訓練3 比較* | ### 結果 - GPT-2の書いた文章と人間の書いた文章の判別: 正答率 57.9% - GPT-3の書いた文章と人間の書いた文章の判別: 正答率 49.9% **→ 約半数の文章は見分けがついていない** - このようにGPT-2→GPT-3にモデルが変わると正答率は10pt近く減少するにも関わらず、「自信を持って回答した率」は有意に減少しない **→ 生成モデルの能力を過小評価している** <!-- - <img width="625" alt="スクリーンショット 2023-02-13 16 58 23" src="https://user-images.githubusercontent.com/68265677/218401956-f36e0032-1c52-4a7f-bedc-1ac7153959fa.png"> - ↑ appendix tab.5 の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧 --> | <img width="550" alt="tab5" src="https://user-images.githubusercontent.com/68265677/218401956-f36e0032-1c52-4a7f-bedc-1ac7153959fa.png"> | |:--:| | *appendix tab.5 の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧* | - 訓練ありのとき、訓練なしのときよりも正答率はわずかに向上。しかし、統計的な有意差はナシ - 一応、3つの訓練の中で最も効果的だったのは、訓練内容②の例示 ## 残された課題・議論 - タイトルの通り 'Human' Is Not Gold であることが実証されたため、言語モデルの評価方法を再検討すべきと提言。代替の方法として挙げられているのは以下 - TuringAdvice: 有用なアドバイスを生成する能力で評価 (Zellers et al., 2021) - RoFT: 評価者に推測ゲームを通して、人間と機械が生成したテキストの境界を決定させる (Dugan et al., 2020) - 生成されたテキストを用いて物語を書かせて評価 (Clark and Smith, 2021. Akoury et al., 2020.) - **論文内では上記のどれが推奨等は言及なく、その検討自体は残された課題と言えよう** - どうしても本論文と同じような条件下で人間による評価が必要な場合は、評価者を例題で訓練することを推奨 - 評価者が暗黙のうちに持っている評価基準が、研究者の意図するものとずれている可能性があるため - 本論文内の実験でも、評価者に対して事前の訓練を施すことで(わずかではあるが)評価精度が向上した - また、評価者に与えた指示と訓練の詳細を論文に含めることを推奨 ## 関連論文 Chmielewski, M., and Kucker, S. C. 2020. "An MTurk Crisis? Shifts in Data Quality and the Impact on Study Results." Social Psychological and Personality Science, 11(4), 464–473. https://doi.org/10.1177/1948550619875149 - 本論文の主張からは少し観点がずれるが、そもそも評価者を集めるときに AMT を使うこと自体がナンセンスなのでは?と提唱している論文 - 日本語解説記事: [MTurkを使った研究はもう信用できないものなのか](https://kuri8ive.hatenablog.com/entry/memo-about-mturk-crisis) ## 参考情報 - [北海道大学調和系工学研究室 平田氏によるサマリー](https://www.slideshare.net/harmonylab/all-thats-human-is-not-gold-evaluating-human-evaluation-of-generated-text) (Slideshare)
1.0
All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text - Elizabeth Clark, Tal August, Sofia Serrano, Nikita Haduong, Suchin Gururangan, and Noah A. Smith. 2022. “All That's 'Human' Is Not Gold: Evaluating Human Evaluation of Generated Text.” arXiv:2107.00061 [cs.CL]. https://arxiv.org/abs/2107.00061. - 言語モデルが生成したテキストの評価を人間が行っているが、そもそもそれ('Human' Is Gold)でよいのか?と警鐘を鳴らす論文 - 人間が書いた文章と機械(GPT-2, 3)が書いた文章の識別タスクを評価者に実施してもらった結果、約半数は見分けがつかないことを実証 - 評価者を訓練すれば区別できるようになるのでは? → 訓練しても正答率は微増に留まり、統計的な有意差は確認されなかった - だから、そもそも評価方法を再考しようねと提言 ## Abstract > Human evaluations are typically considered the gold standard in natural language generation, but as models' fluency improves, how well can evaluators detect and judge machine-generated text? We run a study assessing non-experts' ability to distinguish between human- and machine-authored text (GPT2 and GPT3) in three domains (stories, news articles, and recipes). We find that, without training, evaluators distinguished between GPT3- and human-authored text at random chance level. We explore three approaches for quickly training evaluators to better identify GPT3-authored text (detailed instructions, annotated examples, and paired examples) and find that while evaluators' accuracy improved up to 55%, it did not significantly improve across the three domains. Given the inconsistent results across text domains and the often contradictory reasons evaluators gave for their judgments, we examine the role untrained human evaluations play in NLG evaluation and provide recommendations to NLG researchers for improving human evaluations of text generated from state-of-the-art models. (DeepL翻訳) 自然言語生成において、人間の評価は一般的にゴールドスタンダードと考えられているが、モデルの流暢性が向上するにつれ、評価者は機械が生成したテキストをどれだけ検出し判断できるのだろうか?我々は、3つのドメイン(ストーリー、ニュース記事、レシピ)において、非専門家が人間が作成したテキストと機械が作成したテキスト(GPT2とGPT3)を区別する能力を評価する研究を実施した。その結果、訓練なしでも、評価者はGPT3と人間作成のテキストをランダムな確率レベルで区別することがわかった。また、GPT3が作成したテキストをより適切に識別するために、3つのアプローチ(詳細な説明、注釈付き例、ペア例)を検討し、評価者の精度が最大55%向上するものの、3つのドメインで有意な向上が見られないことを発見しました。テキストドメイン間で一貫性のない結果と、評価者が判断した理由がしばしば矛盾していたことから、訓練されていない人間の評価がNLG評価において果たす役割を検証し、NLG研究者に最先端のモデルから生成されたテキストの人間評価を改善するための提言を行う。 ## コード まとめ作成時点では無し ## 解決した課題/先行研究との比較 GPT-2, 3 と言語モデルの発展につれて、どんどん流暢な文章が生成可能になってきたが、果たして人間は言語モデルが生成した文章と人間が書いた文章を見分けられるだろうか?という疑問からスタート <!-- <img width="411" alt="スクリーンショット 2023-03-14 10 40 38" src="https://user-images.githubusercontent.com/68265677/224870206-e1e00ba1-beac-4949-b7f2-15e3ce971d40.png"> *あなた自身は、これを人間が書いた文章なのかAIが書いた文章なのか見抜けますか?(正解は本論文のFig.1参照)* --> | <img width="500" alt="スクリーンショット 2023-03-14 10 40 38" src="https://user-images.githubusercontent.com/68265677/224870206-e1e00ba1-beac-4949-b7f2-15e3ce971d40.png"> | |:--:| | *あなた自身は、これを人間が書いた文章なのかAIが書いた文章なのか見抜けますか?(正解は本論文のFig.1参照)* | - 人間の言語モデル識別能力に関して調査した以下のような先行研究が存在 - GPT2-L が生成したテキストを、訓練された評価者でも、約30%は「これは人間が書いた」と識別してしまう (Ippolito et al., 2020) - GPT3-davinci が生成したニュース記事を、評価者は約50%の確率で「これは人間が書いた」と識別してしまう (Brown et al., 2020) - 本論文は「評価の設定・データ・参加者を変えて調査してみた」という、上記の派生系の1種としての位置づけ。特にいずれかを否定する趣旨ではない **本論文内で直接解決した課題はなく、今後はこうすべきという提言に留まる(詳細は後述)** ## 評価指標 - 人間が書いた文章と機械(GPT-2, GPT-3)が書いた文章が(当然、各文章を人間と機械のいずれが書いたかは非公開の状態で)計5つ提示されるので、評価者はそれぞれを4段階で評価 + なぜそう思ったかのコメントを残す - Definitely human-written - Possibly human-written - Possibly machine-generated - Definitely machine-generated - 評価者は[Amazon Mechanical Turk (AMT)](https://www.mturk.com/)で収集 - 言語面でのアンバランスが生じないように→米国在住の方に限定 - 雑な評価をしないように→1,000以上のAMTタスクを完了&95%以上の合格率を持つ方に限定 - 2つの条件下でそれぞれ実験 - いずれの条件の場合も評価者一人あたりは計30文章を評価する - 3ドメイン(①物語、②ニュース記事、③料理のレシピ) - 2モデル(GPT-2, GPT-3) - 5つの文章が提示される - ∴ 3 x 2 x 5 = 30 - 評価前の訓練**なし**: 130名 - 評価前の訓練**あり**: 195名(上記130名とは完全に別) <!-- - 訓練内容① インストラクション <img width="418" alt="スクリーンショット 2023-02-13 17 11 51" src="https://user-images.githubusercontent.com/68265677/218404698-d315b143-d47a-4d30-89a1-c9fe289ce4e0.png"> - 訓練内容② 例示 <img width="209" alt="スクリーンショット 2023-02-13 17 11 59" src="https://user-images.githubusercontent.com/68265677/218404669-6dc63ba3-207a-40f9-8d30-3f1d29677cd7.png"> - 訓練内容③ 比較 <img width="324" alt="スクリーンショット 2023-02-13 17 12 15" src="https://user-images.githubusercontent.com/68265677/218404645-495998d3-6fa7-4870-9438-d6b46c80a980.png"> --> | <img width="350" alt="training1" src="https://user-images.githubusercontent.com/68265677/218404698-d315b143-d47a-4d30-89a1-c9fe289ce4e0.png"> | <img width="350" alt="training2" src="https://user-images.githubusercontent.com/68265677/218404669-6dc63ba3-207a-40f9-8d30-3f1d29677cd7.png"> | <img width="350" alt="training3" src="https://user-images.githubusercontent.com/68265677/218404645-495998d3-6fa7-4870-9438-d6b46c80a980.png"> | | :--: | :--: | :--: | | *訓練1 インストラクション* | *訓練2 例示* | *訓練3 比較* | ### 結果 - GPT-2の書いた文章と人間の書いた文章の判別: 正答率 57.9% - GPT-3の書いた文章と人間の書いた文章の判別: 正答率 49.9% **→ 約半数の文章は見分けがついていない** - このようにGPT-2→GPT-3にモデルが変わると正答率は10pt近く減少するにも関わらず、「自信を持って回答した率」は有意に減少しない **→ 生成モデルの能力を過小評価している** <!-- - <img width="625" alt="スクリーンショット 2023-02-13 16 58 23" src="https://user-images.githubusercontent.com/68265677/218401956-f36e0032-1c52-4a7f-bedc-1ac7153959fa.png"> - ↑ appendix tab.5 の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧 --> | <img width="550" alt="tab5" src="https://user-images.githubusercontent.com/68265677/218401956-f36e0032-1c52-4a7f-bedc-1ac7153959fa.png"> | |:--:| | *appendix tab.5 の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧* | - 訓練ありのとき、訓練なしのときよりも正答率はわずかに向上。しかし、統計的な有意差はナシ - 一応、3つの訓練の中で最も効果的だったのは、訓練内容②の例示 ## 残された課題・議論 - タイトルの通り 'Human' Is Not Gold であることが実証されたため、言語モデルの評価方法を再検討すべきと提言。代替の方法として挙げられているのは以下 - TuringAdvice: 有用なアドバイスを生成する能力で評価 (Zellers et al., 2021) - RoFT: 評価者に推測ゲームを通して、人間と機械が生成したテキストの境界を決定させる (Dugan et al., 2020) - 生成されたテキストを用いて物語を書かせて評価 (Clark and Smith, 2021. Akoury et al., 2020.) - **論文内では上記のどれが推奨等は言及なく、その検討自体は残された課題と言えよう** - どうしても本論文と同じような条件下で人間による評価が必要な場合は、評価者を例題で訓練することを推奨 - 評価者が暗黙のうちに持っている評価基準が、研究者の意図するものとずれている可能性があるため - 本論文内の実験でも、評価者に対して事前の訓練を施すことで(わずかではあるが)評価精度が向上した - また、評価者に与えた指示と訓練の詳細を論文に含めることを推奨 ## 関連論文 Chmielewski, M., and Kucker, S. C. 2020. "An MTurk Crisis? Shifts in Data Quality and the Impact on Study Results." Social Psychological and Personality Science, 11(4), 464–473. https://doi.org/10.1177/1948550619875149 - 本論文の主張からは少し観点がずれるが、そもそも評価者を集めるときに AMT を使うこと自体がナンセンスなのでは?と提唱している論文 - 日本語解説記事: [MTurkを使った研究はもう信用できないものなのか](https://kuri8ive.hatenablog.com/entry/memo-about-mturk-crisis) ## 参考情報 - [北海道大学調和系工学研究室 平田氏によるサマリー](https://www.slideshare.net/harmonylab/all-thats-human-is-not-gold-evaluating-human-evaluation-of-generated-text) (Slideshare)
process
all that s human is not gold evaluating human evaluation of generated text elizabeth clark tal august sofia serrano nikita haduong suchin gururangan and noah a smith “all that s human is not gold evaluating human evaluation of generated text ” arxiv 言語モデルが生成したテキストの評価を人間が行っているが、そもそもそれ( human is gold)でよいのか?と警鐘を鳴らす論文 人間が書いた文章と機械(gpt )が書いた文章の識別タスクを評価者に実施してもらった結果、約半数は見分けがつかないことを実証 評価者を訓練すれば区別できるようになるのでは? → 訓練しても正答率は微増に留まり、統計的な有意差は確認されなかった だから、そもそも評価方法を再考しようねと提言 abstract human evaluations are typically considered the gold standard in natural language generation but as models fluency improves how well can evaluators detect and judge machine generated text we run a study assessing non experts ability to distinguish between human and machine authored text and in three domains stories news articles and recipes we find that without training evaluators distinguished between and human authored text at random chance level we explore three approaches for quickly training evaluators to better identify authored text detailed instructions annotated examples and paired examples and find that while evaluators accuracy improved up to it did not significantly improve across the three domains given the inconsistent results across text domains and the often contradictory reasons evaluators gave for their judgments we examine the role untrained human evaluations play in nlg evaluation and provide recommendations to nlg researchers for improving human evaluations of text generated from state of the art models deepl翻訳 自然言語生成において、人間の評価は一般的にゴールドスタンダードと考えられているが、モデルの流暢性が向上するにつれ、評価者は機械が生成したテキストをどれだけ検出し判断できるのだろうか?我々は、 (ストーリー、ニュース記事、レシピ)において、非専門家が人間が作成したテキストと機械が作成したテキスト( )を区別する能力を評価する研究を実施した。その結果、訓練なしでも、 。また、 、 (詳細な説明、注釈付き例、ペア例)を検討し、 向上するものの、 。テキストドメイン間で一貫性のない結果と、評価者が判断した理由がしばしば矛盾していたことから、訓練されていない人間の評価がnlg評価において果たす役割を検証し、nlg研究者に最先端のモデルから生成されたテキストの人間評価を改善するための提言を行う。 コード まとめ作成時点では無し 解決した課題 先行研究との比較 gpt と言語モデルの発展につれて、どんどん流暢な文章が生成可能になってきたが、果たして人間は言語モデルが生成した文章と人間が書いた文章を見分けられるだろうか?という疑問からスタート img width alt スクリーンショット src あなた自身は、これを人間が書いた文章なのかaiが書いた文章なのか見抜けますか?(正解は本論文のfig ) img width alt スクリーンショット src あなた自身は、これを人間が書いた文章なのかaiが書いた文章なのか見抜けますか?(正解は本論文のfig ) 人間の言語モデル識別能力に関して調査した以下のような先行研究が存在 l が生成したテキストを、訓練された評価者でも、 は「これは人間が書いた」と識別してしまう ippolito et al davinci が生成したニュース記事を、 の確率で「これは人間が書いた」と識別してしまう brown et al 本論文は「評価の設定・データ・参加者を変えて調査してみた」という、 。特にいずれかを否定する趣旨ではない 本論文内で直接解決した課題はなく、今後はこうすべきという提言に留まる(詳細は後述) 評価指標 人間が書いた文章と機械(gpt gpt )が書いた文章が(当然、各文章を人間と機械のいずれが書いたかは非公開の状態で) 、 なぜそう思ったかのコメントを残す definitely human written possibly human written possibly machine generated definitely machine generated 評価者は 言語面でのアンバランスが生じないように→米国在住の方に限定 雑な評価をしないように→ & 以上の合格率を持つ方に限定 (①物語、②ニュース記事、③料理のレシピ) (gpt gpt ) ∴ x x 評価前の訓練 なし 評価前の訓練 あり ( ) 訓練内容① インストラクション img width alt スクリーンショット src 訓練内容② 例示 img width alt スクリーンショット src 訓練内容③ 比較 img width alt src img width alt src img width alt src インストラクション 例示 比較 結果 gpt 正答率 gpt 正答率 → 約半数の文章は見分けがついていない このようにgpt →gpt 、「自信を持って回答した率」は有意に減少しない → 生成モデルの能力を過小評価している img width alt スクリーンショット src ↑ appendix tab の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧 img width alt src appendix tab の和訳。機械が書いた文章に対して「これは人間が書いた文章だ」を思った評価者による、なぜそう思ったかのコメント一覧 訓練ありのとき、訓練なしのときよりも正答率はわずかに向上。しかし、統計的な有意差はナシ 一応、 、訓練内容②の例示 残された課題・議論 タイトルの通り human is not gold であることが実証されたため、言語モデルの評価方法を再検討すべきと提言。代替の方法として挙げられているのは以下 turingadvice 有用なアドバイスを生成する能力で評価 zellers et al roft 評価者に推測ゲームを通して、人間と機械が生成したテキストの境界を決定させる dugan et al 生成されたテキストを用いて物語を書かせて評価 clark and smith akoury et al 論文内では上記のどれが推奨等は言及なく、その検討自体は残された課題と言えよう どうしても本論文と同じような条件下で人間による評価が必要な場合は、評価者を例題で訓練することを推奨 評価者が暗黙のうちに持っている評価基準が、研究者の意図するものとずれている可能性があるため 本論文内の実験でも、評価者に対して事前の訓練を施すことで(わずかではあるが)評価精度が向上した また、評価者に与えた指示と訓練の詳細を論文に含めることを推奨 関連論文 chmielewski m and kucker s c an mturk crisis shifts in data quality and the impact on study results social psychological and personality science – 本論文の主張からは少し観点がずれるが、そもそも評価者を集めるときに amt を使うこと自体がナンセンスなのでは?と提唱している論文 日本語解説記事 参考情報 slideshare
1
166,503
12,959,551,852
IssuesEvent
2020-07-20 13:10:23
RPTools/maptool
https://api.github.com/repos/RPTools/maptool
closed
add lumens support to Sight (Light already has it)
feature tested
**Describe the bug** Lights are currently capable of expressing "negative lumens", causing darkness. There's a note in the Help text of the Campaign Property's **Light** tab that says the dark vision **Sight** is equivalent to 100 lumens. The problem is that I have a creature that can cast darkness at lumens=-100 (per the **Light** tab) but needs to be able to see through it and can't (because the **Sight** tab doesn't have a "lumens" keyword). **Expected behavior** Since the **Light** tab allows negative lumens, the **Sight** tab should allow the definition of vision types that can see through it. **MapTool Info** - Version: 1.6.1 **Desktop (please complete the following information):** - OS: macOS - Version 10.15 **Additional context** I haven't looked at the code (I have no development environment installed, currently) but this may or may not be a lot of work. Since the **Light** tab says darkvision (which should be any vision type with the **r** modifier) effectively has "lumens=100", I'm hoping there's already a field in the structure and it just needs the proper parsing in the **Sight** tab to be active. 🤞
1.0
add lumens support to Sight (Light already has it) - **Describe the bug** Lights are currently capable of expressing "negative lumens", causing darkness. There's a note in the Help text of the Campaign Property's **Light** tab that says the dark vision **Sight** is equivalent to 100 lumens. The problem is that I have a creature that can cast darkness at lumens=-100 (per the **Light** tab) but needs to be able to see through it and can't (because the **Sight** tab doesn't have a "lumens" keyword). **Expected behavior** Since the **Light** tab allows negative lumens, the **Sight** tab should allow the definition of vision types that can see through it. **MapTool Info** - Version: 1.6.1 **Desktop (please complete the following information):** - OS: macOS - Version 10.15 **Additional context** I haven't looked at the code (I have no development environment installed, currently) but this may or may not be a lot of work. Since the **Light** tab says darkvision (which should be any vision type with the **r** modifier) effectively has "lumens=100", I'm hoping there's already a field in the structure and it just needs the proper parsing in the **Sight** tab to be active. 🤞
non_process
add lumens support to sight light already has it describe the bug lights are currently capable of expressing negative lumens causing darkness there s a note in the help text of the campaign property s light tab that says the dark vision sight is equivalent to lumens the problem is that i have a creature that can cast darkness at lumens per the light tab but needs to be able to see through it and can t because the sight tab doesn t have a lumens keyword expected behavior since the light tab allows negative lumens the sight tab should allow the definition of vision types that can see through it maptool info version desktop please complete the following information os macos version additional context i haven t looked at the code i have no development environment installed currently but this may or may not be a lot of work since the light tab says darkvision which should be any vision type with the r modifier effectively has lumens i m hoping there s already a field in the structure and it just needs the proper parsing in the sight tab to be active 🤞
0
711
3,203,583,846
IssuesEvent
2015-10-02 19:47:30
johnhbenetech/listTest
https://api.github.com/repos/johnhbenetech/listTest
opened
product.Trachoma_Detection
feature.Data_Collection feature.Sensor process.Social_Service_Referral technology.Hardware technology.Narrow_Band_Communication technology.Software user.Poverty_Global_South vertical.Healthcare
1 Impact 1 Need 1 Mission fit 1 Legalilty 0 Money: current 1 Money: future 0 Cost 0 Talent 0 Technical Risk 1 Champion 0 Community: users 0 Community: partners 0 Portfolio fit 1 Exit strategies 0 Data/proof
1.0
product.Trachoma_Detection - 1 Impact 1 Need 1 Mission fit 1 Legalilty 0 Money: current 1 Money: future 0 Cost 0 Talent 0 Technical Risk 1 Champion 0 Community: users 0 Community: partners 0 Portfolio fit 1 Exit strategies 0 Data/proof
process
product trachoma detection impact need mission fit legalilty money current money future cost talent technical risk champion community users community partners portfolio fit exit strategies data proof
1
13,395
15,867,641,673
IssuesEvent
2021-04-08 17:09:41
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
ARM A32/T32 strb.w ldrb.w data acces size of undefined1 somtimes, undefined2 others times
Feature: Processor/ARM Type: Question
**1. DAT_c28f9080** strb.w ldrb.w are marking the data as undefined1 ![image](https://user-images.githubusercontent.com/4013370/113871080-d3e67200-97b2-11eb-9fe4-e0d6e2717af3.png) ![image](https://user-images.githubusercontent.com/4013370/113871595-5bcc7c00-97b3-11eb-9b81-60f353a5e272.png) **2. DAT_c34eb5aa** ldrh.w access data as undefined2 ![image](https://user-images.githubusercontent.com/4013370/113872117-df866880-97b3-11eb-8105-29a9b5adff34.png) any idea why the size mismatch btw the two data ?
1.0
ARM A32/T32 strb.w ldrb.w data acces size of undefined1 somtimes, undefined2 others times - **1. DAT_c28f9080** strb.w ldrb.w are marking the data as undefined1 ![image](https://user-images.githubusercontent.com/4013370/113871080-d3e67200-97b2-11eb-9fe4-e0d6e2717af3.png) ![image](https://user-images.githubusercontent.com/4013370/113871595-5bcc7c00-97b3-11eb-9b81-60f353a5e272.png) **2. DAT_c34eb5aa** ldrh.w access data as undefined2 ![image](https://user-images.githubusercontent.com/4013370/113872117-df866880-97b3-11eb-8105-29a9b5adff34.png) any idea why the size mismatch btw the two data ?
process
arm strb w ldrb w data acces size of somtimes others times dat strb w ldrb w are marking the data as dat ldrh w access data as any idea why the size mismatch btw the two data
1
3,982
6,911,479,685
IssuesEvent
2017-11-28 08:31:22
pingcap/tikv
https://api.github.com/repos/pingcap/tikv
closed
adjust coprocessor and scheduler slow timer
coprocessor enhancement scheduler
Now the slow timer for the coprocessor and scheduler operation contains the wait time + execution time. If one operation is too slow, it will cause following operations slow too, and the slow log is very difficult for us to see whether the operation is the real slow operation. We should use the slow timer for the execution time, use metric for the wait time is enough.
1.0
adjust coprocessor and scheduler slow timer - Now the slow timer for the coprocessor and scheduler operation contains the wait time + execution time. If one operation is too slow, it will cause following operations slow too, and the slow log is very difficult for us to see whether the operation is the real slow operation. We should use the slow timer for the execution time, use metric for the wait time is enough.
process
adjust coprocessor and scheduler slow timer now the slow timer for the coprocessor and scheduler operation contains the wait time execution time if one operation is too slow it will cause following operations slow too and the slow log is very difficult for us to see whether the operation is the real slow operation we should use the slow timer for the execution time use metric for the wait time is enough
1
696,238
23,892,796,936
IssuesEvent
2022-09-08 12:47:57
bedita/manager
https://api.github.com/repos/bedita/manager
opened
Admin - Roles and modules access
feature Priority - Normal UI / UX
BEM provides modules accesses per role with the configuration `AccessControl`: An example: ``` 'AccessControl' => [ 'manager' => [ 'hidden' => ['objects'], 'readonly' => ['documents'], ], 'guest' => [ 'hidden' => ['objects', 'users'], 'readonly' => ['documents'], ], ], ``` Admin users should be able to change `AccessControl` through an an `Admin - Roles and modules access` page, with an UI like the following: https://xd.adobe.com/view/48d5c8b0-ea5e-4199-9165-591641d31b3f-1d86/screen/f498cac4-19db-4b0e-82b4-957e5e39eec8/ Note: reuse the pattern of overriding the configuration in `configs` via proper API calls.
1.0
Admin - Roles and modules access - BEM provides modules accesses per role with the configuration `AccessControl`: An example: ``` 'AccessControl' => [ 'manager' => [ 'hidden' => ['objects'], 'readonly' => ['documents'], ], 'guest' => [ 'hidden' => ['objects', 'users'], 'readonly' => ['documents'], ], ], ``` Admin users should be able to change `AccessControl` through an an `Admin - Roles and modules access` page, with an UI like the following: https://xd.adobe.com/view/48d5c8b0-ea5e-4199-9165-591641d31b3f-1d86/screen/f498cac4-19db-4b0e-82b4-957e5e39eec8/ Note: reuse the pattern of overriding the configuration in `configs` via proper API calls.
non_process
admin roles and modules access bem provides modules accesses per role with the configuration accesscontrol an example accesscontrol manager hidden readonly guest hidden readonly admin users should be able to change accesscontrol through an an admin roles and modules access page with an ui like the following note reuse the pattern of overriding the configuration in configs via proper api calls
0
41,693
10,762,733,544
IssuesEvent
2019-11-01 00:50:24
randombit/botan
https://api.github.com/repos/randombit/botan
closed
Building with jom/nmake only works if using MSVC
build problem
Cannot build the library because there is no libbotan-2.a ![image](https://user-images.githubusercontent.com/8018219/67022932-a8794e80-f10a-11e9-9365-8275cede6fef.png)
1.0
Building with jom/nmake only works if using MSVC - Cannot build the library because there is no libbotan-2.a ![image](https://user-images.githubusercontent.com/8018219/67022932-a8794e80-f10a-11e9-9365-8275cede6fef.png)
non_process
building with jom nmake only works if using msvc cannot build the library because there is no libbotan a
0
109,489
9,381,798,949
IssuesEvent
2019-04-04 20:34:18
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
ccl/logictestccl: TestCCLLogic failed under stress
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/c6df752eefe4609b8a5bbada0955f79a2cfb790e Parameters: ``` TAGS= GOFLAGS=-parallel=4 ``` To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=TestCCLLogic PKG=github.com/cockroachdb/cockroach/pkg/ccl/logictestccl TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1217865&tab=buildLog ``` === RUN TestCCLLogic --- FAIL: TestCCLLogic (47.49s) test_log_scope.go:81: test logs captured to: /tmp/logTestCCLLogic865338793 test_log_scope.go:62: use -show-logs to present logs inline === RUN TestCCLLogic/5node-dist-opt --- FAIL: TestCCLLogic/5node-dist-opt (0.00s) === RUN TestCCLLogic/5node-dist-opt/zone === PAUSE TestCCLLogic/5node-dist-opt/zone === CONT TestCCLLogic/5node-dist-opt/zone --- done: testdata/logic_test/zone with config 5node-dist-opt: 24 tests, 0 failures --- total: 274 tests, 0 failures test logs left over in: /tmp/logTestCCLLogic865338793 --- FAIL: TestCCLLogic/5node-dist-opt/zone (45.87s) logic.go:1536: condition failed to evaluate within 45s: testdata/logic_test/zone:151: EXPLAIN SELECT * FROM t WHERE k=10 expected: scan · · · table t@tertiary · spans /10-/11 but found (query options: "retry") : scan · · · table t@primary · spans /10-/10/# goroutine 3387 [running]: runtime/debug.Stack(0xa7a358200, 0xc004197a70, 0x39250e0) /usr/local/go/src/runtime/debug/stack.go:24 +0xa7 github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon(0x3992e20, 0xc002a16000, 0xc002ec3d60) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/soon.go:49 +0x103 github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).processSubtest(0xc0010cf680, 0x0, 0x0, 0xc00315d3b0, 0x0, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, ...) /go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:1536 +0x1da2 github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).processTestFile(0xc0010cf680, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, 0x0, 0x3219045, 0x2, 0x3219045, ...) /go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:1186 +0x698 github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).runFile(0xc0010cf680, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, 0x0, 0x3219045, 0x2, 0x3219045, ...) /go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:2085 +0xe1 github.com/cockroachdb/cockroach/pkg/sql/logictest.RunLogicTest.func1.1(0xc002a16000) /go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:2196 +0x1cc testing.tRunner(0xc002a16000, 0xc0029b74a0) /usr/local/go/src/testing/testing.go:827 +0xbf created by testing.(*T).Run /usr/local/go/src/testing/testing.go:878 +0x35c ```
1.0
ccl/logictestccl: TestCCLLogic failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/c6df752eefe4609b8a5bbada0955f79a2cfb790e Parameters: ``` TAGS= GOFLAGS=-parallel=4 ``` To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=TestCCLLogic PKG=github.com/cockroachdb/cockroach/pkg/ccl/logictestccl TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1217865&tab=buildLog ``` === RUN TestCCLLogic --- FAIL: TestCCLLogic (47.49s) test_log_scope.go:81: test logs captured to: /tmp/logTestCCLLogic865338793 test_log_scope.go:62: use -show-logs to present logs inline === RUN TestCCLLogic/5node-dist-opt --- FAIL: TestCCLLogic/5node-dist-opt (0.00s) === RUN TestCCLLogic/5node-dist-opt/zone === PAUSE TestCCLLogic/5node-dist-opt/zone === CONT TestCCLLogic/5node-dist-opt/zone --- done: testdata/logic_test/zone with config 5node-dist-opt: 24 tests, 0 failures --- total: 274 tests, 0 failures test logs left over in: /tmp/logTestCCLLogic865338793 --- FAIL: TestCCLLogic/5node-dist-opt/zone (45.87s) logic.go:1536: condition failed to evaluate within 45s: testdata/logic_test/zone:151: EXPLAIN SELECT * FROM t WHERE k=10 expected: scan · · · table t@tertiary · spans /10-/11 but found (query options: "retry") : scan · · · table t@primary · spans /10-/10/# goroutine 3387 [running]: runtime/debug.Stack(0xa7a358200, 0xc004197a70, 0x39250e0) /usr/local/go/src/runtime/debug/stack.go:24 +0xa7 github.com/cockroachdb/cockroach/pkg/testutils.SucceedsSoon(0x3992e20, 0xc002a16000, 0xc002ec3d60) /go/src/github.com/cockroachdb/cockroach/pkg/testutils/soon.go:49 +0x103 github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).processSubtest(0xc0010cf680, 0x0, 0x0, 0xc00315d3b0, 0x0, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, ...) /go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:1536 +0x1da2 github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).processTestFile(0xc0010cf680, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, 0x0, 0x3219045, 0x2, 0x3219045, ...) /go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:1186 +0x698 github.com/cockroachdb/cockroach/pkg/sql/logictest.(*logicTest).runFile(0xc0010cf680, 0xc0002d6260, 0x18, 0x322d72f, 0xe, 0x5, 0x0, 0x3219045, 0x2, 0x3219045, ...) /go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:2085 +0xe1 github.com/cockroachdb/cockroach/pkg/sql/logictest.RunLogicTest.func1.1(0xc002a16000) /go/src/github.com/cockroachdb/cockroach/pkg/sql/logictest/logic.go:2196 +0x1cc testing.tRunner(0xc002a16000, 0xc0029b74a0) /usr/local/go/src/testing/testing.go:827 +0xbf created by testing.(*T).Run /usr/local/go/src/testing/testing.go:878 +0x35c ```
non_process
ccl logictestccl testccllogic failed under stress sha parameters tags goflags parallel to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests testccllogic pkg github com cockroachdb cockroach pkg ccl logictestccl testtimeout stressflags maxtime timeout tee tmp stress log failed test run testccllogic fail testccllogic test log scope go test logs captured to tmp test log scope go use show logs to present logs inline run testccllogic dist opt fail testccllogic dist opt run testccllogic dist opt zone pause testccllogic dist opt zone cont testccllogic dist opt zone done testdata logic test zone with config dist opt tests failures total tests failures test logs left over in tmp fail testccllogic dist opt zone logic go condition failed to evaluate within testdata logic test zone explain select from t where k expected scan · · · table t tertiary · spans but found query options retry scan · · · table t primary · spans goroutine runtime debug stack usr local go src runtime debug stack go github com cockroachdb cockroach pkg testutils succeedssoon go src github com cockroachdb cockroach pkg testutils soon go github com cockroachdb cockroach pkg sql logictest logictest processsubtest go src github com cockroachdb cockroach pkg sql logictest logic go github com cockroachdb cockroach pkg sql logictest logictest processtestfile go src github com cockroachdb cockroach pkg sql logictest logic go github com cockroachdb cockroach pkg sql logictest logictest runfile go src github com cockroachdb cockroach pkg sql logictest logic go github com cockroachdb cockroach pkg sql logictest runlogictest go src github com cockroachdb cockroach pkg sql logictest logic go testing trunner usr local go src testing testing go created by testing t run usr local go src testing testing go
0
148,954
19,560,751,759
IssuesEvent
2022-01-03 15:54:23
shaimael/Webgoat
https://api.github.com/repos/shaimael/Webgoat
opened
CVE-2021-28170 (Medium) detected in jakarta.el-3.0.3.jar
security vulnerability
## CVE-2021-28170 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jakarta.el-3.0.3.jar</b></p></summary> <p>Jakarta Expression Language provides a specification document, API, reference implementation and TCK that describes an expression language for Java applications.</p> <p>Library home page: <a href="https://projects.eclipse.org/projects/ee4j.el">https://projects.eclipse.org/projects/ee4j.el</a></p> <p>Path to dependency file: /webgoat-lessons/vulnerable-components/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-validation-2.4.3.jar (Root Library) - :x: **jakarta.el-3.0.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/shaimael/Webgoat/commit/06d8d0b5bb8a459ba3d47f61a64fe00c62662d81">06d8d0b5bb8a459ba3d47f61a64fe00c62662d81</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Jakarta Expression Language implementation 3.0.3 and earlier, a bug in the ELParserTokenManager enables invalid EL expressions to be evaluated as if they were valid. <p>Publish Date: 2021-05-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28170>CVE-2021-28170</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://securitylab.github.com/advisories/GHSL-2020-021-jakarta-el/">https://securitylab.github.com/advisories/GHSL-2020-021-jakarta-el/</a></p> <p>Release Date: 2021-05-26</p> <p>Fix Resolution: org.glassfish:jakarta.el:3.0.3.jbossorg-3, 3.0.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.glassfish","packageName":"jakarta.el","packageVersion":"3.0.3","packageFilePaths":["/webgoat-lessons/vulnerable-components/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-validation:2.4.3;org.glassfish:jakarta.el:3.0.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.glassfish:jakarta.el:3.0.3.jbossorg-3, 3.0.4","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-28170","vulnerabilityDetails":"In the Jakarta Expression Language implementation 3.0.3 and earlier, a bug in the ELParserTokenManager enables invalid EL expressions to be evaluated as if they were valid.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28170","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-28170 (Medium) detected in jakarta.el-3.0.3.jar - ## CVE-2021-28170 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jakarta.el-3.0.3.jar</b></p></summary> <p>Jakarta Expression Language provides a specification document, API, reference implementation and TCK that describes an expression language for Java applications.</p> <p>Library home page: <a href="https://projects.eclipse.org/projects/ee4j.el">https://projects.eclipse.org/projects/ee4j.el</a></p> <p>Path to dependency file: /webgoat-lessons/vulnerable-components/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar,/home/wss-scanner/.m2/repository/org/glassfish/jakarta.el/3.0.3/jakarta.el-3.0.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-validation-2.4.3.jar (Root Library) - :x: **jakarta.el-3.0.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/shaimael/Webgoat/commit/06d8d0b5bb8a459ba3d47f61a64fe00c62662d81">06d8d0b5bb8a459ba3d47f61a64fe00c62662d81</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the Jakarta Expression Language implementation 3.0.3 and earlier, a bug in the ELParserTokenManager enables invalid EL expressions to be evaluated as if they were valid. <p>Publish Date: 2021-05-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28170>CVE-2021-28170</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://securitylab.github.com/advisories/GHSL-2020-021-jakarta-el/">https://securitylab.github.com/advisories/GHSL-2020-021-jakarta-el/</a></p> <p>Release Date: 2021-05-26</p> <p>Fix Resolution: org.glassfish:jakarta.el:3.0.3.jbossorg-3, 3.0.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.glassfish","packageName":"jakarta.el","packageVersion":"3.0.3","packageFilePaths":["/webgoat-lessons/vulnerable-components/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-validation:2.4.3;org.glassfish:jakarta.el:3.0.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.glassfish:jakarta.el:3.0.3.jbossorg-3, 3.0.4","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2021-28170","vulnerabilityDetails":"In the Jakarta Expression Language implementation 3.0.3 and earlier, a bug in the ELParserTokenManager enables invalid EL expressions to be evaluated as if they were valid.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28170","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_process
cve medium detected in jakarta el jar cve medium severity vulnerability vulnerable library jakarta el jar jakarta expression language provides a specification document api reference implementation and tck that describes an expression language for java applications library home page a href path to dependency file webgoat lessons vulnerable components pom xml path to vulnerable library home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar home wss scanner repository org glassfish jakarta el jakarta el jar dependency hierarchy spring boot starter validation jar root library x jakarta el jar vulnerable library found in head commit a href found in base branch main vulnerability details in the jakarta expression language implementation and earlier a bug in the elparsertokenmanager enables invalid el expressions to be evaluated as if they were valid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org glassfish jakarta el jbossorg isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org springframework boot spring boot starter validation org glassfish jakarta el isminimumfixversionavailable true minimumfixversion org glassfish jakarta el jbossorg isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails in the jakarta expression language implementation and earlier a bug in the elparsertokenmanager enables invalid el expressions to be evaluated as if they were valid vulnerabilityurl
0
262,688
8,272,348,338
IssuesEvent
2018-09-16 19:12:16
letelete/Sleep-Cycle-Alarm
https://api.github.com/repos/letelete/Sleep-Cycle-Alarm
closed
Save wake up at list content after setting up one
bug high priority medium
List content gets reset when user leaves wake up at tab
1.0
Save wake up at list content after setting up one - List content gets reset when user leaves wake up at tab
non_process
save wake up at list content after setting up one list content gets reset when user leaves wake up at tab
0
15,592
19,718,105,577
IssuesEvent
2022-01-13 13:04:35
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
opened
Add support in Application frontend for BPMN gateways
area/process kind/user-story
<!-- Internal template. If you're external and want to suggest a user story, please use the Feature request template --> ## Description After backend now supports gateways frontend need to be updated to do the same ## Considerations - We need to update App API so that app frontend knows that there are multiple flows out of the current task. Example. Go back to data from confirmation and confirm - We need to design - ### Ops requirements <!-- Are there any requirements for monitoring? What is being built and what could go wrong? Are there any requirements related to backup? --> ## Acceptance criteria <!-- Describe criteria here (i.e. What is allowed/not allowed (negative tesing), validations, error messages and warnings etc.) --> ## Specification tasks - [ ] Design GUI for process navigation - [ ] - [ ] Development tasks are defined <!-- Add other spec tasks here using checkboxes --> ## Development tasks <!-- Add dev tasks here using checkboxes --> ## Test <!-- Add test tasks here using checkboxes --> ## Definition of done <!-- This chekclist is to verify that this issue meets [DoD](https://digdir-digitalisering.atlassian.net/wiki/spaces/T3KP/pages/5049246/Definition+of+Done) (Only for project members) before closing. --> - [ ] Documentation (docs.altinn.studio) is updated (if relevant) - [ ] Technical documentation - [ ] User documentation - [ ] QA - [ ] Manual test is complete (if relevant) - [ ] Automated test is implemented (if relevant) - [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
1.0
Add support in Application frontend for BPMN gateways - <!-- Internal template. If you're external and want to suggest a user story, please use the Feature request template --> ## Description After backend now supports gateways frontend need to be updated to do the same ## Considerations - We need to update App API so that app frontend knows that there are multiple flows out of the current task. Example. Go back to data from confirmation and confirm - We need to design - ### Ops requirements <!-- Are there any requirements for monitoring? What is being built and what could go wrong? Are there any requirements related to backup? --> ## Acceptance criteria <!-- Describe criteria here (i.e. What is allowed/not allowed (negative tesing), validations, error messages and warnings etc.) --> ## Specification tasks - [ ] Design GUI for process navigation - [ ] - [ ] Development tasks are defined <!-- Add other spec tasks here using checkboxes --> ## Development tasks <!-- Add dev tasks here using checkboxes --> ## Test <!-- Add test tasks here using checkboxes --> ## Definition of done <!-- This chekclist is to verify that this issue meets [DoD](https://digdir-digitalisering.atlassian.net/wiki/spaces/T3KP/pages/5049246/Definition+of+Done) (Only for project members) before closing. --> - [ ] Documentation (docs.altinn.studio) is updated (if relevant) - [ ] Technical documentation - [ ] User documentation - [ ] QA - [ ] Manual test is complete (if relevant) - [ ] Automated test is implemented (if relevant) - [ ] All tasks in this userstory are closed (i.e. remaining tasks are moved to other user stories or marked obsolete)
process
add support in application frontend for bpmn gateways description after backend now supports gateways frontend need to be updated to do the same considerations we need to update app api so that app frontend knows that there are multiple flows out of the current task example go back to data from confirmation and confirm we need to design ops requirements are there any requirements for monitoring what is being built and what could go wrong are there any requirements related to backup acceptance criteria specification tasks design gui for process navigation development tasks are defined development tasks test definition of done documentation docs altinn studio is updated if relevant technical documentation user documentation qa manual test is complete if relevant automated test is implemented if relevant all tasks in this userstory are closed i e remaining tasks are moved to other user stories or marked obsolete
1
120,823
4,794,840,303
IssuesEvent
2016-10-31 22:20:09
cyberpwnn/GlacialRealms
https://api.github.com/repos/cyberpwnn/GlacialRealms
closed
Mod+ could kill players on Islands
20 Minute ETA bug general high priority
Anyone who bypassed the Island Protection rules could attack others and become combat tagged and would end up killing both the player they accidentally hit and themselves.
1.0
Mod+ could kill players on Islands - Anyone who bypassed the Island Protection rules could attack others and become combat tagged and would end up killing both the player they accidentally hit and themselves.
non_process
mod could kill players on islands anyone who bypassed the island protection rules could attack others and become combat tagged and would end up killing both the player they accidentally hit and themselves
0
97,265
3,987,907,469
IssuesEvent
2016-05-09 07:23:53
iacopodeenosee/EasyScreenCast
https://api.github.com/repos/iacopodeenosee/EasyScreenCast
closed
Error: No property show_all on this GObject GstDeviceMonitor
PRIORITY: normal STATUS: confirmed TYPE: duplicate TYPE: support
In Firefox 45, Gnome 3.16.4, I hit the button to install EasyScreenCast and affirmed a subsequent Firefox inquiry, and a red Error mark appeared in place of the install button. Then I followed instructions and pressed Alt+F2, issued the r command, reloaded the installation page, and once more chose to install it. This time the red Error mark reappeared, with a tools button to the left. I pressed the tools button and got the following note: ``` Error: No property show_all on this GObject GstDeviceMonitor Stack trace: HelperWebcam<._init@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/utilwebcam.js:38 wrapper@resource:///org/gnome/gjs/modules/lang.js:169 _Base.prototype._construct@resource:///org/gnome/gjs/modules/lang.js:110 Class.prototype._construct/newClass@resource:///org/gnome/gjs/modules/lang.js:204 EasyScreenCastSettingsWidget<._init@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/prefs.js:90 wrapper@resource:///org/gnome/gjs/modules/lang.js:169 buildPrefsWidget@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/prefs.js:730 Application<._selectExtension@resource:///org/gnome/shell/extensionPrefs/main.js:95 wrapper@resource:///org/gnome/gjs/modules/lang.js:169 Application<._onCommandLine@resource:///org/gnome/shell/extensionPrefs/main.js:259 wrapper@resource:///org/gnome/gjs/modules/lang.js:169 main@resource:///org/gnome/shell/extensionPrefs/main.js:397 @<main>:1 ```
1.0
Error: No property show_all on this GObject GstDeviceMonitor - In Firefox 45, Gnome 3.16.4, I hit the button to install EasyScreenCast and affirmed a subsequent Firefox inquiry, and a red Error mark appeared in place of the install button. Then I followed instructions and pressed Alt+F2, issued the r command, reloaded the installation page, and once more chose to install it. This time the red Error mark reappeared, with a tools button to the left. I pressed the tools button and got the following note: ``` Error: No property show_all on this GObject GstDeviceMonitor Stack trace: HelperWebcam<._init@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/utilwebcam.js:38 wrapper@resource:///org/gnome/gjs/modules/lang.js:169 _Base.prototype._construct@resource:///org/gnome/gjs/modules/lang.js:110 Class.prototype._construct/newClass@resource:///org/gnome/gjs/modules/lang.js:204 EasyScreenCastSettingsWidget<._init@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/prefs.js:90 wrapper@resource:///org/gnome/gjs/modules/lang.js:169 buildPrefsWidget@/home/pbwolf/.local/share/gnome-shell/extensions/EasyScreenCast@iacopodeenosee.gmail.com/prefs.js:730 Application<._selectExtension@resource:///org/gnome/shell/extensionPrefs/main.js:95 wrapper@resource:///org/gnome/gjs/modules/lang.js:169 Application<._onCommandLine@resource:///org/gnome/shell/extensionPrefs/main.js:259 wrapper@resource:///org/gnome/gjs/modules/lang.js:169 main@resource:///org/gnome/shell/extensionPrefs/main.js:397 @<main>:1 ```
non_process
error no property show all on this gobject gstdevicemonitor in firefox gnome i hit the button to install easyscreencast and affirmed a subsequent firefox inquiry and a red error mark appeared in place of the install button then i followed instructions and pressed alt issued the r command reloaded the installation page and once more chose to install it this time the red error mark reappeared with a tools button to the left i pressed the tools button and got the following note error no property show all on this gobject gstdevicemonitor stack trace helperwebcam init home pbwolf local share gnome shell extensions easyscreencast iacopodeenosee gmail com utilwebcam js wrapper resource org gnome gjs modules lang js base prototype construct resource org gnome gjs modules lang js class prototype construct newclass resource org gnome gjs modules lang js easyscreencastsettingswidget init home pbwolf local share gnome shell extensions easyscreencast iacopodeenosee gmail com prefs js wrapper resource org gnome gjs modules lang js buildprefswidget home pbwolf local share gnome shell extensions easyscreencast iacopodeenosee gmail com prefs js application selectextension resource org gnome shell extensionprefs main js wrapper resource org gnome gjs modules lang js application oncommandline resource org gnome shell extensionprefs main js wrapper resource org gnome gjs modules lang js main resource org gnome shell extensionprefs main js
0
110,070
13,896,450,391
IssuesEvent
2020-10-19 17:12:57
JeschkeLab/DeerLab
https://api.github.com/repos/JeschkeLab/DeerLab
opened
Simplify parameter boundaries interface for fisignal
design
Another remnant of the old MATLAB interface. In `fitsignal` the lower/upper boundaries of the different parameter subsets are specified as a single keyword ````python fitsignal(__, lb = [lb_dd,lb_bg,lb_ex], ub = [ub_dd,ub_bg,ub_ex]) ```` Since one seldom needs to re-define all boundaries for all subsets, it would make sense to simplify this to multiple keywords ````python fitsignal(__, lb_dd = lb_dd, lb_bg = lb_bg lb_ex = lb_ex) ```` this way a quick assignment can be done without needing to worry about the order of the parameter subsets.
1.0
Simplify parameter boundaries interface for fisignal - Another remnant of the old MATLAB interface. In `fitsignal` the lower/upper boundaries of the different parameter subsets are specified as a single keyword ````python fitsignal(__, lb = [lb_dd,lb_bg,lb_ex], ub = [ub_dd,ub_bg,ub_ex]) ```` Since one seldom needs to re-define all boundaries for all subsets, it would make sense to simplify this to multiple keywords ````python fitsignal(__, lb_dd = lb_dd, lb_bg = lb_bg lb_ex = lb_ex) ```` this way a quick assignment can be done without needing to worry about the order of the parameter subsets.
non_process
simplify parameter boundaries interface for fisignal another remnant of the old matlab interface in fitsignal the lower upper boundaries of the different parameter subsets are specified as a single keyword python fitsignal lb ub since one seldom needs to re define all boundaries for all subsets it would make sense to simplify this to multiple keywords python fitsignal lb dd lb dd lb bg lb bg lb ex lb ex this way a quick assignment can be done without needing to worry about the order of the parameter subsets
0
19,540
25,858,235,558
IssuesEvent
2022-12-13 15:11:16
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
Move samples to client library repositories
type: process
This is a list of all the samples projects that should likely be moved to client library repositories: - [x] [asset](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/asset/cloud-client) - [x] [automl](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/automl) - [x] [bigquery](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery) - [x] [bigquery_datatransfer](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery/transfer) - [x] [bigquery_storage](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery_storage/to_dataframe) - [x] [bigtable](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigtable) - https://github.com/googleapis/python-bigtable/pull/38 - [ ] [cloud-sql](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/cloud-sql) - [ ] [compute](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/compute) - [x] [container_registry](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/container_registry/container_analysise) - [x] [datacatalog](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datacatalog/cloud-client) - [x] [datalabeling](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datalabeling) - [x] [dataproc](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dataproc) - [ ] [datastore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datastore) - [x] [dialogflow](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dialogflow/cloud-client) - [x] [dlp](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dlp) - [ ] [dns](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dns/api) - [x] [document](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/document/cloud-client) - [x] [error_reporting](error_reporting) - [ ] [firestore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/firestore/cloud-client) - [ ] [iot](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/iot/api-client) - [ ] [jobs](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/jobs) - [x] [kms](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/kms) - https://github.com/googleapis/python-kms/pull/14 - [x] [language](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/language) - [x] [logging](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/logging/cloud-client) - [x] [media-translation](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/media-translation/cloud-client) - [ ] [memorystore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/memorystore/redis) - [ ] [monitoring](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/monitoring/api/v3) - [x] [pubsub](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/pubsub) - [x] [scheduler](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/scheduler) - [x] [secretmanager](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/secretmanager/api-client) - [x] [spanner](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/spanner/cloud-client) - [x] [speech](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/speech) - [ ] [storage](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/storage) - [x] [tables](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/tables/automl) (technically automl, unclear if these should be merged) - [x] [tasks](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/tasks) - [x] [texttospeech](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/texttospeech/cloud-client) - #3845 - [ ] [trace](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/trace) - [x] [translate](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/translate) - [x] [video](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/video/cloud-client) - [x] [vision](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/vision)
1.0
Move samples to client library repositories - This is a list of all the samples projects that should likely be moved to client library repositories: - [x] [asset](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/asset/cloud-client) - [x] [automl](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/automl) - [x] [bigquery](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery) - [x] [bigquery_datatransfer](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery/transfer) - [x] [bigquery_storage](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigquery_storage/to_dataframe) - [x] [bigtable](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/bigtable) - https://github.com/googleapis/python-bigtable/pull/38 - [ ] [cloud-sql](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/cloud-sql) - [ ] [compute](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/compute) - [x] [container_registry](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/container_registry/container_analysise) - [x] [datacatalog](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datacatalog/cloud-client) - [x] [datalabeling](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datalabeling) - [x] [dataproc](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dataproc) - [ ] [datastore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/datastore) - [x] [dialogflow](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dialogflow/cloud-client) - [x] [dlp](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dlp) - [ ] [dns](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/dns/api) - [x] [document](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/document/cloud-client) - [x] [error_reporting](error_reporting) - [ ] [firestore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/firestore/cloud-client) - [ ] [iot](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/iot/api-client) - [ ] [jobs](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/jobs) - [x] [kms](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/kms) - https://github.com/googleapis/python-kms/pull/14 - [x] [language](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/language) - [x] [logging](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/logging/cloud-client) - [x] [media-translation](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/media-translation/cloud-client) - [ ] [memorystore](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/memorystore/redis) - [ ] [monitoring](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/monitoring/api/v3) - [x] [pubsub](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/pubsub) - [x] [scheduler](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/scheduler) - [x] [secretmanager](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/secretmanager/api-client) - [x] [spanner](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/spanner/cloud-client) - [x] [speech](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/speech) - [ ] [storage](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/storage) - [x] [tables](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/tables/automl) (technically automl, unclear if these should be merged) - [x] [tasks](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/tasks) - [x] [texttospeech](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/texttospeech/cloud-client) - #3845 - [ ] [trace](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/trace) - [x] [translate](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/translate) - [x] [video](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/video/cloud-client) - [x] [vision](https://github.com/GoogleCloudPlatform/python-docs-samples/tree/master/vision)
process
move samples to client library repositories this is a list of all the samples projects that should likely be moved to client library repositories error reporting technically automl unclear if these should be merged
1
120,755
4,793,622,089
IssuesEvent
2016-10-31 18:40:54
CogRob/workspace
https://api.github.com/repos/CogRob/workspace
opened
Continuous Integration
P2 - NORMAL PRIORITY
We should use CI to ensure all the test passes before merging to master.
1.0
Continuous Integration - We should use CI to ensure all the test passes before merging to master.
non_process
continuous integration we should use ci to ensure all the test passes before merging to master
0
219,585
24,501,586,532
IssuesEvent
2022-10-10 13:13:08
nidhi7598/linux-3.0.35
https://api.github.com/repos/nidhi7598/linux-3.0.35
opened
CVE-2019-13631 (Medium) detected in linux-stable-rtv3.8.6
security vulnerability
## CVE-2019-13631 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In parse_hid_report_descriptor in drivers/input/tablet/gtco.c in the Linux kernel through 5.2.1, a malicious USB device can send an HID report that triggers an out-of-bounds write during generation of debugging messages. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13631>CVE-2019-13631</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-13631">https://nvd.nist.gov/vuln/detail/CVE-2019-13631</a></p> <p>Release Date: 2019-07-26</p> <p>Fix Resolution: linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-13631 (Medium) detected in linux-stable-rtv3.8.6 - ## CVE-2019-13631 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35/commit/4cc6d4a22f88b8effe1090492c1a242ce587b492">4cc6d4a22f88b8effe1090492c1a242ce587b492</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/tablet/gtco.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In parse_hid_report_descriptor in drivers/input/tablet/gtco.c in the Linux kernel through 5.2.1, a malicious USB device can send an HID report that triggers an out-of-bounds write during generation of debugging messages. <p>Publish Date: 2019-07-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13631>CVE-2019-13631</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-13631">https://nvd.nist.gov/vuln/detail/CVE-2019-13631</a></p> <p>Release Date: 2019-07-26</p> <p>Fix Resolution: linux-yocto - 5.4.20+gitAUTOINC+c11911d4d1_f4d7dbafb1,4.8.26+gitAUTOINC+1c60e003c7_27efc3ba68</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers input tablet gtco c drivers input tablet gtco c drivers input tablet gtco c vulnerability details in parse hid report descriptor in drivers input tablet gtco c in the linux kernel through a malicious usb device can send an hid report that triggers an out of bounds write during generation of debugging messages publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution linux yocto gitautoinc gitautoinc step up your open source security game with mend
0
5,850
8,676,158,610
IssuesEvent
2018-11-30 13:17:21
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
opened
Opening RocksDB on partition install
broker enhancement logstreams stream processor
I think it would be nice if we could create an RocksDB instance per Partition, this would mean that we open the RocksDB on Partition install and give the LogBlockIndex (if the LogBlockIndex is allready migrated to RocksDB) and StreamProcessor access to the database, maybe via dependency. There could exist an Service/Actor, which creates periodically checkpoints from the database. But we have to make sure that this causes no inconsistencies if an actor writes to RocksDB and an checkpoint is concurrently created. Possible solutions for this would be to suspend the other actors for this time of checkpoint creation OR they run all in the same actor OR maybe this is even no problem we have to find this out. This solution would give us a separation of the RocksDB creation, data snapshoting and Event processing, because the StreamProcessorController is only interested in reading and writing events. We could deleted then the whole Snapshot stuff.
1.0
Opening RocksDB on partition install - I think it would be nice if we could create an RocksDB instance per Partition, this would mean that we open the RocksDB on Partition install and give the LogBlockIndex (if the LogBlockIndex is allready migrated to RocksDB) and StreamProcessor access to the database, maybe via dependency. There could exist an Service/Actor, which creates periodically checkpoints from the database. But we have to make sure that this causes no inconsistencies if an actor writes to RocksDB and an checkpoint is concurrently created. Possible solutions for this would be to suspend the other actors for this time of checkpoint creation OR they run all in the same actor OR maybe this is even no problem we have to find this out. This solution would give us a separation of the RocksDB creation, data snapshoting and Event processing, because the StreamProcessorController is only interested in reading and writing events. We could deleted then the whole Snapshot stuff.
process
opening rocksdb on partition install i think it would be nice if we could create an rocksdb instance per partition this would mean that we open the rocksdb on partition install and give the logblockindex if the logblockindex is allready migrated to rocksdb and streamprocessor access to the database maybe via dependency there could exist an service actor which creates periodically checkpoints from the database but we have to make sure that this causes no inconsistencies if an actor writes to rocksdb and an checkpoint is concurrently created possible solutions for this would be to suspend the other actors for this time of checkpoint creation or they run all in the same actor or maybe this is even no problem we have to find this out this solution would give us a separation of the rocksdb creation data snapshoting and event processing because the streamprocessorcontroller is only interested in reading and writing events we could deleted then the whole snapshot stuff
1
255,633
19,318,921,753
IssuesEvent
2021-12-14 01:39:42
sillsdev/ptx2pdf
https://api.github.com/repos/sillsdev/ptx2pdf
closed
Add an option to have the header in a second language/script
priority enhancement bug Fixed? Documentation
Some regions of the world (Biblica South Asia, + BSI India) use 2 different languages or scripts for the header. Currently this is not possible, but perhaps it could be made possible. Consideration also has to be given to digit styles, styling of text, size and fonts etc.
1.0
Add an option to have the header in a second language/script - Some regions of the world (Biblica South Asia, + BSI India) use 2 different languages or scripts for the header. Currently this is not possible, but perhaps it could be made possible. Consideration also has to be given to digit styles, styling of text, size and fonts etc.
non_process
add an option to have the header in a second language script some regions of the world biblica south asia bsi india use different languages or scripts for the header currently this is not possible but perhaps it could be made possible consideration also has to be given to digit styles styling of text size and fonts etc
0
9,986
13,034,505,825
IssuesEvent
2020-07-28 08:49:18
threefoldtech/js-sdk
https://api.github.com/repos/threefoldtech/js-sdk
closed
stellar client activate_through_friendbot no longer works
priority_critical process_wontfix type_bug
It gives a bad request and the documentation does not state a traling slash after the absolute url The trailing slash can not be removed when using requests: https://stackoverflow.com/questions/27248521/python-http-request-url-end-slash also, when using curl: `curl "https://friendbot.stellar.org?addr=GA47YZA3PKFUZMPLQ3B5F2E3CJIB57TGGU7SPCQT2WAEYKN766PWIMB3"` curl: (60) SSL certificate problem: certificate has expired More details here: https://curl.haxx.se/docs/sslcerts.html curl failed to verify the legitimacy of the server and therefore could not establish a secure connection to it. To learn more about this situation and how to fix it, please visit the web page mentioned above.
1.0
stellar client activate_through_friendbot no longer works - It gives a bad request and the documentation does not state a traling slash after the absolute url The trailing slash can not be removed when using requests: https://stackoverflow.com/questions/27248521/python-http-request-url-end-slash also, when using curl: `curl "https://friendbot.stellar.org?addr=GA47YZA3PKFUZMPLQ3B5F2E3CJIB57TGGU7SPCQT2WAEYKN766PWIMB3"` curl: (60) SSL certificate problem: certificate has expired More details here: https://curl.haxx.se/docs/sslcerts.html curl failed to verify the legitimacy of the server and therefore could not establish a secure connection to it. To learn more about this situation and how to fix it, please visit the web page mentioned above.
process
stellar client activate through friendbot no longer works it gives a bad request and the documentation does not state a traling slash after the absolute url the trailing slash can not be removed when using requests also when using curl curl curl ssl certificate problem certificate has expired more details here curl failed to verify the legitimacy of the server and therefore could not establish a secure connection to it to learn more about this situation and how to fix it please visit the web page mentioned above
1
324,962
24,027,281,428
IssuesEvent
2022-09-15 12:34:39
vuetifyjs/vuetify
https://api.github.com/repos/vuetifyjs/vuetify
closed
[Bug Report][2.6.10] docs(VDataIterator): wrong slot type
T: documentation
### Environment **Vuetify Version:** 2.6.10 **Vue Version:** 2.7.10 **Browsers:** Edge 105.0.1343.27 **OS:** Windows 10 ### Steps to reproduce ![image](https://user-images.githubusercontent.com/40021217/190403552-f585156a-d696-4466-aff1-5ceed0ed14cc.png) ### Expected Behavior `#default` is a scoped slot ### Actual Behavior Both slot `#default` and scoped slot `#default` exist in docs. ### Reproduction Link https://vuetifyjs.com/en/api/v-data-iterator/#slots <!-- generated by vuetify-issue-helper. DO NOT REMOVE -->
1.0
[Bug Report][2.6.10] docs(VDataIterator): wrong slot type - ### Environment **Vuetify Version:** 2.6.10 **Vue Version:** 2.7.10 **Browsers:** Edge 105.0.1343.27 **OS:** Windows 10 ### Steps to reproduce ![image](https://user-images.githubusercontent.com/40021217/190403552-f585156a-d696-4466-aff1-5ceed0ed14cc.png) ### Expected Behavior `#default` is a scoped slot ### Actual Behavior Both slot `#default` and scoped slot `#default` exist in docs. ### Reproduction Link https://vuetifyjs.com/en/api/v-data-iterator/#slots <!-- generated by vuetify-issue-helper. DO NOT REMOVE -->
non_process
docs vdataiterator wrong slot type environment vuetify version vue version browsers edge os windows steps to reproduce expected behavior default is a scoped slot actual behavior both slot default and scoped slot default exist in docs reproduction link
0
19,067
25,088,875,527
IssuesEvent
2022-11-08 03:28:24
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
[Mirror] make 4.4
P2 type: process team-OSS mirror request
### Please list the URLs of the archives you'd like to mirror: Please mirror http://ftpmirror.gnu.org/gnu/make/make-4.4.tar.gz for `rules_foreign_cc`
1.0
[Mirror] make 4.4 - ### Please list the URLs of the archives you'd like to mirror: Please mirror http://ftpmirror.gnu.org/gnu/make/make-4.4.tar.gz for `rules_foreign_cc`
process
make please list the urls of the archives you d like to mirror please mirror for rules foreign cc
1
16,546
21,568,598,986
IssuesEvent
2022-05-02 04:17:56
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Crime Lab S.F.
suggested title in process
Title: Crime Lab S.F. Type (film/tv show): TV show Film or show in which it appears: Monk Is the parent film/show streaming anywhere? Peacock, Amazon Prime About when in the parent film/show does it appear? Season 2, Episode 12, "Mr. Monk and the TV Star" Actual footage of the film/show can be seen (yes/no)? Yes Cast: Brad Terry Link to Monk wiki: https://monk.fandom.com/wiki/Mr._Monk_and_the_TV_Star
1.0
Add Crime Lab S.F. - Title: Crime Lab S.F. Type (film/tv show): TV show Film or show in which it appears: Monk Is the parent film/show streaming anywhere? Peacock, Amazon Prime About when in the parent film/show does it appear? Season 2, Episode 12, "Mr. Monk and the TV Star" Actual footage of the film/show can be seen (yes/no)? Yes Cast: Brad Terry Link to Monk wiki: https://monk.fandom.com/wiki/Mr._Monk_and_the_TV_Star
process
add crime lab s f title crime lab s f type film tv show tv show film or show in which it appears monk is the parent film show streaming anywhere peacock amazon prime about when in the parent film show does it appear season episode mr monk and the tv star actual footage of the film show can be seen yes no yes cast brad terry link to monk wiki
1
19,840
26,242,571,521
IssuesEvent
2023-01-05 12:47:38
vesoft-inc/nebula
https://api.github.com/repos/vesoft-inc/nebula
closed
with USER role can do balance data
type/bug severity/major auto-sync affects/master process/done
as title, however in doc, balance data only god can do ![image](https://user-images.githubusercontent.com/106510322/207784865-e7cf04cb-9720-4aa0-aa9a-3e765e1215a2.png) ![image](https://user-images.githubusercontent.com/106510322/207784972-4e85d891-74cc-44a2-93c8-81c48bb2bb18.png) while --enable_authorize=true ![image](https://user-images.githubusercontent.com/106510322/207784197-bf442cc9-481a-4faa-82fc-4fc6016b374b.png) <!-- Please check the [FAQ](https://docs.nebula-graph.com.cn/master/20.appendix/0.FAQ/) documentation and old issues before raising an issue in case someone has asked the same question that you are asking. --> balace leader also can do with user role **Describe the bug (__required__)** <!-- A clear and concise description of what the bug is. --> **Your Environments (__required__)** ![image](https://user-images.githubusercontent.com/106510322/207785405-8cc44275-09de-48dc-9191-d64981c0df3f.png) nebula-ent-nightly, commit id : f1e5ec3 **How To Reproduce(__required__)** Steps to reproduce the behavior: ``` 1. CREATE USER user1 WITH PASSWORD 'nebula'; 2. grant role User on yx_test to user1 3. login in with user1; 4. do: balance data; ``` **Expected behavior** report error ,only god role can do balance data
1.0
with USER role can do balance data - as title, however in doc, balance data only god can do ![image](https://user-images.githubusercontent.com/106510322/207784865-e7cf04cb-9720-4aa0-aa9a-3e765e1215a2.png) ![image](https://user-images.githubusercontent.com/106510322/207784972-4e85d891-74cc-44a2-93c8-81c48bb2bb18.png) while --enable_authorize=true ![image](https://user-images.githubusercontent.com/106510322/207784197-bf442cc9-481a-4faa-82fc-4fc6016b374b.png) <!-- Please check the [FAQ](https://docs.nebula-graph.com.cn/master/20.appendix/0.FAQ/) documentation and old issues before raising an issue in case someone has asked the same question that you are asking. --> balace leader also can do with user role **Describe the bug (__required__)** <!-- A clear and concise description of what the bug is. --> **Your Environments (__required__)** ![image](https://user-images.githubusercontent.com/106510322/207785405-8cc44275-09de-48dc-9191-d64981c0df3f.png) nebula-ent-nightly, commit id : f1e5ec3 **How To Reproduce(__required__)** Steps to reproduce the behavior: ``` 1. CREATE USER user1 WITH PASSWORD 'nebula'; 2. grant role User on yx_test to user1 3. login in with user1; 4. do: balance data; ``` **Expected behavior** report error ,only god role can do balance data
process
with user role can do balance data as title however in doc, balance data only god can do while enable authorize true balace leader also can do with user role describe the bug required your environments required nebula ent nightly commit id how to reproduce required steps to reproduce the behavior create user with password nebula grant role user on yx test to login in with do balance data expected behavior report error only god role can do balance data
1
11,298
14,103,847,985
IssuesEvent
2020-11-06 10:54:27
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
GO:0097754 Name clathrin-mediated membrane bending
PomBase cellular processes low priority
missing parent GO:0048268 clathrin coat assembly > or at least there is no link between the membrane bending part and the broader pathway
1.0
GO:0097754 Name clathrin-mediated membrane bending - missing parent GO:0048268 clathrin coat assembly > or at least there is no link between the membrane bending part and the broader pathway
process
go name clathrin mediated membrane bending missing parent go clathrin coat assembly or at least there is no link between the membrane bending part and the broader pathway
1
642,637
20,908,855,414
IssuesEvent
2022-03-24 07:05:53
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
esp32 build error while applicaton in T2 topology
bug priority: low platform: ESP32
`**Describe the bug** build error when create an application in T2 topology based on esp32 Please also mention any information which could help others to understand the problem you're facing: - What target platform are you using? usually create an application using T2 topology described in https://docs.zephyrproject.org/latest/guides/west/workspaces.html - What have you tried to diagnose or workaround this issue? - no idea on this **To Reproduce** Steps to reproduce the behavior: 1. create an application using T2 topology you could follow the instructions bellow: ``` $ git clone https://gitee.com/fwatch/twatch $ cd twatch $ west init -l app/ $ west update # update modules $ west espressif update # update submodules ``` 2. build samples/hell_word ``` west build -p auto -b esp32 -s app ``` 3. See error ``` [84/85] Linking C executable bootloader.elf [85/85] Generating binary image from built executable esptool.py v3.1-dev Merged 1 ELF section Generated D:/workspace/twatch/build/esp-idf/build/bootloader/bootloader.bin [159/169] Linking C executable zephyr\zephyr_pre0.elf FAILED: zephyr/zephyr_pre0.elf zephyr/zephyr_pre0.map D:/workspace/twatch/build/zephyr/zephyr_pre0.map cmd.exe /C "cd . && C:\Users\gouqs\.espressif\tools\zephyr\xtensa-esp32-elf\bin\xtensa-esp32-elf-gcc.exe zephyr/CMakeFiles/zephyr_pre0.dir/misc/empty_file.c.obj -o zephyr\zephyr_pre0.elf zephyr/CMakeFiles/offsets.dir/./arch/xtensa/core/offsets/offsets.c.obj -Wl,-T zephyr/linker_zephyr_pre0.cmd -Wl,-Map=D:/workspace/twatch/build/zephyr/zephyr_pre0.map -Wl,--whole-archive app/libapp.a zephyr/libzephyr.a zephyr/arch/common/libarch__common.a zephyr/arch/arch/xtensa/core/libarch__xtensa__core.a zephyr/lib/libc/minimal/liblib__libc__minimal.a zephyr/lib/posix/liblib__posix.a zephyr/drivers/interrupt_controller/libdrivers__interrupt_controller.a zephyr/drivers/clock_control/libdrivers__clock_control.a zephyr/drivers/console/libdrivers__console.a zephyr/drivers/gpio/libdrivers__gpio.a zephyr/drivers/i2c/libdrivers__i2c.a zephyr/drivers/serial/libdrivers__serial.a zephyr/drivers/timer/libdrivers__timer.a -Wl,--no-whole-archive zephyr/kernel/libkernel.a -L"c:/users/gouqs/.espressif/tools/zephyr/xtensa-esp32-elf/bin/../lib/gcc/xtensa-esp32-elf/8.4.0" -LD:/workspace/twatch/build/zephyr -lgcc zephyr/arch/common/libisr_tables.a -lgcc -no-pie -Wl,--gc-sections -Wl,--build-id=none -Wl,--sort-common=descending -Wl,--sort-section=alignment -Wl,-u,_OffsetAbsSyms -Wl,-u,_ConfigAbsSyms -nostdlib -static -Wl,-X -Wl,-N -Wl,--orphan-handling=warn -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/src/linker/esp32.rom.alias.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.api.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.libgcc.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.redefined.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp32/ld/esp32.peripherals.ld && cmd.exe /C "cd /D D:\workspace\twatch\build\zephyr && "C:\Program Files\CMake\bin\cmake.exe" -E echo "" c:/users/gouqs/.espressif/tools/zephyr/xtensa-esp32-elf/bin/../lib/gcc/xtensa-esp32-elf/8.4.0/../../../../xtensa-esp32-elf/bin/ld.exe: zephyr/drivers/serial/libdrivers__serial.a(uart_esp32.c.obj):(.literal.uart_esp32_configure+0x0): undefined reference to `uart_periph_signal' collect2.exe: error: ld returned 1 exit status ninja: build stopped: subcommand failed. FATAL ERROR: command exited with status 1: 'C:\Program Files\CMake\bin\cmake.EXE' --build 'D:\workspace\twatch\build' ``` **Expected behavior** A clear and concise description of what you expected to happen. there is no error when using T2 topology **Impact** **Logs and console output** **Environment (please complete the following information):** - OS: (Windows 11 build successful with T1 topology ) **Additional context**
1.0
esp32 build error while applicaton in T2 topology - `**Describe the bug** build error when create an application in T2 topology based on esp32 Please also mention any information which could help others to understand the problem you're facing: - What target platform are you using? usually create an application using T2 topology described in https://docs.zephyrproject.org/latest/guides/west/workspaces.html - What have you tried to diagnose or workaround this issue? - no idea on this **To Reproduce** Steps to reproduce the behavior: 1. create an application using T2 topology you could follow the instructions bellow: ``` $ git clone https://gitee.com/fwatch/twatch $ cd twatch $ west init -l app/ $ west update # update modules $ west espressif update # update submodules ``` 2. build samples/hell_word ``` west build -p auto -b esp32 -s app ``` 3. See error ``` [84/85] Linking C executable bootloader.elf [85/85] Generating binary image from built executable esptool.py v3.1-dev Merged 1 ELF section Generated D:/workspace/twatch/build/esp-idf/build/bootloader/bootloader.bin [159/169] Linking C executable zephyr\zephyr_pre0.elf FAILED: zephyr/zephyr_pre0.elf zephyr/zephyr_pre0.map D:/workspace/twatch/build/zephyr/zephyr_pre0.map cmd.exe /C "cd . && C:\Users\gouqs\.espressif\tools\zephyr\xtensa-esp32-elf\bin\xtensa-esp32-elf-gcc.exe zephyr/CMakeFiles/zephyr_pre0.dir/misc/empty_file.c.obj -o zephyr\zephyr_pre0.elf zephyr/CMakeFiles/offsets.dir/./arch/xtensa/core/offsets/offsets.c.obj -Wl,-T zephyr/linker_zephyr_pre0.cmd -Wl,-Map=D:/workspace/twatch/build/zephyr/zephyr_pre0.map -Wl,--whole-archive app/libapp.a zephyr/libzephyr.a zephyr/arch/common/libarch__common.a zephyr/arch/arch/xtensa/core/libarch__xtensa__core.a zephyr/lib/libc/minimal/liblib__libc__minimal.a zephyr/lib/posix/liblib__posix.a zephyr/drivers/interrupt_controller/libdrivers__interrupt_controller.a zephyr/drivers/clock_control/libdrivers__clock_control.a zephyr/drivers/console/libdrivers__console.a zephyr/drivers/gpio/libdrivers__gpio.a zephyr/drivers/i2c/libdrivers__i2c.a zephyr/drivers/serial/libdrivers__serial.a zephyr/drivers/timer/libdrivers__timer.a -Wl,--no-whole-archive zephyr/kernel/libkernel.a -L"c:/users/gouqs/.espressif/tools/zephyr/xtensa-esp32-elf/bin/../lib/gcc/xtensa-esp32-elf/8.4.0" -LD:/workspace/twatch/build/zephyr -lgcc zephyr/arch/common/libisr_tables.a -lgcc -no-pie -Wl,--gc-sections -Wl,--build-id=none -Wl,--sort-common=descending -Wl,--sort-section=alignment -Wl,-u,_OffsetAbsSyms -Wl,-u,_ConfigAbsSyms -nostdlib -static -Wl,-X -Wl,-N -Wl,--orphan-handling=warn -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/src/linker/esp32.rom.alias.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.api.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.libgcc.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp_rom/esp32/ld/esp32.rom.redefined.ld -TD:/workspace/twatch/modules/hal/espressif/zephyr/esp32/../../components/esp32/ld/esp32.peripherals.ld && cmd.exe /C "cd /D D:\workspace\twatch\build\zephyr && "C:\Program Files\CMake\bin\cmake.exe" -E echo "" c:/users/gouqs/.espressif/tools/zephyr/xtensa-esp32-elf/bin/../lib/gcc/xtensa-esp32-elf/8.4.0/../../../../xtensa-esp32-elf/bin/ld.exe: zephyr/drivers/serial/libdrivers__serial.a(uart_esp32.c.obj):(.literal.uart_esp32_configure+0x0): undefined reference to `uart_periph_signal' collect2.exe: error: ld returned 1 exit status ninja: build stopped: subcommand failed. FATAL ERROR: command exited with status 1: 'C:\Program Files\CMake\bin\cmake.EXE' --build 'D:\workspace\twatch\build' ``` **Expected behavior** A clear and concise description of what you expected to happen. there is no error when using T2 topology **Impact** **Logs and console output** **Environment (please complete the following information):** - OS: (Windows 11 build successful with T1 topology ) **Additional context**
non_process
build error while applicaton in topology describe the bug build error when create an application in topology based on please also mention any information which could help others to understand the problem you re facing what target platform are you using usually create an application using topology described in what have you tried to diagnose or workaround this issue no idea on this to reproduce steps to reproduce the behavior create an application using topology you could follow the instructions bellow git clone cd twatch west init l app west update update modules west espressif update update submodules build samples hell word west build p auto b s app see error linking c executable bootloader elf generating binary image from built executable esptool py dev merged elf section generated d workspace twatch build esp idf build bootloader bootloader bin linking c executable zephyr zephyr elf failed zephyr zephyr elf zephyr zephyr map d workspace twatch build zephyr zephyr map cmd exe c cd c users gouqs espressif tools zephyr xtensa elf bin xtensa elf gcc exe zephyr cmakefiles zephyr dir misc empty file c obj o zephyr zephyr elf zephyr cmakefiles offsets dir arch xtensa core offsets offsets c obj wl t zephyr linker zephyr cmd wl map d workspace twatch build zephyr zephyr map wl whole archive app libapp a zephyr libzephyr a zephyr arch common libarch common a zephyr arch arch xtensa core libarch xtensa core a zephyr lib libc minimal liblib libc minimal a zephyr lib posix liblib posix a zephyr drivers interrupt controller libdrivers interrupt controller a zephyr drivers clock control libdrivers clock control a zephyr drivers console libdrivers console a zephyr drivers gpio libdrivers gpio a zephyr drivers libdrivers a zephyr drivers serial libdrivers serial a zephyr drivers timer libdrivers timer a wl no whole archive zephyr kernel libkernel a l c users gouqs espressif tools zephyr xtensa elf bin lib gcc xtensa elf ld workspace twatch build zephyr lgcc zephyr arch common libisr tables a lgcc no pie wl gc sections wl build id none wl sort common descending wl sort section alignment wl u offsetabssyms wl u configabssyms nostdlib static wl x wl n wl orphan handling warn td workspace twatch modules hal espressif zephyr src linker rom alias ld td workspace twatch modules hal espressif zephyr components esp rom ld rom ld td workspace twatch modules hal espressif zephyr components esp rom ld rom api ld td workspace twatch modules hal espressif zephyr components esp rom ld rom libgcc ld td workspace twatch modules hal espressif zephyr components esp rom ld rom redefined ld td workspace twatch modules hal espressif zephyr components ld peripherals ld cmd exe c cd d d workspace twatch build zephyr c program files cmake bin cmake exe e echo c users gouqs espressif tools zephyr xtensa elf bin lib gcc xtensa elf xtensa elf bin ld exe zephyr drivers serial libdrivers serial a uart c obj literal uart configure undefined reference to uart periph signal exe error ld returned exit status ninja build stopped subcommand failed fatal error command exited with status c program files cmake bin cmake exe build d workspace twatch build expected behavior a clear and concise description of what you expected to happen there is no error when using topology impact logs and console output environment please complete the following information os windows build successful with topology additional context
0
16,610
21,671,089,029
IssuesEvent
2022-05-08 00:39:42
veritablequandary/senatebot
https://api.github.com/repos/veritablequandary/senatebot
closed
[FR] Welcome Menu - Specific Guild
feature inprocess
**Description** When users join the server and are sent the welcome menu, they should be given the option to contact the recruiters for a specific guild. That selection will notify that guild's recruiters in the recruitment channel. **Suggested By** Brax Nightfighter
1.0
[FR] Welcome Menu - Specific Guild - **Description** When users join the server and are sent the welcome menu, they should be given the option to contact the recruiters for a specific guild. That selection will notify that guild's recruiters in the recruitment channel. **Suggested By** Brax Nightfighter
process
welcome menu specific guild description when users join the server and are sent the welcome menu they should be given the option to contact the recruiters for a specific guild that selection will notify that guild s recruiters in the recruitment channel suggested by brax nightfighter
1
361,733
25,350,082,628
IssuesEvent
2022-11-19 17:09:22
ahmed-nazir/Capstone
https://api.github.com/repos/ahmed-nazir/Capstone
closed
[DOC] - VnV Plan Section 6.1, 6.3
documentation
**What document(s) are being updated?** VnVPlan.tex **What is being added/updated? (optional)** Brief description of documentation change
1.0
[DOC] - VnV Plan Section 6.1, 6.3 - **What document(s) are being updated?** VnVPlan.tex **What is being added/updated? (optional)** Brief description of documentation change
non_process
vnv plan section what document s are being updated vnvplan tex what is being added updated optional brief description of documentation change
0
287,424
24,829,141,599
IssuesEvent
2022-10-26 00:37:51
jessicalewinter/Utilities
https://api.github.com/repos/jessicalewinter/Utilities
opened
Migrar módulos do +Sources/Utilities/UIKit/Extensions/UITableView+Dequeuable.swift
testing-modules
Esse módulo possui TestDoubles em sua interface e deve ser migrado para um novo módulo de Testing
1.0
Migrar módulos do +Sources/Utilities/UIKit/Extensions/UITableView+Dequeuable.swift - Esse módulo possui TestDoubles em sua interface e deve ser migrado para um novo módulo de Testing
non_process
migrar módulos do sources utilities uikit extensions uitableview dequeuable swift esse módulo possui testdoubles em sua interface e deve ser migrado para um novo módulo de testing
0
7,705
10,799,971,794
IssuesEvent
2019-11-06 13:21:22
prisma-labs/issues
https://api.github.com/repos/prisma-labs/issues
opened
Create a Project Statuses Board
type/process
We need a table that lists the primary projects we are or have worked on and what there status is. We sometimes get asked questions about why some project is is inactive or what its plans are. - Make it a new notion table in the root `Labs` page - Put it below the roadmap - Entries should include `nexus` `nexus-prisma` `graphql-playground` `pumpkins` - Columns should include `name: string` `status: Status` `repo: Link` `reason: string` - `Status` can be single select of `active this quarter` | `inactive this quarter` | `indefinitely inactive` - `Reason` should provide an explanation for the current status (if needed) - Idea: Consider using emojis or emoji+text for each `status`. For example: 🛫|🛬|✈️ or maybe weather themed ☀️|⛈|🌤|☁️
1.0
Create a Project Statuses Board - We need a table that lists the primary projects we are or have worked on and what there status is. We sometimes get asked questions about why some project is is inactive or what its plans are. - Make it a new notion table in the root `Labs` page - Put it below the roadmap - Entries should include `nexus` `nexus-prisma` `graphql-playground` `pumpkins` - Columns should include `name: string` `status: Status` `repo: Link` `reason: string` - `Status` can be single select of `active this quarter` | `inactive this quarter` | `indefinitely inactive` - `Reason` should provide an explanation for the current status (if needed) - Idea: Consider using emojis or emoji+text for each `status`. For example: 🛫|🛬|✈️ or maybe weather themed ☀️|⛈|🌤|☁️
process
create a project statuses board we need a table that lists the primary projects we are or have worked on and what there status is we sometimes get asked questions about why some project is is inactive or what its plans are make it a new notion table in the root labs page put it below the roadmap entries should include nexus nexus prisma graphql playground pumpkins columns should include name string status status repo link reason string status can be single select of active this quarter inactive this quarter indefinitely inactive reason should provide an explanation for the current status if needed idea consider using emojis or emoji text for each status for example 🛫 🛬 ✈️ or maybe weather themed ☀️ ⛈ 🌤 ☁️
1
10,643
13,446,196,249
IssuesEvent
2020-09-08 12:37:02
MHRA/products
https://api.github.com/repos/MHRA/products
closed
PARs - Allow deleting already added products & substances
EPIC - PARs process
When the user enters multiple products and/or substances they should be allowed to delete them. Designs: - https://app.zeplin.io/project/5dd51ae21205c944f8c1d35b/screen/5ebbfcb92c78c074bef52070 - https://app.zeplin.io/project/5dd51ae21205c944f8c1d35b/screen/5ebe9acb07baf73069b9be30
1.0
PARs - Allow deleting already added products & substances - When the user enters multiple products and/or substances they should be allowed to delete them. Designs: - https://app.zeplin.io/project/5dd51ae21205c944f8c1d35b/screen/5ebbfcb92c78c074bef52070 - https://app.zeplin.io/project/5dd51ae21205c944f8c1d35b/screen/5ebe9acb07baf73069b9be30
process
pars allow deleting already added products substances when the user enters multiple products and or substances they should be allowed to delete them designs
1
6,107
8,966,962,329
IssuesEvent
2019-01-29 01:09:38
knative/serving
https://api.github.com/repos/knative/serving
closed
Simplify the traffic.go
area/API kind/cleanup kind/process
<!-- Pro-tip: You can leave this block commented, and it still works! Select the appropriate areas for your issue: /area API Classify what kind of issue this is: /kind cleanup /kind process --> ## Proposal Currently the route/traffic/traffic.go is quite complex, but it doesn't have to be so, since we only permit a single revision for a target, so the whole flattening process can be rewritten with fewer lines. - simplify the data structures - simplify the algorithms - remove the unnecessary code - fix the linter errors with stutter named structures
1.0
Simplify the traffic.go - <!-- Pro-tip: You can leave this block commented, and it still works! Select the appropriate areas for your issue: /area API Classify what kind of issue this is: /kind cleanup /kind process --> ## Proposal Currently the route/traffic/traffic.go is quite complex, but it doesn't have to be so, since we only permit a single revision for a target, so the whole flattening process can be rewritten with fewer lines. - simplify the data structures - simplify the algorithms - remove the unnecessary code - fix the linter errors with stutter named structures
process
simplify the traffic go pro tip you can leave this block commented and it still works select the appropriate areas for your issue area api classify what kind of issue this is kind cleanup kind process proposal currently the route traffic traffic go is quite complex but it doesn t have to be so since we only permit a single revision for a target so the whole flattening process can be rewritten with fewer lines simplify the data structures simplify the algorithms remove the unnecessary code fix the linter errors with stutter named structures
1
7,569
10,684,225,854
IssuesEvent
2019-10-22 09:59:58
CGAL/cgal
https://api.github.com/repos/CGAL/cgal
closed
Bad results after MST normal orientation and Poisson reconstruction
Pkg::Point_set_processing_3 question
[Keypoints.txt](https://github.com/CGAL/cgal/files/1888743/Keypoints.txt) _Please use the following template to help us solving your issue._ ## Issue Details I want to estimate normals of point cloud using pca_estimate_normals, but i found it generates very bad result compare with pcl using same parameters(k neighbor ) test file is attached in the attachment:please change the file extension to ply ## Source Code CGAL::pca_estimate_normals<CGAL::Sequential_tag>(points.begin(), points.end(), CGAL::First_of_pair_property_map<PointVectorPair>(), CGAL::Second_of_pair_property_map<PointVectorPair>(), nb_neighbors_pca_normals); ## Environment * Operating system (Windows/Mac/Linux, 32/64 bits):windows 10 * Compiler: visual studio 2017 * Release or debug mode: release * Specific flags used (if any): * CGAL version:4.11.1 * Boost version:1.66 * Other libraries versions if used (Eigen, TBB, etc.):eigen tbb qt
1.0
Bad results after MST normal orientation and Poisson reconstruction - [Keypoints.txt](https://github.com/CGAL/cgal/files/1888743/Keypoints.txt) _Please use the following template to help us solving your issue._ ## Issue Details I want to estimate normals of point cloud using pca_estimate_normals, but i found it generates very bad result compare with pcl using same parameters(k neighbor ) test file is attached in the attachment:please change the file extension to ply ## Source Code CGAL::pca_estimate_normals<CGAL::Sequential_tag>(points.begin(), points.end(), CGAL::First_of_pair_property_map<PointVectorPair>(), CGAL::Second_of_pair_property_map<PointVectorPair>(), nb_neighbors_pca_normals); ## Environment * Operating system (Windows/Mac/Linux, 32/64 bits):windows 10 * Compiler: visual studio 2017 * Release or debug mode: release * Specific flags used (if any): * CGAL version:4.11.1 * Boost version:1.66 * Other libraries versions if used (Eigen, TBB, etc.):eigen tbb qt
process
bad results after mst normal orientation and poisson reconstruction please use the following template to help us solving your issue issue details i want to estimate normals of point cloud using pca estimate normals but i found it generates very bad result compare with pcl using same parameters k neighbor test file is attached in the attachment please change the file extension to ply source code cgal pca estimate normals points begin points end cgal first of pair property map cgal second of pair property map nb neighbors pca normals environment operating system windows mac linux bits windows compiler visual studio release or debug mode release specific flags used if any cgal version boost version other libraries versions if used eigen tbb etc eigen tbb qt
1
9,927
12,964,607,225
IssuesEvent
2020-07-20 20:47:14
crim-ca/weaver
https://api.github.com/repos/crim-ca/weaver
opened
[BUG] wrong workflow_input_value usage
CWL OGC job process/workflow process/wps3 triage/bug
**Describe the bug** `workflow_input_value` can have multiple types (eg: href, float, int) but `wps3_process.execute()` expects it to be a href. When it isn't and href, `wps3_process.execute()` fails. Either when trying `workflow_input_value["location"]` when `workflow_input_value` is a float (which isn't subscriptable), or later on, during the `.startswith()` validation, which fails since type mismatch. **To Reproduce** 1. Run a workflow, for eg: `pytest tests/functional/test_ems_end2end.py -k "test_workflow_subset_picker"` 2. Notice the TypeError message in the logs (see Screenshots section) **Expected behavior** Having data other than `href` parsed as `data`, to avoid type mismatch. **Screenshots** ``` weaver_1 | File "/opt/local/src/weaver/weaver/processes/wps_workflow.py", line 455, in execute weaver_1 | self.results = self.wps_process.execute(self.builder.job, self.outdir, self.expected_outputs) weaver_1 | File "/opt/local/src/weaver/weaver/processes/wps3_process.py", line 256, in execute weaver_1 | execute_req_input_val: workflow_input_value["location"]}) weaver_1 | TypeError: 'float' object is not subscriptable ``` **Context (please complete the following information):** - OS: linux - Instance: local EMS - Version 1.13.1 Note that although `test_workflow_subset_picker` is not currently working, in the past this test passed successfully on a local EMS.
2.0
[BUG] wrong workflow_input_value usage - **Describe the bug** `workflow_input_value` can have multiple types (eg: href, float, int) but `wps3_process.execute()` expects it to be a href. When it isn't and href, `wps3_process.execute()` fails. Either when trying `workflow_input_value["location"]` when `workflow_input_value` is a float (which isn't subscriptable), or later on, during the `.startswith()` validation, which fails since type mismatch. **To Reproduce** 1. Run a workflow, for eg: `pytest tests/functional/test_ems_end2end.py -k "test_workflow_subset_picker"` 2. Notice the TypeError message in the logs (see Screenshots section) **Expected behavior** Having data other than `href` parsed as `data`, to avoid type mismatch. **Screenshots** ``` weaver_1 | File "/opt/local/src/weaver/weaver/processes/wps_workflow.py", line 455, in execute weaver_1 | self.results = self.wps_process.execute(self.builder.job, self.outdir, self.expected_outputs) weaver_1 | File "/opt/local/src/weaver/weaver/processes/wps3_process.py", line 256, in execute weaver_1 | execute_req_input_val: workflow_input_value["location"]}) weaver_1 | TypeError: 'float' object is not subscriptable ``` **Context (please complete the following information):** - OS: linux - Instance: local EMS - Version 1.13.1 Note that although `test_workflow_subset_picker` is not currently working, in the past this test passed successfully on a local EMS.
process
wrong workflow input value usage describe the bug workflow input value can have multiple types eg href float int but process execute expects it to be a href when it isn t and href process execute fails either when trying workflow input value when workflow input value is a float which isn t subscriptable or later on during the startswith validation which fails since type mismatch to reproduce run a workflow for eg pytest tests functional test ems py k test workflow subset picker notice the typeerror message in the logs see screenshots section expected behavior having data other than href parsed as data to avoid type mismatch screenshots weaver file opt local src weaver weaver processes wps workflow py line in execute weaver self results self wps process execute self builder job self outdir self expected outputs weaver file opt local src weaver weaver processes process py line in execute weaver execute req input val workflow input value weaver typeerror float object is not subscriptable context please complete the following information os linux instance local ems version note that although test workflow subset picker is not currently working in the past this test passed successfully on a local ems
1
385,355
26,633,104,122
IssuesEvent
2023-01-24 19:27:48
0x192/universal-android-debloater
https://api.github.com/repos/0x192/universal-android-debloater
opened
Samsung App Separation official docs
package::documentation
**Your phone**: \<phone model\> **Packages documentation to update:** ``` com.samsung.android.appseparation ... ``` ## Documentation change ### com.samsung.android.appseparation **List**: Pending :arrow_right: OEM **Removal**: Expert (no diff) **Current description** > (none) **Proposed description** > https://docs.samsungknox.com/admin/knox-platform-for-enterprise/separated-apps.htm I know, it's not very helpful, but it's better than nothing
1.0
Samsung App Separation official docs - **Your phone**: \<phone model\> **Packages documentation to update:** ``` com.samsung.android.appseparation ... ``` ## Documentation change ### com.samsung.android.appseparation **List**: Pending :arrow_right: OEM **Removal**: Expert (no diff) **Current description** > (none) **Proposed description** > https://docs.samsungknox.com/admin/knox-platform-for-enterprise/separated-apps.htm I know, it's not very helpful, but it's better than nothing
non_process
samsung app separation official docs your phone packages documentation to update com samsung android appseparation documentation change com samsung android appseparation list pending arrow right oem removal expert no diff current description none proposed description i know it s not very helpful but it s better than nothing
0
26,327
5,243,702,331
IssuesEvent
2017-01-31 21:25:53
symfony/symfony-docs
https://api.github.com/repos/symfony/symfony-docs
closed
Document the deprecation of case-insensitive service ids
DependencyInjection hasPR Missing Documentation
Deprecated in 3.3 and remove in 4.0. See https://github.com/symfony/symfony/pull/21223 and http://symfony.com/blog/new-in-symfony-3-3-dependency-injection-deprecations
1.0
Document the deprecation of case-insensitive service ids - Deprecated in 3.3 and remove in 4.0. See https://github.com/symfony/symfony/pull/21223 and http://symfony.com/blog/new-in-symfony-3-3-dependency-injection-deprecations
non_process
document the deprecation of case insensitive service ids deprecated in and remove in see and
0
15,326
19,433,139,326
IssuesEvent
2021-12-21 14:16:00
threefoldtech/tfchain
https://api.github.com/repos/threefoldtech/tfchain
closed
change uncle filter in pallet authorship config for NPOS chain
process_wontfix
FitlerUncle is currently implemented as an empty tuple `()`, which is a dummy handler. As per the pallet_authorship code docs, we should probably use the also provided `OnePerAuthorPerHeight`. Link: https://github.com/paritytech/substrate/blob/49a4103f4bfef55be20a5c6d26e18ff3003c3353/frame/authorship/src/lib.rs#L55-L57
1.0
change uncle filter in pallet authorship config for NPOS chain - FitlerUncle is currently implemented as an empty tuple `()`, which is a dummy handler. As per the pallet_authorship code docs, we should probably use the also provided `OnePerAuthorPerHeight`. Link: https://github.com/paritytech/substrate/blob/49a4103f4bfef55be20a5c6d26e18ff3003c3353/frame/authorship/src/lib.rs#L55-L57
process
change uncle filter in pallet authorship config for npos chain fitleruncle is currently implemented as an empty tuple which is a dummy handler as per the pallet authorship code docs we should probably use the also provided oneperauthorperheight link
1
29,234
23,837,141,723
IssuesEvent
2022-09-06 07:11:01
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
opened
RFC: noUnusedParameters yes/no?
needs: discussion tool: monorepo infrastructure
## Background In the root directory's `tsconfig.base.json` we have the [noUnusedParameters](https://www.typescriptlang.org/tsconfig#noUnusedParameters) turned on - however as of present moment this rule is not applied to woocommerce-admin or the packages since they have their own `tsconfig.json` which do not extend from the root `tsconfig.base.json` When this rule is applied, there were some violations of this rule, which look like: ![image](https://user-images.githubusercontent.com/27843274/188568837-975e73bc-922c-4163-aac2-6c8865b0faa3.png) In most instances, it can be inferred that the unused parameter serves as documentation for potential future usage. Is there a better solution aside from 1. Disabling the rule entirely 2. Overriding each instance with a @ts-ignore
1.0
RFC: noUnusedParameters yes/no? - ## Background In the root directory's `tsconfig.base.json` we have the [noUnusedParameters](https://www.typescriptlang.org/tsconfig#noUnusedParameters) turned on - however as of present moment this rule is not applied to woocommerce-admin or the packages since they have their own `tsconfig.json` which do not extend from the root `tsconfig.base.json` When this rule is applied, there were some violations of this rule, which look like: ![image](https://user-images.githubusercontent.com/27843274/188568837-975e73bc-922c-4163-aac2-6c8865b0faa3.png) In most instances, it can be inferred that the unused parameter serves as documentation for potential future usage. Is there a better solution aside from 1. Disabling the rule entirely 2. Overriding each instance with a @ts-ignore
non_process
rfc nounusedparameters yes no background in the root directory s tsconfig base json we have the turned on however as of present moment this rule is not applied to woocommerce admin or the packages since they have their own tsconfig json which do not extend from the root tsconfig base json when this rule is applied there were some violations of this rule which look like in most instances it can be inferred that the unused parameter serves as documentation for potential future usage is there a better solution aside from disabling the rule entirely overriding each instance with a ts ignore
0
12,778
15,163,015,843
IssuesEvent
2021-02-12 11:30:55
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
closed
Create interface(s) for Flue Gas property calculation methods
Calculator Process Heating
Calculations methods for O2 and Excess Air take an untyped/any input object. Needs a joint interface or separate for each (Mass/Volume) input. Methods also exist in Air-heating calc (see for example). ![image.png](https://images.zenhubusercontent.com/5e4547eef6a311c23c81ce81/a5a011dd-247c-45d1-a774-16d6d0c6fba4)
1.0
Create interface(s) for Flue Gas property calculation methods - Calculations methods for O2 and Excess Air take an untyped/any input object. Needs a joint interface or separate for each (Mass/Volume) input. Methods also exist in Air-heating calc (see for example). ![image.png](https://images.zenhubusercontent.com/5e4547eef6a311c23c81ce81/a5a011dd-247c-45d1-a774-16d6d0c6fba4)
process
create interface s for flue gas property calculation methods calculations methods for and excess air take an untyped any input object needs a joint interface or separate for each mass volume input methods also exist in air heating calc see for example
1
757,780
26,528,071,769
IssuesEvent
2023-01-19 10:17:56
consta-design-system/uikit
https://api.github.com/repos/consta-design-system/uikit
opened
Table: Вертикальные связи между уровнями
feature 🔥🔥 priority
У нас возникла следующая потребность: в иерархических таблицах нам очень нужны вертикальные связи между уровнями таблиц. Желаемый вид отобразили на макете, скриншот во вложении. ![image](https://user-images.githubusercontent.com/75978078/213416356-b94aed99-0db7-4e2f-b2e9-17e85b7bbdc1.png)
1.0
Table: Вертикальные связи между уровнями - У нас возникла следующая потребность: в иерархических таблицах нам очень нужны вертикальные связи между уровнями таблиц. Желаемый вид отобразили на макете, скриншот во вложении. ![image](https://user-images.githubusercontent.com/75978078/213416356-b94aed99-0db7-4e2f-b2e9-17e85b7bbdc1.png)
non_process
table вертикальные связи между уровнями у нас возникла следующая потребность в иерархических таблицах нам очень нужны вертикальные связи между уровнями таблиц желаемый вид отобразили на макете скриншот во вложении
0
21,297
28,495,290,970
IssuesEvent
2023-04-18 13:50:40
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
System hangs when a php interpreter is executed via Process
Bug Process Status: Needs Review
### Symfony version(s) affected 6.2.6 ### Description Server setup: PHP: 8.2.2 Server API: CGI/FastCGI When I run a PHP script via Process, Symfony creates lots of `/usr/lib/cgi-bin/php8.2` processes, which consume all RAM, server hangs and the OOM killer terminates all Apache instances. ### How to reproduce 1. Create a test project: ``` composer create-project symfony/skeleton:"6.2.*" testing cd testing composer require symfony/apache-pack composer require symfony/process ``` 2. `config/routes.yaml` : ``` controllers: resource: path: ../src/Controller/ namespace: App\Controller type: attribute test: path: /test controller: App\Controller\Test::test ``` 3. `src/Controller/Test.php`: ``` <?php namespace App\Controller; use Symfony\Component\HttpFoundation\Response; use Symfony\Component\Process\PhpExecutableFinder; use Symfony\Component\Process\Exception\ProcessFailedException; use Symfony\Component\Process\Process; class Test { public function test(): Response { $phpBinaryPath = (new PhpExecutableFinder())->find(); $process = new Process([$phpBinaryPath, './test.php']); try { $process->run(); if (!$process->isSuccessful()) { throw new ProcessFailedException($process); } } catch (\Exception $e) { throw new ProcessFailedException($process); } return new Response( 'result: ' . $process->getOutput() ); } } ``` 4. `public/test.php`: ``` <?php echo 'OK'; ``` Now open the path `/test` in your web browser. Symfony starts lots of `/usr/lib/cgi-bin/php8.2` processes, which consume all memory and CPU. I supposed that the issue might be related to HTTP headers, generated by the CGI version. However, this does not help - `new Process([$phpBinaryPath, '-q', './test.php'])`. If I run either command directly on the server, all is good - `/usr/lib/cgi-bin/php8.2 ./test.php` / `/usr/lib/cgi-bin/php8.2 -q ./test.php`. If I hardcode the path to CLI PHP (`$phpBinaryPath = '/bin/php';`) that test script works correctly. ### Possible Solution `PhpExecutableFinder` should probably return the CLI version instead of CGI. ### Additional Context _No response_
1.0
System hangs when a php interpreter is executed via Process - ### Symfony version(s) affected 6.2.6 ### Description Server setup: PHP: 8.2.2 Server API: CGI/FastCGI When I run a PHP script via Process, Symfony creates lots of `/usr/lib/cgi-bin/php8.2` processes, which consume all RAM, server hangs and the OOM killer terminates all Apache instances. ### How to reproduce 1. Create a test project: ``` composer create-project symfony/skeleton:"6.2.*" testing cd testing composer require symfony/apache-pack composer require symfony/process ``` 2. `config/routes.yaml` : ``` controllers: resource: path: ../src/Controller/ namespace: App\Controller type: attribute test: path: /test controller: App\Controller\Test::test ``` 3. `src/Controller/Test.php`: ``` <?php namespace App\Controller; use Symfony\Component\HttpFoundation\Response; use Symfony\Component\Process\PhpExecutableFinder; use Symfony\Component\Process\Exception\ProcessFailedException; use Symfony\Component\Process\Process; class Test { public function test(): Response { $phpBinaryPath = (new PhpExecutableFinder())->find(); $process = new Process([$phpBinaryPath, './test.php']); try { $process->run(); if (!$process->isSuccessful()) { throw new ProcessFailedException($process); } } catch (\Exception $e) { throw new ProcessFailedException($process); } return new Response( 'result: ' . $process->getOutput() ); } } ``` 4. `public/test.php`: ``` <?php echo 'OK'; ``` Now open the path `/test` in your web browser. Symfony starts lots of `/usr/lib/cgi-bin/php8.2` processes, which consume all memory and CPU. I supposed that the issue might be related to HTTP headers, generated by the CGI version. However, this does not help - `new Process([$phpBinaryPath, '-q', './test.php'])`. If I run either command directly on the server, all is good - `/usr/lib/cgi-bin/php8.2 ./test.php` / `/usr/lib/cgi-bin/php8.2 -q ./test.php`. If I hardcode the path to CLI PHP (`$phpBinaryPath = '/bin/php';`) that test script works correctly. ### Possible Solution `PhpExecutableFinder` should probably return the CLI version instead of CGI. ### Additional Context _No response_
process
system hangs when a php interpreter is executed via process symfony version s affected description server setup php server api cgi fastcgi when i run a php script via process symfony creates lots of usr lib cgi bin processes which consume all ram server hangs and the oom killer terminates all apache instances how to reproduce create a test project composer create project symfony skeleton testing cd testing composer require symfony apache pack composer require symfony process config routes yaml controllers resource path src controller namespace app controller type attribute test path test controller app controller test test src controller test php php namespace app controller use symfony component httpfoundation response use symfony component process phpexecutablefinder use symfony component process exception processfailedexception use symfony component process process class test public function test response phpbinarypath new phpexecutablefinder find process new process try process run if process issuccessful throw new processfailedexception process catch exception e throw new processfailedexception process return new response result process getoutput public test php php echo ok now open the path test in your web browser symfony starts lots of usr lib cgi bin processes which consume all memory and cpu i supposed that the issue might be related to http headers generated by the cgi version however this does not help new process if i run either command directly on the server all is good usr lib cgi bin test php usr lib cgi bin q test php if i hardcode the path to cli php phpbinarypath bin php that test script works correctly possible solution phpexecutablefinder should probably return the cli version instead of cgi additional context no response
1
17,184
22,766,639,236
IssuesEvent
2022-07-08 05:31:24
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
opened
Obsoletion notice: GO:0039611 suppression by virus of host translation initiation factor activity
obsoletion multi-species process
Dear all, The proposal has been made to obsolete GO:0039611 suppression by virus of host translation initiation factor activity. The reason for obsoletion is that this a a molecular function represented as a biological process. The term will be replaced by 'GO:0039606 suppression by virus of host translation initiation' . There were 5 experimental annotations (updated). There is one keyword mapping that can also be moved to 'GO:0039606 suppression by virus of host translation initiation', KW-1075 Inhibition of host translation factors by virus. This term is not in any subsets. You can comment on the ticket: Thanks, Pascale
1.0
Obsoletion notice: GO:0039611 suppression by virus of host translation initiation factor activity - Dear all, The proposal has been made to obsolete GO:0039611 suppression by virus of host translation initiation factor activity. The reason for obsoletion is that this a a molecular function represented as a biological process. The term will be replaced by 'GO:0039606 suppression by virus of host translation initiation' . There were 5 experimental annotations (updated). There is one keyword mapping that can also be moved to 'GO:0039606 suppression by virus of host translation initiation', KW-1075 Inhibition of host translation factors by virus. This term is not in any subsets. You can comment on the ticket: Thanks, Pascale
process
obsoletion notice go suppression by virus of host translation initiation factor activity dear all the proposal has been made to obsolete go suppression by virus of host translation initiation factor activity the reason for obsoletion is that this a a molecular function represented as a biological process the term will be replaced by go suppression by virus of host translation initiation there were experimental annotations updated there is one keyword mapping that can also be moved to go suppression by virus of host translation initiation kw inhibition of host translation factors by virus this term is not in any subsets you can comment on the ticket thanks pascale
1