Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
15,337
| 19,472,409,961
|
IssuesEvent
|
2021-12-24 05:01:21
|
home-climate-control/dz
|
https://api.github.com/repos/home-climate-control/dz
|
opened
|
Ignore 1-Wire sensor departure, just react to timeout
|
fault tolerance 1-wire reactive process control
|
### Existing Behavior
1-Wire sensor departure immediately propagates to the control system and makes it act prematurely, as jitter and short blackouts are possible.
### Desired Behavior
If/when a 1-Wire sensor departs, the departure is not propagated down the control flow immediately, but only after a given timeout when the sensor is considered stale.
### Corollary
Hardware failure notifications must be handled outside of control flow.
|
1.0
|
Ignore 1-Wire sensor departure, just react to timeout - ### Existing Behavior
1-Wire sensor departure immediately propagates to the control system and makes it act prematurely, as jitter and short blackouts are possible.
### Desired Behavior
If/when a 1-Wire sensor departs, the departure is not propagated down the control flow immediately, but only after a given timeout when the sensor is considered stale.
### Corollary
Hardware failure notifications must be handled outside of control flow.
|
process
|
ignore wire sensor departure just react to timeout existing behavior wire sensor departure immediately propagates to the control system and makes it act prematurely as jitter and short blackouts are possible desired behavior if when a wire sensor departs the departure is not propagated down the control flow immediately but only after a given timeout when the sensor is considered stale corollary hardware failure notifications must be handled outside of control flow
| 1
|
18,918
| 24,865,200,149
|
IssuesEvent
|
2022-10-27 11:22:38
|
alphagov/govuk-design-system
|
https://api.github.com/repos/alphagov/govuk-design-system
|
closed
|
Secure community facilitators for Design System Day 2022 workshops and breakout sessions
|
🕔 weeks process shared ownership
|
## What
Approach the community to facilitate workshops and breakout sessions on Design System Day 2022.
### Workshops
90-minute sessions, usually involves a presentation and audience activities
5-7 facilitators needed to run their workshop on both days
### Breakout sessions
50-minute sessions, freeform — can be a discussion, activity, anything they'd like
14-20 facilitators needed over both days (breakout sessions are not repeated)
## Why
So we can began planning the workshops and let attendees know what to expect.
## Who needs to work on this
TBC
## Who needs to review this
## Dependencies
- [x] A theme for the event has been agreed
- [x] Decision needed: How can people propose ideas and volunteer to facilitate?
- [x] Clarification needed: what's facilitators need to do ([see proposed schedule](https://govukdesignsystem.padlet.org/ImHuYorks/vntxtz9vv3do3uiz))
## Done when
Call for facilitators (week of 22 August?)
- [x] Drafted comms
- [x] Comms reviewed by at least one team member
- [x] Published on appropriate platforms
Selection process
- [x] Gather and organise potential facilitators
- [x] Choose facilitators
For the event (done by 7 September)
- [x] Confirm with facilitators
- [x] Schedule facilitators, working with participant comms group
|
1.0
|
Secure community facilitators for Design System Day 2022 workshops and breakout sessions - ## What
Approach the community to facilitate workshops and breakout sessions on Design System Day 2022.
### Workshops
90-minute sessions, usually involves a presentation and audience activities
5-7 facilitators needed to run their workshop on both days
### Breakout sessions
50-minute sessions, freeform — can be a discussion, activity, anything they'd like
14-20 facilitators needed over both days (breakout sessions are not repeated)
## Why
So we can began planning the workshops and let attendees know what to expect.
## Who needs to work on this
TBC
## Who needs to review this
## Dependencies
- [x] A theme for the event has been agreed
- [x] Decision needed: How can people propose ideas and volunteer to facilitate?
- [x] Clarification needed: what's facilitators need to do ([see proposed schedule](https://govukdesignsystem.padlet.org/ImHuYorks/vntxtz9vv3do3uiz))
## Done when
Call for facilitators (week of 22 August?)
- [x] Drafted comms
- [x] Comms reviewed by at least one team member
- [x] Published on appropriate platforms
Selection process
- [x] Gather and organise potential facilitators
- [x] Choose facilitators
For the event (done by 7 September)
- [x] Confirm with facilitators
- [x] Schedule facilitators, working with participant comms group
|
process
|
secure community facilitators for design system day workshops and breakout sessions what approach the community to facilitate workshops and breakout sessions on design system day workshops minute sessions usually involves a presentation and audience activities facilitators needed to run their workshop on both days breakout sessions minute sessions freeform — can be a discussion activity anything they d like facilitators needed over both days breakout sessions are not repeated why so we can began planning the workshops and let attendees know what to expect who needs to work on this tbc who needs to review this dependencies a theme for the event has been agreed decision needed how can people propose ideas and volunteer to facilitate clarification needed what s facilitators need to do done when call for facilitators week of august drafted comms comms reviewed by at least one team member published on appropriate platforms selection process gather and organise potential facilitators choose facilitators for the event done by september confirm with facilitators schedule facilitators working with participant comms group
| 1
|
8,048
| 11,220,764,528
|
IssuesEvent
|
2020-01-07 16:25:43
|
code4romania/expert-consultation-api
|
https://api.github.com/repos/code4romania/expert-consultation-api
|
opened
|
Unable to load document path that contains spaces
|
bug document processing documents good first issue help wanted
|
When uploading a new document in the platform, of the document URL contains spaces, the document upload fails with the following message:
{
"i18nErrors": null,
"i18nFieldErrors": null,
**"additionalInfo": "Index 0 out of bounds for length 0"**
}
The initial DTO for creating the new document is the following:
{
"dateOfDevelopment": "2020-01-01T16:02:03.790Z",
"dateOfReceipt": "2020-01-01T16:02:03.790Z",
"documentInitializer": "Ministerul Dezvoltarii",
"documentNumber": 876345354,
"documentTitle": "Document de test 2",
"documentType": "LEGE",
"filePath": **"/Users/radustefanescu/legal_consult_uploads/MmujDzZFpwregulament_grupuri_de_lucru_pdf_1520247820 (1).pdf"**
}
|
1.0
|
Unable to load document path that contains spaces - When uploading a new document in the platform, of the document URL contains spaces, the document upload fails with the following message:
{
"i18nErrors": null,
"i18nFieldErrors": null,
**"additionalInfo": "Index 0 out of bounds for length 0"**
}
The initial DTO for creating the new document is the following:
{
"dateOfDevelopment": "2020-01-01T16:02:03.790Z",
"dateOfReceipt": "2020-01-01T16:02:03.790Z",
"documentInitializer": "Ministerul Dezvoltarii",
"documentNumber": 876345354,
"documentTitle": "Document de test 2",
"documentType": "LEGE",
"filePath": **"/Users/radustefanescu/legal_consult_uploads/MmujDzZFpwregulament_grupuri_de_lucru_pdf_1520247820 (1).pdf"**
}
|
process
|
unable to load document path that contains spaces when uploading a new document in the platform of the document url contains spaces the document upload fails with the following message null null additionalinfo index out of bounds for length the initial dto for creating the new document is the following dateofdevelopment dateofreceipt documentinitializer ministerul dezvoltarii documentnumber documenttitle document de test documenttype lege filepath users radustefanescu legal consult uploads mmujdzzfpwregulament grupuri de lucru pdf pdf
| 1
|
15,731
| 19,905,560,906
|
IssuesEvent
|
2022-01-25 12:26:03
|
prometheus-community/windows_exporter
|
https://api.github.com/repos/prometheus-community/windows_exporter
|
closed
|
How to write a process restart alarm rule?
|
question collector/process
|
A process restart can be indicated by windows_process_start_time. If the value of this indicator changes, it indicates that the process is restarted. For example: windows_process_start_time{process="ACDG",instance="10.35.236.40"}, what if this cannot be greater than or less than?
|
1.0
|
How to write a process restart alarm rule? - A process restart can be indicated by windows_process_start_time. If the value of this indicator changes, it indicates that the process is restarted. For example: windows_process_start_time{process="ACDG",instance="10.35.236.40"}, what if this cannot be greater than or less than?
|
process
|
how to write a process restart alarm rule a process restart can be indicated by windows process start time if the value of this indicator changes it indicates that the process is restarted for example windows process start time process acdg instance what if this cannot be greater than or less than
| 1
|
75,790
| 9,886,000,605
|
IssuesEvent
|
2019-06-25 05:07:27
|
getgauge/taiko
|
https://api.github.com/repos/getgauge/taiko
|
closed
|
Need to fix taiko doc
|
documentation
|
Observation:
1. Under overview there are few texts which appears as link we need to rectify. Example: sample project
<img width="894" alt="Screen Shot 2019-06-14 at 3 46 31 PM" src="https://user-images.githubusercontent.com/46309201/59582980-88536780-90f7-11e9-8c23-b4ad51b2519b.png">
2. cosmetic issue:
<img width="841" alt="Screen Shot 2019-06-14 at 3 47 15 PM" src="https://user-images.githubusercontent.com/46309201/59583006-986b4700-90f7-11e9-9e27-1c89516cd28f.png">
3. ```TAIKO_EMULATE_NETWORK```
env vaiable is misssing under https://taiko-preview.gauge.org/?#taiko-env-variables
<img width="820" alt="Screen Shot 2019-06-14 at 3 52 06 PM" src="https://user-images.githubusercontent.com/46309201/59583015-a4ef9f80-90f7-11e9-9145-3eac57fab593.png">
|
1.0
|
Need to fix taiko doc -
Observation:
1. Under overview there are few texts which appears as link we need to rectify. Example: sample project
<img width="894" alt="Screen Shot 2019-06-14 at 3 46 31 PM" src="https://user-images.githubusercontent.com/46309201/59582980-88536780-90f7-11e9-8c23-b4ad51b2519b.png">
2. cosmetic issue:
<img width="841" alt="Screen Shot 2019-06-14 at 3 47 15 PM" src="https://user-images.githubusercontent.com/46309201/59583006-986b4700-90f7-11e9-9e27-1c89516cd28f.png">
3. ```TAIKO_EMULATE_NETWORK```
env vaiable is misssing under https://taiko-preview.gauge.org/?#taiko-env-variables
<img width="820" alt="Screen Shot 2019-06-14 at 3 52 06 PM" src="https://user-images.githubusercontent.com/46309201/59583015-a4ef9f80-90f7-11e9-9145-3eac57fab593.png">
|
non_process
|
need to fix taiko doc observation under overview there are few texts which appears as link we need to rectify example sample project img width alt screen shot at pm src cosmetic issue img width alt screen shot at pm src taiko emulate network env vaiable is misssing under img width alt screen shot at pm src
| 0
|
15,624
| 19,770,228,787
|
IssuesEvent
|
2022-01-17 09:17:19
|
skellig-framework/skellig-core
|
https://api.github.com/repos/skellig-framework/skellig-core
|
closed
|
Modify date and time from current date
|
processing
|
Add ability to modify current date and time, by adding a function which can go after `now()`, ex:
`now().plusDays(1)`
|
1.0
|
Modify date and time from current date - Add ability to modify current date and time, by adding a function which can go after `now()`, ex:
`now().plusDays(1)`
|
process
|
modify date and time from current date add ability to modify current date and time by adding a function which can go after now ex now plusdays
| 1
|
45,635
| 24,147,423,805
|
IssuesEvent
|
2022-09-21 20:08:35
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Expression.Compile performance regression on .NET7 RC1
|
tenet-performance area-VM-coreclr
|
<!--This is just a template - feel free to delete any and all of it and replace as appropriate.-->
### Description
Running simple benchmark to test performance of Expression.Compile shows regressions on .NET7 RC1
<details>
<summary>Benchmark code</summary>
```c#
using System;
using System.Linq.Expressions;
using BenchmarkDotNet.Attributes;
using BenchmarkDotNet.Configs;
using BenchmarkDotNet.Jobs;
using BenchmarkDotNet.Running;
BenchmarkRunner.Run<Benchmarks>();
[MemoryDiagnoser(false)]
[SimpleJob(RuntimeMoniker.Net60)]
[SimpleJob(RuntimeMoniker.Net70)]
[GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByJob)]
public class Benchmarks
{
private static readonly Expression<Func<MyClass>> _factory = FactoryCreator<MyClass>.GetFactory();
[Benchmark]
public Func<MyClass> CompileFactory()
{
return _factory.Compile();
}
}
public static class FactoryCreator<T>
{
public static Expression<Func<T>> GetFactory()
{
var ctorBody = Expression.New(typeof(T).GetConstructor(Type.EmptyTypes)!);
return Expression.Lambda<Func<T>>(ctorBody);
}
}
public class MyClass {}
```
</details>
### Regression?
Yes. On .NET 6 this is ~50% faster
### Data
``` ini
BenchmarkDotNet=v0.13.2, OS=Windows 11 (10.0.22000.978/21H2)
AMD Ryzen 5 3600, 1 CPU, 12 logical and 6 physical cores
.NET SDK=7.0.100-rc.1.22431.12
[Host] : .NET 7.0.0 (7.0.22.42610), X64 RyuJIT AVX2
.NET 6.0 : .NET 6.0.9 (6.0.922.41905), X64 RyuJIT AVX2
.NET 7.0 : .NET 7.0.0 (7.0.22.42610), X64 RyuJIT AVX2
```
| Method | Job | Runtime | Mean | Error | StdDev | Allocated |
|--------------- |--------- |--------- |---------:|---------:|---------:|----------:|
| CompileFactory | .NET 6.0 | .NET 6.0 | 33.51 μs | 0.361 μs | 0.337 μs | 3.66 KB |
| | | | | | | |
| CompileFactory | .NET 7.0 | .NET 7.0 | 51.86 μs | 0.223 μs | 0.208 μs | 3.69 KB |
<!--
* Please include any benchmark results, images of graphs, timings or measurements, or callstacks that are relevant.
* If possible please include text as text rather than images (so it shows up in searches).
* If applicable please include before and after measurements.
* There is helpful information about measuring code in this repo [here](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md).
-->
|
True
|
Expression.Compile performance regression on .NET7 RC1 - <!--This is just a template - feel free to delete any and all of it and replace as appropriate.-->
### Description
Running simple benchmark to test performance of Expression.Compile shows regressions on .NET7 RC1
<details>
<summary>Benchmark code</summary>
```c#
using System;
using System.Linq.Expressions;
using BenchmarkDotNet.Attributes;
using BenchmarkDotNet.Configs;
using BenchmarkDotNet.Jobs;
using BenchmarkDotNet.Running;
BenchmarkRunner.Run<Benchmarks>();
[MemoryDiagnoser(false)]
[SimpleJob(RuntimeMoniker.Net60)]
[SimpleJob(RuntimeMoniker.Net70)]
[GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByJob)]
public class Benchmarks
{
private static readonly Expression<Func<MyClass>> _factory = FactoryCreator<MyClass>.GetFactory();
[Benchmark]
public Func<MyClass> CompileFactory()
{
return _factory.Compile();
}
}
public static class FactoryCreator<T>
{
public static Expression<Func<T>> GetFactory()
{
var ctorBody = Expression.New(typeof(T).GetConstructor(Type.EmptyTypes)!);
return Expression.Lambda<Func<T>>(ctorBody);
}
}
public class MyClass {}
```
</details>
### Regression?
Yes. On .NET 6 this is ~50% faster
### Data
``` ini
BenchmarkDotNet=v0.13.2, OS=Windows 11 (10.0.22000.978/21H2)
AMD Ryzen 5 3600, 1 CPU, 12 logical and 6 physical cores
.NET SDK=7.0.100-rc.1.22431.12
[Host] : .NET 7.0.0 (7.0.22.42610), X64 RyuJIT AVX2
.NET 6.0 : .NET 6.0.9 (6.0.922.41905), X64 RyuJIT AVX2
.NET 7.0 : .NET 7.0.0 (7.0.22.42610), X64 RyuJIT AVX2
```
| Method | Job | Runtime | Mean | Error | StdDev | Allocated |
|--------------- |--------- |--------- |---------:|---------:|---------:|----------:|
| CompileFactory | .NET 6.0 | .NET 6.0 | 33.51 μs | 0.361 μs | 0.337 μs | 3.66 KB |
| | | | | | | |
| CompileFactory | .NET 7.0 | .NET 7.0 | 51.86 μs | 0.223 μs | 0.208 μs | 3.69 KB |
<!--
* Please include any benchmark results, images of graphs, timings or measurements, or callstacks that are relevant.
* If possible please include text as text rather than images (so it shows up in searches).
* If applicable please include before and after measurements.
* There is helpful information about measuring code in this repo [here](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md).
-->
|
non_process
|
expression compile performance regression on description running simple benchmark to test performance of expression compile shows regressions on benchmark code c using system using system linq expressions using benchmarkdotnet attributes using benchmarkdotnet configs using benchmarkdotnet jobs using benchmarkdotnet running benchmarkrunner run public class benchmarks private static readonly expression factory factorycreator getfactory public func compilefactory return factory compile public static class factorycreator public static expression getfactory var ctorbody expression new typeof t getconstructor type emptytypes return expression lambda ctorbody public class myclass regression yes on net this is faster data ini benchmarkdotnet os windows amd ryzen cpu logical and physical cores net sdk rc net ryujit net net ryujit net net ryujit method job runtime mean error stddev allocated compilefactory net net μs μs μs kb compilefactory net net μs μs μs kb please include any benchmark results images of graphs timings or measurements or callstacks that are relevant if possible please include text as text rather than images so it shows up in searches if applicable please include before and after measurements there is helpful information about measuring code in this repo
| 0
|
195,180
| 6,904,978,482
|
IssuesEvent
|
2017-11-27 03:45:06
|
AIE-2017-Yr1-Group1/cultist-game
|
https://api.github.com/repos/AIE-2017-Yr1-Group1/cultist-game
|
closed
|
Cultists are firing at Demorial's feet
|
bug priority-1
|
Probably because the lookat fix for offset fireballs is looking at Demorial position, which is 0,0,0 at his feet.
|
1.0
|
Cultists are firing at Demorial's feet - Probably because the lookat fix for offset fireballs is looking at Demorial position, which is 0,0,0 at his feet.
|
non_process
|
cultists are firing at demorial s feet probably because the lookat fix for offset fireballs is looking at demorial position which is at his feet
| 0
|
160,052
| 20,092,552,107
|
IssuesEvent
|
2022-02-06 01:28:25
|
PGreaneyLYIT/easybuggy4django
|
https://api.github.com/repos/PGreaneyLYIT/easybuggy4django
|
opened
|
CVE-2019-8331 (Medium) detected in bootstrap-3.3.7.min.js
|
security vulnerability
|
## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /templates/base.html</p>
<p>Path to vulnerable library: /templates/base.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/PGreaneyLYIT/easybuggy4django/commit/5403b8fbcea4b699ce64d05146aeaa76d1062d89">5403b8fbcea4b699ce64d05146aeaa76d1062d89</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-8331 (Medium) detected in bootstrap-3.3.7.min.js - ## CVE-2019-8331 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /templates/base.html</p>
<p>Path to vulnerable library: /templates/base.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/PGreaneyLYIT/easybuggy4django/commit/5403b8fbcea4b699ce64d05146aeaa76d1062d89">5403b8fbcea4b699ce64d05146aeaa76d1062d89</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute.
<p>Publish Date: 2019-02-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p>
<p>Release Date: 2019-02-20</p>
<p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file templates base html path to vulnerable library templates base html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap bootstrap sass step up your open source security game with whitesource
| 0
|
4,361
| 7,260,514,255
|
IssuesEvent
|
2018-02-18 10:53:23
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[FEATURE] [processing] New algorithm to compute geometry by expression
|
Processing User Manual
|
This algorithm updates existing geometries (or creates new geometries) for input features by use of a QGIS expression. This allows complex geometry modifications which can utilise all the
flexibility of the QGIS expression engine to manipulate and create geometries for output features.
see https://github.com/qgis/QGIS/pull/3692/commits/12ab6c6f6f77aea5cf5e31e9af81723d35a263cb
|
1.0
|
[FEATURE] [processing] New algorithm to compute geometry by expression - This algorithm updates existing geometries (or creates new geometries) for input features by use of a QGIS expression. This allows complex geometry modifications which can utilise all the
flexibility of the QGIS expression engine to manipulate and create geometries for output features.
see https://github.com/qgis/QGIS/pull/3692/commits/12ab6c6f6f77aea5cf5e31e9af81723d35a263cb
|
process
|
new algorithm to compute geometry by expression this algorithm updates existing geometries or creates new geometries for input features by use of a qgis expression this allows complex geometry modifications which can utilise all the flexibility of the qgis expression engine to manipulate and create geometries for output features see
| 1
|
776,863
| 27,264,724,273
|
IssuesEvent
|
2023-02-22 17:09:43
|
ascheid/itsg33-pbmm-issue-gen
|
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
|
opened
|
SA-11(5): Developer Security Testing And Evaluation | Penetration Testing / Analysis
|
Priority: P3 ITSG-33 Suggested Assignment: IT Projects Class: Management Control: SA-11
|
# Control Definition
DEVELOPER SECURITY TESTING AND EVALUATION | PENETRATION TESTING / ANALYSIS
The organization requires the developer of the information system, system component, or information system service to perform penetration testing at [Assignment: organization-defined breadth/depth] and with [Assignment: organization-defined constraints].
# Class
Management
# Supplemental Guidance
Penetration testing is an assessment methodology in which assessors, using all available information technology product and/or information system documentation (e.g., product/system design specifications, source code, and administrator/operator manuals) and working under specific constraints, attempt to circumvent implemented security features of information technology products and information systems. Penetration testing can include, for example, white, gray, or black box testing with analyses performed by skilled security professionals simulating adversary actions. The objective of penetration testing is to uncover potential vulnerabilities in information technology products and information systems resulting from implementation errors, configuration faults, or other operational deployment weaknesses or deficiencies. Penetration tests can be performed in conjunction with automated and manual code reviews to provide greater levels of analysis than would ordinarily be possible.
# General Guide
Apply to boundary and other security critical components. For COTS products require 3rd party evaluation such as Common Criteria.
# Suggested Assignment
IT Projects
# Support Teams
IT Security Function
|
1.0
|
SA-11(5): Developer Security Testing And Evaluation | Penetration Testing / Analysis - # Control Definition
DEVELOPER SECURITY TESTING AND EVALUATION | PENETRATION TESTING / ANALYSIS
The organization requires the developer of the information system, system component, or information system service to perform penetration testing at [Assignment: organization-defined breadth/depth] and with [Assignment: organization-defined constraints].
# Class
Management
# Supplemental Guidance
Penetration testing is an assessment methodology in which assessors, using all available information technology product and/or information system documentation (e.g., product/system design specifications, source code, and administrator/operator manuals) and working under specific constraints, attempt to circumvent implemented security features of information technology products and information systems. Penetration testing can include, for example, white, gray, or black box testing with analyses performed by skilled security professionals simulating adversary actions. The objective of penetration testing is to uncover potential vulnerabilities in information technology products and information systems resulting from implementation errors, configuration faults, or other operational deployment weaknesses or deficiencies. Penetration tests can be performed in conjunction with automated and manual code reviews to provide greater levels of analysis than would ordinarily be possible.
# General Guide
Apply to boundary and other security critical components. For COTS products require 3rd party evaluation such as Common Criteria.
# Suggested Assignment
IT Projects
# Support Teams
IT Security Function
|
non_process
|
sa developer security testing and evaluation penetration testing analysis control definition developer security testing and evaluation penetration testing analysis the organization requires the developer of the information system system component or information system service to perform penetration testing at and with class management supplemental guidance penetration testing is an assessment methodology in which assessors using all available information technology product and or information system documentation e g product system design specifications source code and administrator operator manuals and working under specific constraints attempt to circumvent implemented security features of information technology products and information systems penetration testing can include for example white gray or black box testing with analyses performed by skilled security professionals simulating adversary actions the objective of penetration testing is to uncover potential vulnerabilities in information technology products and information systems resulting from implementation errors configuration faults or other operational deployment weaknesses or deficiencies penetration tests can be performed in conjunction with automated and manual code reviews to provide greater levels of analysis than would ordinarily be possible general guide apply to boundary and other security critical components for cots products require party evaluation such as common criteria suggested assignment it projects support teams it security function
| 0
|
5,848
| 8,672,991,394
|
IssuesEvent
|
2018-11-30 00:14:57
|
googleapis/google-cloud-java
|
https://api.github.com/repos/googleapis/google-cloud-java
|
closed
|
Setup and fix Java 11 tests
|
type: process
|
Java 11 is an LTS version scheduled for GA 2018-09-25.
Also, we need to fix the build on Java 11.
|
1.0
|
Setup and fix Java 11 tests - Java 11 is an LTS version scheduled for GA 2018-09-25.
Also, we need to fix the build on Java 11.
|
process
|
setup and fix java tests java is an lts version scheduled for ga also we need to fix the build on java
| 1
|
768,699
| 26,976,316,511
|
IssuesEvent
|
2023-02-09 09:49:40
|
MattTheLegoman/RealmsInExile
|
https://api.github.com/repos/MattTheLegoman/RealmsInExile
|
opened
|
GUI in culture screen to show descriptions
|
priority: low gui
|
Tweak GUI in culture screen to show cultural descriptions.
|
1.0
|
GUI in culture screen to show descriptions - Tweak GUI in culture screen to show cultural descriptions.
|
non_process
|
gui in culture screen to show descriptions tweak gui in culture screen to show cultural descriptions
| 0
|
674,165
| 23,041,692,335
|
IssuesEvent
|
2022-07-23 08:31:37
|
tsunamods-codes/7th-Heaven
|
https://api.github.com/repos/tsunamods-codes/7th-Heaven
|
closed
|
Incorrect folder priority because of naming.
|
help wanted priority/P2
|
I have been doing some tests. I use my Models Fusion mod which includes Field/Battle/World models, normally with al the files needed to create the model. I decided to add Ninostyle Chibi models (IRO) v0.245 also to 7H. So, my order is:
- Ninostyle Chibis (for Field)
- Models Fusion (basically are Chaos models, for Field)
- Other mods...
Well, the thing is that it does not matter the order I put my Models Fusion in 7H, I always see the Chaos models in game. After some checking, I detected that Ninostyle Chibis has the models in the folder **"char.lgp"**, and I have the models in the folder **"char"**.
So, as the priority is by folder naming (and they are literally different), "my" folder, "char", which has higher priority (alphabetically goes before than "char.lgp"), loads its models ALWAYS first, don't matter the order the mods are in the "My Mods" tab.
I have checked this extracting Ninostyle IRO and changing its folder to "char", then it worked correctly. But of course, a modder should not depend of the folder naming, not at least for the same type of files.
Would it be possible to treat "char/char.lgp" or "battle/battle.lgp" or "world/world_us.lgp" to be of the same level?
|
1.0
|
Incorrect folder priority because of naming. - I have been doing some tests. I use my Models Fusion mod which includes Field/Battle/World models, normally with al the files needed to create the model. I decided to add Ninostyle Chibi models (IRO) v0.245 also to 7H. So, my order is:
- Ninostyle Chibis (for Field)
- Models Fusion (basically are Chaos models, for Field)
- Other mods...
Well, the thing is that it does not matter the order I put my Models Fusion in 7H, I always see the Chaos models in game. After some checking, I detected that Ninostyle Chibis has the models in the folder **"char.lgp"**, and I have the models in the folder **"char"**.
So, as the priority is by folder naming (and they are literally different), "my" folder, "char", which has higher priority (alphabetically goes before than "char.lgp"), loads its models ALWAYS first, don't matter the order the mods are in the "My Mods" tab.
I have checked this extracting Ninostyle IRO and changing its folder to "char", then it worked correctly. But of course, a modder should not depend of the folder naming, not at least for the same type of files.
Would it be possible to treat "char/char.lgp" or "battle/battle.lgp" or "world/world_us.lgp" to be of the same level?
|
non_process
|
incorrect folder priority because of naming i have been doing some tests i use my models fusion mod which includes field battle world models normally with al the files needed to create the model i decided to add ninostyle chibi models iro also to so my order is ninostyle chibis for field models fusion basically are chaos models for field other mods well the thing is that it does not matter the order i put my models fusion in i always see the chaos models in game after some checking i detected that ninostyle chibis has the models in the folder char lgp and i have the models in the folder char so as the priority is by folder naming and they are literally different my folder char which has higher priority alphabetically goes before than char lgp loads its models always first don t matter the order the mods are in the my mods tab i have checked this extracting ninostyle iro and changing its folder to char then it worked correctly but of course a modder should not depend of the folder naming not at least for the same type of files would it be possible to treat char char lgp or battle battle lgp or world world us lgp to be of the same level
| 0
|
5,222
| 8,026,315,844
|
IssuesEvent
|
2018-07-27 03:13:45
|
turnkeylinux/tracker
|
https://api.github.com/repos/turnkeylinux/tracker
|
opened
|
Develop user friendly instructions (script?) for updating 3rd party PHP packages
|
mambo processmaker sitracker ushahidi vtiger
|
A few TurnKey appliances that include 3rd party PHP apps are not compatible with PHP7. In those instances, we have included PHP5.6 from Ondřej Surý's [third party Debian repo](https://deb.sury.org/).
That's all well and good, but we have not included these third party repos in the auto security updates. Whilst we are sure that Ondřej will do his best to not break things, it's unreasonable to expect a single person working in their own time to be able to match the level of testing and assurance that the Debian Security Team provide.
So with all that in mind, the only "newb friendly" way to update the PHP5.6 packages would be to do an `apt-get update && apt-get upgrade`. Whilst that will likely work fine, regressions can sometimes occur in non-security packages and IMO are best avoided in a production environment, unless thouroughly tested first.
So IMO we should ideally provide users with an easy way to upgrade just the third party PHP packages that are installed from Ondřej's repo. A confconsole plugin or even a downloadable script would be ideal, however copy/paste-able instruction would do.
|
1.0
|
Develop user friendly instructions (script?) for updating 3rd party PHP packages - A few TurnKey appliances that include 3rd party PHP apps are not compatible with PHP7. In those instances, we have included PHP5.6 from Ondřej Surý's [third party Debian repo](https://deb.sury.org/).
That's all well and good, but we have not included these third party repos in the auto security updates. Whilst we are sure that Ondřej will do his best to not break things, it's unreasonable to expect a single person working in their own time to be able to match the level of testing and assurance that the Debian Security Team provide.
So with all that in mind, the only "newb friendly" way to update the PHP5.6 packages would be to do an `apt-get update && apt-get upgrade`. Whilst that will likely work fine, regressions can sometimes occur in non-security packages and IMO are best avoided in a production environment, unless thouroughly tested first.
So IMO we should ideally provide users with an easy way to upgrade just the third party PHP packages that are installed from Ondřej's repo. A confconsole plugin or even a downloadable script would be ideal, however copy/paste-able instruction would do.
|
process
|
develop user friendly instructions script for updating party php packages a few turnkey appliances that include party php apps are not compatible with in those instances we have included from ondřej surý s that s all well and good but we have not included these third party repos in the auto security updates whilst we are sure that ondřej will do his best to not break things it s unreasonable to expect a single person working in their own time to be able to match the level of testing and assurance that the debian security team provide so with all that in mind the only newb friendly way to update the packages would be to do an apt get update apt get upgrade whilst that will likely work fine regressions can sometimes occur in non security packages and imo are best avoided in a production environment unless thouroughly tested first so imo we should ideally provide users with an easy way to upgrade just the third party php packages that are installed from ondřej s repo a confconsole plugin or even a downloadable script would be ideal however copy paste able instruction would do
| 1
|
47,334
| 24,956,753,360
|
IssuesEvent
|
2022-11-01 12:28:39
|
deeplearning4j/deeplearning4j
|
https://api.github.com/repos/deeplearning4j/deeplearning4j
|
closed
|
DL4J: implement MLKDNNLSTMHelper.backpropGradient + enable helper
|
Performance DL4J
|
https://github.com/eclipse/deeplearning4j/blob/master/deeplearning4j/deeplearning4j-nn/src/main/java/org/deeplearning4j/nn/layers/mkldnn/MKLDNNLSTMHelper.java#L34-L42
Now that LSTMLayer backprop has been merged (with MKLDNN support), we should finally add this to libnd4j and enable the helper in the DL4J LSTM layer:
https://github.com/eclipse/deeplearning4j/blob/master/deeplearning4j/deeplearning4j-nn/src/main/java/org/deeplearning4j/nn/layers/recurrent/LSTM.java#L79-L87
|
True
|
DL4J: implement MLKDNNLSTMHelper.backpropGradient + enable helper - https://github.com/eclipse/deeplearning4j/blob/master/deeplearning4j/deeplearning4j-nn/src/main/java/org/deeplearning4j/nn/layers/mkldnn/MKLDNNLSTMHelper.java#L34-L42
Now that LSTMLayer backprop has been merged (with MKLDNN support), we should finally add this to libnd4j and enable the helper in the DL4J LSTM layer:
https://github.com/eclipse/deeplearning4j/blob/master/deeplearning4j/deeplearning4j-nn/src/main/java/org/deeplearning4j/nn/layers/recurrent/LSTM.java#L79-L87
|
non_process
|
implement mlkdnnlstmhelper backpropgradient enable helper now that lstmlayer backprop has been merged with mkldnn support we should finally add this to and enable the helper in the lstm layer
| 0
|
54,258
| 23,217,111,397
|
IssuesEvent
|
2022-08-02 14:53:53
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
closed
|
Complete Citybase integration with Knack action items
|
Service: Dev Service: Apps Product: Banners
|
Complete [Standard Checkout document](https://docs.google.com/document/d/1jl6j3CHr1UWnCmfe3uyJK7cuPUqJIuZC/edit)
**Action items from session 1**
- [x] Section 1: Service Fees- Mateo/Diana to confirm fees are passed to customer at time of payment, or if Austin is paying fees for customers.
In issue #8268 we documented the stakeholders comments:
- Having the customer pay the service fees is the right way to go.
- If possible on the payment page, I just ask that we be transparent about the difference in costs between e-check and credit card fees.
- [x] Section 3: Min/Max- Taejin to review payment processing limits with ATD to confirm upperbound set with the payment processor
- [ ] Section 6: Custom Attributes- Mateo/Diana to work on confirming the description, subdescription, and any additional backend attributes that need to be passed along
- [x] Section 7: Cancel URL- Erin to provide TST urls; NOTE: the listed URLS in the doc are for PROD
- [x] Section 8: Return URL- Mateo/Diana to provide desired "payment success page" URL on the Austin site.
- [ ] Section 9: Confirmation Page Content- Mateo/Diana to provide Agency support phone # (if so desired by the department stakeholders)
- [x] Taejin to provide Citywide T&C to Austin Transport for review (see attachment 2 to this email)
[Terms and Conditions](https://drive.google.com/drive/folders/1GMhQjN-I4mojBMz1SDMJM0-tdTuZwl7n)
- [x] Taejin and Stephanie to coordinate on when to schedule a follow up session to:
o Discuss/close out the remaining sections listed above
o Discuss approach for knack integration (dependent on ATD meeting with Austin Library Team to review their build)
Scheduled 8/2
|
2.0
|
Complete Citybase integration with Knack action items - Complete [Standard Checkout document](https://docs.google.com/document/d/1jl6j3CHr1UWnCmfe3uyJK7cuPUqJIuZC/edit)
**Action items from session 1**
- [x] Section 1: Service Fees- Mateo/Diana to confirm fees are passed to customer at time of payment, or if Austin is paying fees for customers.
In issue #8268 we documented the stakeholders comments:
- Having the customer pay the service fees is the right way to go.
- If possible on the payment page, I just ask that we be transparent about the difference in costs between e-check and credit card fees.
- [x] Section 3: Min/Max- Taejin to review payment processing limits with ATD to confirm upperbound set with the payment processor
- [ ] Section 6: Custom Attributes- Mateo/Diana to work on confirming the description, subdescription, and any additional backend attributes that need to be passed along
- [x] Section 7: Cancel URL- Erin to provide TST urls; NOTE: the listed URLS in the doc are for PROD
- [x] Section 8: Return URL- Mateo/Diana to provide desired "payment success page" URL on the Austin site.
- [ ] Section 9: Confirmation Page Content- Mateo/Diana to provide Agency support phone # (if so desired by the department stakeholders)
- [x] Taejin to provide Citywide T&C to Austin Transport for review (see attachment 2 to this email)
[Terms and Conditions](https://drive.google.com/drive/folders/1GMhQjN-I4mojBMz1SDMJM0-tdTuZwl7n)
- [x] Taejin and Stephanie to coordinate on when to schedule a follow up session to:
o Discuss/close out the remaining sections listed above
o Discuss approach for knack integration (dependent on ATD meeting with Austin Library Team to review their build)
Scheduled 8/2
|
non_process
|
complete citybase integration with knack action items complete action items from session section service fees mateo diana to confirm fees are passed to customer at time of payment or if austin is paying fees for customers in issue we documented the stakeholders comments having the customer pay the service fees is the right way to go if possible on the payment page i just ask that we be transparent about the difference in costs between e check and credit card fees section min max taejin to review payment processing limits with atd to confirm upperbound set with the payment processor section custom attributes mateo diana to work on confirming the description subdescription and any additional backend attributes that need to be passed along section cancel url erin to provide tst urls note the listed urls in the doc are for prod section return url mateo diana to provide desired payment success page url on the austin site section confirmation page content mateo diana to provide agency support phone if so desired by the department stakeholders taejin to provide citywide t c to austin transport for review see attachment to this email taejin and stephanie to coordinate on when to schedule a follow up session to o discuss close out the remaining sections listed above o discuss approach for knack integration dependent on atd meeting with austin library team to review their build scheduled
| 0
|
215,110
| 24,126,433,296
|
IssuesEvent
|
2022-09-21 01:09:57
|
smb-h/Estates-price-prediction
|
https://api.github.com/repos/smb-h/Estates-price-prediction
|
opened
|
CVE-2022-36027 (Medium) detected in tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl
|
security vulnerability
|
## CVE-2022-36027 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/73/a3/142f73d0e076f5582fd8da29c68af0413bf529933eed09f86a8857fab0d6/tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/73/a3/142f73d0e076f5582fd8da29c68af0413bf529933eed09f86a8857fab0d6/tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/smb-h/Estates-price-prediction/commit/43d8dec55efbdc71655c52119862fee409624fda">43d8dec55efbdc71655c52119862fee409624fda</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. When converting transposed convolutions using per-channel weight quantization the converter segfaults and crashes the Python process. We have patched the issue in GitHub commit aa0b852a4588cea4d36b74feb05d93055540b450. The fix will be included in TensorFlow 2.10.0. We will also cherrypick this commit on TensorFlow 2.9.1, TensorFlow 2.8.1, and TensorFlow 2.7.2, as these are also affected and still in supported range. There are no known workarounds for this issue.
<p>Publish Date: 2022-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36027>CVE-2022-36027</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79h2-q768-fpxr">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79h2-q768-fpxr</a></p>
<p>Release Date: 2022-09-16</p>
<p>Fix Resolution: tensorflow - 2.7.2,2.8.1,2.9.1,2.10.0, tensorflow-cpu - 2.7.2,2.8.1,2.9.1,2.10.0, tensorflow-gpu - 2.7.2,2.8.1,2.9.1,2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-36027 (Medium) detected in tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl - ## CVE-2022-36027 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/73/a3/142f73d0e076f5582fd8da29c68af0413bf529933eed09f86a8857fab0d6/tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/73/a3/142f73d0e076f5582fd8da29c68af0413bf529933eed09f86a8857fab0d6/tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/smb-h/Estates-price-prediction/commit/43d8dec55efbdc71655c52119862fee409624fda">43d8dec55efbdc71655c52119862fee409624fda</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. When converting transposed convolutions using per-channel weight quantization the converter segfaults and crashes the Python process. We have patched the issue in GitHub commit aa0b852a4588cea4d36b74feb05d93055540b450. The fix will be included in TensorFlow 2.10.0. We will also cherrypick this commit on TensorFlow 2.9.1, TensorFlow 2.8.1, and TensorFlow 2.7.2, as these are also affected and still in supported range. There are no known workarounds for this issue.
<p>Publish Date: 2022-09-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36027>CVE-2022-36027</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79h2-q768-fpxr">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79h2-q768-fpxr</a></p>
<p>Release Date: 2022-09-16</p>
<p>Fix Resolution: tensorflow - 2.7.2,2.8.1,2.9.1,2.10.0, tensorflow-cpu - 2.7.2,2.8.1,2.9.1,2.10.0, tensorflow-gpu - 2.7.2,2.8.1,2.9.1,2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x tensorflow whl vulnerable library found in head commit a href found in base branch main vulnerability details tensorflow is an open source platform for machine learning when converting transposed convolutions using per channel weight quantization the converter segfaults and crashes the python process we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend
| 0
|
11,794
| 14,620,804,822
|
IssuesEvent
|
2020-12-22 20:24:48
|
pacificclimate/quail
|
https://api.github.com/repos/pacificclimate/quail
|
closed
|
Get available indices by name
|
process
|
## Description
This function takes a climdexInput object as input and returns the names of all the indices which may be computed or, if \code{get.function.names} is TRUE (the default), the names of the functions corresponding to the indices.
## Function to wrap
[`climdex.get.available.indices`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L1315)
|
1.0
|
Get available indices by name - ## Description
This function takes a climdexInput object as input and returns the names of all the indices which may be computed or, if \code{get.function.names} is TRUE (the default), the names of the functions corresponding to the indices.
## Function to wrap
[`climdex.get.available.indices`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L1315)
|
process
|
get available indices by name description this function takes a climdexinput object as input and returns the names of all the indices which may be computed or if code get function names is true the default the names of the functions corresponding to the indices function to wrap
| 1
|
230,319
| 25,464,164,546
|
IssuesEvent
|
2022-11-25 01:01:37
|
pactflow/example-bi-directional-provider-restassured
|
https://api.github.com/repos/pactflow/example-bi-directional-provider-restassured
|
opened
|
CVE-2022-45868 (High) detected in h2-1.4.200.jar
|
security vulnerability
|
## CVE-2022-45868 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.200.jar</b></p></summary>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /m.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.200.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/pactflow/example-bi-directional-provider-restassured/commit/f89cca989323a3ac73aac3d278cab02a868ac80b">f89cca989323a3ac73aac3d278cab02a868ac80b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The web-based admin console in H2 Database Engine through 2.1.214 can be started via the CLI with the argument -webAdminPassword, which allows the user to specify the password in cleartext for the web admin console. Consequently, a local user (or an attacker that has obtained local access through some means) would be able to discover the password by listing processes and their arguments. NOTE: the vendor states "This is not a vulnerability of H2 Console ... Passwords should never be passed on the command line and every qualified DBA or system administrator is expected to know that."
<p>Publish Date: 2022-11-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45868>CVE-2022-45868</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
True
|
CVE-2022-45868 (High) detected in h2-1.4.200.jar - ## CVE-2022-45868 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.200.jar</b></p></summary>
<p>H2 Database Engine</p>
<p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /m.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.200.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/pactflow/example-bi-directional-provider-restassured/commit/f89cca989323a3ac73aac3d278cab02a868ac80b">f89cca989323a3ac73aac3d278cab02a868ac80b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The web-based admin console in H2 Database Engine through 2.1.214 can be started via the CLI with the argument -webAdminPassword, which allows the user to specify the password in cleartext for the web admin console. Consequently, a local user (or an attacker that has obtained local access through some means) would be able to discover the password by listing processes and their arguments. NOTE: the vendor states "This is not a vulnerability of H2 Console ... Passwords should never be passed on the command line and every qualified DBA or system administrator is expected to know that."
<p>Publish Date: 2022-11-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45868>CVE-2022-45868</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
non_process
|
cve high detected in jar cve high severity vulnerability vulnerable library jar database engine library home page a href path to dependency file build gradle path to vulnerable library m jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details the web based admin console in database engine through can be started via the cli with the argument webadminpassword which allows the user to specify the password in cleartext for the web admin console consequently a local user or an attacker that has obtained local access through some means would be able to discover the password by listing processes and their arguments note the vendor states this is not a vulnerability of console passwords should never be passed on the command line and every qualified dba or system administrator is expected to know that publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href
| 0
|
22,501
| 31,494,396,015
|
IssuesEvent
|
2023-08-31 00:15:37
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
multiple parametric masks
|
feature: new scope: UI scope: image processing controversial no-issue-activity
|
**Is your feature request related to a problem? Please describe.**
Occasionally one would like to select a union of various properties that can be described by parametric masks (eg skin color and blue for eyes, nothing else).
**Describe the solution you'd like**
Multiple parametric masks in a module instance, to be combined under the same [rules](https://docs.darktable.org/usermanual/3.6/darkroom/masking-and-blending/masks/drawn-and-parametric/) as drawn elements and parametric masks.
**Alternatives**
For some transformations, it is possible to work around this problem by using multiple module instances with the parametric mask elements that would form the union above. However, that becomes tedious when information would be lost by a transformation (eg desaturate some colors), or the transformations are nonlinear.
For example, to desaturate everything but skin tones and blue, one would currently desaturate from red to skin tones, then from skin tones to blue, then from blue onwards. This requires 3 module instances, but all that changes is the mask.
Naturally, drawn masks can be a workaround too and are very powerful, but address a different problem.
**Additional context**
[Discussion](https://discuss.pixls.us/t/combine-parametric-masks/27711).
|
1.0
|
multiple parametric masks - **Is your feature request related to a problem? Please describe.**
Occasionally one would like to select a union of various properties that can be described by parametric masks (eg skin color and blue for eyes, nothing else).
**Describe the solution you'd like**
Multiple parametric masks in a module instance, to be combined under the same [rules](https://docs.darktable.org/usermanual/3.6/darkroom/masking-and-blending/masks/drawn-and-parametric/) as drawn elements and parametric masks.
**Alternatives**
For some transformations, it is possible to work around this problem by using multiple module instances with the parametric mask elements that would form the union above. However, that becomes tedious when information would be lost by a transformation (eg desaturate some colors), or the transformations are nonlinear.
For example, to desaturate everything but skin tones and blue, one would currently desaturate from red to skin tones, then from skin tones to blue, then from blue onwards. This requires 3 module instances, but all that changes is the mask.
Naturally, drawn masks can be a workaround too and are very powerful, but address a different problem.
**Additional context**
[Discussion](https://discuss.pixls.us/t/combine-parametric-masks/27711).
|
process
|
multiple parametric masks is your feature request related to a problem please describe occasionally one would like to select a union of various properties that can be described by parametric masks eg skin color and blue for eyes nothing else describe the solution you d like multiple parametric masks in a module instance to be combined under the same as drawn elements and parametric masks alternatives for some transformations it is possible to work around this problem by using multiple module instances with the parametric mask elements that would form the union above however that becomes tedious when information would be lost by a transformation eg desaturate some colors or the transformations are nonlinear for example to desaturate everything but skin tones and blue one would currently desaturate from red to skin tones then from skin tones to blue then from blue onwards this requires module instances but all that changes is the mask naturally drawn masks can be a workaround too and are very powerful but address a different problem additional context
| 1
|
47,117
| 7,307,492,269
|
IssuesEvent
|
2018-02-28 03:03:57
|
adafruit/circuitpython
|
https://api.github.com/repos/adafruit/circuitpython
|
closed
|
Fix warnings in sphinx documentation build
|
documentation
|
There are unconnected pages, ambiguous references, and warnings about the new `templates/replace.inc` file, which came in with MPy v1.9.2
|
1.0
|
Fix warnings in sphinx documentation build - There are unconnected pages, ambiguous references, and warnings about the new `templates/replace.inc` file, which came in with MPy v1.9.2
|
non_process
|
fix warnings in sphinx documentation build there are unconnected pages ambiguous references and warnings about the new templates replace inc file which came in with mpy
| 0
|
95,315
| 10,877,574,960
|
IssuesEvent
|
2019-11-16 11:21:04
|
tobiasanker/SakuraTree
|
https://api.github.com/repos/tobiasanker/SakuraTree
|
closed
|
Add usage of multiple blossom-subtypes
|
documentation feature / enhancement
|
## Feature-request
### Description
Blossoms can have only one subtype at the moment. It is only a problem of the compiler, because its not completely updated for the current parser-version.
### Possible Implementation
Update `convertBlossom`-method in `SakuraCompiler` to iterate over the subtypes.
|
1.0
|
Add usage of multiple blossom-subtypes - ## Feature-request
### Description
Blossoms can have only one subtype at the moment. It is only a problem of the compiler, because its not completely updated for the current parser-version.
### Possible Implementation
Update `convertBlossom`-method in `SakuraCompiler` to iterate over the subtypes.
|
non_process
|
add usage of multiple blossom subtypes feature request description blossoms can have only one subtype at the moment it is only a problem of the compiler because its not completely updated for the current parser version possible implementation update convertblossom method in sakuracompiler to iterate over the subtypes
| 0
|
257,865
| 19,533,836,895
|
IssuesEvent
|
2021-12-30 23:38:20
|
mjskay/ggdist
|
https://api.github.com/repos/mjskay/ggdist
|
closed
|
Use new @examplesIf tag in roxygen examples
|
documentation
|
For conditional examples (e.g. ones depending on suggested packages), switch to using `@examplesIf`
|
1.0
|
Use new @examplesIf tag in roxygen examples - For conditional examples (e.g. ones depending on suggested packages), switch to using `@examplesIf`
|
non_process
|
use new examplesif tag in roxygen examples for conditional examples e g ones depending on suggested packages switch to using examplesif
| 0
|
2,317
| 5,139,972,878
|
IssuesEvent
|
2017-01-12 02:18:01
|
vuejs/vue-loader
|
https://api.github.com/repos/vuejs/vue-loader
|
closed
|
External Pug template not working if it's filename is the same as the component's one.
|
pre-processor
|
I got this code in a `Hello.vue` component ;
```
<template lang="pug" src="./Hello..pug"></template>
<script>
export default {
name: 'hello',
.....
}
}
</script>
<style lang='sass' src='./style.sass'></style>
```
That doesn't work.
I noticed that it only appens when template name is the exact same as the component's one.
If for instance template name is sufixed like `Hello.tpl.pug`, that works.
|
1.0
|
External Pug template not working if it's filename is the same as the component's one. - I got this code in a `Hello.vue` component ;
```
<template lang="pug" src="./Hello..pug"></template>
<script>
export default {
name: 'hello',
.....
}
}
</script>
<style lang='sass' src='./style.sass'></style>
```
That doesn't work.
I noticed that it only appens when template name is the exact same as the component's one.
If for instance template name is sufixed like `Hello.tpl.pug`, that works.
|
process
|
external pug template not working if it s filename is the same as the component s one i got this code in a hello vue component export default name hello that doesn t work i noticed that it only appens when template name is the exact same as the component s one if for instance template name is sufixed like hello tpl pug that works
| 1
|
20,862
| 27,645,510,056
|
IssuesEvent
|
2023-03-10 22:29:14
|
cse442-at-ub/project_s23-cinco
|
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
|
opened
|
Create routing within React to navigate different pages when the corresponding button is clicked
|
Processing Task Sprint 2
|
*Task Test*
test 1:
-go to the homepage: in your editor of choice, in the project folder, type "npm start" in the terminal to open up the homepage in the browser.
- click on the login button and ensure it takes you to the login page.
test 2:
-go to the homepage: in your editor of choice, in the project folder, type "npm start" in the terminal to open up the homepage in the browser.
- click on the signup button and ensure it takes you to the signup page.
|
1.0
|
Create routing within React to navigate different pages when the corresponding button is clicked - *Task Test*
test 1:
-go to the homepage: in your editor of choice, in the project folder, type "npm start" in the terminal to open up the homepage in the browser.
- click on the login button and ensure it takes you to the login page.
test 2:
-go to the homepage: in your editor of choice, in the project folder, type "npm start" in the terminal to open up the homepage in the browser.
- click on the signup button and ensure it takes you to the signup page.
|
process
|
create routing within react to navigate different pages when the corresponding button is clicked task test test go to the homepage in your editor of choice in the project folder type npm start in the terminal to open up the homepage in the browser click on the login button and ensure it takes you to the login page test go to the homepage in your editor of choice in the project folder type npm start in the terminal to open up the homepage in the browser click on the signup button and ensure it takes you to the signup page
| 1
|
7,697
| 18,893,353,318
|
IssuesEvent
|
2021-11-15 15:23:41
|
Krakenus00/EzNote
|
https://api.github.com/repos/Krakenus00/EzNote
|
opened
|
Unite entities at the Domain-Layer
|
architecture
|
It`s some changes in architecture. We need to move the same entities from Business-Layer and Data-Access to Domain-Layer
|
1.0
|
Unite entities at the Domain-Layer - It`s some changes in architecture. We need to move the same entities from Business-Layer and Data-Access to Domain-Layer
|
non_process
|
unite entities at the domain layer it s some changes in architecture we need to move the same entities from business layer and data access to domain layer
| 0
|
9,089
| 4,413,688,691
|
IssuesEvent
|
2016-08-13 01:02:30
|
facebook/osquery
|
https://api.github.com/repos/facebook/osquery
|
closed
|
Flaky tests: DaemonTests::test_5_daemon_sigint variance in return code
|
build/test test error
|
See:
```
5/9 Test #5: python_test_osqueryd .............***Failed 33.75 sec
.I0721 12:22:59.882045 26907 options.cpp:61] Verbose logging enabled by config option
I0721 12:22:59.882532 26907 daemon.cpp:38] Not starting the distributed query service: Distributed query service not enabled.
...FI0721 12:23:29.781241 26946 options.cpp:61] Verbose logging enabled by config option
I0721 12:23:29.781740 26946 daemon.cpp:38] Not starting the distributed query service: Distributed query service not enabled.
.
======================================================================
FAIL: test_5_daemon_sigint (__main__.DaemonTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/osquery/jenkins/workspace/osqueryPullRequestBuild/TargetSystem/centos7/tools/tests/test_base.py", line 455, in wrapper
raise exceptions[0][0]
AssertionError: -2 != 130
----------------------------------------------------------------------
Ran 6 tests in 33.695s
FAILED (failures=1)
Test (attempt 1) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437)
Test (attempt 2) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437)
Test (attempt 3) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437)
```
For an example see: https://jenkins.osquery.io/job/osqueryPullRequestBuild/2912/TargetSystem=centos7/console
|
1.0
|
Flaky tests: DaemonTests::test_5_daemon_sigint variance in return code - See:
```
5/9 Test #5: python_test_osqueryd .............***Failed 33.75 sec
.I0721 12:22:59.882045 26907 options.cpp:61] Verbose logging enabled by config option
I0721 12:22:59.882532 26907 daemon.cpp:38] Not starting the distributed query service: Distributed query service not enabled.
...FI0721 12:23:29.781241 26946 options.cpp:61] Verbose logging enabled by config option
I0721 12:23:29.781740 26946 daemon.cpp:38] Not starting the distributed query service: Distributed query service not enabled.
.
======================================================================
FAIL: test_5_daemon_sigint (__main__.DaemonTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/osquery/jenkins/workspace/osqueryPullRequestBuild/TargetSystem/centos7/tools/tests/test_base.py", line 455, in wrapper
raise exceptions[0][0]
AssertionError: -2 != 130
----------------------------------------------------------------------
Ran 6 tests in 33.695s
FAILED (failures=1)
Test (attempt 1) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437)
Test (attempt 2) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437)
Test (attempt 3) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437)
```
For an example see: https://jenkins.osquery.io/job/osqueryPullRequestBuild/2912/TargetSystem=centos7/console
|
non_process
|
flaky tests daemontests test daemon sigint variance in return code see test python test osqueryd failed sec options cpp verbose logging enabled by config option daemon cpp not starting the distributed query service distributed query service not enabled options cpp verbose logging enabled by config option daemon cpp not starting the distributed query service distributed query service not enabled fail test daemon sigint main daemontests traceback most recent call last file home osquery jenkins workspace osquerypullrequestbuild targetsystem tools tests test base py line in wrapper raise exceptions assertionerror ran tests in failed failures test attempt daemontests test daemon sigint failed test base py test attempt daemontests test daemon sigint failed test base py test attempt daemontests test daemon sigint failed test base py for an example see
| 0
|
243,852
| 26,290,654,331
|
IssuesEvent
|
2023-01-08 11:20:39
|
yaeljacobs67/fs-agent
|
https://api.github.com/repos/yaeljacobs67/fs-agent
|
opened
|
WS-2016-7112 (Medium) detected in spring-context-4.3.1.RELEASE.jar
|
security vulnerability
|
## WS-2016-7112 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-context-4.3.1.RELEASE.jar</b></p></summary>
<p>Spring Context</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /fs-agent/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-context/4.3.1.RELEASE/spring-context-4.3.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- whitesource-analysis-via-18.12.1.204.jar (Root Library)
- whitesource-utilities-0.0.1.jar
- spring-web-4.3.1.RELEASE.jar
- :x: **spring-context-4.3.1.RELEASE.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework, versions 3.0.0.RELEASE through 3.2.17.RELEASE, 4.0.0.RELEASE through 4.2.7.RELEASE and 4.3.0.RELEASE through 4.3.1.RELEASE are vulnerable to Stack-based Buffer Overflow, which allows an authenticated attacker to crash the application when giving CronSequenceGenerator a reversed range in the “minutes” or “hours” fields.
<p>Publish Date: 2021-09-23
<p>URL: <a href=https://github.com/spring-projects/spring-framework/commit/e431624e8472b3b53d1a0c4528bf736c612f1fd9>WS-2016-7112</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2016-07-14</p>
<p>Fix Resolution: org.springframework:spring-context:3.2.18.RELEASE,4.2.8.RELEASE,4.3.2.RELEASE,5.0.0.RELEASE</p>
</p>
</details>
<p></p>
|
True
|
WS-2016-7112 (Medium) detected in spring-context-4.3.1.RELEASE.jar - ## WS-2016-7112 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-context-4.3.1.RELEASE.jar</b></p></summary>
<p>Spring Context</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /fs-agent/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-context/4.3.1.RELEASE/spring-context-4.3.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- whitesource-analysis-via-18.12.1.204.jar (Root Library)
- whitesource-utilities-0.0.1.jar
- spring-web-4.3.1.RELEASE.jar
- :x: **spring-context-4.3.1.RELEASE.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework, versions 3.0.0.RELEASE through 3.2.17.RELEASE, 4.0.0.RELEASE through 4.2.7.RELEASE and 4.3.0.RELEASE through 4.3.1.RELEASE are vulnerable to Stack-based Buffer Overflow, which allows an authenticated attacker to crash the application when giving CronSequenceGenerator a reversed range in the “minutes” or “hours” fields.
<p>Publish Date: 2021-09-23
<p>URL: <a href=https://github.com/spring-projects/spring-framework/commit/e431624e8472b3b53d1a0c4528bf736c612f1fd9>WS-2016-7112</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2016-07-14</p>
<p>Fix Resolution: org.springframework:spring-context:3.2.18.RELEASE,4.2.8.RELEASE,4.3.2.RELEASE,5.0.0.RELEASE</p>
</p>
</details>
<p></p>
|
non_process
|
ws medium detected in spring context release jar ws medium severity vulnerability vulnerable library spring context release jar spring context library home page a href path to dependency file fs agent pom xml path to vulnerable library root repository org springframework spring context release spring context release jar dependency hierarchy whitesource analysis via jar root library whitesource utilities jar spring web release jar x spring context release jar vulnerable library vulnerability details in spring framework versions release through release release through release and release through release are vulnerable to stack based buffer overflow which allows an authenticated attacker to crash the application when giving cronsequencegenerator a reversed range in the “minutes” or “hours” fields publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution org springframework spring context release release release release
| 0
|
7,148
| 10,289,323,736
|
IssuesEvent
|
2019-08-27 15:35:53
|
Holmeyoung/blog-comment
|
https://api.github.com/repos/Holmeyoung/blog-comment
|
opened
|
python 多线程与多进程 | 一只羊的碎碎念
|
/python-threading-multiprocessing/ Gitalk
|
https://www.holmeyoung.com/python-threading-multiprocessing/
多进程进程池阻塞与非阻塞 异步进程池(非阻塞)123456789101112131415161718from multiprocessing import Pooldef test(i): print (i)if __name__=="__main__": pool = Pool(processes=10) for i in xrange(500): ''' For循环中执行步骤:
|
1.0
|
python 多线程与多进程 | 一只羊的碎碎念 - https://www.holmeyoung.com/python-threading-multiprocessing/
多进程进程池阻塞与非阻塞 异步进程池(非阻塞)123456789101112131415161718from multiprocessing import Pooldef test(i): print (i)if __name__=="__main__": pool = Pool(processes=10) for i in xrange(500): ''' For循环中执行步骤:
|
process
|
python 多线程与多进程 一只羊的碎碎念 多进程进程池阻塞与非阻塞 异步进程池(非阻塞) multiprocessing import pooldef test i print i if name main pool pool processes for i in xrange for循环中执行步骤:
| 1
|
111,677
| 14,121,332,300
|
IssuesEvent
|
2020-11-09 01:39:42
|
rohmishra/personal_website
|
https://api.github.com/repos/rohmishra/personal_website
|
closed
|
General improvement and changes.
|
Design Needed enhancement help wanted no-issue-activity
|
**This report is tracking multiple changes to improve the design of website.**
1. ~Fix "My projects" text on dark mode chrome. #38~ ( Fixed in #43 )
2. Change background design for designs pages
3. Improve the cards for projects.
4. Show more details for each project (pending design)
5. Prepare for transition to rmishra.me
6. Possibly disable contact me section till database is moved out of m-lab.
7. Add About me page #13
8. Add better CSS animations
|
1.0
|
General improvement and changes. - **This report is tracking multiple changes to improve the design of website.**
1. ~Fix "My projects" text on dark mode chrome. #38~ ( Fixed in #43 )
2. Change background design for designs pages
3. Improve the cards for projects.
4. Show more details for each project (pending design)
5. Prepare for transition to rmishra.me
6. Possibly disable contact me section till database is moved out of m-lab.
7. Add About me page #13
8. Add better CSS animations
|
non_process
|
general improvement and changes this report is tracking multiple changes to improve the design of website fix my projects text on dark mode chrome fixed in change background design for designs pages improve the cards for projects show more details for each project pending design prepare for transition to rmishra me possibly disable contact me section till database is moved out of m lab add about me page add better css animations
| 0
|
303,827
| 26,230,700,791
|
IssuesEvent
|
2023-01-04 23:44:26
|
ImagingDataCommons/IDC-WebApp
|
https://api.github.com/repos/ImagingDataCommons/IDC-WebApp
|
closed
|
In test and prod tier, There is a warning message for BMI when a cohorts with BMI filters is saved.
|
bug testing needed testing passed
|
Note - I was unable to determine if this issue is actually a bug (We discussed this in our weekly meeting and i was suggested to open a ticket). Please feel free to delete the ticket if its not a bug.
Summary - On the IDC webapp, When I select TCGA collection and BMI filter to create a cohort, I see the message that says Invalid operator seen for attribute 'bmi'; default of OR used instead. (Like in the image below) But I am unable to find where the OR is being used. It still shows AND in the filter definition and also in the filter URL,

To reproduce this issue in test,
Select TCGA collections and related filter BMI (eg overweight)
Save the cohort to see the warning on top of the page.
|
2.0
|
In test and prod tier, There is a warning message for BMI when a cohorts with BMI filters is saved. - Note - I was unable to determine if this issue is actually a bug (We discussed this in our weekly meeting and i was suggested to open a ticket). Please feel free to delete the ticket if its not a bug.
Summary - On the IDC webapp, When I select TCGA collection and BMI filter to create a cohort, I see the message that says Invalid operator seen for attribute 'bmi'; default of OR used instead. (Like in the image below) But I am unable to find where the OR is being used. It still shows AND in the filter definition and also in the filter URL,

To reproduce this issue in test,
Select TCGA collections and related filter BMI (eg overweight)
Save the cohort to see the warning on top of the page.
|
non_process
|
in test and prod tier there is a warning message for bmi when a cohorts with bmi filters is saved note i was unable to determine if this issue is actually a bug we discussed this in our weekly meeting and i was suggested to open a ticket please feel free to delete the ticket if its not a bug summary on the idc webapp when i select tcga collection and bmi filter to create a cohort i see the message that says invalid operator seen for attribute bmi default of or used instead like in the image below but i am unable to find where the or is being used it still shows and in the filter definition and also in the filter url to reproduce this issue in test select tcga collections and related filter bmi eg overweight save the cohort to see the warning on top of the page
| 0
|
6,748
| 6,584,002,898
|
IssuesEvent
|
2017-09-13 08:35:08
|
vmware/docker-volume-vsphere
|
https://api.github.com/repos/vmware/docker-volume-vsphere
|
closed
|
Improve CI/CD log collection
|
component/ci-infrastructure kind/enhancement
|
- [ ] Dump relevant information for both successful and failed runs
- [ ] Post logs to a location where developers can download it
- [ ] Improve logging output in CI/CD
|
1.0
|
Improve CI/CD log collection - - [ ] Dump relevant information for both successful and failed runs
- [ ] Post logs to a location where developers can download it
- [ ] Improve logging output in CI/CD
|
non_process
|
improve ci cd log collection dump relevant information for both successful and failed runs post logs to a location where developers can download it improve logging output in ci cd
| 0
|
5,105
| 7,883,610,503
|
IssuesEvent
|
2018-06-27 06:08:31
|
ropensci/onboarding-meta
|
https://api.github.com/repos/ropensci/onboarding-meta
|
closed
|
Submitters not using reviews
|
process
|
What do we do if submitters don't want to make changes that reviewers/editors suggest?
I think we have to be firm about this. If submitter gives good reasons for not making changes reviewers/editors suggest that's fine (and it should be item by item response too), but if there's a blanket statement that they just don't want to make the changes, I think we need to reject the submission.
I think this scenario could arise from a submitter simply wanting to be under rOpenSci but not wanting to do the work.
|
1.0
|
Submitters not using reviews - What do we do if submitters don't want to make changes that reviewers/editors suggest?
I think we have to be firm about this. If submitter gives good reasons for not making changes reviewers/editors suggest that's fine (and it should be item by item response too), but if there's a blanket statement that they just don't want to make the changes, I think we need to reject the submission.
I think this scenario could arise from a submitter simply wanting to be under rOpenSci but not wanting to do the work.
|
process
|
submitters not using reviews what do we do if submitters don t want to make changes that reviewers editors suggest i think we have to be firm about this if submitter gives good reasons for not making changes reviewers editors suggest that s fine and it should be item by item response too but if there s a blanket statement that they just don t want to make the changes i think we need to reject the submission i think this scenario could arise from a submitter simply wanting to be under ropensci but not wanting to do the work
| 1
|
17,312
| 23,134,091,686
|
IssuesEvent
|
2022-07-28 13:01:41
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
opened
|
Cross compile qnx, Bazel does not produce shared libraries. Release 5.2.0 - $MONTH $YEAR
|
P1 type: process release team-OSS
|
**Description of the bug:**
When I cross compile QNX, no dynamic library is generated. At this time, after I add
`feature( name = "supports_dynamic_linker", enabled = True, )`, it prompts me that ld cannot be found. then i create a ld soft connection `ln aarch64-unknown-nto-qnx7.1.0-ld ld`, this problem is solved.i'm surprised that i specified ld in the tool_path like this `tool_path( name = "ld", path = QNX_HOST + "/usr/bin/aarch64-unknown-nto-qnx7.1.0-ld",),`, but it doesn't seem to work,How can I solve this problem gracefully。
**Which operating system are you running Bazel on?**
Linux.
**What is the output of bazel info release?**
5.2.0
**Have you found anything relevant by searching the web?**
[When I cross compile the example , Bazel does not produce shared libraries](https://github.com/bazelbuild/bazel/issues/7637)
|
1.0
|
Cross compile qnx, Bazel does not produce shared libraries. Release 5.2.0 - $MONTH $YEAR - **Description of the bug:**
When I cross compile QNX, no dynamic library is generated. At this time, after I add
`feature( name = "supports_dynamic_linker", enabled = True, )`, it prompts me that ld cannot be found. then i create a ld soft connection `ln aarch64-unknown-nto-qnx7.1.0-ld ld`, this problem is solved.i'm surprised that i specified ld in the tool_path like this `tool_path( name = "ld", path = QNX_HOST + "/usr/bin/aarch64-unknown-nto-qnx7.1.0-ld",),`, but it doesn't seem to work,How can I solve this problem gracefully。
**Which operating system are you running Bazel on?**
Linux.
**What is the output of bazel info release?**
5.2.0
**Have you found anything relevant by searching the web?**
[When I cross compile the example , Bazel does not produce shared libraries](https://github.com/bazelbuild/bazel/issues/7637)
|
process
|
cross compile qnx bazel does not produce shared libraries release month year description of the bug when i cross compile qnx no dynamic library is generated at this time after i add feature name supports dynamic linker enabled true , it prompts me that ld cannot be found then i create a ld soft connection ln unknown nto ld ld this problem is solved i m surprised that i specified ld in the tool path like this tool path name ld path qnx host usr bin unknown nto ld but it doesn t seem to work,how can i solve this problem gracefully。 which operating system are you running bazel on linux what is the output of bazel info release have you found anything relevant by searching the web
| 1
|
7,132
| 10,278,464,312
|
IssuesEvent
|
2019-08-25 14:39:06
|
nextmoov/nextmoov
|
https://api.github.com/repos/nextmoov/nextmoov
|
closed
|
Standard entrypoint: yarn start, yarn start:production
|
#Dev Tools & Processes
|
# Standard EntryPoint
```yarn start``` Run the project in *Development*
```yarn build``` Build the static assets to be served in production in /dist
# Archived Discussion
I would like to standardize the entrypoint for Front & Back project, for both Production & Development.
I hereby propose that all our project (back & front) follow this convention:
```yarn start``` Run the project in *Development*
```yarn serve``` Run the project in *Production* (serve precompiled assets in /dist)
#### Why not yarn start in Production?
Create-React-App (and actually a lot of other framework) use ```yarn start``` as the entrypoint for Development — I propose that we don't override this behaviour
|
1.0
|
Standard entrypoint: yarn start, yarn start:production - # Standard EntryPoint
```yarn start``` Run the project in *Development*
```yarn build``` Build the static assets to be served in production in /dist
# Archived Discussion
I would like to standardize the entrypoint for Front & Back project, for both Production & Development.
I hereby propose that all our project (back & front) follow this convention:
```yarn start``` Run the project in *Development*
```yarn serve``` Run the project in *Production* (serve precompiled assets in /dist)
#### Why not yarn start in Production?
Create-React-App (and actually a lot of other framework) use ```yarn start``` as the entrypoint for Development — I propose that we don't override this behaviour
|
process
|
standard entrypoint yarn start yarn start production standard entrypoint yarn start run the project in development yarn build build the static assets to be served in production in dist archived discussion i would like to standardize the entrypoint for front back project for both production development i hereby propose that all our project back front follow this convention yarn start run the project in development yarn serve run the project in production serve precompiled assets in dist why not yarn start in production create react app and actually a lot of other framework use yarn start as the entrypoint for development — i propose that we don t override this behaviour
| 1
|
54,708
| 3,071,123,078
|
IssuesEvent
|
2015-08-19 09:58:13
|
pavel-pimenov/flylinkdc-r5xx
|
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
|
closed
|
Некорректно работает перенос папки в очереди загрузки
|
bug Component-Logic imported Priority-Critical Usability
|
_From [rain.bipper@gmail.com](https://code.google.com/u/rain.bipper@gmail.com/) on November 06, 2009 12:26:45_
если в очереди загрузки выбрать "переместить" папку (подчеркиваю! именно
папку, а не файлы), то после перемещения в конечной папке будут все файлы
из исходной папки, а сама папка не создается.
Например: в очереди есть папка1, которой есть N файлов. делаю переместить
папка1 в папка2. в результате N файлов сыпется прямо в папка2, подпапка не
создается!
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=49_
|
1.0
|
Некорректно работает перенос папки в очереди загрузки - _From [rain.bipper@gmail.com](https://code.google.com/u/rain.bipper@gmail.com/) on November 06, 2009 12:26:45_
если в очереди загрузки выбрать "переместить" папку (подчеркиваю! именно
папку, а не файлы), то после перемещения в конечной папке будут все файлы
из исходной папки, а сама папка не создается.
Например: в очереди есть папка1, которой есть N файлов. делаю переместить
папка1 в папка2. в результате N файлов сыпется прямо в папка2, подпапка не
создается!
_Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=49_
|
non_process
|
некорректно работает перенос папки в очереди загрузки from on november если в очереди загрузки выбрать переместить папку подчеркиваю именно папку а не файлы то после перемещения в конечной папке будут все файлы из исходной папки а сама папка не создается например в очереди есть которой есть n файлов делаю переместить в в результате n файлов сыпется прямо в подпапка не создается original issue
| 0
|
12,638
| 15,016,625,529
|
IssuesEvent
|
2021-02-01 09:49:39
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
closed
|
newly deployed solutions are not listed on 'marketplace' deployed solutions.
|
process_wontfix type_bug
|
just deployed taiga and its not listed on marketplace deployed solution
<img width="1433" alt="Screenshot 2021-01-28 at 13 42 40" src="https://user-images.githubusercontent.com/43240801/106140257-d0a98700-616e-11eb-90dd-543a786022ea.png">
although it's shown on taiga/my workloads
<img width="1404" alt="Screenshot 2021-01-28 at 13 42 34" src="https://user-images.githubusercontent.com/43240801/106140389-ffbff880-616e-11eb-8ce4-fd72aeabe6b1.png">
|
1.0
|
newly deployed solutions are not listed on 'marketplace' deployed solutions. - just deployed taiga and its not listed on marketplace deployed solution
<img width="1433" alt="Screenshot 2021-01-28 at 13 42 40" src="https://user-images.githubusercontent.com/43240801/106140257-d0a98700-616e-11eb-90dd-543a786022ea.png">
although it's shown on taiga/my workloads
<img width="1404" alt="Screenshot 2021-01-28 at 13 42 34" src="https://user-images.githubusercontent.com/43240801/106140389-ffbff880-616e-11eb-8ce4-fd72aeabe6b1.png">
|
process
|
newly deployed solutions are not listed on marketplace deployed solutions just deployed taiga and its not listed on marketplace deployed solution img width alt screenshot at src although it s shown on taiga my workloads img width alt screenshot at src
| 1
|
20,053
| 26,541,065,203
|
IssuesEvent
|
2023-01-19 19:18:31
|
Azure/azure-sdk-tools
|
https://api.github.com/repos/Azure/azure-sdk-tools
|
closed
|
APIView integration from Cadl PR
|
Epic APIView Central-EngSys Cadl WS: Process Tools & Automation
|
When a user creates a PR in the specs repo and is using Cadl, make sure there is an APIView created for it (as we currently do for SDK changes)
|
1.0
|
APIView integration from Cadl PR - When a user creates a PR in the specs repo and is using Cadl, make sure there is an APIView created for it (as we currently do for SDK changes)
|
process
|
apiview integration from cadl pr when a user creates a pr in the specs repo and is using cadl make sure there is an apiview created for it as we currently do for sdk changes
| 1
|
107,706
| 13,503,651,480
|
IssuesEvent
|
2020-09-13 14:30:52
|
rubyforgood/casa
|
https://api.github.com/repos/rubyforgood/casa
|
closed
|
Make edit profile accessible by dropdown menu in All CASA Admin dashboard
|
:cityscape: Multitenancy :globe_with_meridians: All CASA Admin :paintbrush: Design Good First Issue Help Wanted Priority: High Ruby For Good 🎃 Fall 2020
|
Part of epic #527
**What type of user is this for? [volunteer/admin/supervisor/all or All CASA Admin]**
All CASA Admin
**Where does/should this occur?**
In the All CASA Admin dashboard (dependent on #628)
**Description**
All CASA Admin needs an entry point to their Edit Profile page. It should be accessible by clicking on their email in the upper right hand corner of their user dashboard. When they click, "Edit Profile" should appear as the only option in a drop down menu.
Log in as any other user to see how this should look. <a href="https://github.com/rubyforgood/casa/blob/master/db/seeds.rb>Login instructions here.</a>
**Screenshots**
It should look like this:

Relates to issues: #627, #628, #630
|
1.0
|
Make edit profile accessible by dropdown menu in All CASA Admin dashboard - Part of epic #527
**What type of user is this for? [volunteer/admin/supervisor/all or All CASA Admin]**
All CASA Admin
**Where does/should this occur?**
In the All CASA Admin dashboard (dependent on #628)
**Description**
All CASA Admin needs an entry point to their Edit Profile page. It should be accessible by clicking on their email in the upper right hand corner of their user dashboard. When they click, "Edit Profile" should appear as the only option in a drop down menu.
Log in as any other user to see how this should look. <a href="https://github.com/rubyforgood/casa/blob/master/db/seeds.rb>Login instructions here.</a>
**Screenshots**
It should look like this:

Relates to issues: #627, #628, #630
|
non_process
|
make edit profile accessible by dropdown menu in all casa admin dashboard part of epic what type of user is this for all casa admin where does should this occur in the all casa admin dashboard dependent on description all casa admin needs an entry point to their edit profile page it should be accessible by clicking on their email in the upper right hand corner of their user dashboard when they click edit profile should appear as the only option in a drop down menu log in as any other user to see how this should look screenshots it should look like this relates to issues
| 0
|
14,772
| 18,049,104,580
|
IssuesEvent
|
2021-09-19 12:29:34
|
openstreetmap-polska/gugik2osm
|
https://api.github.com/repos/openstreetmap-polska/gugik2osm
|
closed
|
Budynki w budowie
|
data_processing
|
Aktualnie budynki w budowie są importowane jako zbudowane.
Proponuję je oznaczać jako building=construction, ew. pomijać.


|
1.0
|
Budynki w budowie - Aktualnie budynki w budowie są importowane jako zbudowane.
Proponuję je oznaczać jako building=construction, ew. pomijać.


|
process
|
budynki w budowie aktualnie budynki w budowie są importowane jako zbudowane proponuję je oznaczać jako building construction ew pomijać
| 1
|
84,822
| 16,563,155,407
|
IssuesEvent
|
2021-05-29 00:08:45
|
alefragnani/vscode-pascal-formatter
|
https://api.github.com/repos/alefragnani/vscode-pascal-formatter
|
closed
|
[FEATURE] - Support Workspace Trust API (limited)
|
enhancement vscode api adoption
|
<!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
Just like https://github.com/alefragnani/vscode-bookmarks/issues/430
The extension will have **limited** support, not allowing **indicate the formatter app path** on untrusted workspaces.
|
1.0
|
[FEATURE] - Support Workspace Trust API (limited) - <!-- Please search existing issues to avoid creating duplicates. -->
<!-- Describe the feature you'd like. -->
Just like https://github.com/alefragnani/vscode-bookmarks/issues/430
The extension will have **limited** support, not allowing **indicate the formatter app path** on untrusted workspaces.
|
non_process
|
support workspace trust api limited just like the extension will have limited support not allowing indicate the formatter app path on untrusted workspaces
| 0
|
93,613
| 3,906,621,301
|
IssuesEvent
|
2016-04-19 09:31:57
|
leoncastillejos/sonar
|
https://api.github.com/repos/leoncastillejos/sonar
|
closed
|
Sonar Web authentication and encryption
|
action:in_progress priority:high type:enhancement
|
Right now, Sonar web is cleartext, and does not ask for a password, therefore anyone is able to create, delete, view and modify machines and alerts. This is not ideal.
|
1.0
|
Sonar Web authentication and encryption - Right now, Sonar web is cleartext, and does not ask for a password, therefore anyone is able to create, delete, view and modify machines and alerts. This is not ideal.
|
non_process
|
sonar web authentication and encryption right now sonar web is cleartext and does not ask for a password therefore anyone is able to create delete view and modify machines and alerts this is not ideal
| 0
|
70
| 2,523,723,824
|
IssuesEvent
|
2015-01-20 13:06:53
|
sysown/proxysql-0.2
|
https://api.github.com/repos/sysown/proxysql-0.2
|
opened
|
Implement table stats_mysql_commands_counters to store counters about commands
|
ADMIN cxx_db cxx_pa development enhancement MYSQL PROTOCOL QUERY PROCESSOR STATISTICS
|
Table needs to be created in statsdb
Optionally it can also store total execution time
|
1.0
|
Implement table stats_mysql_commands_counters to store counters about commands - Table needs to be created in statsdb
Optionally it can also store total execution time
|
process
|
implement table stats mysql commands counters to store counters about commands table needs to be created in statsdb optionally it can also store total execution time
| 1
|
518,850
| 15,035,991,207
|
IssuesEvent
|
2021-02-02 14:46:04
|
StatCan/daaas
|
https://api.github.com/repos/StatCan/daaas
|
closed
|
Support Construction Starts project by investigating running fastai/pytorch jobs in distributed multi-GPU setup
|
component/kubeflow current-sprint kind/task priority/soon size/XL triage/support
|
Goals:
* demonstrate running pytorch jobs in distributed multi-gpu setup on platform
* demonstrate running fastai jobs in distributed multi-gpu setup on platform
* advise construction starts project on how to achieve above multi-gpu setups to scale up their network training
* build a tutorial/guide notebooks for distributed multi-gpu jobs
|
1.0
|
Support Construction Starts project by investigating running fastai/pytorch jobs in distributed multi-GPU setup - Goals:
* demonstrate running pytorch jobs in distributed multi-gpu setup on platform
* demonstrate running fastai jobs in distributed multi-gpu setup on platform
* advise construction starts project on how to achieve above multi-gpu setups to scale up their network training
* build a tutorial/guide notebooks for distributed multi-gpu jobs
|
non_process
|
support construction starts project by investigating running fastai pytorch jobs in distributed multi gpu setup goals demonstrate running pytorch jobs in distributed multi gpu setup on platform demonstrate running fastai jobs in distributed multi gpu setup on platform advise construction starts project on how to achieve above multi gpu setups to scale up their network training build a tutorial guide notebooks for distributed multi gpu jobs
| 0
|
14,258
| 9,221,057,571
|
IssuesEvent
|
2019-03-11 19:00:00
|
OriginProtocol/origin
|
https://api.github.com/repos/OriginProtocol/origin
|
opened
|
Shipping Information
|
dapp discussion html/css javascript p1 security transaction flow ui/ux
|
This is a placeholder for discussing the long-requested ability for a buyer to add a shipping address when making an offer. This hasn't been done yet because we obviously don't want to store unencrypted, sensitive information publicly. And the messaging infrastructure was not stable enough to rely on for this use.
|
True
|
Shipping Information - This is a placeholder for discussing the long-requested ability for a buyer to add a shipping address when making an offer. This hasn't been done yet because we obviously don't want to store unencrypted, sensitive information publicly. And the messaging infrastructure was not stable enough to rely on for this use.
|
non_process
|
shipping information this is a placeholder for discussing the long requested ability for a buyer to add a shipping address when making an offer this hasn t been done yet because we obviously don t want to store unencrypted sensitive information publicly and the messaging infrastructure was not stable enough to rely on for this use
| 0
|
199,284
| 22,693,267,196
|
IssuesEvent
|
2022-07-05 01:05:37
|
PGreaneyLYIT/easybuggy4django
|
https://api.github.com/repos/PGreaneyLYIT/easybuggy4django
|
opened
|
CVE-2022-34265 (High) detected in Django-2.0.3-py3-none-any.whl
|
security vulnerability
|
## CVE-2022-34265 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.0.3-py3-none-any.whl</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl">https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl</a></p>
<p>Path to dependency file: /tmp/ws-scm/easybuggy4django</p>
<p>Path to vulnerable library: /tmp/ws-scm/easybuggy4django,/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-2.0.3-py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/PGreaneyLYIT/easybuggy4django/commit/5403b8fbcea4b699ce64d05146aeaa76d1062d89">5403b8fbcea4b699ce64d05146aeaa76d1062d89</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in Django 3.2 before 3.2.14 and 4.0 before 4.0.6. The Trunc() and Extract() database functions are subject to SQL injection if untrusted data is used as a kind/lookup_name value. Applications that constrain the lookup name and kind choice to a known safe list are unaffected.
<p>Publish Date: 2022-07-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34265>CVE-2022-34265</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.djangoproject.com/weblog/2022/jul/04/security-releases/">https://www.djangoproject.com/weblog/2022/jul/04/security-releases/</a></p>
<p>Release Date: 2022-06-22</p>
<p>Fix Resolution: Django - 3.2.14,4.0.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-34265 (High) detected in Django-2.0.3-py3-none-any.whl - ## CVE-2022-34265 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.0.3-py3-none-any.whl</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl">https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl</a></p>
<p>Path to dependency file: /tmp/ws-scm/easybuggy4django</p>
<p>Path to vulnerable library: /tmp/ws-scm/easybuggy4django,/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-2.0.3-py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/PGreaneyLYIT/easybuggy4django/commit/5403b8fbcea4b699ce64d05146aeaa76d1062d89">5403b8fbcea4b699ce64d05146aeaa76d1062d89</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in Django 3.2 before 3.2.14 and 4.0 before 4.0.6. The Trunc() and Extract() database functions are subject to SQL injection if untrusted data is used as a kind/lookup_name value. Applications that constrain the lookup name and kind choice to a known safe list are unaffected.
<p>Publish Date: 2022-07-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34265>CVE-2022-34265</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.djangoproject.com/weblog/2022/jul/04/security-releases/">https://www.djangoproject.com/weblog/2022/jul/04/security-releases/</a></p>
<p>Release Date: 2022-06-22</p>
<p>Fix Resolution: Django - 3.2.14,4.0.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in django none any whl cve high severity vulnerability vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file tmp ws scm path to vulnerable library tmp ws scm requirements txt dependency hierarchy x django none any whl vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in django before and before the trunc and extract database functions are subject to sql injection if untrusted data is used as a kind lookup name value applications that constrain the lookup name and kind choice to a known safe list are unaffected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution django step up your open source security game with mend
| 0
|
664,690
| 22,285,284,120
|
IssuesEvent
|
2022-06-11 14:22:09
|
magento/magento2
|
https://api.github.com/repos/magento/magento2
|
closed
|
jQuery UI Slider and SelectMenu Mapping is not correct in requirejs-config.js
|
Issue: Confirmed Reproduced on 2.4.x Progress: PR in progress Priority: P2 Area: Framework Reported on 2.4.4
|
<!---
Thank you for contributing to Magento.
To help us process this issue we recommend that you add the following information:
- Summary of the issue,
- Information on your environment,
- Steps to reproduce,
- Expected and actual results,
Fields marked with (*) are required. Please don't remove the template.
Please also have a look at our guidelines article before adding a new issue https://github.com/magento/magento2/wiki/Issue-reporting-guidelines
-->
### Preconditions (*)
<!---
Please provide as detailed information about your environment as possible.
For example Magento version, tag, HEAD, PHP & MySQL version, etc..
-->
On Magento 2.4.4 release, jQuery UI Slider and SelectMenu Mapping is not correct in vendor/magento/module-theme/view/base/requirejs-config.js. This causes an error on frontend as jQuery slider and selectmenu modules do not get loaded and we see an error saying "$j(..).slider is not a function".
### Steps to reproduce (*)
<!---
It is important to provide a set of clear steps to reproduce this bug.
If relevant please include code samples
-->
Open vendor/magento/module-theme/view/base/requirejs-config.js
On line 42 you will notice the incorrect mapping -
'jquery-ui-modules/slider': 'jquery/ui-modules/widgets/selectmenu',
### Expected result (*)
<!--- Tell us what should happen -->
The correct mapping should be -
'jquery-ui-modules/selectmenu': 'jquery/ui-modules/widgets/selectmenu',
'jquery-ui-modules/slider': 'jquery/ui-modules/widgets/slider',
### Actual result (*)
<!--- Tell us what happens instead -->
This causes an error on frontend where jQuery slider is used and shows an error saying "$j(..).slider is not a function".
<img width="1222" alt="Screenshot 2022-04-19 at 4 38 23 PM" src="https://user-images.githubusercontent.com/22843013/163991117-2a3bcc65-4215-4c43-81ea-991cd8d7e21e.png">
|
1.0
|
jQuery UI Slider and SelectMenu Mapping is not correct in requirejs-config.js - <!---
Thank you for contributing to Magento.
To help us process this issue we recommend that you add the following information:
- Summary of the issue,
- Information on your environment,
- Steps to reproduce,
- Expected and actual results,
Fields marked with (*) are required. Please don't remove the template.
Please also have a look at our guidelines article before adding a new issue https://github.com/magento/magento2/wiki/Issue-reporting-guidelines
-->
### Preconditions (*)
<!---
Please provide as detailed information about your environment as possible.
For example Magento version, tag, HEAD, PHP & MySQL version, etc..
-->
On Magento 2.4.4 release, jQuery UI Slider and SelectMenu Mapping is not correct in vendor/magento/module-theme/view/base/requirejs-config.js. This causes an error on frontend as jQuery slider and selectmenu modules do not get loaded and we see an error saying "$j(..).slider is not a function".
### Steps to reproduce (*)
<!---
It is important to provide a set of clear steps to reproduce this bug.
If relevant please include code samples
-->
Open vendor/magento/module-theme/view/base/requirejs-config.js
On line 42 you will notice the incorrect mapping -
'jquery-ui-modules/slider': 'jquery/ui-modules/widgets/selectmenu',
### Expected result (*)
<!--- Tell us what should happen -->
The correct mapping should be -
'jquery-ui-modules/selectmenu': 'jquery/ui-modules/widgets/selectmenu',
'jquery-ui-modules/slider': 'jquery/ui-modules/widgets/slider',
### Actual result (*)
<!--- Tell us what happens instead -->
This causes an error on frontend where jQuery slider is used and shows an error saying "$j(..).slider is not a function".
<img width="1222" alt="Screenshot 2022-04-19 at 4 38 23 PM" src="https://user-images.githubusercontent.com/22843013/163991117-2a3bcc65-4215-4c43-81ea-991cd8d7e21e.png">
|
non_process
|
jquery ui slider and selectmenu mapping is not correct in requirejs config js thank you for contributing to magento to help us process this issue we recommend that you add the following information summary of the issue information on your environment steps to reproduce expected and actual results fields marked with are required please don t remove the template please also have a look at our guidelines article before adding a new issue preconditions please provide as detailed information about your environment as possible for example magento version tag head php mysql version etc on magento release jquery ui slider and selectmenu mapping is not correct in vendor magento module theme view base requirejs config js this causes an error on frontend as jquery slider and selectmenu modules do not get loaded and we see an error saying j slider is not a function steps to reproduce it is important to provide a set of clear steps to reproduce this bug if relevant please include code samples open vendor magento module theme view base requirejs config js on line you will notice the incorrect mapping jquery ui modules slider jquery ui modules widgets selectmenu expected result the correct mapping should be jquery ui modules selectmenu jquery ui modules widgets selectmenu jquery ui modules slider jquery ui modules widgets slider actual result this causes an error on frontend where jquery slider is used and shows an error saying j slider is not a function img width alt screenshot at pm src
| 0
|
2,055
| 2,603,975,613
|
IssuesEvent
|
2015-02-24 19:01:25
|
chrsmith/nishazi6
|
https://api.github.com/repos/chrsmith/nishazi6
|
opened
|
沈阳治疗湿尤需要多长时间
|
auto-migrated Priority-Medium Type-Defect
|
```
沈阳治疗湿尤需要多长时间〓沈陽軍區政治部醫院性病〓TEL��
�024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治�
��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共�
��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、
保健、醫療、科研康復為一體的綜合性醫院。是國家首批公��
�甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大�
��、東南大學等知名高等院校的教學醫院。曾被中國人民解放
軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立��
�體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:17
|
1.0
|
沈阳治疗湿尤需要多长时间 - ```
沈阳治疗湿尤需要多长时间〓沈陽軍區政治部醫院性病〓TEL��
�024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治�
��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共�
��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、
保健、醫療、科研康復為一體的綜合性醫院。是國家首批公��
�甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大�
��、東南大學等知名高等院校的教學醫院。曾被中國人民解放
軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立��
�體二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:17
|
non_process
|
沈阳治疗湿尤需要多长时间 沈阳治疗湿尤需要多长时间〓沈陽軍區政治部醫院性病〓tel�� � 〓 , � ��。 。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 original issue reported on code google com by gmail com on jun at
| 0
|
52,893
| 10,959,744,161
|
IssuesEvent
|
2019-11-27 12:06:05
|
Regalis11/Barotrauma
|
https://api.github.com/repos/Regalis11/Barotrauma
|
closed
|
[0.9.5.1] [Issue] [Graphics] Volume Icons on crewlist don't indicate volume.
|
Bug Code
|
Right now the new voice chat VOIP volume icon is a dot that extends into a bunch of curved lines to indicate volume. While it properly works in the lobby, only the dot shows up in the ingame crew list in the top left. Someone probably just forgot to port over the new system.
|
1.0
|
[0.9.5.1] [Issue] [Graphics] Volume Icons on crewlist don't indicate volume. - Right now the new voice chat VOIP volume icon is a dot that extends into a bunch of curved lines to indicate volume. While it properly works in the lobby, only the dot shows up in the ingame crew list in the top left. Someone probably just forgot to port over the new system.
|
non_process
|
volume icons on crewlist don t indicate volume right now the new voice chat voip volume icon is a dot that extends into a bunch of curved lines to indicate volume while it properly works in the lobby only the dot shows up in the ingame crew list in the top left someone probably just forgot to port over the new system
| 0
|
20,672
| 27,335,245,132
|
IssuesEvent
|
2023-02-26 05:18:32
|
python/cpython
|
https://api.github.com/repos/python/cpython
|
closed
|
Process and Thread resource recycling issue
|
type-bug expert-multiprocessing
|
# Summary
<!--
如果一个子进程为了提供服务而创建,子进程的父进程如果创建一个线程去轮询对应服务,则当前的进程默认退出机制会先释放子进程,导致线程轮询服务的线程raise Error
-->
For a child process with daemon=False creates the multiprocessing.Manager object, if it creates a thread for polling SyncManager-created Lock/Event objects, the default recycling mechanism will first release the grandchild process corresponding to the created multiprocessing.Manager object, which can interrupt the polling thread and raise Error.
# Description
<!--
测试代码如下:
-->
The test code is as follows:
```python
import time
import queue
import threading
import multiprocessing
class GlobalMgr(threading.Thread):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
self.mgr = multiprocessing.Manager()
self.task_q = self.mgr.Queue()
self.stop_event = threading.Event()
def run(self):
while not self.stop_event.is_set():
try:
self.task_q.get(timeout=0.001)
except queue.Empty as e:
continue
def subprocess():
mgr = GlobalMgr(daemon=True)
mgr.start()
time.sleep(1)
if __name__ == "__main__":
process = multiprocessing.Process(target=subprocess, daemon=False)
process.start()
process.join()
```
<!--
上述测试代码会抛出如下异常
-->
The above test code will throw the following exception:
```python
Traceback (most recent call last):
File "/usr/lib/python3.10/threading.py", line 1009, in _bootstrap_inner
multiprocessing/process.py 318 source exit
self.run()
File "/home/hanjiangtao/workspace/program_learning/python_workspace/mp_resource_manager.py", line 18, in run
self.task_q.get(timeout=0.001)
File "<string>", line 2, in get
File "/usr/lib/python3.10/multiprocessing/managers.py", line 833, in _callmethod
raise convert_to_error(kind, result)
multiprocessing.managers.RemoteError:
---------------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python3.10/multiprocessing/managers.py", line 260, in serve_client
self.id_to_local_proxy_obj[ident]
KeyError: '7fcff61b4d00'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/lib/python3.10/multiprocessing/managers.py", line 262, in serve_client
raise ke
File "/usr/lib/python3.10/multiprocessing/managers.py", line 256, in serve_client
obj, exposed, gettypeid = id_to_obj[ident]
KeyError: '7fcff61b4d00'
---------------------------------------------------------------------------
```
<!--
经过debug,发现上述异常抛出的根因是process子进程创建并start后,BaseProcess._bootstrap被调用,在程序资源释放也就是process对象退出时,走了BaseProcess._bootstrap中finally分支,导致提前调用util._exit_function,而util._exit_function方法会将BaseManger.start方法中注册的BaseManager._finalize_manager提前调用,导致thread还在运行时,SyncManager对象已经被释放,并抛出异常
-->
After debugging, it was found that the main cause of the above exception was the invocation of BaseProcess._bootstrap after the subprocess was created and started.While releasing the program resources, that is, when the process object try to exit, the finally brach in BaseProcess._bootstrap will be taken. Which leds to calling util._exit_function in advance. The util._exit_function method will futher call the BaseManager._finalize_manager registered in the BaseManger.start method in advance, and this will trigger the exception during the runtime of thread for the accidental release of SyncManager object.</br>
<!--
但是,multiprocessing.util模块在被import的时候,已经将util._exit_function注册至atexit模块中,所以本身在子进程退出的时候,util._exit_function已经会被被动调用,那么我们是否可以认为在BaseProcess._bootstrap的finally分支的主动调用是非必需的?
-->
However, the multiprocessing.util module registers util._exit_function in the atexit module automatically when it is imported, which means that util._exit_function will be called naturally while subprocess exits. Consequently, can we consider the active invocation of BaseProcess._bootstrap's finally branch as unnecessary?</br>
<!--
当然,通过学习multiprocessing.util模块的代码,我们可以通过如下方案进行规避:
-->
Of course, after learning the multiprocessing.util module, we can avoid the issue by the following:</br>
```python
import time
import queue
import threading
import multiprocessing
class GlobalMgr(threading.Thread):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
self.mgr = multiprocessing.Manager()
self.task_q = self.mgr.Queue()
self.stop_event = threading.Event()
def run(self):
while not self.stop_event.is_set():
try:
self.task_q.get(timeout=0.001)
except queue.Empty as e:
continue
def release(mgr):
mgr.stop_event.set()
mgr.join()
def subprocess():
mgr = GlobalMgr(daemon=True)
mgr.start()
# Actively registers destructors for high-priority _exit_function calls
multiprocessing.util.Finalize(None, release, args=(mgr,), exitpriority=100)
time.sleep(1)
if __name__ == "__main__":
process = multiprocessing.Process(target=subprocess, daemon=False)
process.start()
process.join()
```
<!--
但是multiprocessing.util.Finalize方法不是模块外部可见方法,所以我认为这只算是个规避的方法。
-->
The multiprocessing.util.Finalize module is not an externally visible method, so it's only a hedge.'
# Environment
> Python 3.10.4</br>
> Ubuntu x86_64</br>
|
1.0
|
Process and Thread resource recycling issue - # Summary
<!--
如果一个子进程为了提供服务而创建,子进程的父进程如果创建一个线程去轮询对应服务,则当前的进程默认退出机制会先释放子进程,导致线程轮询服务的线程raise Error
-->
For a child process with daemon=False creates the multiprocessing.Manager object, if it creates a thread for polling SyncManager-created Lock/Event objects, the default recycling mechanism will first release the grandchild process corresponding to the created multiprocessing.Manager object, which can interrupt the polling thread and raise Error.
# Description
<!--
测试代码如下:
-->
The test code is as follows:
```python
import time
import queue
import threading
import multiprocessing
class GlobalMgr(threading.Thread):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
self.mgr = multiprocessing.Manager()
self.task_q = self.mgr.Queue()
self.stop_event = threading.Event()
def run(self):
while not self.stop_event.is_set():
try:
self.task_q.get(timeout=0.001)
except queue.Empty as e:
continue
def subprocess():
mgr = GlobalMgr(daemon=True)
mgr.start()
time.sleep(1)
if __name__ == "__main__":
process = multiprocessing.Process(target=subprocess, daemon=False)
process.start()
process.join()
```
<!--
上述测试代码会抛出如下异常
-->
The above test code will throw the following exception:
```python
Traceback (most recent call last):
File "/usr/lib/python3.10/threading.py", line 1009, in _bootstrap_inner
multiprocessing/process.py 318 source exit
self.run()
File "/home/hanjiangtao/workspace/program_learning/python_workspace/mp_resource_manager.py", line 18, in run
self.task_q.get(timeout=0.001)
File "<string>", line 2, in get
File "/usr/lib/python3.10/multiprocessing/managers.py", line 833, in _callmethod
raise convert_to_error(kind, result)
multiprocessing.managers.RemoteError:
---------------------------------------------------------------------------
Traceback (most recent call last):
File "/usr/lib/python3.10/multiprocessing/managers.py", line 260, in serve_client
self.id_to_local_proxy_obj[ident]
KeyError: '7fcff61b4d00'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/lib/python3.10/multiprocessing/managers.py", line 262, in serve_client
raise ke
File "/usr/lib/python3.10/multiprocessing/managers.py", line 256, in serve_client
obj, exposed, gettypeid = id_to_obj[ident]
KeyError: '7fcff61b4d00'
---------------------------------------------------------------------------
```
<!--
经过debug,发现上述异常抛出的根因是process子进程创建并start后,BaseProcess._bootstrap被调用,在程序资源释放也就是process对象退出时,走了BaseProcess._bootstrap中finally分支,导致提前调用util._exit_function,而util._exit_function方法会将BaseManger.start方法中注册的BaseManager._finalize_manager提前调用,导致thread还在运行时,SyncManager对象已经被释放,并抛出异常
-->
After debugging, it was found that the main cause of the above exception was the invocation of BaseProcess._bootstrap after the subprocess was created and started.While releasing the program resources, that is, when the process object try to exit, the finally brach in BaseProcess._bootstrap will be taken. Which leds to calling util._exit_function in advance. The util._exit_function method will futher call the BaseManager._finalize_manager registered in the BaseManger.start method in advance, and this will trigger the exception during the runtime of thread for the accidental release of SyncManager object.</br>
<!--
但是,multiprocessing.util模块在被import的时候,已经将util._exit_function注册至atexit模块中,所以本身在子进程退出的时候,util._exit_function已经会被被动调用,那么我们是否可以认为在BaseProcess._bootstrap的finally分支的主动调用是非必需的?
-->
However, the multiprocessing.util module registers util._exit_function in the atexit module automatically when it is imported, which means that util._exit_function will be called naturally while subprocess exits. Consequently, can we consider the active invocation of BaseProcess._bootstrap's finally branch as unnecessary?</br>
<!--
当然,通过学习multiprocessing.util模块的代码,我们可以通过如下方案进行规避:
-->
Of course, after learning the multiprocessing.util module, we can avoid the issue by the following:</br>
```python
import time
import queue
import threading
import multiprocessing
class GlobalMgr(threading.Thread):
def __init__(self, *args, **kwargs):
super().__init__(*args, **kwargs)
self.mgr = multiprocessing.Manager()
self.task_q = self.mgr.Queue()
self.stop_event = threading.Event()
def run(self):
while not self.stop_event.is_set():
try:
self.task_q.get(timeout=0.001)
except queue.Empty as e:
continue
def release(mgr):
mgr.stop_event.set()
mgr.join()
def subprocess():
mgr = GlobalMgr(daemon=True)
mgr.start()
# Actively registers destructors for high-priority _exit_function calls
multiprocessing.util.Finalize(None, release, args=(mgr,), exitpriority=100)
time.sleep(1)
if __name__ == "__main__":
process = multiprocessing.Process(target=subprocess, daemon=False)
process.start()
process.join()
```
<!--
但是multiprocessing.util.Finalize方法不是模块外部可见方法,所以我认为这只算是个规避的方法。
-->
The multiprocessing.util.Finalize module is not an externally visible method, so it's only a hedge.'
# Environment
> Python 3.10.4</br>
> Ubuntu x86_64</br>
|
process
|
process and thread resource recycling issue summary 如果一个子进程为了提供服务而创建,子进程的父进程如果创建一个线程去轮询对应服务,则当前的进程默认退出机制会先释放子进程,导致线程轮询服务的线程raise error for a child process with daemon false creates the multiprocessing manager object if it creates a thread for polling syncmanager created lock event objects the default recycling mechanism will first release the grandchild process corresponding to the created multiprocessing manager object which can interrupt the polling thread and raise error description 测试代码如下 the test code is as follows python import time import queue import threading import multiprocessing class globalmgr threading thread def init self args kwargs super init args kwargs self mgr multiprocessing manager self task q self mgr queue self stop event threading event def run self while not self stop event is set try self task q get timeout except queue empty as e continue def subprocess mgr globalmgr daemon true mgr start time sleep if name main process multiprocessing process target subprocess daemon false process start process join 上述测试代码会抛出如下异常 the above test code will throw the following exception python traceback most recent call last file usr lib threading py line in bootstrap inner multiprocessing process py source exit self run file home hanjiangtao workspace program learning python workspace mp resource manager py line in run self task q get timeout file line in get file usr lib multiprocessing managers py line in callmethod raise convert to error kind result multiprocessing managers remoteerror traceback most recent call last file usr lib multiprocessing managers py line in serve client self id to local proxy obj keyerror during handling of the above exception another exception occurred traceback most recent call last file usr lib multiprocessing managers py line in serve client raise ke file usr lib multiprocessing managers py line in serve client obj exposed gettypeid id to obj keyerror 经过debug,发现上述异常抛出的根因是process子进程创建并start后 baseprocess bootstrap被调用,在程序资源释放也就是process对象退出时,走了baseprocess bootstrap中finally分支,导致提前调用util exit function,而util exit function方法会将basemanger start方法中注册的basemanager finalize manager提前调用,导致thread还在运行时,syncmanager对象已经被释放,并抛出异常 after debugging it was found that the main cause of the above exception was the invocation of baseprocess bootstrap after the subprocess was created and started while releasing the program resources that is when the process object try to exit the finally brach in baseprocess bootstrap will be taken which leds to calling util exit function in advance the util exit function method will futher call the basemanager finalize manager registered in the basemanger start method in advance and this will trigger the exception during the runtime of thread for the accidental release of syncmanager object 但是,multiprocessing util模块在被import的时候,已经将util exit function注册至atexit模块中,所以本身在子进程退出的时候,util exit function已经会被被动调用,那么我们是否可以认为在baseprocess bootstrap的finally分支的主动调用是非必需的? however the multiprocessing util module registers util exit function in the atexit module automatically when it is imported which means that util exit function will be called naturally while subprocess exits consequently can we consider the active invocation of baseprocess bootstrap s finally branch as unnecessary 当然,通过学习multiprocessing util模块的代码,我们可以通过如下方案进行规避: of course after learning the multiprocessing util module we can avoid the issue by the following python import time import queue import threading import multiprocessing class globalmgr threading thread def init self args kwargs super init args kwargs self mgr multiprocessing manager self task q self mgr queue self stop event threading event def run self while not self stop event is set try self task q get timeout except queue empty as e continue def release mgr mgr stop event set mgr join def subprocess mgr globalmgr daemon true mgr start actively registers destructors for high priority exit function calls multiprocessing util finalize none release args mgr exitpriority time sleep if name main process multiprocessing process target subprocess daemon false process start process join 但是multiprocessing util finalize方法不是模块外部可见方法,所以我认为这只算是个规避的方法。 the multiprocessing util finalize module is not an externally visible method so it s only a hedge environment python ubuntu
| 1
|
7,039
| 10,197,375,067
|
IssuesEvent
|
2019-08-13 00:04:22
|
Ultimate-Hosts-Blacklist/whitelist
|
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
|
closed
|
False positive
|
whitelisting process
|
*@macuser666 commented on Aug 12, 2019, 12:33 PM UTC:*
hi
the following site is blocked
[www.bcee.lu](http://www.bcee.lu)
bcee.lu
It's a legimate bank website of Banque et Caisse d'Épargne de l'État.
*This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Ultimate.Hosts.Blacklist#519](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/519).*
|
1.0
|
False positive - *@macuser666 commented on Aug 12, 2019, 12:33 PM UTC:*
hi
the following site is blocked
[www.bcee.lu](http://www.bcee.lu)
bcee.lu
It's a legimate bank website of Banque et Caisse d'Épargne de l'État.
*This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Ultimate.Hosts.Blacklist#519](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/519).*
|
process
|
false positive commented on aug pm utc hi the following site is blocked bcee lu it s a legimate bank website of banque et caisse d épargne de l état this issue was moved by from
| 1
|
197,833
| 14,944,836,392
|
IssuesEvent
|
2021-01-26 02:28:28
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
test_cholesky_solve_batched_many_batches_cuda_complex128 has cuda illegal memory access
|
high priority module: crash module: linear algebra module: tests triaged
|
## 🐛 Bug
test_cholesky_solve_batched_many_batches_cuda_complex128 has cuda illegal memory access. https://github.com/pytorch/pytorch/pull/47047 might be related.
## To Reproduce
Steps to reproduce the behavior:
```
$ PYTORCH_TEST_WITH_SLOW=1 python test/test_linalg.py -v -k test_cholesky_solve_batched_many_batches_cuda_complex128
test_cholesky_solve_batched_many_batches_cuda_complex128 (__main__.TestLinalgCUDA) ... CUDA runtime error: an illegal memory access was encountered (700) in magma_zpotrf_batched at /home/xwang/Developer/magma-2.5.3/src/zpotrf_batched.cpp:234
CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:945
CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:946
CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:947
ERROR
======================================================================
ERROR: test_cholesky_solve_batched_many_batches_cuda_complex128 (__main__.TestLinalgCUDA)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_utils.py", line 864, in wrapper
method(*args, **kwargs)
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 273, in instantiated_test
result = test_fn(self, *args)
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_utils.py", line 542, in wrapper
fn(*args, **kwargs)
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 545, in dep_fn
return fn(slf, device, *args, **kwargs)
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 545, in dep_fn
return fn(slf, device, *args, **kwargs)
File "test/test_linalg.py", line 1906, in test_cholesky_solve_batched_many_batches
b, A, L = self.cholesky_solve_test_helper(A_dims, b_dims, upper, device, dtype)
File "test/test_linalg.py", line 1845, in cholesky_solve_test_helper
L = torch.cholesky(A, upper=upper)
RuntimeError: CUDA error: an illegal memory access was encountered
Exception raised from magmaCholeskyBatched<c10::complex<double> > at /home/xwang/Developer/pytorch/aten/src/ATen/native/cuda/BatchLinearAlgebra.cu:653 (most recent call first):
frame #0: c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >) + 0x68 (0x7efde8c50828 in /home/xwang/Developer/pytorch/torch/lib/libc10.so)
frame #1: void at::native::magmaCholeskyBatched<c10::complex<double> >(magma_uplo_t, int, c10::complex<double>**, int, int*, int, at::native::MAGMAQueue const&) + 0x155 (0x7efde9fffbd5 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so)
frame #2: at::native::_cholesky_helper_cuda(at::Tensor const&, bool) + 0x1709 (0x7efdea017d19 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so)
frame #3: <unknown function> + 0x3336c13 (0x7efdebfdcc13 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so)
frame #4: <unknown function> + 0x3336c84 (0x7efdebfdcc84 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so)
frame #5: at::_cholesky_helper(at::Tensor const&, bool) + 0x116 (0x7efdfde476a6 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #6: at::native::cholesky(at::Tensor const&, bool) + 0xa9 (0x7efdfd8b2c99 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #7: <unknown function> + 0x1c388f3 (0x7efdfe0398f3 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #8: <unknown function> + 0x1c38964 (0x7efdfe039964 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #9: at::cholesky(at::Tensor const&, bool) + 0x116 (0x7efdfde46bc6 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #10: <unknown function> + 0x31ab9c4 (0x7efdff5ac9c4 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #11: <unknown function> + 0x31abcb4 (0x7efdff5accb4 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #12: at::Tensor::cholesky(bool) const + 0x116 (0x7efdfe1a7696 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #13: <unknown function> + 0x734e6d (0x7efe04b8de6d in /home/xwang/Developer/pytorch/torch/lib/libtorch_python.so)
<omitting python frames>
----------------------------------------------------------------------
Ran 1 test in 2.616s
FAILED (errors=1)
```
## Expected behavior
No fail
## Environment
```
Collecting environment information...
PyTorch version: 1.8.0a0+533c837
Is debug build: False
CUDA used to build PyTorch: 11.1
ROCM used to build PyTorch: N/A
OS: Manjaro Linux (x86_64)
GCC version: (GCC) 10.2.0
Clang version: Could not collect
CMake version: version 3.18.4
Python version: 3.8 (64-bit runtime)
Is CUDA available: True
CUDA runtime version: Could not collect
GPU models and configuration:
GPU 0: GeForce RTX 2070 SUPER
GPU 1: GeForce GTX 1070 Ti
Nvidia driver version: 455.38
cuDNN version: Probably one of the following:
/usr/lib/libcudnn.so.8.0.5
/usr/lib/libcudnn_adv_infer.so.8.0.5
/usr/lib/libcudnn_adv_train.so.8.0.5
/usr/lib/libcudnn_cnn_infer.so.8.0.5
/usr/lib/libcudnn_cnn_train.so.8.0.5
/usr/lib/libcudnn_ops_infer.so.8.0.5
/usr/lib/libcudnn_ops_train.so.8.0.5
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.19.4
[pip3] pytorch-ignite==0.4.2
[pip3] torch==1.8.0a0
[pip3] torchvision==0.9.0a0+74de51d
[conda] Could not collect
```
## Additional context
Seems to be MAGMA-related.
CC @ptrblck @mruberry
cc @ezyang @gchanan @zou3519 @bdhirsh @jianyuh @nikitaved @pearu @mruberry @heitorschueroff @walterddr @VitalyFedyunin @IvanYashchuk
|
1.0
|
test_cholesky_solve_batched_many_batches_cuda_complex128 has cuda illegal memory access - ## 🐛 Bug
test_cholesky_solve_batched_many_batches_cuda_complex128 has cuda illegal memory access. https://github.com/pytorch/pytorch/pull/47047 might be related.
## To Reproduce
Steps to reproduce the behavior:
```
$ PYTORCH_TEST_WITH_SLOW=1 python test/test_linalg.py -v -k test_cholesky_solve_batched_many_batches_cuda_complex128
test_cholesky_solve_batched_many_batches_cuda_complex128 (__main__.TestLinalgCUDA) ... CUDA runtime error: an illegal memory access was encountered (700) in magma_zpotrf_batched at /home/xwang/Developer/magma-2.5.3/src/zpotrf_batched.cpp:234
CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:945
CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:946
CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:947
ERROR
======================================================================
ERROR: test_cholesky_solve_batched_many_batches_cuda_complex128 (__main__.TestLinalgCUDA)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_utils.py", line 864, in wrapper
method(*args, **kwargs)
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 273, in instantiated_test
result = test_fn(self, *args)
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_utils.py", line 542, in wrapper
fn(*args, **kwargs)
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 545, in dep_fn
return fn(slf, device, *args, **kwargs)
File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 545, in dep_fn
return fn(slf, device, *args, **kwargs)
File "test/test_linalg.py", line 1906, in test_cholesky_solve_batched_many_batches
b, A, L = self.cholesky_solve_test_helper(A_dims, b_dims, upper, device, dtype)
File "test/test_linalg.py", line 1845, in cholesky_solve_test_helper
L = torch.cholesky(A, upper=upper)
RuntimeError: CUDA error: an illegal memory access was encountered
Exception raised from magmaCholeskyBatched<c10::complex<double> > at /home/xwang/Developer/pytorch/aten/src/ATen/native/cuda/BatchLinearAlgebra.cu:653 (most recent call first):
frame #0: c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >) + 0x68 (0x7efde8c50828 in /home/xwang/Developer/pytorch/torch/lib/libc10.so)
frame #1: void at::native::magmaCholeskyBatched<c10::complex<double> >(magma_uplo_t, int, c10::complex<double>**, int, int*, int, at::native::MAGMAQueue const&) + 0x155 (0x7efde9fffbd5 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so)
frame #2: at::native::_cholesky_helper_cuda(at::Tensor const&, bool) + 0x1709 (0x7efdea017d19 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so)
frame #3: <unknown function> + 0x3336c13 (0x7efdebfdcc13 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so)
frame #4: <unknown function> + 0x3336c84 (0x7efdebfdcc84 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so)
frame #5: at::_cholesky_helper(at::Tensor const&, bool) + 0x116 (0x7efdfde476a6 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #6: at::native::cholesky(at::Tensor const&, bool) + 0xa9 (0x7efdfd8b2c99 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #7: <unknown function> + 0x1c388f3 (0x7efdfe0398f3 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #8: <unknown function> + 0x1c38964 (0x7efdfe039964 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #9: at::cholesky(at::Tensor const&, bool) + 0x116 (0x7efdfde46bc6 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #10: <unknown function> + 0x31ab9c4 (0x7efdff5ac9c4 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #11: <unknown function> + 0x31abcb4 (0x7efdff5accb4 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #12: at::Tensor::cholesky(bool) const + 0x116 (0x7efdfe1a7696 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so)
frame #13: <unknown function> + 0x734e6d (0x7efe04b8de6d in /home/xwang/Developer/pytorch/torch/lib/libtorch_python.so)
<omitting python frames>
----------------------------------------------------------------------
Ran 1 test in 2.616s
FAILED (errors=1)
```
## Expected behavior
No fail
## Environment
```
Collecting environment information...
PyTorch version: 1.8.0a0+533c837
Is debug build: False
CUDA used to build PyTorch: 11.1
ROCM used to build PyTorch: N/A
OS: Manjaro Linux (x86_64)
GCC version: (GCC) 10.2.0
Clang version: Could not collect
CMake version: version 3.18.4
Python version: 3.8 (64-bit runtime)
Is CUDA available: True
CUDA runtime version: Could not collect
GPU models and configuration:
GPU 0: GeForce RTX 2070 SUPER
GPU 1: GeForce GTX 1070 Ti
Nvidia driver version: 455.38
cuDNN version: Probably one of the following:
/usr/lib/libcudnn.so.8.0.5
/usr/lib/libcudnn_adv_infer.so.8.0.5
/usr/lib/libcudnn_adv_train.so.8.0.5
/usr/lib/libcudnn_cnn_infer.so.8.0.5
/usr/lib/libcudnn_cnn_train.so.8.0.5
/usr/lib/libcudnn_ops_infer.so.8.0.5
/usr/lib/libcudnn_ops_train.so.8.0.5
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.19.4
[pip3] pytorch-ignite==0.4.2
[pip3] torch==1.8.0a0
[pip3] torchvision==0.9.0a0+74de51d
[conda] Could not collect
```
## Additional context
Seems to be MAGMA-related.
CC @ptrblck @mruberry
cc @ezyang @gchanan @zou3519 @bdhirsh @jianyuh @nikitaved @pearu @mruberry @heitorschueroff @walterddr @VitalyFedyunin @IvanYashchuk
|
non_process
|
test cholesky solve batched many batches cuda has cuda illegal memory access 🐛 bug test cholesky solve batched many batches cuda has cuda illegal memory access might be related to reproduce steps to reproduce the behavior pytorch test with slow python test test linalg py v k test cholesky solve batched many batches cuda test cholesky solve batched many batches cuda main testlinalgcuda cuda runtime error an illegal memory access was encountered in magma zpotrf batched at home xwang developer magma src zpotrf batched cpp cuda runtime error an illegal memory access was encountered in magma queue destroy internal at home xwang developer magma interface cuda interface cpp cuda runtime error an illegal memory access was encountered in magma queue destroy internal at home xwang developer magma interface cuda interface cpp cuda runtime error an illegal memory access was encountered in magma queue destroy internal at home xwang developer magma interface cuda interface cpp error error test cholesky solve batched many batches cuda main testlinalgcuda traceback most recent call last file home xwang developer pytorch torch testing internal common utils py line in wrapper method args kwargs file home xwang developer pytorch torch testing internal common device type py line in instantiated test result test fn self args file home xwang developer pytorch torch testing internal common utils py line in wrapper fn args kwargs file home xwang developer pytorch torch testing internal common device type py line in dep fn return fn slf device args kwargs file home xwang developer pytorch torch testing internal common device type py line in dep fn return fn slf device args kwargs file test test linalg py line in test cholesky solve batched many batches b a l self cholesky solve test helper a dims b dims upper device dtype file test test linalg py line in cholesky solve test helper l torch cholesky a upper upper runtimeerror cuda error an illegal memory access was encountered exception raised from magmacholeskybatched at home xwang developer pytorch aten src aten native cuda batchlinearalgebra cu most recent call first frame error error sourcelocation std basic string std allocator in home xwang developer pytorch torch lib so frame void at native magmacholeskybatched magma uplo t int complex int int int at native magmaqueue const in home xwang developer pytorch torch lib libtorch cuda so frame at native cholesky helper cuda at tensor const bool in home xwang developer pytorch torch lib libtorch cuda so frame in home xwang developer pytorch torch lib libtorch cuda so frame in home xwang developer pytorch torch lib libtorch cuda so frame at cholesky helper at tensor const bool in home xwang developer pytorch torch lib libtorch cpu so frame at native cholesky at tensor const bool in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch cpu so frame at cholesky at tensor const bool in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch cpu so frame at tensor cholesky bool const in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch python so ran test in failed errors expected behavior no fail environment collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os manjaro linux gcc version gcc clang version could not collect cmake version version python version bit runtime is cuda available true cuda runtime version could not collect gpu models and configuration gpu geforce rtx super gpu geforce gtx ti nvidia driver version cudnn version probably one of the following usr lib libcudnn so usr lib libcudnn adv infer so usr lib libcudnn adv train so usr lib libcudnn cnn infer so usr lib libcudnn cnn train so usr lib libcudnn ops infer so usr lib libcudnn ops train so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy pytorch ignite torch torchvision could not collect additional context seems to be magma related cc ptrblck mruberry cc ezyang gchanan bdhirsh jianyuh nikitaved pearu mruberry heitorschueroff walterddr vitalyfedyunin ivanyashchuk
| 0
|
163,811
| 12,745,623,589
|
IssuesEvent
|
2020-06-26 14:33:47
|
eclipse/openj9
|
https://api.github.com/repos/eclipse/openj9
|
closed
|
samplingObjectAllocation.soae001 failed, expected 1+ but got: 0
|
comp:gc test failure
|
https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/3
https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_x86-64_windows_Nightly_testList_0/10
cmdLineTester_jvmtitests_soae_3
```
Testing: soae001
Test start time: 2020/06/05 04:04:16 Central Standard Time
Running command: "C:/Users/jenkins/workspace/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/openjdkbinary/j2sdk-image\\bin\\java" -Xcompressedrefs -Xjit -Xgcpolicy:gencon -Xdump -agentlib:jvmtitest=test:soae001 -cp "C:/Users/jenkins/workspace/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/openjdk-tests///..//jvmtest\functional\cmdLineTests\jvmtitests\jvmtitest.jar" com.ibm.jvmti.tests.util.TestRunner
Time spent starting: 15 milliseconds
Time spent executing: 781 milliseconds
Test result: FAILED
Output from test:
[OUT] *** Testing [1/1]: testDefaultInterval
[OUT] Allocated a byte array with size 524288
[OUT] com.ibm.jvmti.tests.samplingObjectAllocation.soae001.check() failed, expected 1+ but got: 0
[OUT] *** Test took 15 milliseconds
[OUT] FAILED
[OUT]
>> Success condition was not found: [Return code: 0]
```
https://github.com/eclipse/openj9/compare/589eaa5...5f0390e
I expect this is caused by https://github.com/eclipse/openj9/pull/9745
|
1.0
|
samplingObjectAllocation.soae001 failed, expected 1+ but got: 0 - https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/3
https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_x86-64_windows_Nightly_testList_0/10
cmdLineTester_jvmtitests_soae_3
```
Testing: soae001
Test start time: 2020/06/05 04:04:16 Central Standard Time
Running command: "C:/Users/jenkins/workspace/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/openjdkbinary/j2sdk-image\\bin\\java" -Xcompressedrefs -Xjit -Xgcpolicy:gencon -Xdump -agentlib:jvmtitest=test:soae001 -cp "C:/Users/jenkins/workspace/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/openjdk-tests///..//jvmtest\functional\cmdLineTests\jvmtitests\jvmtitest.jar" com.ibm.jvmti.tests.util.TestRunner
Time spent starting: 15 milliseconds
Time spent executing: 781 milliseconds
Test result: FAILED
Output from test:
[OUT] *** Testing [1/1]: testDefaultInterval
[OUT] Allocated a byte array with size 524288
[OUT] com.ibm.jvmti.tests.samplingObjectAllocation.soae001.check() failed, expected 1+ but got: 0
[OUT] *** Test took 15 milliseconds
[OUT] FAILED
[OUT]
>> Success condition was not found: [Return code: 0]
```
https://github.com/eclipse/openj9/compare/589eaa5...5f0390e
I expect this is caused by https://github.com/eclipse/openj9/pull/9745
|
non_process
|
samplingobjectallocation failed expected but got cmdlinetester jvmtitests soae testing test start time central standard time running command c users jenkins workspace test sanity functional windows testlist openjdkbinary image bin java xcompressedrefs xjit xgcpolicy gencon xdump agentlib jvmtitest test cp c users jenkins workspace test sanity functional windows testlist openjdk tests jvmtest functional cmdlinetests jvmtitests jvmtitest jar com ibm jvmti tests util testrunner time spent starting milliseconds time spent executing milliseconds test result failed output from test testing testdefaultinterval allocated a byte array with size com ibm jvmti tests samplingobjectallocation check failed expected but got test took milliseconds failed success condition was not found i expect this is caused by
| 0
|
401,361
| 11,789,200,533
|
IssuesEvent
|
2020-03-17 16:45:42
|
aces/Loris
|
https://api.github.com/repos/aces/Loris
|
closed
|
[New Module] Electrophysiology browser
|
Feature Priority: High
|
List of bugs found in the current set up:
- Task Name seems to be hardcoded to FaceHousCheck. Fix that so it takes that information from the physiological_parameter_file table (TaskName is one of the JSON file field). In jsx/components/eeg_session_panels.js and js/eeg_session_view.js
|
1.0
|
[New Module] Electrophysiology browser - List of bugs found in the current set up:
- Task Name seems to be hardcoded to FaceHousCheck. Fix that so it takes that information from the physiological_parameter_file table (TaskName is one of the JSON file field). In jsx/components/eeg_session_panels.js and js/eeg_session_view.js
|
non_process
|
electrophysiology browser list of bugs found in the current set up task name seems to be hardcoded to facehouscheck fix that so it takes that information from the physiological parameter file table taskname is one of the json file field in jsx components eeg session panels js and js eeg session view js
| 0
|
17,102
| 22,622,771,145
|
IssuesEvent
|
2022-06-30 08:03:13
|
2i2c-org/infrastructure
|
https://api.github.com/repos/2i2c-org/infrastructure
|
closed
|
Define some first-line and second-line support processes
|
:label: team-process
|
### Background and proposal
There are often cases where our support process is under-documented. For example, a few questions that people weren't sure how to answer:
- How should we prioritize support requests?
- What should we do if a request is not immediately "closeable"? Or if it requires ongoing follow-up work?
- How can we communicate our inability to fulfill a request?
- What kind of communication should we use throughout the support process?
We should document some rough guidelines for these common questions, and also provide references to documentation about how to carry out first-line support.
### Implementation guide and constraints
Another way to approach this is to ask "what are some common support situations, and what should we do in each situation?" We can draw from our experiences thus far to agree on some team practices to follow.
### Updates and ongoing work
_Add items below as we learn more_
- [x] Do some research into support processes at other orgs (see refs below)
- [x] #1155
- [x] #1154
### Refs
- [Distributed communication and incident response write-up](https://queue.acm.org/detail.cfm?id=3380779)
- [Example support process from Sarah](https://github.com/2i2c-org/infrastructure/issues/1068#issuecomment-1063138772)
- [Example support process from Chris](https://github.com/2i2c-org/infrastructure/issues/1068#issuecomment-1063516429)
- [Wikimedia clinic duty](https://wikitech.wikimedia.org/wiki/SRE/Clinic_Duty#Responsibilities)
- We have a semi-related issue to this here: https://github.com/2i2c-org/infrastructure/issues/1118
|
1.0
|
Define some first-line and second-line support processes - ### Background and proposal
There are often cases where our support process is under-documented. For example, a few questions that people weren't sure how to answer:
- How should we prioritize support requests?
- What should we do if a request is not immediately "closeable"? Or if it requires ongoing follow-up work?
- How can we communicate our inability to fulfill a request?
- What kind of communication should we use throughout the support process?
We should document some rough guidelines for these common questions, and also provide references to documentation about how to carry out first-line support.
### Implementation guide and constraints
Another way to approach this is to ask "what are some common support situations, and what should we do in each situation?" We can draw from our experiences thus far to agree on some team practices to follow.
### Updates and ongoing work
_Add items below as we learn more_
- [x] Do some research into support processes at other orgs (see refs below)
- [x] #1155
- [x] #1154
### Refs
- [Distributed communication and incident response write-up](https://queue.acm.org/detail.cfm?id=3380779)
- [Example support process from Sarah](https://github.com/2i2c-org/infrastructure/issues/1068#issuecomment-1063138772)
- [Example support process from Chris](https://github.com/2i2c-org/infrastructure/issues/1068#issuecomment-1063516429)
- [Wikimedia clinic duty](https://wikitech.wikimedia.org/wiki/SRE/Clinic_Duty#Responsibilities)
- We have a semi-related issue to this here: https://github.com/2i2c-org/infrastructure/issues/1118
|
process
|
define some first line and second line support processes background and proposal there are often cases where our support process is under documented for example a few questions that people weren t sure how to answer how should we prioritize support requests what should we do if a request is not immediately closeable or if it requires ongoing follow up work how can we communicate our inability to fulfill a request what kind of communication should we use throughout the support process we should document some rough guidelines for these common questions and also provide references to documentation about how to carry out first line support implementation guide and constraints another way to approach this is to ask what are some common support situations and what should we do in each situation we can draw from our experiences thus far to agree on some team practices to follow updates and ongoing work add items below as we learn more do some research into support processes at other orgs see refs below refs we have a semi related issue to this here
| 1
|
136,184
| 18,722,443,415
|
IssuesEvent
|
2021-11-03 13:17:12
|
ioana-github-enterprise/simplebuild
|
https://api.github.com/repos/ioana-github-enterprise/simplebuild
|
opened
|
CVE-2019-16335 (High) detected in jackson-databind-2.7.9.jar
|
security vulnerability
|
## CVE-2019-16335 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: simplebuild/build.gradle</p>
<p>Path to vulnerable library: le/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9/a4c0b14c7dd85bdf4d25da074e90a10fa4b9b88b/jackson-databind-2.7.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ioana-github-enterprise/simplebuild/commit/c99f55cc0f24d3137d4aa52ce1de8552a9d07579">c99f55cc0f24d3137d4aa52ce1de8552a9d07579</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariDataSource. This is a different vulnerability than CVE-2019-14540.
<p>Publish Date: 2019-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16335>CVE-2019-16335</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/blob/master/release-notes/VERSION-2.x">https://github.com/FasterXML/jackson-databind/blob/master/release-notes/VERSION-2.x</a></p>
<p>Release Date: 2020-10-20</p>
<p>Fix Resolution: 2.9.10</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.9","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16335","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariDataSource. This is a different vulnerability than CVE-2019-14540.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16335","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-16335 (High) detected in jackson-databind-2.7.9.jar - ## CVE-2019-16335 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: simplebuild/build.gradle</p>
<p>Path to vulnerable library: le/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9/a4c0b14c7dd85bdf4d25da074e90a10fa4b9b88b/jackson-databind-2.7.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.7.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ioana-github-enterprise/simplebuild/commit/c99f55cc0f24d3137d4aa52ce1de8552a9d07579">c99f55cc0f24d3137d4aa52ce1de8552a9d07579</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariDataSource. This is a different vulnerability than CVE-2019-14540.
<p>Publish Date: 2019-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16335>CVE-2019-16335</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/blob/master/release-notes/VERSION-2.x">https://github.com/FasterXML/jackson-databind/blob/master/release-notes/VERSION-2.x</a></p>
<p>Release Date: 2020-10-20</p>
<p>Fix Resolution: 2.9.10</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.9","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16335","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariDataSource. This is a different vulnerability than CVE-2019-14540.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16335","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_process
|
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file simplebuild build gradle path to vulnerable library le caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikaridatasource this is a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikaridatasource this is a different vulnerability than cve vulnerabilityurl
| 0
|
14,433
| 17,483,498,373
|
IssuesEvent
|
2021-08-09 07:52:05
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
[API Proposal]: System.Diagnostics.Process.WaitForExit() should return the Process to allow chaining of calls
|
api-suggestion area-System.Diagnostics.Process untriaged
|
### Background and motivation
This would allow method calls to be chained and reduce the number of lines of code required for certain actions.
### API Proposal
Before
```C#
namespace System.Diagnostics
{
public class Process
{
public void WaitForExit();
}
}
```
After:
```C#
namespace System.Diagnostics
{
public class Process
{
public Process WaitForExit();
}
}
```
### API Usage
Before:
```C#
Process p = Process.Start(exeFilePath);
p.WaitForExit();
if (p.ExitCode == 1)
{
// Handle the error somehow
}
```
After:
```C#
if (Process.Start(exeFilePath).WaitForExit().ExitCode == 1)
{
// Handle the error somehow
}
```
### Risks
It currently returns void, so I'm not sure if this would cause any issues with backwards compatibility.
Since if someone is already using the above 'Before' code it would continue working without any changes.
|
1.0
|
[API Proposal]: System.Diagnostics.Process.WaitForExit() should return the Process to allow chaining of calls - ### Background and motivation
This would allow method calls to be chained and reduce the number of lines of code required for certain actions.
### API Proposal
Before
```C#
namespace System.Diagnostics
{
public class Process
{
public void WaitForExit();
}
}
```
After:
```C#
namespace System.Diagnostics
{
public class Process
{
public Process WaitForExit();
}
}
```
### API Usage
Before:
```C#
Process p = Process.Start(exeFilePath);
p.WaitForExit();
if (p.ExitCode == 1)
{
// Handle the error somehow
}
```
After:
```C#
if (Process.Start(exeFilePath).WaitForExit().ExitCode == 1)
{
// Handle the error somehow
}
```
### Risks
It currently returns void, so I'm not sure if this would cause any issues with backwards compatibility.
Since if someone is already using the above 'Before' code it would continue working without any changes.
|
process
|
system diagnostics process waitforexit should return the process to allow chaining of calls background and motivation this would allow method calls to be chained and reduce the number of lines of code required for certain actions api proposal before c namespace system diagnostics public class process public void waitforexit after c namespace system diagnostics public class process public process waitforexit api usage before c process p process start exefilepath p waitforexit if p exitcode handle the error somehow after c if process start exefilepath waitforexit exitcode handle the error somehow risks it currently returns void so i m not sure if this would cause any issues with backwards compatibility since if someone is already using the above before code it would continue working without any changes
| 1
|
99,941
| 16,470,788,375
|
IssuesEvent
|
2021-05-23 11:19:40
|
axe-api/axe-api
|
https://api.github.com/repos/axe-api/axe-api
|
closed
|
Relationship queries
|
enhancement security
|
We need to add `with` feature to the **Queries**.
## Examples
`api/users?with=posts{id|title|comments{content}}`
## Dependencies
We should update the documentations.
|
True
|
Relationship queries - We need to add `with` feature to the **Queries**.
## Examples
`api/users?with=posts{id|title|comments{content}}`
## Dependencies
We should update the documentations.
|
non_process
|
relationship queries we need to add with feature to the queries examples api users with posts id title comments content dependencies we should update the documentations
| 0
|
518,712
| 15,033,172,642
|
IssuesEvent
|
2021-02-02 11:07:54
|
YangCatalog/backend
|
https://api.github.com/repos/YangCatalog/backend
|
closed
|
Reading log from .gz files
|
Priority: Critical bug
|
since we are compressing log files on daily basis we need to be able to read them back programmatically as well
|
1.0
|
Reading log from .gz files - since we are compressing log files on daily basis we need to be able to read them back programmatically as well
|
non_process
|
reading log from gz files since we are compressing log files on daily basis we need to be able to read them back programmatically as well
| 0
|
41,853
| 10,683,695,202
|
IssuesEvent
|
2019-10-22 08:54:49
|
jOOQ/jOOQ
|
https://api.github.com/repos/jOOQ/jOOQ
|
closed
|
Regression in Firebird's SUBSTRING() implementation
|
C: DB: Firebird C: Functionality E: All Editions P: High R: Fixed T: Defect
|
We're currently generating bad SQL for `substring` in Firebird:
```sql
select
substring('abcde' from 1),
substring('abcde' from 1 for2),
substring('abcde' from 3),
substring('abcde' from 3 for2)
from RDB$DATABASE
```
The regression was introduced by this, or a related change:
https://github.com/jOOQ/jOOQ/commit/b2d1119e24f33984c17af649ead532848f5c11ac
|
1.0
|
Regression in Firebird's SUBSTRING() implementation - We're currently generating bad SQL for `substring` in Firebird:
```sql
select
substring('abcde' from 1),
substring('abcde' from 1 for2),
substring('abcde' from 3),
substring('abcde' from 3 for2)
from RDB$DATABASE
```
The regression was introduced by this, or a related change:
https://github.com/jOOQ/jOOQ/commit/b2d1119e24f33984c17af649ead532848f5c11ac
|
non_process
|
regression in firebird s substring implementation we re currently generating bad sql for substring in firebird sql select substring abcde from substring abcde from substring abcde from substring abcde from from rdb database the regression was introduced by this or a related change
| 0
|
760,717
| 26,653,997,110
|
IssuesEvent
|
2023-01-25 15:34:47
|
KingGizzard/Ballotbox
|
https://api.github.com/repos/KingGizzard/Ballotbox
|
closed
|
Build Agent 3
|
high priority
|
Agent 3 is the data request agent!
You can see the dynamics of agent 3 in the [figma diagram here](https://www.figma.com/file/SkUdZjkenjeh5WQvJIe9vz/BallotBox?node-id=0%3A1&t=Gy4phXY2zIQ3Scjy-0)
Agent 3 should be able to:
- Request data from the Ballotbox.sol contract, and provide an email to be contacted through
- Access IPFS data and save data locally
! - when the encryption methods for the [Oracle](https://github.com/KingGizzard/Ballotbox/issues/1) and [Agent 2](https://github.com/KingGizzard/Ballotbox/issues/4) are finished, then Agent 3 should be able to:
- Decrypt IPFS data
|
1.0
|
Build Agent 3 - Agent 3 is the data request agent!
You can see the dynamics of agent 3 in the [figma diagram here](https://www.figma.com/file/SkUdZjkenjeh5WQvJIe9vz/BallotBox?node-id=0%3A1&t=Gy4phXY2zIQ3Scjy-0)
Agent 3 should be able to:
- Request data from the Ballotbox.sol contract, and provide an email to be contacted through
- Access IPFS data and save data locally
! - when the encryption methods for the [Oracle](https://github.com/KingGizzard/Ballotbox/issues/1) and [Agent 2](https://github.com/KingGizzard/Ballotbox/issues/4) are finished, then Agent 3 should be able to:
- Decrypt IPFS data
|
non_process
|
build agent agent is the data request agent you can see the dynamics of agent in the agent should be able to request data from the ballotbox sol contract and provide an email to be contacted through access ipfs data and save data locally when the encryption methods for the and are finished then agent should be able to decrypt ipfs data
| 0
|
12,246
| 14,744,109,121
|
IssuesEvent
|
2021-01-07 14:51:23
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Holiday Fee Bug - Child accounts
|
anc-process anp-0.5 ant-bug ant-enhancement has attachment
|
In GitLab by @kdjstudios on Jan 2, 2020, 14:15
**Submitted by:** Kyle
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/10125583
**Server:** ALL
**Client/Site:** NA
**Account:** NA
**Issue:**
During a phone call with Leah we believe we found an issue with the holiday fee settings with parent child accounts.
We setup a parent account (056-MX3177) with the Holiday fee enabled and a child account (056-M20004) with it disabled and a zero value. When we generated the draft invoice it was showing two line items for the holiday fee. It should only be charging the one as the child account has it disabled.

|
1.0
|
Holiday Fee Bug - Child accounts - In GitLab by @kdjstudios on Jan 2, 2020, 14:15
**Submitted by:** Kyle
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/10125583
**Server:** ALL
**Client/Site:** NA
**Account:** NA
**Issue:**
During a phone call with Leah we believe we found an issue with the holiday fee settings with parent child accounts.
We setup a parent account (056-MX3177) with the Holiday fee enabled and a child account (056-M20004) with it disabled and a zero value. When we generated the draft invoice it was showing two line items for the holiday fee. It should only be charging the one as the child account has it disabled.

|
process
|
holiday fee bug child accounts in gitlab by kdjstudios on jan submitted by kyle helpdesk server all client site na account na issue during a phone call with leah we believe we found an issue with the holiday fee settings with parent child accounts we setup a parent account with the holiday fee enabled and a child account with it disabled and a zero value when we generated the draft invoice it was showing two line items for the holiday fee it should only be charging the one as the child account has it disabled uploads image png
| 1
|
23,672
| 2,660,221,062
|
IssuesEvent
|
2015-03-19 04:05:47
|
dartsim/dart
|
https://api.github.com/repos/dartsim/dart
|
closed
|
Bug in FreeJoint helper
|
Kind: Bug Priority: High
|
* File: dart/dynamics/FreeJoint.cpp
* Functions compromised: convertToPositions (line 61) and converToTransform (line 70)
* Problem description: I am generating a set of transformations with rotations around Z with angles in the range [0,2*PI], increasing at each step. I use these Transformations to set the joint positions of a freeJoint (using the same formula as used in convertToPositions).
The joint positions I get are wrong for some angles. The joint positions correspond to the intended Transformation rotated 90 around Z. This happens usually after half of the range (either if I use [0,2PI] or [-PI,PI]. An example in which the bug happens is copied at the end of this issue. In this example I use:
Input transform:
```
-0.815214 -0.57916 0 0.437796
-0.57916 0.815214 0 0.488175
0 0 -1 1.03
0 0 0 1
```
Positions obtained using convertToPositions():
```
-0.954925 -2.99295 -0 0.437796 0.488175 1.03
```
Output transform obtained using convertToTransform() (should be the same as input transform):
```
-0.815214 0.57916 -5.39747e-16 0.437796
0.57916 0.815214 1.72211e-16 0.488175
5.39747e-16 -1.72211e-16 -1 1.03
6.95317e-310 1.38338e-322 6.92085e-310 2.12999e-317
```
Notice the flipped signs in x an y columns.
The problem described above is related to convertToPositions. I also noticed that in convertToTransform, there is a line missing: Tf.setIdentity(). Otherwise, the last row of Tf will be filled with garbage (as in the example above).
* Final note: Please, if somebody can fix this quickly I would appreciate it. I spent most of today searching for the bug in my own code until I realized that the bug was (very possibly and unless I am mistaken) in dart so I would like to continue with my work once the bug is fixed. Thanks!
PS.- Bug test file:
```cpp
#include <dart/math/Geometry.h>
#include <iostream>
Eigen::Vector6d convertToPositions(const Eigen::Isometry3d& _tf);
Eigen::Isometry3d convertToTransform(const Eigen::Vector6d& _positions);
int main( int argc, char* argv[] ) {
Eigen::Isometry3d Tf;
Eigen::VectorXd q;
Eigen::Isometry3d Tf_back;
Tf.setIdentity();
Tf.translation() << 0.1, 0.2, 0.3;
double dang = 10*M_PI/180.0;
int n = (int) (2*M_PI /dang);
Tf.matrix() << -0.815214, -0.57916, -0, 0.437796,
-0.57916, 0.815214, 0, 0.488175,
0, 0, -1, 1.03,
0, 0, 0, 1;
double det = Tf.linear().determinant();
std::cout << "Determinant of test TF: "<< det << std::endl;
/*
for( size_t i = 0; i < n; ++i ) {
Eigen::Matrix3d rot;
rot = Eigen::AngleAxisd( i*dang, Eigen::Vector3d(0,0,1) );
Tf.linear() = rot;
*/
// Get position with helper function
q = convertToPositions( Tf );
// Get translation back with helper function
Tf_back = convertToTransform( q );
//std::cout << "i: "<< i << std::endl;
std::cout << "\n ** Tf original: \n"<< Tf.matrix() << std::endl;
std::cout << "\n ** Positions from Tf: \n"<< q.transpose() << std::endl;
std::cout << "\n ** Tf from q: \n" << Tf_back.matrix() << std::endl;
std::cout << "\n\n" << std::endl;
/*
}
*/
}
//==============================================================================
Eigen::Vector6d convertToPositions(const Eigen::Isometry3d& _tf)
{
Eigen::Vector6d x;
x.head<3>() = dart::math::logMap(_tf.linear());
x.tail<3>() = _tf.translation();
return x;
}
//==============================================================================
Eigen::Isometry3d convertToTransform(
const Eigen::Vector6d& _positions)
{
Eigen::Isometry3d tf;
tf.linear() = dart::math::expMapRot(_positions.head<3>());
tf.translation() = _positions.tail<3>();
return tf;
}
```
|
1.0
|
Bug in FreeJoint helper - * File: dart/dynamics/FreeJoint.cpp
* Functions compromised: convertToPositions (line 61) and converToTransform (line 70)
* Problem description: I am generating a set of transformations with rotations around Z with angles in the range [0,2*PI], increasing at each step. I use these Transformations to set the joint positions of a freeJoint (using the same formula as used in convertToPositions).
The joint positions I get are wrong for some angles. The joint positions correspond to the intended Transformation rotated 90 around Z. This happens usually after half of the range (either if I use [0,2PI] or [-PI,PI]. An example in which the bug happens is copied at the end of this issue. In this example I use:
Input transform:
```
-0.815214 -0.57916 0 0.437796
-0.57916 0.815214 0 0.488175
0 0 -1 1.03
0 0 0 1
```
Positions obtained using convertToPositions():
```
-0.954925 -2.99295 -0 0.437796 0.488175 1.03
```
Output transform obtained using convertToTransform() (should be the same as input transform):
```
-0.815214 0.57916 -5.39747e-16 0.437796
0.57916 0.815214 1.72211e-16 0.488175
5.39747e-16 -1.72211e-16 -1 1.03
6.95317e-310 1.38338e-322 6.92085e-310 2.12999e-317
```
Notice the flipped signs in x an y columns.
The problem described above is related to convertToPositions. I also noticed that in convertToTransform, there is a line missing: Tf.setIdentity(). Otherwise, the last row of Tf will be filled with garbage (as in the example above).
* Final note: Please, if somebody can fix this quickly I would appreciate it. I spent most of today searching for the bug in my own code until I realized that the bug was (very possibly and unless I am mistaken) in dart so I would like to continue with my work once the bug is fixed. Thanks!
PS.- Bug test file:
```cpp
#include <dart/math/Geometry.h>
#include <iostream>
Eigen::Vector6d convertToPositions(const Eigen::Isometry3d& _tf);
Eigen::Isometry3d convertToTransform(const Eigen::Vector6d& _positions);
int main( int argc, char* argv[] ) {
Eigen::Isometry3d Tf;
Eigen::VectorXd q;
Eigen::Isometry3d Tf_back;
Tf.setIdentity();
Tf.translation() << 0.1, 0.2, 0.3;
double dang = 10*M_PI/180.0;
int n = (int) (2*M_PI /dang);
Tf.matrix() << -0.815214, -0.57916, -0, 0.437796,
-0.57916, 0.815214, 0, 0.488175,
0, 0, -1, 1.03,
0, 0, 0, 1;
double det = Tf.linear().determinant();
std::cout << "Determinant of test TF: "<< det << std::endl;
/*
for( size_t i = 0; i < n; ++i ) {
Eigen::Matrix3d rot;
rot = Eigen::AngleAxisd( i*dang, Eigen::Vector3d(0,0,1) );
Tf.linear() = rot;
*/
// Get position with helper function
q = convertToPositions( Tf );
// Get translation back with helper function
Tf_back = convertToTransform( q );
//std::cout << "i: "<< i << std::endl;
std::cout << "\n ** Tf original: \n"<< Tf.matrix() << std::endl;
std::cout << "\n ** Positions from Tf: \n"<< q.transpose() << std::endl;
std::cout << "\n ** Tf from q: \n" << Tf_back.matrix() << std::endl;
std::cout << "\n\n" << std::endl;
/*
}
*/
}
//==============================================================================
Eigen::Vector6d convertToPositions(const Eigen::Isometry3d& _tf)
{
Eigen::Vector6d x;
x.head<3>() = dart::math::logMap(_tf.linear());
x.tail<3>() = _tf.translation();
return x;
}
//==============================================================================
Eigen::Isometry3d convertToTransform(
const Eigen::Vector6d& _positions)
{
Eigen::Isometry3d tf;
tf.linear() = dart::math::expMapRot(_positions.head<3>());
tf.translation() = _positions.tail<3>();
return tf;
}
```
|
non_process
|
bug in freejoint helper file dart dynamics freejoint cpp functions compromised converttopositions line and convertotransform line problem description i am generating a set of transformations with rotations around z with angles in the range increasing at each step i use these transformations to set the joint positions of a freejoint using the same formula as used in converttopositions the joint positions i get are wrong for some angles the joint positions correspond to the intended transformation rotated around z this happens usually after half of the range either if i use or an example in which the bug happens is copied at the end of this issue in this example i use input transform positions obtained using converttopositions output transform obtained using converttotransform should be the same as input transform notice the flipped signs in x an y columns the problem described above is related to converttopositions i also noticed that in converttotransform there is a line missing tf setidentity otherwise the last row of tf will be filled with garbage as in the example above final note please if somebody can fix this quickly i would appreciate it i spent most of today searching for the bug in my own code until i realized that the bug was very possibly and unless i am mistaken in dart so i would like to continue with my work once the bug is fixed thanks ps bug test file cpp include include eigen converttopositions const eigen tf eigen converttotransform const eigen positions int main int argc char argv eigen tf eigen vectorxd q eigen tf back tf setidentity tf translation double dang m pi int n int m pi dang tf matrix double det tf linear determinant std cout determinant of test tf det std endl for size t i i n i eigen rot rot eigen angleaxisd i dang eigen tf linear rot get position with helper function q converttopositions tf get translation back with helper function tf back converttotransform q std cout i i std endl std cout n tf original n tf matrix std endl std cout n positions from tf n q transpose std endl std cout n tf from q n tf back matrix std endl std cout n n std endl eigen converttopositions const eigen tf eigen x x head dart math logmap tf linear x tail tf translation return x eigen converttotransform const eigen positions eigen tf tf linear dart math expmaprot positions head tf translation positions tail return tf
| 0
|
16,500
| 21,481,344,694
|
IssuesEvent
|
2022-04-26 18:03:37
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
opened
|
Iterate on intake form
|
intake-process-pct
|
# User story
As a Platform team member, I need a way of submitting a request to the Content team, so that I can request Platform Content team support for Platform website pages.
# Tasks
- [ ] Revise intake form in YAML
- [ ] Confirm that form submissions are appearing on documentation triage board
# Acceptance criteria
- Ticket is modified and available thru the VA.gov-team repo
# Reference info
- [Documentation triage board](https://github.com/department-of-veterans-affairs/va.gov-team/projects/16)
- [Process steps for ad hoc documentaiton support requests (draft)](https://vfs.atlassian.net/wiki/spaces/~672346535/pages/2141487170/Process+steps+for+ad-hoc+documentation+support+requests+draft)
- [Jill's question in #github_information](https://dsva.slack.com/archives/CU1E4CX9U/p1650908057955009)
|
1.0
|
Iterate on intake form - # User story
As a Platform team member, I need a way of submitting a request to the Content team, so that I can request Platform Content team support for Platform website pages.
# Tasks
- [ ] Revise intake form in YAML
- [ ] Confirm that form submissions are appearing on documentation triage board
# Acceptance criteria
- Ticket is modified and available thru the VA.gov-team repo
# Reference info
- [Documentation triage board](https://github.com/department-of-veterans-affairs/va.gov-team/projects/16)
- [Process steps for ad hoc documentaiton support requests (draft)](https://vfs.atlassian.net/wiki/spaces/~672346535/pages/2141487170/Process+steps+for+ad-hoc+documentation+support+requests+draft)
- [Jill's question in #github_information](https://dsva.slack.com/archives/CU1E4CX9U/p1650908057955009)
|
process
|
iterate on intake form user story as a platform team member i need a way of submitting a request to the content team so that i can request platform content team support for platform website pages tasks revise intake form in yaml confirm that form submissions are appearing on documentation triage board acceptance criteria ticket is modified and available thru the va gov team repo reference info
| 1
|
362,443
| 25,376,305,556
|
IssuesEvent
|
2022-11-21 14:21:26
|
redhat-plumbers-in-action/advanced-issue-labeler
|
https://api.github.com/repos/redhat-plumbers-in-action/advanced-issue-labeler
|
opened
|
Make it more visible in README which options don't work together :memo:
|
type: documentation
|
### Type of issue
other
### Description
_No response_
### Describe the solution you'd like
_No response_
|
1.0
|
Make it more visible in README which options don't work together :memo: - ### Type of issue
other
### Description
_No response_
### Describe the solution you'd like
_No response_
|
non_process
|
make it more visible in readme which options don t work together memo type of issue other description no response describe the solution you d like no response
| 0
|
68,199
| 17,191,958,982
|
IssuesEvent
|
2021-07-16 12:19:54
|
angular/angular-cli
|
https://api.github.com/repos/angular/angular-cli
|
closed
|
ng serve fails after some modification to the source code - module.buildInfo.jsonData TypeError: Cannot read property 'jsonData' of undefined
|
comp: devkit/build-angular freq1: low severity3: broken type: bug/fix
|
<!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅
Oh hi there! 😄
To expedite issue processing please search open and closed issues before submitting a new one.
Existing issues often contain information about workarounds, resolution, or progress updates.
🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅-->
# 🐞 Bug report
### Command (mark with an `x`)
<!-- Can you pin-point the command or commands that are effected by this bug? -->
<!-- ✍️edit: -->
- [ ] new
- [ ] build
- [x] serve
- [ ] test
- [ ] e2e
- [ ] generate
- [ ] add
- [ ] update
- [ ] lint
- [ ] extract-i18n
- [ ] run
- [ ] config
- [ ] help
- [ ] version
- [ ] doc
### Is this a regression?
<!-- Did this behavior use to work in the previous version? -->
<!-- ✍️-->
Yes, the previous version in which this bug was not present was: 12.1.1
### Description
<!-- ✍️-->
Run ng serve like allways
<pre>
<code>
Initial Chunk Files | Names | Size
vendor.js | vendor | 17.10 MB
main.js | main | 7.20 MB
scripts.js | scripts | 1.01 MB
styles.css | styles | 846.56 kB
polyfills.js | polyfills | 406.35 kB
runtime.js | runtime | 13.86 kB
| Initial Total | 26.54 MB
Lazy Chunk Files | Names | Size
src_app_ceramicos-la-plata_administracion_administracion_module_ts.js | - | 4.83 MB
default-src_app_ceramicos-la-plata_venta_venta_module_ts.js | - | 2.79 MB
src_app_madeco_venta_venta_module_ts.js | - | 1.87 MB
src_app_madeco_produccion_produccion_module_ts.js | - | 1.62 MB
default-src_app_madeco_common-components_common-components_module_ts.js | - | 1.39 MB
src_app_madeco_proveedores_proveedores_module_ts.js | - | 1.17 MB
src_app_ventas_ventas_module_ts.js | - | 1.16 MB
src_app_madeco_banco_bancos_module_ts.js | - | 917.62 kB
src_app_ceramicos-la-plata_logistica_logistica_module_ts.js | - | 893.92 kB
src_app_contabilidad_contabilidad_module_ts.js | - | 781.78 kB
default-src_app_madeco_custom-directives_custom-directives_module_ts-src_app_main_service_con-9c34b3.js | - | 557.74 kB
src_app_ceramicos-la-plata_banco_banco_module_ts.js | - | 482.17 kB
src_app_ceramicos-la-plata_logistica-handhelds_logistica-handhelds_module_ts.js | - | 386.45 kB
src_app_ceramicos-la-plata_planeamiento_planeamiento_module_ts.js | - | 385.68 kB
src_app_ceramicos-la-plata_grupo_grupo_module_ts.js | - | 302.96 kB
src_app_ceramicos-la-plata_facturacion_facturacion_module_ts.js | - | 277.80 kB
src_app_shared-components_lector-qr-web-worker_qr-decoder_worker_ts.js | - | 253.25 kB
src_app_ceramicos-la-plata_etiquetas_etiquetas_module_ts.js | - | 192.78 kB
default-src_app_ceramicos-la-plata_venta_caja-ventas_caja-ventas_component_ts-src_app_ceramic-ecd1a6.js | - | 172.79 kB
node_modules_firebase_messaging_dist_index_esm_js.js | - | 147.01 kB
src_app_administracion_administracion_module_ts.js | - | 107.53 kB
default-src_app_madeco_logistica_logistica_module_ts.js | - | 103.91 kB
src_app_ceramicos-la-plata_tablero_tablero_module_ts.js | - | 55.45 kB
default-src_app_ceramicos-la-plata_administracion_ingreso-egreso-motivo_ingreso-egreso-motivo-f547ef.js | - | 33.06 kB
default-src_app_ceramicos-la-plata_logistica_local-storage_local-storage-clp-logistica-strategy_ts.js | - | 23.66 kB
src_app_configuracion_configuracion_module_ts.js | - | 19.57 kB
common.js | common | 9.92 kB
</code>
</pre>
## 🔬 Minimal Reproduction
Modify some code then the error shows up after the rebuild of a lazy chunk
## 🔥 Exception or Error
<pre>
<code>
Generating browser application bundles.../home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/json/JsonGenerator.js:119
let data = module.buildInfo.jsonData;
^
TypeError: Cannot read property 'jsonData' of undefined
at JsonGenerator.getSize (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/json/JsonGenerator.js:119:31)
at NormalModule.size (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModule.js:1224:43)
at NormalModule.cleanupForCache (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModule.js:364:9)
at NormalModuleFactory.cleanupForCache (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModuleFactory.js:670:11)
at Compiler._cleanupLastNormalModuleFactory (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:383:34)
at Compiler.createNormalModuleFactory (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1049:8)
at Compiler.newCompilationParams (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1071:30)
at Compiler.compile (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1082:23)
at /home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Watching.js:188:19
at Hook.eval [as callAsync] (eval at create (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:22:1)
</code></pre>
## 🌍 Your Environment
<pre><code>
<!-- run `ng version` and paste output below -->
<!-- ✍️-->
_ _ ____ _ ___
/ \ _ __ __ _ _ _| | __ _ _ __ / ___| | |_ _|
/ △ \ | '_ \ / _` | | | | |/ _` | '__| | | | | | |
/ ___ \| | | | (_| | |_| | | (_| | | | |___| |___ | |
/_/ \_\_| |_|\__, |\__,_|_|\__,_|_| \____|_____|___|
|___/
Angular CLI: 12.1.2
Node: 14.17.0
Package Manager: npm 7.19.1
OS: linux x64
Angular: 12.1.2
... animations, cli, common, compiler, compiler-cli, core, forms
... language-service, platform-browser, platform-browser-dynamic
... router, service-worker
Package Version
---------------------------------------------------------
@angular-devkit/architect 0.1201.2
@angular-devkit/build-angular 12.1.2
@angular-devkit/build-webpack 0.1201.2
@angular-devkit/core 12.1.2
@angular-devkit/schematics 12.1.2
@angular/fire 6.1.5
@schematics/angular 12.1.2
ng-packagr 12.1.2
rxjs 6.6.7
typescript 4.3.5
webpack 5.44.0
</code></pre>
**ng build works fine**
|
1.0
|
ng serve fails after some modification to the source code - module.buildInfo.jsonData TypeError: Cannot read property 'jsonData' of undefined - <!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅
Oh hi there! 😄
To expedite issue processing please search open and closed issues before submitting a new one.
Existing issues often contain information about workarounds, resolution, or progress updates.
🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅-->
# 🐞 Bug report
### Command (mark with an `x`)
<!-- Can you pin-point the command or commands that are effected by this bug? -->
<!-- ✍️edit: -->
- [ ] new
- [ ] build
- [x] serve
- [ ] test
- [ ] e2e
- [ ] generate
- [ ] add
- [ ] update
- [ ] lint
- [ ] extract-i18n
- [ ] run
- [ ] config
- [ ] help
- [ ] version
- [ ] doc
### Is this a regression?
<!-- Did this behavior use to work in the previous version? -->
<!-- ✍️-->
Yes, the previous version in which this bug was not present was: 12.1.1
### Description
<!-- ✍️-->
Run ng serve like allways
<pre>
<code>
Initial Chunk Files | Names | Size
vendor.js | vendor | 17.10 MB
main.js | main | 7.20 MB
scripts.js | scripts | 1.01 MB
styles.css | styles | 846.56 kB
polyfills.js | polyfills | 406.35 kB
runtime.js | runtime | 13.86 kB
| Initial Total | 26.54 MB
Lazy Chunk Files | Names | Size
src_app_ceramicos-la-plata_administracion_administracion_module_ts.js | - | 4.83 MB
default-src_app_ceramicos-la-plata_venta_venta_module_ts.js | - | 2.79 MB
src_app_madeco_venta_venta_module_ts.js | - | 1.87 MB
src_app_madeco_produccion_produccion_module_ts.js | - | 1.62 MB
default-src_app_madeco_common-components_common-components_module_ts.js | - | 1.39 MB
src_app_madeco_proveedores_proveedores_module_ts.js | - | 1.17 MB
src_app_ventas_ventas_module_ts.js | - | 1.16 MB
src_app_madeco_banco_bancos_module_ts.js | - | 917.62 kB
src_app_ceramicos-la-plata_logistica_logistica_module_ts.js | - | 893.92 kB
src_app_contabilidad_contabilidad_module_ts.js | - | 781.78 kB
default-src_app_madeco_custom-directives_custom-directives_module_ts-src_app_main_service_con-9c34b3.js | - | 557.74 kB
src_app_ceramicos-la-plata_banco_banco_module_ts.js | - | 482.17 kB
src_app_ceramicos-la-plata_logistica-handhelds_logistica-handhelds_module_ts.js | - | 386.45 kB
src_app_ceramicos-la-plata_planeamiento_planeamiento_module_ts.js | - | 385.68 kB
src_app_ceramicos-la-plata_grupo_grupo_module_ts.js | - | 302.96 kB
src_app_ceramicos-la-plata_facturacion_facturacion_module_ts.js | - | 277.80 kB
src_app_shared-components_lector-qr-web-worker_qr-decoder_worker_ts.js | - | 253.25 kB
src_app_ceramicos-la-plata_etiquetas_etiquetas_module_ts.js | - | 192.78 kB
default-src_app_ceramicos-la-plata_venta_caja-ventas_caja-ventas_component_ts-src_app_ceramic-ecd1a6.js | - | 172.79 kB
node_modules_firebase_messaging_dist_index_esm_js.js | - | 147.01 kB
src_app_administracion_administracion_module_ts.js | - | 107.53 kB
default-src_app_madeco_logistica_logistica_module_ts.js | - | 103.91 kB
src_app_ceramicos-la-plata_tablero_tablero_module_ts.js | - | 55.45 kB
default-src_app_ceramicos-la-plata_administracion_ingreso-egreso-motivo_ingreso-egreso-motivo-f547ef.js | - | 33.06 kB
default-src_app_ceramicos-la-plata_logistica_local-storage_local-storage-clp-logistica-strategy_ts.js | - | 23.66 kB
src_app_configuracion_configuracion_module_ts.js | - | 19.57 kB
common.js | common | 9.92 kB
</code>
</pre>
## 🔬 Minimal Reproduction
Modify some code then the error shows up after the rebuild of a lazy chunk
## 🔥 Exception or Error
<pre>
<code>
Generating browser application bundles.../home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/json/JsonGenerator.js:119
let data = module.buildInfo.jsonData;
^
TypeError: Cannot read property 'jsonData' of undefined
at JsonGenerator.getSize (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/json/JsonGenerator.js:119:31)
at NormalModule.size (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModule.js:1224:43)
at NormalModule.cleanupForCache (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModule.js:364:9)
at NormalModuleFactory.cleanupForCache (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModuleFactory.js:670:11)
at Compiler._cleanupLastNormalModuleFactory (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:383:34)
at Compiler.createNormalModuleFactory (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1049:8)
at Compiler.newCompilationParams (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1071:30)
at Compiler.compile (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1082:23)
at /home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Watching.js:188:19
at Hook.eval [as callAsync] (eval at create (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:22:1)
</code></pre>
## 🌍 Your Environment
<pre><code>
<!-- run `ng version` and paste output below -->
<!-- ✍️-->
_ _ ____ _ ___
/ \ _ __ __ _ _ _| | __ _ _ __ / ___| | |_ _|
/ △ \ | '_ \ / _` | | | | |/ _` | '__| | | | | | |
/ ___ \| | | | (_| | |_| | | (_| | | | |___| |___ | |
/_/ \_\_| |_|\__, |\__,_|_|\__,_|_| \____|_____|___|
|___/
Angular CLI: 12.1.2
Node: 14.17.0
Package Manager: npm 7.19.1
OS: linux x64
Angular: 12.1.2
... animations, cli, common, compiler, compiler-cli, core, forms
... language-service, platform-browser, platform-browser-dynamic
... router, service-worker
Package Version
---------------------------------------------------------
@angular-devkit/architect 0.1201.2
@angular-devkit/build-angular 12.1.2
@angular-devkit/build-webpack 0.1201.2
@angular-devkit/core 12.1.2
@angular-devkit/schematics 12.1.2
@angular/fire 6.1.5
@schematics/angular 12.1.2
ng-packagr 12.1.2
rxjs 6.6.7
typescript 4.3.5
webpack 5.44.0
</code></pre>
**ng build works fine**
|
non_process
|
ng serve fails after some modification to the source code module buildinfo jsondata typeerror cannot read property jsondata of undefined 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 oh hi there 😄 to expedite issue processing please search open and closed issues before submitting a new one existing issues often contain information about workarounds resolution or progress updates 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 🐞 bug report command mark with an x new build serve test generate add update lint extract run config help version doc is this a regression yes the previous version in which this bug was not present was description run ng serve like allways initial chunk files names size vendor js vendor mb main js main mb scripts js scripts mb styles css styles kb polyfills js polyfills kb runtime js runtime kb initial total mb lazy chunk files names size src app ceramicos la plata administracion administracion module ts js mb default src app ceramicos la plata venta venta module ts js mb src app madeco venta venta module ts js mb src app madeco produccion produccion module ts js mb default src app madeco common components common components module ts js mb src app madeco proveedores proveedores module ts js mb src app ventas ventas module ts js mb src app madeco banco bancos module ts js kb src app ceramicos la plata logistica logistica module ts js kb src app contabilidad contabilidad module ts js kb default src app madeco custom directives custom directives module ts src app main service con js kb src app ceramicos la plata banco banco module ts js kb src app ceramicos la plata logistica handhelds logistica handhelds module ts js kb src app ceramicos la plata planeamiento planeamiento module ts js kb src app ceramicos la plata grupo grupo module ts js kb src app ceramicos la plata facturacion facturacion module ts js kb src app shared components lector qr web worker qr decoder worker ts js kb src app ceramicos la plata etiquetas etiquetas module ts js kb default src app ceramicos la plata venta caja ventas caja ventas component ts src app ceramic js kb node modules firebase messaging dist index esm js js kb src app administracion administracion module ts js kb default src app madeco logistica logistica module ts js kb src app ceramicos la plata tablero tablero module ts js kb default src app ceramicos la plata administracion ingreso egreso motivo ingreso egreso motivo js kb default src app ceramicos la plata logistica local storage local storage clp logistica strategy ts js kb src app configuracion configuracion module ts js kb common js common kb 🔬 minimal reproduction modify some code then the error shows up after the rebuild of a lazy chunk 🔥 exception or error generating browser application bundles home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib json jsongenerator js let data module buildinfo jsondata typeerror cannot read property jsondata of undefined at jsongenerator getsize home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib json jsongenerator js at normalmodule size home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib normalmodule js at normalmodule cleanupforcache home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib normalmodule js at normalmodulefactory cleanupforcache home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib normalmodulefactory js at compiler cleanuplastnormalmodulefactory home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib compiler js at compiler createnormalmodulefactory home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib compiler js at compiler newcompilationparams home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib compiler js at compiler compile home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib compiler js at home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib watching js at hook eval eval at create home naxs trabajo naxscloud naxscloudweb node modules tapable lib hookcodefactory js 🌍 your environment △ angular cli node package manager npm os linux angular animations cli common compiler compiler cli core forms language service platform browser platform browser dynamic router service worker package version angular devkit architect angular devkit build angular angular devkit build webpack angular devkit core angular devkit schematics angular fire schematics angular ng packagr rxjs typescript webpack ng build works fine
| 0
|
219,553
| 7,343,624,460
|
IssuesEvent
|
2018-03-07 12:02:53
|
trimstray/pvimport
|
https://api.github.com/repos/trimstray/pvimport
|
opened
|
Limiting disk usage during import.
|
Priority: Critical Status: In Progress Type: Maintenance
|
Next Release: **[testing](https://github.com/trimstray/pvimport/tree/testing)**
Status: **In Progress**
Limiting disk usage during import:
- use **CFQ scheduler**
- use **ionice** to change io scheduling class and priority
|
1.0
|
Limiting disk usage during import. - Next Release: **[testing](https://github.com/trimstray/pvimport/tree/testing)**
Status: **In Progress**
Limiting disk usage during import:
- use **CFQ scheduler**
- use **ionice** to change io scheduling class and priority
|
non_process
|
limiting disk usage during import next release status in progress limiting disk usage during import use cfq scheduler use ionice to change io scheduling class and priority
| 0
|
431,055
| 12,474,556,762
|
IssuesEvent
|
2020-05-29 09:52:05
|
StrangeLoopGames/EcoIssues
|
https://api.github.com/repos/StrangeLoopGames/EcoIssues
|
closed
|
Don't show removed titles in lists
|
Priority: Medium Status: Fixed
|

1) It clutters up the list without any use.
2) In case of transfering a deed by mistake to one of the removed titles, an admin needs to reclaim it, as it's stuck otherwise.
I'd have them removed from all such lists, actually.
|
1.0
|
Don't show removed titles in lists - 
1) It clutters up the list without any use.
2) In case of transfering a deed by mistake to one of the removed titles, an admin needs to reclaim it, as it's stuck otherwise.
I'd have them removed from all such lists, actually.
|
non_process
|
don t show removed titles in lists it clutters up the list without any use in case of transfering a deed by mistake to one of the removed titles an admin needs to reclaim it as it s stuck otherwise i d have them removed from all such lists actually
| 0
|
14,187
| 17,070,591,552
|
IssuesEvent
|
2021-07-07 12:54:05
|
opensrp/web
|
https://api.github.com/repos/opensrp/web
|
opened
|
[FHIR Road Map] Descriptive error messages for all forms
|
FHIR compatibility
|
## Issue Description
add descriptive error messages to forms. Replace generic 'error occurred' messages in all forms with descriptive messages.
## Resources
### FHIR Resources to reference
## Please share any other relevant information about the change request
---
Remember:
1. To alert the team in the [#opensrp-web](https://onaio.slack.com/archives/C01R82Y9N58) channel.
2. To follow up on delivery. You're in-charge.
|
True
|
[FHIR Road Map] Descriptive error messages for all forms - ## Issue Description
add descriptive error messages to forms. Replace generic 'error occurred' messages in all forms with descriptive messages.
## Resources
### FHIR Resources to reference
## Please share any other relevant information about the change request
---
Remember:
1. To alert the team in the [#opensrp-web](https://onaio.slack.com/archives/C01R82Y9N58) channel.
2. To follow up on delivery. You're in-charge.
|
non_process
|
descriptive error messages for all forms issue description add descriptive error messages to forms replace generic error occurred messages in all forms with descriptive messages resources fhir resources to reference please share any other relevant information about the change request remember to alert the team in the channel to follow up on delivery you re in charge
| 0
|
8,791
| 11,908,170,982
|
IssuesEvent
|
2020-03-31 00:10:51
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Change location of temporary executables for geoprocessing tools
|
Feature Request Processing
|
**Feature description.**
<!-- A clear and concise description of what you want to happen. Ex. QGIS would rock even more if [...] -->
I am not sure if I missed something somewhere, but it would be amazing to add the option to decide where the intermediate geoprocessing executables are stored and executed.
We would like to use QGIS (with SAGA + GRASS + GDAL) at a university, and the fact that some intermediate executables are created inside the user's temporary folder means that we can't whitelist them.
We end up having this type of error message:
> "This program is blocked by group policy. For more information, contact your system administrator."
In a GRASS example, the offending executable would be:
> C:\Users\<username>\AppData\Local\Temp\processing_670dd4b4fc494c41a3761314d2eb317a\grassdata\grass_batch_job.cmd
In the case of SAGA, it seems it tries to run a .bat file.
We can't whitelist every user's entire `AppData\Local\Temp` folder, and the unpredictable naming of the folders does not allow to whitelist a more precise path.
Could there be an option to define where all those intermediate executables will be going and executed from? For example, an extra option in Options > Processing > General
**Additional context**
Tested with QGIS 3.6.3 and 3.8.3 on Windows 10.
|
1.0
|
Change location of temporary executables for geoprocessing tools - **Feature description.**
<!-- A clear and concise description of what you want to happen. Ex. QGIS would rock even more if [...] -->
I am not sure if I missed something somewhere, but it would be amazing to add the option to decide where the intermediate geoprocessing executables are stored and executed.
We would like to use QGIS (with SAGA + GRASS + GDAL) at a university, and the fact that some intermediate executables are created inside the user's temporary folder means that we can't whitelist them.
We end up having this type of error message:
> "This program is blocked by group policy. For more information, contact your system administrator."
In a GRASS example, the offending executable would be:
> C:\Users\<username>\AppData\Local\Temp\processing_670dd4b4fc494c41a3761314d2eb317a\grassdata\grass_batch_job.cmd
In the case of SAGA, it seems it tries to run a .bat file.
We can't whitelist every user's entire `AppData\Local\Temp` folder, and the unpredictable naming of the folders does not allow to whitelist a more precise path.
Could there be an option to define where all those intermediate executables will be going and executed from? For example, an extra option in Options > Processing > General
**Additional context**
Tested with QGIS 3.6.3 and 3.8.3 on Windows 10.
|
process
|
change location of temporary executables for geoprocessing tools feature description i am not sure if i missed something somewhere but it would be amazing to add the option to decide where the intermediate geoprocessing executables are stored and executed we would like to use qgis with saga grass gdal at a university and the fact that some intermediate executables are created inside the user s temporary folder means that we can t whitelist them we end up having this type of error message this program is blocked by group policy for more information contact your system administrator in a grass example the offending executable would be c users appdata local temp processing grassdata grass batch job cmd in the case of saga it seems it tries to run a bat file we can t whitelist every user s entire appdata local temp folder and the unpredictable naming of the folders does not allow to whitelist a more precise path could there be an option to define where all those intermediate executables will be going and executed from for example an extra option in options processing general additional context tested with qgis and on windows
| 1
|
12,468
| 14,703,397,671
|
IssuesEvent
|
2021-01-04 14:59:29
|
fourMs/MGT-python
|
https://api.github.com/repos/fourMs/MGT-python
|
closed
|
Tkinter on Ubuntu
|
compatibility
|
Just discovered that installing the Tkinter package is slightly trickier on Ubuntu. In `musicalgestures`, in the cropping pre-processing step, when using manual cropping, I used tkinter to get the screensize and scale the image (_utils.py: `get_screen_resolution_scaled()`: *Gets the scaled screen resolution. Respects display scaling on high DPI displays.*), so it does not overflow the screen. On Ubuntu [there seems to be a problem](https://stackoverflow.com/questions/6084416/tkinter-module-not-found-on-ubuntu) with Tkinter by default. Using `sudo apt-get install python3-tk` solved it for me. But this means the installation process should include a similar step on Linux.
|
True
|
Tkinter on Ubuntu - Just discovered that installing the Tkinter package is slightly trickier on Ubuntu. In `musicalgestures`, in the cropping pre-processing step, when using manual cropping, I used tkinter to get the screensize and scale the image (_utils.py: `get_screen_resolution_scaled()`: *Gets the scaled screen resolution. Respects display scaling on high DPI displays.*), so it does not overflow the screen. On Ubuntu [there seems to be a problem](https://stackoverflow.com/questions/6084416/tkinter-module-not-found-on-ubuntu) with Tkinter by default. Using `sudo apt-get install python3-tk` solved it for me. But this means the installation process should include a similar step on Linux.
|
non_process
|
tkinter on ubuntu just discovered that installing the tkinter package is slightly trickier on ubuntu in musicalgestures in the cropping pre processing step when using manual cropping i used tkinter to get the screensize and scale the image utils py get screen resolution scaled gets the scaled screen resolution respects display scaling on high dpi displays so it does not overflow the screen on ubuntu with tkinter by default using sudo apt get install tk solved it for me but this means the installation process should include a similar step on linux
| 0
|
15,693
| 19,848,101,388
|
IssuesEvent
|
2022-01-21 09:12:44
|
ooi-data/CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_statistics_recovered
|
https://api.github.com/repos/ooi-data/CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_statistics_recovered
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:12:43.462049.
## Details
Flow name: `CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_statistics_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T09:12:43.462049.
## Details
Flow name: `CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_statistics_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__
return self.func(self.array)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask
data = np.asarray(data, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host wavss a dcl statistics recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
725,419
| 24,961,611,792
|
IssuesEvent
|
2022-11-01 15:57:18
|
bcgov/entity
|
https://api.github.com/repos/bcgov/entity
|
closed
|
Can't add registration to MHR table
|
bug Priority1 Assets
|
Issue:
1. Logged in as staff, I want to add an existing registration (Active, Historical, Exempt) to my table.
2. When I search for an existing registration, example MHR "082086", I see a modal saying the registration exists
3. When I click "Add Registration to my table", I am served an error modal and am not able to add it to my table.

|
1.0
|
Can't add registration to MHR table - Issue:
1. Logged in as staff, I want to add an existing registration (Active, Historical, Exempt) to my table.
2. When I search for an existing registration, example MHR "082086", I see a modal saying the registration exists
3. When I click "Add Registration to my table", I am served an error modal and am not able to add it to my table.

|
non_process
|
can t add registration to mhr table issue logged in as staff i want to add an existing registration active historical exempt to my table when i search for an existing registration example mhr i see a modal saying the registration exists when i click add registration to my table i am served an error modal and am not able to add it to my table
| 0
|
80,957
| 15,609,560,455
|
IssuesEvent
|
2021-03-19 12:04:24
|
fac21/week2--Http-Project--AJMS
|
https://api.github.com/repos/fac21/week2--Http-Project--AJMS
|
closed
|
Tidy up commented out code
|
code review
|
There is a lot of commented out code, if you don't need it delete it 🧚♀️🧚♀️🧚♀️ it will make reading it easier for other people!
|
1.0
|
Tidy up commented out code - There is a lot of commented out code, if you don't need it delete it 🧚♀️🧚♀️🧚♀️ it will make reading it easier for other people!
|
non_process
|
tidy up commented out code there is a lot of commented out code if you don t need it delete it 🧚♀️🧚♀️🧚♀️ it will make reading it easier for other people
| 0
|
1,140
| 3,626,972,173
|
IssuesEvent
|
2016-02-10 05:13:16
|
worldspawn/mascis
|
https://api.github.com/repos/worldspawn/mascis
|
opened
|
Support Delete...Where
|
enhancement linq-expression-parser postgres-language-processor query-api t-sql-language-processor
|
- [ ] Linq Expression Parser
- [ ] T-Sql Language Processor
- [ ] Postgres Language Processor
- [ ] Query Api
|
2.0
|
Support Delete...Where - - [ ] Linq Expression Parser
- [ ] T-Sql Language Processor
- [ ] Postgres Language Processor
- [ ] Query Api
|
process
|
support delete where linq expression parser t sql language processor postgres language processor query api
| 1
|
11,823
| 7,719,337,037
|
IssuesEvent
|
2018-05-23 19:04:36
|
RoaringBitmap/CRoaring
|
https://api.github.com/repos/RoaringBitmap/CRoaring
|
opened
|
Add "roaring_bitmap_remove_many" function
|
help wanted performance
|
We would like to have a ``roaring_bitmap_remove_many`` function which allows you to remove from a bitmap all of the elements from an array. This can be done currently with repeated calls to ``roaring_bitmap_remove``. But it would be more convenient and efficient to have a dedicated function.
|
True
|
Add "roaring_bitmap_remove_many" function - We would like to have a ``roaring_bitmap_remove_many`` function which allows you to remove from a bitmap all of the elements from an array. This can be done currently with repeated calls to ``roaring_bitmap_remove``. But it would be more convenient and efficient to have a dedicated function.
|
non_process
|
add roaring bitmap remove many function we would like to have a roaring bitmap remove many function which allows you to remove from a bitmap all of the elements from an array this can be done currently with repeated calls to roaring bitmap remove but it would be more convenient and efficient to have a dedicated function
| 0
|
17,396
| 23,211,708,876
|
IssuesEvent
|
2022-08-02 10:43:18
|
apache/arrow-rs
|
https://api.github.com/repos/apache/arrow-rs
|
opened
|
arduino/setup-protoc@v1 is not allowed to be used in apache/arrow-rs
|
bug development-process
|
**Describe the bug**
<!--
A clear and concise description of what the bug is.
-->
https://github.com/apache/arrow-rs/actions/runs/2781430582
**To Reproduce**
<!--
Steps to reproduce the behavior:
-->
**Expected behavior**
<!--
A clear and concise description of what you expected to happen.
-->
**Additional context**
<!--
Add any other context about the problem here.
-->
|
1.0
|
arduino/setup-protoc@v1 is not allowed to be used in apache/arrow-rs - **Describe the bug**
<!--
A clear and concise description of what the bug is.
-->
https://github.com/apache/arrow-rs/actions/runs/2781430582
**To Reproduce**
<!--
Steps to reproduce the behavior:
-->
**Expected behavior**
<!--
A clear and concise description of what you expected to happen.
-->
**Additional context**
<!--
Add any other context about the problem here.
-->
|
process
|
arduino setup protoc is not allowed to be used in apache arrow rs describe the bug a clear and concise description of what the bug is to reproduce steps to reproduce the behavior expected behavior a clear and concise description of what you expected to happen additional context add any other context about the problem here
| 1
|
9,223
| 12,257,986,462
|
IssuesEvent
|
2020-05-06 14:29:01
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
Upload new PAR
|
EPIC - PARs process HIGH PRIORITY :arrow_double_up: SPIKE :hocho:
|
## User Want
As a Medical Writer in the licensing team
I would like to upload a new PAR
so that it is available on products.mhra.gov.uk
## Acceptance Criteria
### Technical acceptance criteria
- [ ] PAR pdf is in blob storage after upload
- [ ] PAR metadata is attached
- [ ] PAR is in the search index
- [ ] blob/metadata/search index is handled by doc-index-updater
### Customer acceptance criteria
- [ ] Medical writers can upload a file which will then be surfaced on products.mhra.gov.uk
- [ ] The PAR can be linked to one or multiple products (which will have their own PL number)
- [ ] PAR has product name
- [ ] PAR has active substances, something about TMS list (unsure what this is, #no-acronyms)
- [ ] PAR has one or more PLs / NR (unsure what this is) / THR (unsure what this is)
## Data - Potential impact
**Size**
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
1.0
|
Upload new PAR - ## User Want
As a Medical Writer in the licensing team
I would like to upload a new PAR
so that it is available on products.mhra.gov.uk
## Acceptance Criteria
### Technical acceptance criteria
- [ ] PAR pdf is in blob storage after upload
- [ ] PAR metadata is attached
- [ ] PAR is in the search index
- [ ] blob/metadata/search index is handled by doc-index-updater
### Customer acceptance criteria
- [ ] Medical writers can upload a file which will then be surfaced on products.mhra.gov.uk
- [ ] The PAR can be linked to one or multiple products (which will have their own PL number)
- [ ] PAR has product name
- [ ] PAR has active substances, something about TMS list (unsure what this is, #no-acronyms)
- [ ] PAR has one or more PLs / NR (unsure what this is) / THR (unsure what this is)
## Data - Potential impact
**Size**
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
process
|
upload new par user want as a medical writer in the licensing team i would like to upload a new par so that it is available on products mhra gov uk acceptance criteria technical acceptance criteria par pdf is in blob storage after upload par metadata is attached par is in the search index blob metadata search index is handled by doc index updater customer acceptance criteria medical writers can upload a file which will then be surfaced on products mhra gov uk the par can be linked to one or multiple products which will have their own pl number par has product name par has active substances something about tms list unsure what this is no acronyms par has one or more pls nr unsure what this is thr unsure what this is data potential impact size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
| 1
|
81,091
| 23,386,477,324
|
IssuesEvent
|
2022-08-11 14:09:01
|
reapit/foundations
|
https://api.github.com/repos/reapit/foundations
|
closed
|
App Builder making an entity editable launches the page creation wizard
|
feature front-end app-builder
|
**Background context or User story:**
_Selecting editable from the customise panel on a table should launch a pre-populated page creation wizard. Should have entity type and page type table pre-selected_
|
1.0
|
App Builder making an entity editable launches the page creation wizard - **Background context or User story:**
_Selecting editable from the customise panel on a table should launch a pre-populated page creation wizard. Should have entity type and page type table pre-selected_
|
non_process
|
app builder making an entity editable launches the page creation wizard background context or user story selecting editable from the customise panel on a table should launch a pre populated page creation wizard should have entity type and page type table pre selected
| 0
|
143,838
| 5,531,055,357
|
IssuesEvent
|
2017-03-21 05:46:13
|
knipferrc/plate
|
https://api.github.com/repos/knipferrc/plate
|
closed
|
Complete Plate Color Coordination
|
Priority: Medium Type: Feature
|
Attributes in place. Complete a random generation of colors for each unique plate.
|
1.0
|
Complete Plate Color Coordination - Attributes in place. Complete a random generation of colors for each unique plate.
|
non_process
|
complete plate color coordination attributes in place complete a random generation of colors for each unique plate
| 0
|
666
| 3,136,006,190
|
IssuesEvent
|
2015-09-10 17:49:45
|
PHPOffice/PHPWord
|
https://api.github.com/repos/PHPOffice/PHPWord
|
opened
|
Support arrays in TemplateProcessor.setValue method
|
Change Request Template Processor
|
To do:
- arrays in `$subject` to handle headers, footers and main document part at once.
- arrays in `$search` and in `$replacement` to avoid looping on client side.
Requested by @OAFCROB (see https://github.com/PHPOffice/PHPWord/issues/513#issuecomment-136964062).
|
1.0
|
Support arrays in TemplateProcessor.setValue method - To do:
- arrays in `$subject` to handle headers, footers and main document part at once.
- arrays in `$search` and in `$replacement` to avoid looping on client side.
Requested by @OAFCROB (see https://github.com/PHPOffice/PHPWord/issues/513#issuecomment-136964062).
|
process
|
support arrays in templateprocessor setvalue method to do arrays in subject to handle headers footers and main document part at once arrays in search and in replacement to avoid looping on client side requested by oafcrob see
| 1
|
11,203
| 13,957,703,535
|
IssuesEvent
|
2020-10-24 08:13:57
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
BE: Harvesting frequency
|
BE - Belgium Geoportal Harvesting process
|
Dear INSPIRE Geoportal Thematic viewer team,
Dear Helpdesk; we have updated the harvesting frequencies of 2 catalogues over a week ago in the inspire registry (http://inspire-geoportal.ec.europa.eu/INSPIRERegistry/registers/services/allproposals ): * Geocatalogue de la région Bruxelloise : from weekly to daily * Geopunt : from weekly to daily On the thematic geoportal the harvesting frequency has not yet been updated? Does this mean this has to be updated differently ? Thank you in advance Ouns
Best regards,
Ouns Kissiyar
|
1.0
|
BE: Harvesting frequency - Dear INSPIRE Geoportal Thematic viewer team,
Dear Helpdesk; we have updated the harvesting frequencies of 2 catalogues over a week ago in the inspire registry (http://inspire-geoportal.ec.europa.eu/INSPIRERegistry/registers/services/allproposals ): * Geocatalogue de la région Bruxelloise : from weekly to daily * Geopunt : from weekly to daily On the thematic geoportal the harvesting frequency has not yet been updated? Does this mean this has to be updated differently ? Thank you in advance Ouns
Best regards,
Ouns Kissiyar
|
process
|
be harvesting frequency dear inspire geoportal thematic viewer team dear helpdesk we have updated the harvesting frequencies of catalogues over a week ago in the inspire registry geocatalogue de la r eacute gion bruxelloise from weekly to daily geopunt from weekly to daily on the thematic geoportal the harvesting frequency has not yet been updated does this mean this has to be updated differently thank you in advance ouns best regards ouns kissiyar
| 1
|
15,595
| 19,721,759,324
|
IssuesEvent
|
2022-01-13 15:59:49
|
eunseo2/JAVA
|
https://api.github.com/repos/eunseo2/JAVA
|
opened
|
Iterating
|
process
|
> Task :Iterating.main()
1
2
3
total = 6
```java
public static List<User> userList(){
List<User> userList = new ArrayList<>();
userList.add(new User(1,"user1","user1@gmail.com"));
userList.add(new User(2,"user2","user2@gmail.com"));
userList.add(new User(3,"user3","user3@gmail.com"));
return userList;
}
```
user1의 Id 값 + user2의 Id 값 + user3의 Id 값 = 6
|
1.0
|
Iterating - > Task :Iterating.main()
1
2
3
total = 6
```java
public static List<User> userList(){
List<User> userList = new ArrayList<>();
userList.add(new User(1,"user1","user1@gmail.com"));
userList.add(new User(2,"user2","user2@gmail.com"));
userList.add(new User(3,"user3","user3@gmail.com"));
return userList;
}
```
user1의 Id 값 + user2의 Id 값 + user3의 Id 값 = 6
|
process
|
iterating task iterating main total java public static list userlist list userlist new arraylist userlist add new user gmail com userlist add new user gmail com userlist add new user gmail com return userlist id 값 id 값 id 값
| 1
|
17,229
| 22,917,655,674
|
IssuesEvent
|
2022-07-17 07:34:07
|
streamnative/flink
|
https://api.github.com/repos/streamnative/flink
|
closed
|
[SQL connector] Users using value.format should receive validationException, and they are not
|
compute/data-processing
|
Community user reported that when using `value.format` they get a NoSuchElementException. This should not occur as when users specify `value.format`, they should get a ValidationException.
|
1.0
|
[SQL connector] Users using value.format should receive validationException, and they are not - Community user reported that when using `value.format` they get a NoSuchElementException. This should not occur as when users specify `value.format`, they should get a ValidationException.
|
process
|
users using value format should receive validationexception and they are not community user reported that when using value format they get a nosuchelementexception this should not occur as when users specify value format they should get a validationexception
| 1
|
424,505
| 29,117,249,548
|
IssuesEvent
|
2023-05-17 02:48:54
|
pilcrowOnPaper/lucia
|
https://api.github.com/repos/pilcrowOnPaper/lucia
|
closed
|
[Docs-Bug]: Drizzle ORM Sqlite incorrect
|
documentation bug
|
### Describe the bug
In the Drizzle ORM documentation page, the SQLite section has the same code as the mysql section, and refers to functions (i.e. varchar) that are not available in the sqlite adapter (text is used instead).
Current Documentation
```typescript
import { sqliteTable, varchar, integer } from "drizzle-orm/sqlite-core";
const user = mysqlTable("auth_user", {
id: varchar("id", {
length: 15 // change this when using custom user ids
}).primaryKey()
// other user attributes
});
const session = mysqlTable("auth_session", {
id: varchar("id", {
length: 128
}).primaryKey(),
userId: varchar("user_id", {
length: 15
})
.notNull()
.references(() => user.id),
activeExpires: integer("active_expires").notNull(),
idleExpires: integer("idle_expires").notNull()
});
const key = mysqlTable("auth_key", {
id: varchar("id", {
length: 255
}).primaryKey(),
userId: varchar("user_id", {
length: 15
})
.notNull()
.references(() => user.id),
primaryKey: integer().notNull(),
hashedPassword: varchar("hashed_password", {
length: 255
})
});
```
|
1.0
|
[Docs-Bug]: Drizzle ORM Sqlite incorrect - ### Describe the bug
In the Drizzle ORM documentation page, the SQLite section has the same code as the mysql section, and refers to functions (i.e. varchar) that are not available in the sqlite adapter (text is used instead).
Current Documentation
```typescript
import { sqliteTable, varchar, integer } from "drizzle-orm/sqlite-core";
const user = mysqlTable("auth_user", {
id: varchar("id", {
length: 15 // change this when using custom user ids
}).primaryKey()
// other user attributes
});
const session = mysqlTable("auth_session", {
id: varchar("id", {
length: 128
}).primaryKey(),
userId: varchar("user_id", {
length: 15
})
.notNull()
.references(() => user.id),
activeExpires: integer("active_expires").notNull(),
idleExpires: integer("idle_expires").notNull()
});
const key = mysqlTable("auth_key", {
id: varchar("id", {
length: 255
}).primaryKey(),
userId: varchar("user_id", {
length: 15
})
.notNull()
.references(() => user.id),
primaryKey: integer().notNull(),
hashedPassword: varchar("hashed_password", {
length: 255
})
});
```
|
non_process
|
drizzle orm sqlite incorrect describe the bug in the drizzle orm documentation page the sqlite section has the same code as the mysql section and refers to functions i e varchar that are not available in the sqlite adapter text is used instead current documentation typescript import sqlitetable varchar integer from drizzle orm sqlite core const user mysqltable auth user id varchar id length change this when using custom user ids primarykey other user attributes const session mysqltable auth session id varchar id length primarykey userid varchar user id length notnull references user id activeexpires integer active expires notnull idleexpires integer idle expires notnull const key mysqltable auth key id varchar id length primarykey userid varchar user id length notnull references user id primarykey integer notnull hashedpassword varchar hashed password length
| 0
|
5,537
| 8,392,233,582
|
IssuesEvent
|
2018-10-09 17:02:12
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
Do not get goaccess to work with my log file
|
log-processing log/date/time format question
|
Hi, I went through all kind of issues, but I am not able to get the default vcombined to work
THis is an example of a line:
testma.mantagifts.com:443 ip-217-105-39-28.ip.prioritytelecom.net - mantasalesapp [04/Oct/2018:07:53:48 +0200] "POST /sales_app/backups/backup.php HTTP/1.1" 200 427 "-" "Appcelerator Titanium/7.4.0 (iPad/12.0; iOS; en_US;)"
This is my command:
> goaccess /var/log/apache2/access_ssl_testma.mantagifts.com.log --log-format=VCOMBINED
..
Token 'ip-217-105-39-28.ip.prioritytelecom.net' doesn't match specifier '%h'
..
Format Errors - Verify your log/date/time format
I have tried multiple thinkgs (custom log, time etc),but nothing seem to work. Can you help me with this?
Thanks
|
1.0
|
Do not get goaccess to work with my log file - Hi, I went through all kind of issues, but I am not able to get the default vcombined to work
THis is an example of a line:
testma.mantagifts.com:443 ip-217-105-39-28.ip.prioritytelecom.net - mantasalesapp [04/Oct/2018:07:53:48 +0200] "POST /sales_app/backups/backup.php HTTP/1.1" 200 427 "-" "Appcelerator Titanium/7.4.0 (iPad/12.0; iOS; en_US;)"
This is my command:
> goaccess /var/log/apache2/access_ssl_testma.mantagifts.com.log --log-format=VCOMBINED
..
Token 'ip-217-105-39-28.ip.prioritytelecom.net' doesn't match specifier '%h'
..
Format Errors - Verify your log/date/time format
I have tried multiple thinkgs (custom log, time etc),but nothing seem to work. Can you help me with this?
Thanks
|
process
|
do not get goaccess to work with my log file hi i went through all kind of issues but i am not able to get the default vcombined to work this is an example of a line testma mantagifts com ip ip prioritytelecom net mantasalesapp post sales app backups backup php http appcelerator titanium ipad ios en us this is my command goaccess var log access ssl testma mantagifts com log log format vcombined token ip ip prioritytelecom net doesn t match specifier h format errors verify your log date time format i have tried multiple thinkgs custom log time etc but nothing seem to work can you help me with this thanks
| 1
|
2,427
| 5,203,079,979
|
IssuesEvent
|
2017-01-24 11:37:26
|
rogerthat-platform/rogerthat-android-client
|
https://api.github.com/repos/rogerthat-platform/rogerthat-android-client
|
closed
|
Speed up database initialization on fresh installs
|
process_duplicate type_feature
|
Update script from version 0 to version 73 + support for upgrading from 73 to 74, if 74 exists.
|
1.0
|
Speed up database initialization on fresh installs - Update script from version 0 to version 73 + support for upgrading from 73 to 74, if 74 exists.
|
process
|
speed up database initialization on fresh installs update script from version to version support for upgrading from to if exists
| 1
|
6,306
| 9,309,369,444
|
IssuesEvent
|
2019-03-25 16:20:06
|
googleapis/google-cloud-cpp
|
https://api.github.com/repos/googleapis/google-cloud-cpp
|
closed
|
Test driver scripts may miss errors.
|
type: process
|
The driver scripts for tests and examples use expressions like this:
```sh
local foo="blah-$(date +%s)"
```
These are a problem because errors in the shell expansion `$(...)` are not reported in this case. We can use:
```sh
local foo
foo="blah-$(date +%s)"
```
but those do not work with `local -r`. We can replace the "date as a source of random numbers" with `${RANDOM}` in some cases.
Regardless, we should search for such expressions and fix them appropriately.
Thanks to @dopiera for pointing this out.
|
1.0
|
Test driver scripts may miss errors. - The driver scripts for tests and examples use expressions like this:
```sh
local foo="blah-$(date +%s)"
```
These are a problem because errors in the shell expansion `$(...)` are not reported in this case. We can use:
```sh
local foo
foo="blah-$(date +%s)"
```
but those do not work with `local -r`. We can replace the "date as a source of random numbers" with `${RANDOM}` in some cases.
Regardless, we should search for such expressions and fix them appropriately.
Thanks to @dopiera for pointing this out.
|
process
|
test driver scripts may miss errors the driver scripts for tests and examples use expressions like this sh local foo blah date s these are a problem because errors in the shell expansion are not reported in this case we can use sh local foo foo blah date s but those do not work with local r we can replace the date as a source of random numbers with random in some cases regardless we should search for such expressions and fix them appropriately thanks to dopiera for pointing this out
| 1
|
1,674
| 4,312,642,471
|
IssuesEvent
|
2016-07-22 06:56:57
|
Jumpscale/jscockpit
|
https://api.github.com/repos/Jumpscale/jscockpit
|
reopened
|
Install a BP to deploy a CS which is already deployed doesn't update its properties
|
process_wontfix type_bug
|
**Steps**
- Create a CS in du-conv-2 with name "checkdeleteissue". Its ID is 1647

- Create a new Repo in a cockpit and enter this bp which will install a CS with "checkdeleteissue" (same CS in the same env)
```
g8client__dubai:
g8.url: 'du-conv-2.demo.greenitglobe.com'
g8.login: 'gig'
g8.password: 'KrOe6gE9K5nCQdmretfXnj'
g8.account: 'gig'
vdc__checkdeleteissue:
g8.client.name: 'dubai'
maxMemoryCapacity: 2
maxVDiskCapacity: 10
maxCPUCapacity: 2
maxNASCapacity: 20
maxArchiveCapacity: 20
maxNetworkOptTransfer: 5
maxNetworkPeerTransfer: 15
maxNumPublicIP: 1
```
- Execute BP
- Init Repo
- Install

- Check the Repo instance

**Expected result**
Cockpit should check if this CS is already deployed and if yes, It must produce an error message.
|
1.0
|
Install a BP to deploy a CS which is already deployed doesn't update its properties - **Steps**
- Create a CS in du-conv-2 with name "checkdeleteissue". Its ID is 1647

- Create a new Repo in a cockpit and enter this bp which will install a CS with "checkdeleteissue" (same CS in the same env)
```
g8client__dubai:
g8.url: 'du-conv-2.demo.greenitglobe.com'
g8.login: 'gig'
g8.password: 'KrOe6gE9K5nCQdmretfXnj'
g8.account: 'gig'
vdc__checkdeleteissue:
g8.client.name: 'dubai'
maxMemoryCapacity: 2
maxVDiskCapacity: 10
maxCPUCapacity: 2
maxNASCapacity: 20
maxArchiveCapacity: 20
maxNetworkOptTransfer: 5
maxNetworkPeerTransfer: 15
maxNumPublicIP: 1
```
- Execute BP
- Init Repo
- Install

- Check the Repo instance

**Expected result**
Cockpit should check if this CS is already deployed and if yes, It must produce an error message.
|
process
|
install a bp to deploy a cs which is already deployed doesn t update its properties steps create a cs in du conv with name checkdeleteissue its id is create a new repo in a cockpit and enter this bp which will install a cs with checkdeleteissue same cs in the same env dubai url du conv demo greenitglobe com login gig password account gig vdc checkdeleteissue client name dubai maxmemorycapacity maxvdiskcapacity maxcpucapacity maxnascapacity maxarchivecapacity maxnetworkopttransfer maxnetworkpeertransfer maxnumpublicip execute bp init repo install check the repo instance expected result cockpit should check if this cs is already deployed and if yes it must produce an error message
| 1
|
236,193
| 18,074,622,066
|
IssuesEvent
|
2021-09-21 08:29:10
|
jupyterhub/binderhub
|
https://api.github.com/repos/jupyterhub/binderhub
|
closed
|
Going from z2jh to z2bh
|
enhancement documentation
|
Related to #611 and https://github.com/jupyterhub/zero-to-jupyterhub-k8s/issues/643
@aculich and I discussed working through the tutorials in the following order:
1. littlest jupyterhub on GCP
2. zero to jupyterhub
3. zero to binderhub
@betatim and @matthew-brett @LotteWeerts started working through this yesterday and we have some ideas on how to document this.
It would be useful to document the best way to transition from each tutorial so you don't destroy your session each time.
|
1.0
|
Going from z2jh to z2bh - Related to #611 and https://github.com/jupyterhub/zero-to-jupyterhub-k8s/issues/643
@aculich and I discussed working through the tutorials in the following order:
1. littlest jupyterhub on GCP
2. zero to jupyterhub
3. zero to binderhub
@betatim and @matthew-brett @LotteWeerts started working through this yesterday and we have some ideas on how to document this.
It would be useful to document the best way to transition from each tutorial so you don't destroy your session each time.
|
non_process
|
going from to related to and aculich and i discussed working through the tutorials in the following order littlest jupyterhub on gcp zero to jupyterhub zero to binderhub betatim and matthew brett lotteweerts started working through this yesterday and we have some ideas on how to document this it would be useful to document the best way to transition from each tutorial so you don t destroy your session each time
| 0
|
302,559
| 26,150,595,593
|
IssuesEvent
|
2022-12-30 13:00:13
|
nupac/nupac
|
https://api.github.com/repos/nupac/nupac
|
opened
|
Add tests with `--long` flag
|
good first issue priority: low tests
|
### Describe the issue
There are no tests using the `--long` flag, which means we don't test if it raises some errors. There's no need for robust parsing, just test if each command works just fine with the flag — basically `ASSERT_NO_THROW`
### Argumentation
We don't test if the code actually works with this flag
### Screenshots and other helpful media
_No response_
### Additional context
_No response_
|
1.0
|
Add tests with `--long` flag - ### Describe the issue
There are no tests using the `--long` flag, which means we don't test if it raises some errors. There's no need for robust parsing, just test if each command works just fine with the flag — basically `ASSERT_NO_THROW`
### Argumentation
We don't test if the code actually works with this flag
### Screenshots and other helpful media
_No response_
### Additional context
_No response_
|
non_process
|
add tests with long flag describe the issue there are no tests using the long flag which means we don t test if it raises some errors there s no need for robust parsing just test if each command works just fine with the flag — basically assert no throw argumentation we don t test if the code actually works with this flag screenshots and other helpful media no response additional context no response
| 0
|
3,336
| 6,470,674,941
|
IssuesEvent
|
2017-08-17 09:37:15
|
openvstorage/framework
|
https://api.github.com/repos/openvstorage/framework
|
closed
|
Add ability to search for a vdisk based on the name and vpool
|
process_duplicate
|
Currently we only have a limited number of ways to search for vdisks. An awesome, additional way would be to be able to search with the name and vpool of a vdisk. Now we have to loop over all vdisks and compare the name to find the matching disk.
|
1.0
|
Add ability to search for a vdisk based on the name and vpool - Currently we only have a limited number of ways to search for vdisks. An awesome, additional way would be to be able to search with the name and vpool of a vdisk. Now we have to loop over all vdisks and compare the name to find the matching disk.
|
process
|
add ability to search for a vdisk based on the name and vpool currently we only have a limited number of ways to search for vdisks an awesome additional way would be to be able to search with the name and vpool of a vdisk now we have to loop over all vdisks and compare the name to find the matching disk
| 1
|
3,734
| 6,733,143,621
|
IssuesEvent
|
2017-10-18 13:58:51
|
york-region-tpss/stp
|
https://api.github.com/repos/york-region-tpss/stp
|
closed
|
Warranty Assessment Species Analysis - View
|
enhancement process workflow report
|
Create a report to display the species analysis for a certain warranty period
|
1.0
|
Warranty Assessment Species Analysis - View - Create a report to display the species analysis for a certain warranty period
|
process
|
warranty assessment species analysis view create a report to display the species analysis for a certain warranty period
| 1
|
927
| 3,389,328,879
|
IssuesEvent
|
2015-11-30 00:37:42
|
MaretEngineering/MROV
|
https://api.github.com/repos/MaretEngineering/MROV
|
opened
|
Be able to turn off Serial by changing one constant
|
Necessary Addition Processing
|
All the lines that refer to serial should be inside if statements that refer to that constant.
|
1.0
|
Be able to turn off Serial by changing one constant - All the lines that refer to serial should be inside if statements that refer to that constant.
|
process
|
be able to turn off serial by changing one constant all the lines that refer to serial should be inside if statements that refer to that constant
| 1
|
117,941
| 15,193,346,814
|
IssuesEvent
|
2021-02-16 00:27:29
|
darktable-org/darktable
|
https://api.github.com/repos/darktable-org/darktable
|
closed
|
Limit output resolution/size from darkroom
|
feature: redesign no-issue-activity priority: low scope: UI
|
From https://redmine.darktable.org/issues/12351.
When processing noisy images (e.g. ISO 6400), I do not export with full pixel count but limit to a reasonable size that hides the noise a bit. The same holds for slightly unsharp images. That way, the maximum usable image size, e.g. maximum print size, is given by the image resolution and limited such that the technical flaws are not visible for the people viewing the picture.
Unfortunately, the output size/resolution is determined in lighttable mode and the image itself does not carry this information. When I export an old picture again, I maybe forget that it was a blurry shot and therefore it is sent out at e.g. full resolution. Furthermore, when processing a set of images, one always has to export one by one to account for the blurry/noisy shots. It would therefore be great if one could limit the resolution from within darkroom such that is cannot be overridden by the export module (smaller export would still be possible). A new module “limit output size” would be a reasonable way to implement this feature. Within the module, the output size limit could be selected as either (maybe a dropdown to choose among these)
* absolute pixel numbers (input of two numbers, max. horizontal and vertical pixel count; dependent on the aspect ratio, one of the numbers wins, as in the export module),
* megapixel count (input one number, e.g. 10M, and the closest smaller size that would match the aspect ratio is used), or
* factor of the original size (e.g. “2” for downscaling to half the size; factor 2 would therefore mean quarter of the pixel count; or, instead, the factor of the pixel count to avoid the necessity of sqrt(2) input; or maybe both options).
The darkroom view would of course follow the limitation to have a reasonable preview. The module itself would also show the actual output size limit of the image.
I thought a bit more about that idea, and came to the conclusion that there is an easy part and a difficult one. The easy part would be to implement a factor of the original size. This could, in the UI, just be a text entry or a slider. A widget that computes and shows the resulting image size would be convenient. The benefit is that this does not interfere with changes in the crop. It would only require one additional entry in the xmp/data base, which is the scaling factor as a float number.
The other two ways to specify the scaling have the issue that they fail for strong crops, where you may even go below the specified output size. Therefore, for the other two options, the UI would have to reflect what it refers to, the original image size or the crop (a check box named “referred to the crop size” which is initially not checked, and an “upscaling allowed” check box.
The first, easy, solution would already help a lot, so the implementation of this feature would not have to go the entire way at a time.
Slightly related: https://redmine.darktable.org/issues/12234
|
1.0
|
Limit output resolution/size from darkroom - From https://redmine.darktable.org/issues/12351.
When processing noisy images (e.g. ISO 6400), I do not export with full pixel count but limit to a reasonable size that hides the noise a bit. The same holds for slightly unsharp images. That way, the maximum usable image size, e.g. maximum print size, is given by the image resolution and limited such that the technical flaws are not visible for the people viewing the picture.
Unfortunately, the output size/resolution is determined in lighttable mode and the image itself does not carry this information. When I export an old picture again, I maybe forget that it was a blurry shot and therefore it is sent out at e.g. full resolution. Furthermore, when processing a set of images, one always has to export one by one to account for the blurry/noisy shots. It would therefore be great if one could limit the resolution from within darkroom such that is cannot be overridden by the export module (smaller export would still be possible). A new module “limit output size” would be a reasonable way to implement this feature. Within the module, the output size limit could be selected as either (maybe a dropdown to choose among these)
* absolute pixel numbers (input of two numbers, max. horizontal and vertical pixel count; dependent on the aspect ratio, one of the numbers wins, as in the export module),
* megapixel count (input one number, e.g. 10M, and the closest smaller size that would match the aspect ratio is used), or
* factor of the original size (e.g. “2” for downscaling to half the size; factor 2 would therefore mean quarter of the pixel count; or, instead, the factor of the pixel count to avoid the necessity of sqrt(2) input; or maybe both options).
The darkroom view would of course follow the limitation to have a reasonable preview. The module itself would also show the actual output size limit of the image.
I thought a bit more about that idea, and came to the conclusion that there is an easy part and a difficult one. The easy part would be to implement a factor of the original size. This could, in the UI, just be a text entry or a slider. A widget that computes and shows the resulting image size would be convenient. The benefit is that this does not interfere with changes in the crop. It would only require one additional entry in the xmp/data base, which is the scaling factor as a float number.
The other two ways to specify the scaling have the issue that they fail for strong crops, where you may even go below the specified output size. Therefore, for the other two options, the UI would have to reflect what it refers to, the original image size or the crop (a check box named “referred to the crop size” which is initially not checked, and an “upscaling allowed” check box.
The first, easy, solution would already help a lot, so the implementation of this feature would not have to go the entire way at a time.
Slightly related: https://redmine.darktable.org/issues/12234
|
non_process
|
limit output resolution size from darkroom from when processing noisy images e g iso i do not export with full pixel count but limit to a reasonable size that hides the noise a bit the same holds for slightly unsharp images that way the maximum usable image size e g maximum print size is given by the image resolution and limited such that the technical flaws are not visible for the people viewing the picture unfortunately the output size resolution is determined in lighttable mode and the image itself does not carry this information when i export an old picture again i maybe forget that it was a blurry shot and therefore it is sent out at e g full resolution furthermore when processing a set of images one always has to export one by one to account for the blurry noisy shots it would therefore be great if one could limit the resolution from within darkroom such that is cannot be overridden by the export module smaller export would still be possible a new module “limit output size” would be a reasonable way to implement this feature within the module the output size limit could be selected as either maybe a dropdown to choose among these absolute pixel numbers input of two numbers max horizontal and vertical pixel count dependent on the aspect ratio one of the numbers wins as in the export module megapixel count input one number e g and the closest smaller size that would match the aspect ratio is used or factor of the original size e g “ ” for downscaling to half the size factor would therefore mean quarter of the pixel count or instead the factor of the pixel count to avoid the necessity of sqrt input or maybe both options the darkroom view would of course follow the limitation to have a reasonable preview the module itself would also show the actual output size limit of the image i thought a bit more about that idea and came to the conclusion that there is an easy part and a difficult one the easy part would be to implement a factor of the original size this could in the ui just be a text entry or a slider a widget that computes and shows the resulting image size would be convenient the benefit is that this does not interfere with changes in the crop it would only require one additional entry in the xmp data base which is the scaling factor as a float number the other two ways to specify the scaling have the issue that they fail for strong crops where you may even go below the specified output size therefore for the other two options the ui would have to reflect what it refers to the original image size or the crop a check box named “referred to the crop size” which is initially not checked and an “upscaling allowed” check box the first easy solution would already help a lot so the implementation of this feature would not have to go the entire way at a time slightly related
| 0
|
43,319
| 9,417,819,382
|
IssuesEvent
|
2019-04-10 17:40:34
|
grpc/grpc-java
|
https://api.github.com/repos/grpc/grpc-java
|
closed
|
Releases can be uploaded without signing being successful
|
code health
|
We configure signing to be required, but it isn't working for some reason.
|
1.0
|
Releases can be uploaded without signing being successful - We configure signing to be required, but it isn't working for some reason.
|
non_process
|
releases can be uploaded without signing being successful we configure signing to be required but it isn t working for some reason
| 0
|
15,230
| 19,101,357,993
|
IssuesEvent
|
2021-11-29 23:05:26
|
ooi-data/CE09OSSM-MFD37-03-DOSTAD000-telemetered-dosta_abcdjm_ctdbp_dcl_instrument
|
https://api.github.com/repos/ooi-data/CE09OSSM-MFD37-03-DOSTAD000-telemetered-dosta_abcdjm_ctdbp_dcl_instrument
|
closed
|
🛑 Processing failed: ResponseParserError
|
process
|
## Overview
`ResponseParserError` found in `processing_task` task during run ended on 2021-05-05T16:35:55.842847.
## Details
Flow name: `CE09OSSM-MFD37-03-DOSTAD000-telemetered-dosta_abcdjm_ctdbp_dcl_instrument`
Task name: `processing_task`
Error type: `ResponseParserError`
Error message: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed:
b'<?xml version="1.0" encoding="UTF-8"?>\n'
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 452, in _parse_xml_string_to_dom
root = parser.close()
File "<string>", line None
xml.etree.ElementTree.ParseError: no element found: line 2, column 0
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/share/miniconda/envs/harvester/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 71, in processing_task
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 311, in finalize_zarr
source_store.fs.delete(source_store.root, recursive=True)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1151, in delete
return self.rm(path, recursive=recursive, maxdepth=maxdepth)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 72, in wrapper
return sync(self.loop, func, *args, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 53, in sync
raise result[0]
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 20, in _runner
result[0] = await coro
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1510, in _rm
await asyncio.gather(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1498, in _bulk_delete
await self._call_s3("delete_objects", kwargs, Bucket=bucket, Delete=delete_keys)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 252, in _call_s3
raise translate_boto_error(err)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 233, in _call_s3
out = await method(**additional_kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 140, in _make_api_call
http, parsed_response = await self._make_request(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 160, in _make_request
return await self._endpoint.make_request(operation_model, request_dict)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 101, in _send_request
success_response, exception = await self._get_response(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 120, in _get_response
success_response, exception = await self._do_get_response(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 180, in _do_get_response
parsed_response = parser.parse(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 245, in parse
parsed = self._do_parse(response, shape)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 809, in _do_parse
self._add_modeled_parse(response, shape, final_parsed)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 818, in _add_modeled_parse
self._parse_payload(response, shape, member_shapes, final_parsed)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 858, in _parse_payload
original_parsed = self._initial_body_parse(response['body'])
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 944, in _initial_body_parse
return self._parse_xml_string_to_dom(xml_string)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 454, in _parse_xml_string_to_dom
raise ResponseParserError(
botocore.parsers.ResponseParserError: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed:
b'<?xml version="1.0" encoding="UTF-8"?>\n'
```
</details>
|
1.0
|
🛑 Processing failed: ResponseParserError - ## Overview
`ResponseParserError` found in `processing_task` task during run ended on 2021-05-05T16:35:55.842847.
## Details
Flow name: `CE09OSSM-MFD37-03-DOSTAD000-telemetered-dosta_abcdjm_ctdbp_dcl_instrument`
Task name: `processing_task`
Error type: `ResponseParserError`
Error message: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed:
b'<?xml version="1.0" encoding="UTF-8"?>\n'
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 452, in _parse_xml_string_to_dom
root = parser.close()
File "<string>", line None
xml.etree.ElementTree.ParseError: no element found: line 2, column 0
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/usr/share/miniconda/envs/harvester/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 71, in processing_task
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 311, in finalize_zarr
source_store.fs.delete(source_store.root, recursive=True)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1151, in delete
return self.rm(path, recursive=recursive, maxdepth=maxdepth)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 72, in wrapper
return sync(self.loop, func, *args, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 53, in sync
raise result[0]
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 20, in _runner
result[0] = await coro
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1510, in _rm
await asyncio.gather(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1498, in _bulk_delete
await self._call_s3("delete_objects", kwargs, Bucket=bucket, Delete=delete_keys)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 252, in _call_s3
raise translate_boto_error(err)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 233, in _call_s3
out = await method(**additional_kwargs)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 140, in _make_api_call
http, parsed_response = await self._make_request(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 160, in _make_request
return await self._endpoint.make_request(operation_model, request_dict)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 101, in _send_request
success_response, exception = await self._get_response(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 120, in _get_response
success_response, exception = await self._do_get_response(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 180, in _do_get_response
parsed_response = parser.parse(
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 245, in parse
parsed = self._do_parse(response, shape)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 809, in _do_parse
self._add_modeled_parse(response, shape, final_parsed)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 818, in _add_modeled_parse
self._parse_payload(response, shape, member_shapes, final_parsed)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 858, in _parse_payload
original_parsed = self._initial_body_parse(response['body'])
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 944, in _initial_body_parse
return self._parse_xml_string_to_dom(xml_string)
File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 454, in _parse_xml_string_to_dom
raise ResponseParserError(
botocore.parsers.ResponseParserError: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed:
b'<?xml version="1.0" encoding="UTF-8"?>\n'
```
</details>
|
process
|
🛑 processing failed responseparsererror overview responseparsererror found in processing task task during run ended on details flow name telemetered dosta abcdjm ctdbp dcl instrument task name processing task error type responseparsererror error message unable to parse response no element found line column invalid xml received further retries may succeed b n traceback traceback most recent call last file srv conda envs notebook lib site packages botocore parsers py line in parse xml string to dom root parser close file line none xml etree elementtree parseerror no element found line column during handling of the above exception another exception occurred traceback most recent call last file usr share miniconda envs harvester lib site packages ooi harvester processor pipeline py line in processing task file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize zarr source store fs delete source store root recursive true file srv conda envs notebook lib site packages fsspec spec py line in delete return self rm path recursive recursive maxdepth maxdepth file srv conda envs notebook lib site packages fsspec asyn py line in wrapper return sync self loop func args kwargs file srv conda envs notebook lib site packages fsspec asyn py line in sync raise result file srv conda envs notebook lib site packages fsspec asyn py line in runner result await coro file srv conda envs notebook lib site packages core py line in rm await asyncio gather file srv conda envs notebook lib site packages core py line in bulk delete await self call delete objects kwargs bucket bucket delete delete keys file srv conda envs notebook lib site packages core py line in call raise translate boto error err file srv conda envs notebook lib site packages core py line in call out await method additional kwargs file srv conda envs notebook lib site packages aiobotocore client py line in make api call http parsed response await self make request file srv conda envs notebook lib site packages aiobotocore client py line in make request return await self endpoint make request operation model request dict file srv conda envs notebook lib site packages aiobotocore endpoint py line in send request success response exception await self get response file srv conda envs notebook lib site packages aiobotocore endpoint py line in get response success response exception await self do get response file srv conda envs notebook lib site packages aiobotocore endpoint py line in do get response parsed response parser parse file srv conda envs notebook lib site packages botocore parsers py line in parse parsed self do parse response shape file srv conda envs notebook lib site packages botocore parsers py line in do parse self add modeled parse response shape final parsed file srv conda envs notebook lib site packages botocore parsers py line in add modeled parse self parse payload response shape member shapes final parsed file srv conda envs notebook lib site packages botocore parsers py line in parse payload original parsed self initial body parse response file srv conda envs notebook lib site packages botocore parsers py line in initial body parse return self parse xml string to dom xml string file srv conda envs notebook lib site packages botocore parsers py line in parse xml string to dom raise responseparsererror botocore parsers responseparsererror unable to parse response no element found line column invalid xml received further retries may succeed b n
| 1
|
212,285
| 16,437,903,573
|
IssuesEvent
|
2021-05-20 11:22:29
|
ethersphere/bee
|
https://api.github.com/repos/ethersphere/bee
|
closed
|
TestNeighborhoodDepth flakes
|
flaky-test issue
|
```
=== RUN TestNeighborhoodDepth
kademlia_test.go:142: timed out waiting for depth. want 14 got 15
--- FAIL: TestNeighborhoodDepth (4.31s)
```
Probably related to balanced kademlia changes
|
1.0
|
TestNeighborhoodDepth flakes - ```
=== RUN TestNeighborhoodDepth
kademlia_test.go:142: timed out waiting for depth. want 14 got 15
--- FAIL: TestNeighborhoodDepth (4.31s)
```
Probably related to balanced kademlia changes
|
non_process
|
testneighborhooddepth flakes run testneighborhooddepth kademlia test go timed out waiting for depth want got fail testneighborhooddepth probably related to balanced kademlia changes
| 0
|
62,533
| 17,023,941,337
|
IssuesEvent
|
2021-07-03 04:41:07
|
tomhughes/trac-tickets
|
https://api.github.com/repos/tomhughes/trac-tickets
|
closed
|
OSM API v0.6 Reopening closed note bug
|
Component: api Priority: minor Resolution: invalid Type: defect
|
**[Submitted to the original trac issue database at 9.56am, Wednesday, 12th December 2018]**
when using the api `https://api.openstreetmap.org/api/0.6/notes/#id/reopen?text=Comment`,
if the text contains `=` error code 401 will be thrown.
while the comment api `https://api.openstreetmap.org/api/0.6/notes/#id/comment?text=ThisIsANoteComment`
does not have such restrictions
|
1.0
|
OSM API v0.6 Reopening closed note bug - **[Submitted to the original trac issue database at 9.56am, Wednesday, 12th December 2018]**
when using the api `https://api.openstreetmap.org/api/0.6/notes/#id/reopen?text=Comment`,
if the text contains `=` error code 401 will be thrown.
while the comment api `https://api.openstreetmap.org/api/0.6/notes/#id/comment?text=ThisIsANoteComment`
does not have such restrictions
|
non_process
|
osm api reopening closed note bug when using the api if the text contains error code will be thrown while the comment api does not have such restrictions
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.