Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
15,337
19,472,409,961
IssuesEvent
2021-12-24 05:01:21
home-climate-control/dz
https://api.github.com/repos/home-climate-control/dz
opened
Ignore 1-Wire sensor departure, just react to timeout
fault tolerance 1-wire reactive process control
### Existing Behavior 1-Wire sensor departure immediately propagates to the control system and makes it act prematurely, as jitter and short blackouts are possible. ### Desired Behavior If/when a 1-Wire sensor departs, the departure is not propagated down the control flow immediately, but only after a given timeout when the sensor is considered stale. ### Corollary Hardware failure notifications must be handled outside of control flow.
1.0
Ignore 1-Wire sensor departure, just react to timeout - ### Existing Behavior 1-Wire sensor departure immediately propagates to the control system and makes it act prematurely, as jitter and short blackouts are possible. ### Desired Behavior If/when a 1-Wire sensor departs, the departure is not propagated down the control flow immediately, but only after a given timeout when the sensor is considered stale. ### Corollary Hardware failure notifications must be handled outside of control flow.
process
ignore wire sensor departure just react to timeout existing behavior wire sensor departure immediately propagates to the control system and makes it act prematurely as jitter and short blackouts are possible desired behavior if when a wire sensor departs the departure is not propagated down the control flow immediately but only after a given timeout when the sensor is considered stale corollary hardware failure notifications must be handled outside of control flow
1
18,918
24,865,200,149
IssuesEvent
2022-10-27 11:22:38
alphagov/govuk-design-system
https://api.github.com/repos/alphagov/govuk-design-system
closed
Secure community facilitators for Design System Day 2022 workshops and breakout sessions
🕔 weeks process shared ownership
## What Approach the community to facilitate workshops and breakout sessions on Design System Day 2022. ### Workshops 90-minute sessions, usually involves a presentation and audience activities 5-7 facilitators needed to run their workshop on both days ### Breakout sessions 50-minute sessions, freeform — can be a discussion, activity, anything they'd like 14-20 facilitators needed over both days (breakout sessions are not repeated) ## Why So we can began planning the workshops and let attendees know what to expect. ## Who needs to work on this TBC ## Who needs to review this ## Dependencies - [x] A theme for the event has been agreed - [x] Decision needed: How can people propose ideas and volunteer to facilitate? - [x] Clarification needed: what's facilitators need to do ([see proposed schedule](https://govukdesignsystem.padlet.org/ImHuYorks/vntxtz9vv3do3uiz)) ## Done when Call for facilitators (week of 22 August?) - [x] Drafted comms - [x] Comms reviewed by at least one team member - [x] Published on appropriate platforms Selection process - [x] Gather and organise potential facilitators - [x] Choose facilitators For the event (done by 7 September) - [x] Confirm with facilitators - [x] Schedule facilitators, working with participant comms group
1.0
Secure community facilitators for Design System Day 2022 workshops and breakout sessions - ## What Approach the community to facilitate workshops and breakout sessions on Design System Day 2022. ### Workshops 90-minute sessions, usually involves a presentation and audience activities 5-7 facilitators needed to run their workshop on both days ### Breakout sessions 50-minute sessions, freeform — can be a discussion, activity, anything they'd like 14-20 facilitators needed over both days (breakout sessions are not repeated) ## Why So we can began planning the workshops and let attendees know what to expect. ## Who needs to work on this TBC ## Who needs to review this ## Dependencies - [x] A theme for the event has been agreed - [x] Decision needed: How can people propose ideas and volunteer to facilitate? - [x] Clarification needed: what's facilitators need to do ([see proposed schedule](https://govukdesignsystem.padlet.org/ImHuYorks/vntxtz9vv3do3uiz)) ## Done when Call for facilitators (week of 22 August?) - [x] Drafted comms - [x] Comms reviewed by at least one team member - [x] Published on appropriate platforms Selection process - [x] Gather and organise potential facilitators - [x] Choose facilitators For the event (done by 7 September) - [x] Confirm with facilitators - [x] Schedule facilitators, working with participant comms group
process
secure community facilitators for design system day workshops and breakout sessions what approach the community to facilitate workshops and breakout sessions on design system day workshops minute sessions usually involves a presentation and audience activities facilitators needed to run their workshop on both days breakout sessions minute sessions freeform — can be a discussion activity anything they d like facilitators needed over both days breakout sessions are not repeated why so we can began planning the workshops and let attendees know what to expect who needs to work on this tbc who needs to review this dependencies a theme for the event has been agreed decision needed how can people propose ideas and volunteer to facilitate clarification needed what s facilitators need to do done when call for facilitators week of august drafted comms comms reviewed by at least one team member published on appropriate platforms selection process gather and organise potential facilitators choose facilitators for the event done by september confirm with facilitators schedule facilitators working with participant comms group
1
8,048
11,220,764,528
IssuesEvent
2020-01-07 16:25:43
code4romania/expert-consultation-api
https://api.github.com/repos/code4romania/expert-consultation-api
opened
Unable to load document path that contains spaces
bug document processing documents good first issue help wanted
When uploading a new document in the platform, of the document URL contains spaces, the document upload fails with the following message: { "i18nErrors": null, "i18nFieldErrors": null, **"additionalInfo": "Index 0 out of bounds for length 0"** } The initial DTO for creating the new document is the following: { "dateOfDevelopment": "2020-01-01T16:02:03.790Z", "dateOfReceipt": "2020-01-01T16:02:03.790Z", "documentInitializer": "Ministerul Dezvoltarii", "documentNumber": 876345354, "documentTitle": "Document de test 2", "documentType": "LEGE", "filePath": **"/Users/radustefanescu/legal_consult_uploads/MmujDzZFpwregulament_grupuri_de_lucru_pdf_1520247820 (1).pdf"** }
1.0
Unable to load document path that contains spaces - When uploading a new document in the platform, of the document URL contains spaces, the document upload fails with the following message: { "i18nErrors": null, "i18nFieldErrors": null, **"additionalInfo": "Index 0 out of bounds for length 0"** } The initial DTO for creating the new document is the following: { "dateOfDevelopment": "2020-01-01T16:02:03.790Z", "dateOfReceipt": "2020-01-01T16:02:03.790Z", "documentInitializer": "Ministerul Dezvoltarii", "documentNumber": 876345354, "documentTitle": "Document de test 2", "documentType": "LEGE", "filePath": **"/Users/radustefanescu/legal_consult_uploads/MmujDzZFpwregulament_grupuri_de_lucru_pdf_1520247820 (1).pdf"** }
process
unable to load document path that contains spaces when uploading a new document in the platform of the document url contains spaces the document upload fails with the following message null null additionalinfo index out of bounds for length the initial dto for creating the new document is the following dateofdevelopment dateofreceipt documentinitializer ministerul dezvoltarii documentnumber documenttitle document de test documenttype lege filepath users radustefanescu legal consult uploads mmujdzzfpwregulament grupuri de lucru pdf pdf
1
15,731
19,905,560,906
IssuesEvent
2022-01-25 12:26:03
prometheus-community/windows_exporter
https://api.github.com/repos/prometheus-community/windows_exporter
closed
How to write a process restart alarm rule?
question collector/process
A process restart can be indicated by windows_process_start_time. If the value of this indicator changes, it indicates that the process is restarted. For example: windows_process_start_time{process="ACDG",instance="10.35.236.40"}, what if this cannot be greater than or less than?
1.0
How to write a process restart alarm rule? - A process restart can be indicated by windows_process_start_time. If the value of this indicator changes, it indicates that the process is restarted. For example: windows_process_start_time{process="ACDG",instance="10.35.236.40"}, what if this cannot be greater than or less than?
process
how to write a process restart alarm rule a process restart can be indicated by windows process start time if the value of this indicator changes it indicates that the process is restarted for example windows process start time process acdg instance what if this cannot be greater than or less than
1
75,790
9,886,000,605
IssuesEvent
2019-06-25 05:07:27
getgauge/taiko
https://api.github.com/repos/getgauge/taiko
closed
Need to fix taiko doc
documentation
Observation: 1. Under overview there are few texts which appears as link we need to rectify. Example: sample project <img width="894" alt="Screen Shot 2019-06-14 at 3 46 31 PM" src="https://user-images.githubusercontent.com/46309201/59582980-88536780-90f7-11e9-8c23-b4ad51b2519b.png"> 2. cosmetic issue: <img width="841" alt="Screen Shot 2019-06-14 at 3 47 15 PM" src="https://user-images.githubusercontent.com/46309201/59583006-986b4700-90f7-11e9-9e27-1c89516cd28f.png"> 3. ```TAIKO_EMULATE_NETWORK``` env vaiable is misssing under https://taiko-preview.gauge.org/?#taiko-env-variables <img width="820" alt="Screen Shot 2019-06-14 at 3 52 06 PM" src="https://user-images.githubusercontent.com/46309201/59583015-a4ef9f80-90f7-11e9-9145-3eac57fab593.png">
1.0
Need to fix taiko doc - Observation: 1. Under overview there are few texts which appears as link we need to rectify. Example: sample project <img width="894" alt="Screen Shot 2019-06-14 at 3 46 31 PM" src="https://user-images.githubusercontent.com/46309201/59582980-88536780-90f7-11e9-8c23-b4ad51b2519b.png"> 2. cosmetic issue: <img width="841" alt="Screen Shot 2019-06-14 at 3 47 15 PM" src="https://user-images.githubusercontent.com/46309201/59583006-986b4700-90f7-11e9-9e27-1c89516cd28f.png"> 3. ```TAIKO_EMULATE_NETWORK``` env vaiable is misssing under https://taiko-preview.gauge.org/?#taiko-env-variables <img width="820" alt="Screen Shot 2019-06-14 at 3 52 06 PM" src="https://user-images.githubusercontent.com/46309201/59583015-a4ef9f80-90f7-11e9-9145-3eac57fab593.png">
non_process
need to fix taiko doc observation under overview there are few texts which appears as link we need to rectify example sample project img width alt screen shot at pm src cosmetic issue img width alt screen shot at pm src taiko emulate network env vaiable is misssing under img width alt screen shot at pm src
0
15,624
19,770,228,787
IssuesEvent
2022-01-17 09:17:19
skellig-framework/skellig-core
https://api.github.com/repos/skellig-framework/skellig-core
closed
Modify date and time from current date
processing
Add ability to modify current date and time, by adding a function which can go after `now()`, ex: `now().plusDays(1)`
1.0
Modify date and time from current date - Add ability to modify current date and time, by adding a function which can go after `now()`, ex: `now().plusDays(1)`
process
modify date and time from current date add ability to modify current date and time by adding a function which can go after now ex now plusdays
1
45,635
24,147,423,805
IssuesEvent
2022-09-21 20:08:35
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Expression.Compile performance regression on .NET7 RC1
tenet-performance area-VM-coreclr
<!--This is just a template - feel free to delete any and all of it and replace as appropriate.--> ### Description Running simple benchmark to test performance of Expression.Compile shows regressions on .NET7 RC1 <details> <summary>Benchmark code</summary> ```c# using System; using System.Linq.Expressions; using BenchmarkDotNet.Attributes; using BenchmarkDotNet.Configs; using BenchmarkDotNet.Jobs; using BenchmarkDotNet.Running; BenchmarkRunner.Run<Benchmarks>(); [MemoryDiagnoser(false)] [SimpleJob(RuntimeMoniker.Net60)] [SimpleJob(RuntimeMoniker.Net70)] [GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByJob)] public class Benchmarks { private static readonly Expression<Func<MyClass>> _factory = FactoryCreator<MyClass>.GetFactory(); [Benchmark] public Func<MyClass> CompileFactory() { return _factory.Compile(); } } public static class FactoryCreator<T> { public static Expression<Func<T>> GetFactory() { var ctorBody = Expression.New(typeof(T).GetConstructor(Type.EmptyTypes)!); return Expression.Lambda<Func<T>>(ctorBody); } } public class MyClass {} ``` </details> ### Regression? Yes. On .NET 6 this is ~50% faster ### Data ``` ini BenchmarkDotNet=v0.13.2, OS=Windows 11 (10.0.22000.978/21H2) AMD Ryzen 5 3600, 1 CPU, 12 logical and 6 physical cores .NET SDK=7.0.100-rc.1.22431.12 [Host] : .NET 7.0.0 (7.0.22.42610), X64 RyuJIT AVX2 .NET 6.0 : .NET 6.0.9 (6.0.922.41905), X64 RyuJIT AVX2 .NET 7.0 : .NET 7.0.0 (7.0.22.42610), X64 RyuJIT AVX2 ``` | Method | Job | Runtime | Mean | Error | StdDev | Allocated | |--------------- |--------- |--------- |---------:|---------:|---------:|----------:| | CompileFactory | .NET 6.0 | .NET 6.0 | 33.51 μs | 0.361 μs | 0.337 μs | 3.66 KB | | | | | | | | | | CompileFactory | .NET 7.0 | .NET 7.0 | 51.86 μs | 0.223 μs | 0.208 μs | 3.69 KB | <!-- * Please include any benchmark results, images of graphs, timings or measurements, or callstacks that are relevant. * If possible please include text as text rather than images (so it shows up in searches). * If applicable please include before and after measurements. * There is helpful information about measuring code in this repo [here](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md). -->
True
Expression.Compile performance regression on .NET7 RC1 - <!--This is just a template - feel free to delete any and all of it and replace as appropriate.--> ### Description Running simple benchmark to test performance of Expression.Compile shows regressions on .NET7 RC1 <details> <summary>Benchmark code</summary> ```c# using System; using System.Linq.Expressions; using BenchmarkDotNet.Attributes; using BenchmarkDotNet.Configs; using BenchmarkDotNet.Jobs; using BenchmarkDotNet.Running; BenchmarkRunner.Run<Benchmarks>(); [MemoryDiagnoser(false)] [SimpleJob(RuntimeMoniker.Net60)] [SimpleJob(RuntimeMoniker.Net70)] [GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByJob)] public class Benchmarks { private static readonly Expression<Func<MyClass>> _factory = FactoryCreator<MyClass>.GetFactory(); [Benchmark] public Func<MyClass> CompileFactory() { return _factory.Compile(); } } public static class FactoryCreator<T> { public static Expression<Func<T>> GetFactory() { var ctorBody = Expression.New(typeof(T).GetConstructor(Type.EmptyTypes)!); return Expression.Lambda<Func<T>>(ctorBody); } } public class MyClass {} ``` </details> ### Regression? Yes. On .NET 6 this is ~50% faster ### Data ``` ini BenchmarkDotNet=v0.13.2, OS=Windows 11 (10.0.22000.978/21H2) AMD Ryzen 5 3600, 1 CPU, 12 logical and 6 physical cores .NET SDK=7.0.100-rc.1.22431.12 [Host] : .NET 7.0.0 (7.0.22.42610), X64 RyuJIT AVX2 .NET 6.0 : .NET 6.0.9 (6.0.922.41905), X64 RyuJIT AVX2 .NET 7.0 : .NET 7.0.0 (7.0.22.42610), X64 RyuJIT AVX2 ``` | Method | Job | Runtime | Mean | Error | StdDev | Allocated | |--------------- |--------- |--------- |---------:|---------:|---------:|----------:| | CompileFactory | .NET 6.0 | .NET 6.0 | 33.51 μs | 0.361 μs | 0.337 μs | 3.66 KB | | | | | | | | | | CompileFactory | .NET 7.0 | .NET 7.0 | 51.86 μs | 0.223 μs | 0.208 μs | 3.69 KB | <!-- * Please include any benchmark results, images of graphs, timings or measurements, or callstacks that are relevant. * If possible please include text as text rather than images (so it shows up in searches). * If applicable please include before and after measurements. * There is helpful information about measuring code in this repo [here](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md). -->
non_process
expression compile performance regression on description running simple benchmark to test performance of expression compile shows regressions on benchmark code c using system using system linq expressions using benchmarkdotnet attributes using benchmarkdotnet configs using benchmarkdotnet jobs using benchmarkdotnet running benchmarkrunner run public class benchmarks private static readonly expression factory factorycreator getfactory public func compilefactory return factory compile public static class factorycreator public static expression getfactory var ctorbody expression new typeof t getconstructor type emptytypes return expression lambda ctorbody public class myclass regression yes on net this is faster data ini benchmarkdotnet os windows amd ryzen cpu logical and physical cores net sdk rc net ryujit net net ryujit net net ryujit method job runtime mean error stddev allocated compilefactory net net μs μs μs kb compilefactory net net μs μs μs kb please include any benchmark results images of graphs timings or measurements or callstacks that are relevant if possible please include text as text rather than images so it shows up in searches if applicable please include before and after measurements there is helpful information about measuring code in this repo
0
195,180
6,904,978,482
IssuesEvent
2017-11-27 03:45:06
AIE-2017-Yr1-Group1/cultist-game
https://api.github.com/repos/AIE-2017-Yr1-Group1/cultist-game
closed
Cultists are firing at Demorial's feet
bug priority-1
Probably because the lookat fix for offset fireballs is looking at Demorial position, which is 0,0,0 at his feet.
1.0
Cultists are firing at Demorial's feet - Probably because the lookat fix for offset fireballs is looking at Demorial position, which is 0,0,0 at his feet.
non_process
cultists are firing at demorial s feet probably because the lookat fix for offset fireballs is looking at demorial position which is at his feet
0
160,052
20,092,552,107
IssuesEvent
2022-02-06 01:28:25
PGreaneyLYIT/easybuggy4django
https://api.github.com/repos/PGreaneyLYIT/easybuggy4django
opened
CVE-2019-8331 (Medium) detected in bootstrap-3.3.7.min.js
security vulnerability
## CVE-2019-8331 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /templates/base.html</p> <p>Path to vulnerable library: /templates/base.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/PGreaneyLYIT/easybuggy4django/commit/5403b8fbcea4b699ce64d05146aeaa76d1062d89">5403b8fbcea4b699ce64d05146aeaa76d1062d89</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-8331 (Medium) detected in bootstrap-3.3.7.min.js - ## CVE-2019-8331 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /templates/base.html</p> <p>Path to vulnerable library: /templates/base.html</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/PGreaneyLYIT/easybuggy4django/commit/5403b8fbcea4b699ce64d05146aeaa76d1062d89">5403b8fbcea4b699ce64d05146aeaa76d1062d89</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-8331>CVE-2019-8331</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/twbs/bootstrap/pull/28236">https://github.com/twbs/bootstrap/pull/28236</a></p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file templates base html path to vulnerable library templates base html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch master vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap bootstrap sass step up your open source security game with whitesource
0
4,361
7,260,514,255
IssuesEvent
2018-02-18 10:53:23
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[FEATURE] [processing] New algorithm to compute geometry by expression
Processing User Manual
This algorithm updates existing geometries (or creates new geometries) for input features by use of a QGIS expression. This allows complex geometry modifications which can utilise all the flexibility of the QGIS expression engine to manipulate and create geometries for output features. see https://github.com/qgis/QGIS/pull/3692/commits/12ab6c6f6f77aea5cf5e31e9af81723d35a263cb
1.0
[FEATURE] [processing] New algorithm to compute geometry by expression - This algorithm updates existing geometries (or creates new geometries) for input features by use of a QGIS expression. This allows complex geometry modifications which can utilise all the flexibility of the QGIS expression engine to manipulate and create geometries for output features. see https://github.com/qgis/QGIS/pull/3692/commits/12ab6c6f6f77aea5cf5e31e9af81723d35a263cb
process
new algorithm to compute geometry by expression this algorithm updates existing geometries or creates new geometries for input features by use of a qgis expression this allows complex geometry modifications which can utilise all the flexibility of the qgis expression engine to manipulate and create geometries for output features see
1
776,863
27,264,724,273
IssuesEvent
2023-02-22 17:09:43
ascheid/itsg33-pbmm-issue-gen
https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen
opened
SA-11(5): Developer Security Testing And Evaluation | Penetration Testing / Analysis
Priority: P3 ITSG-33 Suggested Assignment: IT Projects Class: Management Control: SA-11
# Control Definition DEVELOPER SECURITY TESTING AND EVALUATION | PENETRATION TESTING / ANALYSIS The organization requires the developer of the information system, system component, or information system service to perform penetration testing at [Assignment: organization-defined breadth/depth] and with [Assignment: organization-defined constraints]. # Class Management # Supplemental Guidance Penetration testing is an assessment methodology in which assessors, using all available information technology product and/or information system documentation (e.g., product/system design specifications, source code, and administrator/operator manuals) and working under specific constraints, attempt to circumvent implemented security features of information technology products and information systems. Penetration testing can include, for example, white, gray, or black box testing with analyses performed by skilled security professionals simulating adversary actions. The objective of penetration testing is to uncover potential vulnerabilities in information technology products and information systems resulting from implementation errors, configuration faults, or other operational deployment weaknesses or deficiencies. Penetration tests can be performed in conjunction with automated and manual code reviews to provide greater levels of analysis than would ordinarily be possible. # General Guide Apply to boundary and other security critical components. For COTS products require 3rd party evaluation such as Common Criteria. # Suggested Assignment IT Projects # Support Teams IT Security Function
1.0
SA-11(5): Developer Security Testing And Evaluation | Penetration Testing / Analysis - # Control Definition DEVELOPER SECURITY TESTING AND EVALUATION | PENETRATION TESTING / ANALYSIS The organization requires the developer of the information system, system component, or information system service to perform penetration testing at [Assignment: organization-defined breadth/depth] and with [Assignment: organization-defined constraints]. # Class Management # Supplemental Guidance Penetration testing is an assessment methodology in which assessors, using all available information technology product and/or information system documentation (e.g., product/system design specifications, source code, and administrator/operator manuals) and working under specific constraints, attempt to circumvent implemented security features of information technology products and information systems. Penetration testing can include, for example, white, gray, or black box testing with analyses performed by skilled security professionals simulating adversary actions. The objective of penetration testing is to uncover potential vulnerabilities in information technology products and information systems resulting from implementation errors, configuration faults, or other operational deployment weaknesses or deficiencies. Penetration tests can be performed in conjunction with automated and manual code reviews to provide greater levels of analysis than would ordinarily be possible. # General Guide Apply to boundary and other security critical components. For COTS products require 3rd party evaluation such as Common Criteria. # Suggested Assignment IT Projects # Support Teams IT Security Function
non_process
sa developer security testing and evaluation penetration testing analysis control definition developer security testing and evaluation penetration testing analysis the organization requires the developer of the information system system component or information system service to perform penetration testing at and with class management supplemental guidance penetration testing is an assessment methodology in which assessors using all available information technology product and or information system documentation e g product system design specifications source code and administrator operator manuals and working under specific constraints attempt to circumvent implemented security features of information technology products and information systems penetration testing can include for example white gray or black box testing with analyses performed by skilled security professionals simulating adversary actions the objective of penetration testing is to uncover potential vulnerabilities in information technology products and information systems resulting from implementation errors configuration faults or other operational deployment weaknesses or deficiencies penetration tests can be performed in conjunction with automated and manual code reviews to provide greater levels of analysis than would ordinarily be possible general guide apply to boundary and other security critical components for cots products require party evaluation such as common criteria suggested assignment it projects support teams it security function
0
5,848
8,672,991,394
IssuesEvent
2018-11-30 00:14:57
googleapis/google-cloud-java
https://api.github.com/repos/googleapis/google-cloud-java
closed
Setup and fix Java 11 tests
type: process
Java 11 is an LTS version scheduled for GA 2018-09-25. Also, we need to fix the build on Java 11.
1.0
Setup and fix Java 11 tests - Java 11 is an LTS version scheduled for GA 2018-09-25. Also, we need to fix the build on Java 11.
process
setup and fix java tests java is an lts version scheduled for ga also we need to fix the build on java
1
768,699
26,976,316,511
IssuesEvent
2023-02-09 09:49:40
MattTheLegoman/RealmsInExile
https://api.github.com/repos/MattTheLegoman/RealmsInExile
opened
GUI in culture screen to show descriptions
priority: low gui
Tweak GUI in culture screen to show cultural descriptions.
1.0
GUI in culture screen to show descriptions - Tweak GUI in culture screen to show cultural descriptions.
non_process
gui in culture screen to show descriptions tweak gui in culture screen to show cultural descriptions
0
674,165
23,041,692,335
IssuesEvent
2022-07-23 08:31:37
tsunamods-codes/7th-Heaven
https://api.github.com/repos/tsunamods-codes/7th-Heaven
closed
Incorrect folder priority because of naming.
help wanted priority/P2
I have been doing some tests. I use my Models Fusion mod which includes Field/Battle/World models, normally with al the files needed to create the model. I decided to add Ninostyle Chibi models (IRO) v0.245 also to 7H. So, my order is: - Ninostyle Chibis (for Field) - Models Fusion (basically are Chaos models, for Field) - Other mods... Well, the thing is that it does not matter the order I put my Models Fusion in 7H, I always see the Chaos models in game. After some checking, I detected that Ninostyle Chibis has the models in the folder **"char.lgp"**, and I have the models in the folder **"char"**. So, as the priority is by folder naming (and they are literally different), "my" folder, "char", which has higher priority (alphabetically goes before than "char.lgp"), loads its models ALWAYS first, don't matter the order the mods are in the "My Mods" tab. I have checked this extracting Ninostyle IRO and changing its folder to "char", then it worked correctly. But of course, a modder should not depend of the folder naming, not at least for the same type of files. Would it be possible to treat "char/char.lgp" or "battle/battle.lgp" or "world/world_us.lgp" to be of the same level?
1.0
Incorrect folder priority because of naming. - I have been doing some tests. I use my Models Fusion mod which includes Field/Battle/World models, normally with al the files needed to create the model. I decided to add Ninostyle Chibi models (IRO) v0.245 also to 7H. So, my order is: - Ninostyle Chibis (for Field) - Models Fusion (basically are Chaos models, for Field) - Other mods... Well, the thing is that it does not matter the order I put my Models Fusion in 7H, I always see the Chaos models in game. After some checking, I detected that Ninostyle Chibis has the models in the folder **"char.lgp"**, and I have the models in the folder **"char"**. So, as the priority is by folder naming (and they are literally different), "my" folder, "char", which has higher priority (alphabetically goes before than "char.lgp"), loads its models ALWAYS first, don't matter the order the mods are in the "My Mods" tab. I have checked this extracting Ninostyle IRO and changing its folder to "char", then it worked correctly. But of course, a modder should not depend of the folder naming, not at least for the same type of files. Would it be possible to treat "char/char.lgp" or "battle/battle.lgp" or "world/world_us.lgp" to be of the same level?
non_process
incorrect folder priority because of naming i have been doing some tests i use my models fusion mod which includes field battle world models normally with al the files needed to create the model i decided to add ninostyle chibi models iro also to so my order is ninostyle chibis for field models fusion basically are chaos models for field other mods well the thing is that it does not matter the order i put my models fusion in i always see the chaos models in game after some checking i detected that ninostyle chibis has the models in the folder char lgp and i have the models in the folder char so as the priority is by folder naming and they are literally different my folder char which has higher priority alphabetically goes before than char lgp loads its models always first don t matter the order the mods are in the my mods tab i have checked this extracting ninostyle iro and changing its folder to char then it worked correctly but of course a modder should not depend of the folder naming not at least for the same type of files would it be possible to treat char char lgp or battle battle lgp or world world us lgp to be of the same level
0
5,222
8,026,315,844
IssuesEvent
2018-07-27 03:13:45
turnkeylinux/tracker
https://api.github.com/repos/turnkeylinux/tracker
opened
Develop user friendly instructions (script?) for updating 3rd party PHP packages
mambo processmaker sitracker ushahidi vtiger
A few TurnKey appliances that include 3rd party PHP apps are not compatible with PHP7. In those instances, we have included PHP5.6 from Ondřej Surý's [third party Debian repo](https://deb.sury.org/). That's all well and good, but we have not included these third party repos in the auto security updates. Whilst we are sure that Ondřej will do his best to not break things, it's unreasonable to expect a single person working in their own time to be able to match the level of testing and assurance that the Debian Security Team provide. So with all that in mind, the only "newb friendly" way to update the PHP5.6 packages would be to do an `apt-get update && apt-get upgrade`. Whilst that will likely work fine, regressions can sometimes occur in non-security packages and IMO are best avoided in a production environment, unless thouroughly tested first. So IMO we should ideally provide users with an easy way to upgrade just the third party PHP packages that are installed from Ondřej's repo. A confconsole plugin or even a downloadable script would be ideal, however copy/paste-able instruction would do.
1.0
Develop user friendly instructions (script?) for updating 3rd party PHP packages - A few TurnKey appliances that include 3rd party PHP apps are not compatible with PHP7. In those instances, we have included PHP5.6 from Ondřej Surý's [third party Debian repo](https://deb.sury.org/). That's all well and good, but we have not included these third party repos in the auto security updates. Whilst we are sure that Ondřej will do his best to not break things, it's unreasonable to expect a single person working in their own time to be able to match the level of testing and assurance that the Debian Security Team provide. So with all that in mind, the only "newb friendly" way to update the PHP5.6 packages would be to do an `apt-get update && apt-get upgrade`. Whilst that will likely work fine, regressions can sometimes occur in non-security packages and IMO are best avoided in a production environment, unless thouroughly tested first. So IMO we should ideally provide users with an easy way to upgrade just the third party PHP packages that are installed from Ondřej's repo. A confconsole plugin or even a downloadable script would be ideal, however copy/paste-able instruction would do.
process
develop user friendly instructions script for updating party php packages a few turnkey appliances that include party php apps are not compatible with in those instances we have included from ondřej surý s that s all well and good but we have not included these third party repos in the auto security updates whilst we are sure that ondřej will do his best to not break things it s unreasonable to expect a single person working in their own time to be able to match the level of testing and assurance that the debian security team provide so with all that in mind the only newb friendly way to update the packages would be to do an apt get update apt get upgrade whilst that will likely work fine regressions can sometimes occur in non security packages and imo are best avoided in a production environment unless thouroughly tested first so imo we should ideally provide users with an easy way to upgrade just the third party php packages that are installed from ondřej s repo a confconsole plugin or even a downloadable script would be ideal however copy paste able instruction would do
1
47,334
24,956,753,360
IssuesEvent
2022-11-01 12:28:39
deeplearning4j/deeplearning4j
https://api.github.com/repos/deeplearning4j/deeplearning4j
closed
DL4J: implement MLKDNNLSTMHelper.backpropGradient + enable helper
Performance DL4J
https://github.com/eclipse/deeplearning4j/blob/master/deeplearning4j/deeplearning4j-nn/src/main/java/org/deeplearning4j/nn/layers/mkldnn/MKLDNNLSTMHelper.java#L34-L42 Now that LSTMLayer backprop has been merged (with MKLDNN support), we should finally add this to libnd4j and enable the helper in the DL4J LSTM layer: https://github.com/eclipse/deeplearning4j/blob/master/deeplearning4j/deeplearning4j-nn/src/main/java/org/deeplearning4j/nn/layers/recurrent/LSTM.java#L79-L87
True
DL4J: implement MLKDNNLSTMHelper.backpropGradient + enable helper - https://github.com/eclipse/deeplearning4j/blob/master/deeplearning4j/deeplearning4j-nn/src/main/java/org/deeplearning4j/nn/layers/mkldnn/MKLDNNLSTMHelper.java#L34-L42 Now that LSTMLayer backprop has been merged (with MKLDNN support), we should finally add this to libnd4j and enable the helper in the DL4J LSTM layer: https://github.com/eclipse/deeplearning4j/blob/master/deeplearning4j/deeplearning4j-nn/src/main/java/org/deeplearning4j/nn/layers/recurrent/LSTM.java#L79-L87
non_process
implement mlkdnnlstmhelper backpropgradient enable helper now that lstmlayer backprop has been merged with mkldnn support we should finally add this to and enable the helper in the lstm layer
0
54,258
23,217,111,397
IssuesEvent
2022-08-02 14:53:53
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Complete Citybase integration with Knack action items
Service: Dev Service: Apps Product: Banners
Complete [Standard Checkout document](https://docs.google.com/document/d/1jl6j3CHr1UWnCmfe3uyJK7cuPUqJIuZC/edit) **Action items from session 1** - [x] Section 1: Service Fees- Mateo/Diana to confirm fees are passed to customer at time of payment, or if Austin is paying fees for customers. In issue #8268 we documented the stakeholders comments: - Having the customer pay the service fees is the right way to go. - If possible on the payment page, I just ask that we be transparent about the difference in costs between e-check and credit card fees. - [x] Section 3: Min/Max- Taejin to review payment processing limits with ATD to confirm upperbound set with the payment processor - [ ] Section 6: Custom Attributes- Mateo/Diana to work on confirming the description, subdescription, and any additional backend attributes that need to be passed along - [x] Section 7: Cancel URL- Erin to provide TST urls; NOTE: the listed URLS in the doc are for PROD - [x] Section 8: Return URL- Mateo/Diana to provide desired "payment success page" URL on the Austin site. - [ ] Section 9: Confirmation Page Content- Mateo/Diana to provide Agency support phone # (if so desired by the department stakeholders) - [x] Taejin to provide Citywide T&C to Austin Transport for review (see attachment 2 to this email) [Terms and Conditions](https://drive.google.com/drive/folders/1GMhQjN-I4mojBMz1SDMJM0-tdTuZwl7n) - [x] Taejin and Stephanie to coordinate on when to schedule a follow up session to: o Discuss/close out the remaining sections listed above o Discuss approach for knack integration (dependent on ATD meeting with Austin Library Team to review their build) Scheduled 8/2
2.0
Complete Citybase integration with Knack action items - Complete [Standard Checkout document](https://docs.google.com/document/d/1jl6j3CHr1UWnCmfe3uyJK7cuPUqJIuZC/edit) **Action items from session 1** - [x] Section 1: Service Fees- Mateo/Diana to confirm fees are passed to customer at time of payment, or if Austin is paying fees for customers. In issue #8268 we documented the stakeholders comments: - Having the customer pay the service fees is the right way to go. - If possible on the payment page, I just ask that we be transparent about the difference in costs between e-check and credit card fees. - [x] Section 3: Min/Max- Taejin to review payment processing limits with ATD to confirm upperbound set with the payment processor - [ ] Section 6: Custom Attributes- Mateo/Diana to work on confirming the description, subdescription, and any additional backend attributes that need to be passed along - [x] Section 7: Cancel URL- Erin to provide TST urls; NOTE: the listed URLS in the doc are for PROD - [x] Section 8: Return URL- Mateo/Diana to provide desired "payment success page" URL on the Austin site. - [ ] Section 9: Confirmation Page Content- Mateo/Diana to provide Agency support phone # (if so desired by the department stakeholders) - [x] Taejin to provide Citywide T&C to Austin Transport for review (see attachment 2 to this email) [Terms and Conditions](https://drive.google.com/drive/folders/1GMhQjN-I4mojBMz1SDMJM0-tdTuZwl7n) - [x] Taejin and Stephanie to coordinate on when to schedule a follow up session to: o Discuss/close out the remaining sections listed above o Discuss approach for knack integration (dependent on ATD meeting with Austin Library Team to review their build) Scheduled 8/2
non_process
complete citybase integration with knack action items complete action items from session section service fees mateo diana to confirm fees are passed to customer at time of payment or if austin is paying fees for customers in issue we documented the stakeholders comments having the customer pay the service fees is the right way to go if possible on the payment page i just ask that we be transparent about the difference in costs between e check and credit card fees section min max taejin to review payment processing limits with atd to confirm upperbound set with the payment processor section custom attributes mateo diana to work on confirming the description subdescription and any additional backend attributes that need to be passed along section cancel url erin to provide tst urls note the listed urls in the doc are for prod section return url mateo diana to provide desired payment success page url on the austin site section confirmation page content mateo diana to provide agency support phone if so desired by the department stakeholders taejin to provide citywide t c to austin transport for review see attachment to this email taejin and stephanie to coordinate on when to schedule a follow up session to o discuss close out the remaining sections listed above o discuss approach for knack integration dependent on atd meeting with austin library team to review their build scheduled
0
215,110
24,126,433,296
IssuesEvent
2022-09-21 01:09:57
smb-h/Estates-price-prediction
https://api.github.com/repos/smb-h/Estates-price-prediction
opened
CVE-2022-36027 (Medium) detected in tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl
security vulnerability
## CVE-2022-36027 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/73/a3/142f73d0e076f5582fd8da29c68af0413bf529933eed09f86a8857fab0d6/tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/73/a3/142f73d0e076f5582fd8da29c68af0413bf529933eed09f86a8857fab0d6/tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/smb-h/Estates-price-prediction/commit/43d8dec55efbdc71655c52119862fee409624fda">43d8dec55efbdc71655c52119862fee409624fda</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. When converting transposed convolutions using per-channel weight quantization the converter segfaults and crashes the Python process. We have patched the issue in GitHub commit aa0b852a4588cea4d36b74feb05d93055540b450. The fix will be included in TensorFlow 2.10.0. We will also cherrypick this commit on TensorFlow 2.9.1, TensorFlow 2.8.1, and TensorFlow 2.7.2, as these are also affected and still in supported range. There are no known workarounds for this issue. <p>Publish Date: 2022-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36027>CVE-2022-36027</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79h2-q768-fpxr">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79h2-q768-fpxr</a></p> <p>Release Date: 2022-09-16</p> <p>Fix Resolution: tensorflow - 2.7.2,2.8.1,2.9.1,2.10.0, tensorflow-cpu - 2.7.2,2.8.1,2.9.1,2.10.0, tensorflow-gpu - 2.7.2,2.8.1,2.9.1,2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-36027 (Medium) detected in tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl - ## CVE-2022-36027 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/73/a3/142f73d0e076f5582fd8da29c68af0413bf529933eed09f86a8857fab0d6/tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/73/a3/142f73d0e076f5582fd8da29c68af0413bf529933eed09f86a8857fab0d6/tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-2.6.3-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/smb-h/Estates-price-prediction/commit/43d8dec55efbdc71655c52119862fee409624fda">43d8dec55efbdc71655c52119862fee409624fda</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. When converting transposed convolutions using per-channel weight quantization the converter segfaults and crashes the Python process. We have patched the issue in GitHub commit aa0b852a4588cea4d36b74feb05d93055540b450. The fix will be included in TensorFlow 2.10.0. We will also cherrypick this commit on TensorFlow 2.9.1, TensorFlow 2.8.1, and TensorFlow 2.7.2, as these are also affected and still in supported range. There are no known workarounds for this issue. <p>Publish Date: 2022-09-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36027>CVE-2022-36027</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79h2-q768-fpxr">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-79h2-q768-fpxr</a></p> <p>Release Date: 2022-09-16</p> <p>Fix Resolution: tensorflow - 2.7.2,2.8.1,2.9.1,2.10.0, tensorflow-cpu - 2.7.2,2.8.1,2.9.1,2.10.0, tensorflow-gpu - 2.7.2,2.8.1,2.9.1,2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy x tensorflow whl vulnerable library found in head commit a href found in base branch main vulnerability details tensorflow is an open source platform for machine learning when converting transposed convolutions using per channel weight quantization the converter segfaults and crashes the python process we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend
0
11,794
14,620,804,822
IssuesEvent
2020-12-22 20:24:48
pacificclimate/quail
https://api.github.com/repos/pacificclimate/quail
closed
Get available indices by name
process
## Description This function takes a climdexInput object as input and returns the names of all the indices which may be computed or, if \code{get.function.names} is TRUE (the default), the names of the functions corresponding to the indices. ## Function to wrap [`climdex.get.available.indices`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L1315)
1.0
Get available indices by name - ## Description This function takes a climdexInput object as input and returns the names of all the indices which may be computed or, if \code{get.function.names} is TRUE (the default), the names of the functions corresponding to the indices. ## Function to wrap [`climdex.get.available.indices`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L1315)
process
get available indices by name description this function takes a climdexinput object as input and returns the names of all the indices which may be computed or if code get function names is true the default the names of the functions corresponding to the indices function to wrap
1
230,319
25,464,164,546
IssuesEvent
2022-11-25 01:01:37
pactflow/example-bi-directional-provider-restassured
https://api.github.com/repos/pactflow/example-bi-directional-provider-restassured
opened
CVE-2022-45868 (High) detected in h2-1.4.200.jar
security vulnerability
## CVE-2022-45868 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.200.jar</b></p></summary> <p>H2 Database Engine</p> <p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /m.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p> <p> Dependency Hierarchy: - :x: **h2-1.4.200.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/pactflow/example-bi-directional-provider-restassured/commit/f89cca989323a3ac73aac3d278cab02a868ac80b">f89cca989323a3ac73aac3d278cab02a868ac80b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The web-based admin console in H2 Database Engine through 2.1.214 can be started via the CLI with the argument -webAdminPassword, which allows the user to specify the password in cleartext for the web admin console. Consequently, a local user (or an attacker that has obtained local access through some means) would be able to discover the password by listing processes and their arguments. NOTE: the vendor states "This is not a vulnerability of H2 Console ... Passwords should never be passed on the command line and every qualified DBA or system administrator is expected to know that." <p>Publish Date: 2022-11-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45868>CVE-2022-45868</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-45868 (High) detected in h2-1.4.200.jar - ## CVE-2022-45868 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.200.jar</b></p></summary> <p>H2 Database Engine</p> <p>Library home page: <a href="https://h2database.com">https://h2database.com</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /m.h2database/h2/1.4.200/f7533fe7cb8e99c87a43d325a77b4b678ad9031a/h2-1.4.200.jar</p> <p> Dependency Hierarchy: - :x: **h2-1.4.200.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/pactflow/example-bi-directional-provider-restassured/commit/f89cca989323a3ac73aac3d278cab02a868ac80b">f89cca989323a3ac73aac3d278cab02a868ac80b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The web-based admin console in H2 Database Engine through 2.1.214 can be started via the CLI with the argument -webAdminPassword, which allows the user to specify the password in cleartext for the web admin console. Consequently, a local user (or an attacker that has obtained local access through some means) would be able to discover the password by listing processes and their arguments. NOTE: the vendor states "This is not a vulnerability of H2 Console ... Passwords should never be passed on the command line and every qualified DBA or system administrator is expected to know that." <p>Publish Date: 2022-11-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-45868>CVE-2022-45868</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_process
cve high detected in jar cve high severity vulnerability vulnerable library jar database engine library home page a href path to dependency file build gradle path to vulnerable library m jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details the web based admin console in database engine through can be started via the cli with the argument webadminpassword which allows the user to specify the password in cleartext for the web admin console consequently a local user or an attacker that has obtained local access through some means would be able to discover the password by listing processes and their arguments note the vendor states this is not a vulnerability of console passwords should never be passed on the command line and every qualified dba or system administrator is expected to know that publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href
0
22,501
31,494,396,015
IssuesEvent
2023-08-31 00:15:37
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
multiple parametric masks
feature: new scope: UI scope: image processing controversial no-issue-activity
**Is your feature request related to a problem? Please describe.** Occasionally one would like to select a union of various properties that can be described by parametric masks (eg skin color and blue for eyes, nothing else). **Describe the solution you'd like** Multiple parametric masks in a module instance, to be combined under the same [rules](https://docs.darktable.org/usermanual/3.6/darkroom/masking-and-blending/masks/drawn-and-parametric/) as drawn elements and parametric masks. **Alternatives** For some transformations, it is possible to work around this problem by using multiple module instances with the parametric mask elements that would form the union above. However, that becomes tedious when information would be lost by a transformation (eg desaturate some colors), or the transformations are nonlinear. For example, to desaturate everything but skin tones and blue, one would currently desaturate from red to skin tones, then from skin tones to blue, then from blue onwards. This requires 3 module instances, but all that changes is the mask. Naturally, drawn masks can be a workaround too and are very powerful, but address a different problem. **Additional context** [Discussion](https://discuss.pixls.us/t/combine-parametric-masks/27711).
1.0
multiple parametric masks - **Is your feature request related to a problem? Please describe.** Occasionally one would like to select a union of various properties that can be described by parametric masks (eg skin color and blue for eyes, nothing else). **Describe the solution you'd like** Multiple parametric masks in a module instance, to be combined under the same [rules](https://docs.darktable.org/usermanual/3.6/darkroom/masking-and-blending/masks/drawn-and-parametric/) as drawn elements and parametric masks. **Alternatives** For some transformations, it is possible to work around this problem by using multiple module instances with the parametric mask elements that would form the union above. However, that becomes tedious when information would be lost by a transformation (eg desaturate some colors), or the transformations are nonlinear. For example, to desaturate everything but skin tones and blue, one would currently desaturate from red to skin tones, then from skin tones to blue, then from blue onwards. This requires 3 module instances, but all that changes is the mask. Naturally, drawn masks can be a workaround too and are very powerful, but address a different problem. **Additional context** [Discussion](https://discuss.pixls.us/t/combine-parametric-masks/27711).
process
multiple parametric masks is your feature request related to a problem please describe occasionally one would like to select a union of various properties that can be described by parametric masks eg skin color and blue for eyes nothing else describe the solution you d like multiple parametric masks in a module instance to be combined under the same as drawn elements and parametric masks alternatives for some transformations it is possible to work around this problem by using multiple module instances with the parametric mask elements that would form the union above however that becomes tedious when information would be lost by a transformation eg desaturate some colors or the transformations are nonlinear for example to desaturate everything but skin tones and blue one would currently desaturate from red to skin tones then from skin tones to blue then from blue onwards this requires module instances but all that changes is the mask naturally drawn masks can be a workaround too and are very powerful but address a different problem additional context
1
47,117
7,307,492,269
IssuesEvent
2018-02-28 03:03:57
adafruit/circuitpython
https://api.github.com/repos/adafruit/circuitpython
closed
Fix warnings in sphinx documentation build
documentation
There are unconnected pages, ambiguous references, and warnings about the new `templates/replace.inc` file, which came in with MPy v1.9.2
1.0
Fix warnings in sphinx documentation build - There are unconnected pages, ambiguous references, and warnings about the new `templates/replace.inc` file, which came in with MPy v1.9.2
non_process
fix warnings in sphinx documentation build there are unconnected pages ambiguous references and warnings about the new templates replace inc file which came in with mpy
0
95,315
10,877,574,960
IssuesEvent
2019-11-16 11:21:04
tobiasanker/SakuraTree
https://api.github.com/repos/tobiasanker/SakuraTree
closed
Add usage of multiple blossom-subtypes
documentation feature / enhancement
## Feature-request ### Description Blossoms can have only one subtype at the moment. It is only a problem of the compiler, because its not completely updated for the current parser-version. ### Possible Implementation Update `convertBlossom`-method in `SakuraCompiler` to iterate over the subtypes.
1.0
Add usage of multiple blossom-subtypes - ## Feature-request ### Description Blossoms can have only one subtype at the moment. It is only a problem of the compiler, because its not completely updated for the current parser-version. ### Possible Implementation Update `convertBlossom`-method in `SakuraCompiler` to iterate over the subtypes.
non_process
add usage of multiple blossom subtypes feature request description blossoms can have only one subtype at the moment it is only a problem of the compiler because its not completely updated for the current parser version possible implementation update convertblossom method in sakuracompiler to iterate over the subtypes
0
257,865
19,533,836,895
IssuesEvent
2021-12-30 23:38:20
mjskay/ggdist
https://api.github.com/repos/mjskay/ggdist
closed
Use new @examplesIf tag in roxygen examples
documentation
For conditional examples (e.g. ones depending on suggested packages), switch to using `@examplesIf`
1.0
Use new @examplesIf tag in roxygen examples - For conditional examples (e.g. ones depending on suggested packages), switch to using `@examplesIf`
non_process
use new examplesif tag in roxygen examples for conditional examples e g ones depending on suggested packages switch to using examplesif
0
2,317
5,139,972,878
IssuesEvent
2017-01-12 02:18:01
vuejs/vue-loader
https://api.github.com/repos/vuejs/vue-loader
closed
External Pug template not working if it's filename is the same as the component's one.
pre-processor
I got this code in a `Hello.vue` component ; ``` <template lang="pug" src="./Hello..pug"></template> <script> export default { name: 'hello', ..... } } </script> <style lang='sass' src='./style.sass'></style> ``` That doesn't work. I noticed that it only appens when template name is the exact same as the component's one. If for instance template name is sufixed like `Hello.tpl.pug`, that works.
1.0
External Pug template not working if it's filename is the same as the component's one. - I got this code in a `Hello.vue` component ; ``` <template lang="pug" src="./Hello..pug"></template> <script> export default { name: 'hello', ..... } } </script> <style lang='sass' src='./style.sass'></style> ``` That doesn't work. I noticed that it only appens when template name is the exact same as the component's one. If for instance template name is sufixed like `Hello.tpl.pug`, that works.
process
external pug template not working if it s filename is the same as the component s one i got this code in a hello vue component export default name hello that doesn t work i noticed that it only appens when template name is the exact same as the component s one if for instance template name is sufixed like hello tpl pug that works
1
20,862
27,645,510,056
IssuesEvent
2023-03-10 22:29:14
cse442-at-ub/project_s23-cinco
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
opened
Create routing within React to navigate different pages when the corresponding button is clicked
Processing Task Sprint 2
*Task Test* test 1: -go to the homepage: in your editor of choice, in the project folder, type "npm start" in the terminal to open up the homepage in the browser. - click on the login button and ensure it takes you to the login page. test 2: -go to the homepage: in your editor of choice, in the project folder, type "npm start" in the terminal to open up the homepage in the browser. - click on the signup button and ensure it takes you to the signup page.
1.0
Create routing within React to navigate different pages when the corresponding button is clicked - *Task Test* test 1: -go to the homepage: in your editor of choice, in the project folder, type "npm start" in the terminal to open up the homepage in the browser. - click on the login button and ensure it takes you to the login page. test 2: -go to the homepage: in your editor of choice, in the project folder, type "npm start" in the terminal to open up the homepage in the browser. - click on the signup button and ensure it takes you to the signup page.
process
create routing within react to navigate different pages when the corresponding button is clicked task test test go to the homepage in your editor of choice in the project folder type npm start in the terminal to open up the homepage in the browser click on the login button and ensure it takes you to the login page test go to the homepage in your editor of choice in the project folder type npm start in the terminal to open up the homepage in the browser click on the signup button and ensure it takes you to the signup page
1
7,697
18,893,353,318
IssuesEvent
2021-11-15 15:23:41
Krakenus00/EzNote
https://api.github.com/repos/Krakenus00/EzNote
opened
Unite entities at the Domain-Layer
architecture
It`s some changes in architecture. We need to move the same entities from Business-Layer and Data-Access to Domain-Layer
1.0
Unite entities at the Domain-Layer - It`s some changes in architecture. We need to move the same entities from Business-Layer and Data-Access to Domain-Layer
non_process
unite entities at the domain layer it s some changes in architecture we need to move the same entities from business layer and data access to domain layer
0
9,089
4,413,688,691
IssuesEvent
2016-08-13 01:02:30
facebook/osquery
https://api.github.com/repos/facebook/osquery
closed
Flaky tests: DaemonTests::test_5_daemon_sigint variance in return code
build/test test error
See: ``` 5/9 Test #5: python_test_osqueryd .............***Failed 33.75 sec .I0721 12:22:59.882045 26907 options.cpp:61] Verbose logging enabled by config option I0721 12:22:59.882532 26907 daemon.cpp:38] Not starting the distributed query service: Distributed query service not enabled. ...FI0721 12:23:29.781241 26946 options.cpp:61] Verbose logging enabled by config option I0721 12:23:29.781740 26946 daemon.cpp:38] Not starting the distributed query service: Distributed query service not enabled. . ====================================================================== FAIL: test_5_daemon_sigint (__main__.DaemonTests) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/osquery/jenkins/workspace/osqueryPullRequestBuild/TargetSystem/centos7/tools/tests/test_base.py", line 455, in wrapper raise exceptions[0][0] AssertionError: -2 != 130 ---------------------------------------------------------------------- Ran 6 tests in 33.695s FAILED (failures=1) Test (attempt 1) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437) Test (attempt 2) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437) Test (attempt 3) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437) ``` For an example see: https://jenkins.osquery.io/job/osqueryPullRequestBuild/2912/TargetSystem=centos7/console
1.0
Flaky tests: DaemonTests::test_5_daemon_sigint variance in return code - See: ``` 5/9 Test #5: python_test_osqueryd .............***Failed 33.75 sec .I0721 12:22:59.882045 26907 options.cpp:61] Verbose logging enabled by config option I0721 12:22:59.882532 26907 daemon.cpp:38] Not starting the distributed query service: Distributed query service not enabled. ...FI0721 12:23:29.781241 26946 options.cpp:61] Verbose logging enabled by config option I0721 12:23:29.781740 26946 daemon.cpp:38] Not starting the distributed query service: Distributed query service not enabled. . ====================================================================== FAIL: test_5_daemon_sigint (__main__.DaemonTests) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/osquery/jenkins/workspace/osqueryPullRequestBuild/TargetSystem/centos7/tools/tests/test_base.py", line 455, in wrapper raise exceptions[0][0] AssertionError: -2 != 130 ---------------------------------------------------------------------- Ran 6 tests in 33.695s FAILED (failures=1) Test (attempt 1) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437) Test (attempt 2) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437) Test (attempt 3) DaemonTests::test_5_daemon_sigint failed: -2 != 130 (test_base.py:437) ``` For an example see: https://jenkins.osquery.io/job/osqueryPullRequestBuild/2912/TargetSystem=centos7/console
non_process
flaky tests daemontests test daemon sigint variance in return code see test python test osqueryd failed sec options cpp verbose logging enabled by config option daemon cpp not starting the distributed query service distributed query service not enabled options cpp verbose logging enabled by config option daemon cpp not starting the distributed query service distributed query service not enabled fail test daemon sigint main daemontests traceback most recent call last file home osquery jenkins workspace osquerypullrequestbuild targetsystem tools tests test base py line in wrapper raise exceptions assertionerror ran tests in failed failures test attempt daemontests test daemon sigint failed test base py test attempt daemontests test daemon sigint failed test base py test attempt daemontests test daemon sigint failed test base py for an example see
0
243,852
26,290,654,331
IssuesEvent
2023-01-08 11:20:39
yaeljacobs67/fs-agent
https://api.github.com/repos/yaeljacobs67/fs-agent
opened
WS-2016-7112 (Medium) detected in spring-context-4.3.1.RELEASE.jar
security vulnerability
## WS-2016-7112 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-context-4.3.1.RELEASE.jar</b></p></summary> <p>Spring Context</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /fs-agent/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-context/4.3.1.RELEASE/spring-context-4.3.1.RELEASE.jar</p> <p> Dependency Hierarchy: - whitesource-analysis-via-18.12.1.204.jar (Root Library) - whitesource-utilities-0.0.1.jar - spring-web-4.3.1.RELEASE.jar - :x: **spring-context-4.3.1.RELEASE.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 3.0.0.RELEASE through 3.2.17.RELEASE, 4.0.0.RELEASE through 4.2.7.RELEASE and 4.3.0.RELEASE through 4.3.1.RELEASE are vulnerable to Stack-based Buffer Overflow, which allows an authenticated attacker to crash the application when giving CronSequenceGenerator a reversed range in the “minutes” or “hours” fields. <p>Publish Date: 2021-09-23 <p>URL: <a href=https://github.com/spring-projects/spring-framework/commit/e431624e8472b3b53d1a0c4528bf736c612f1fd9>WS-2016-7112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2016-07-14</p> <p>Fix Resolution: org.springframework:spring-context:3.2.18.RELEASE,4.2.8.RELEASE,4.3.2.RELEASE,5.0.0.RELEASE</p> </p> </details> <p></p>
True
WS-2016-7112 (Medium) detected in spring-context-4.3.1.RELEASE.jar - ## WS-2016-7112 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-context-4.3.1.RELEASE.jar</b></p></summary> <p>Spring Context</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: /fs-agent/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/org/springframework/spring-context/4.3.1.RELEASE/spring-context-4.3.1.RELEASE.jar</p> <p> Dependency Hierarchy: - whitesource-analysis-via-18.12.1.204.jar (Root Library) - whitesource-utilities-0.0.1.jar - spring-web-4.3.1.RELEASE.jar - :x: **spring-context-4.3.1.RELEASE.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 3.0.0.RELEASE through 3.2.17.RELEASE, 4.0.0.RELEASE through 4.2.7.RELEASE and 4.3.0.RELEASE through 4.3.1.RELEASE are vulnerable to Stack-based Buffer Overflow, which allows an authenticated attacker to crash the application when giving CronSequenceGenerator a reversed range in the “minutes” or “hours” fields. <p>Publish Date: 2021-09-23 <p>URL: <a href=https://github.com/spring-projects/spring-framework/commit/e431624e8472b3b53d1a0c4528bf736c612f1fd9>WS-2016-7112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2016-07-14</p> <p>Fix Resolution: org.springframework:spring-context:3.2.18.RELEASE,4.2.8.RELEASE,4.3.2.RELEASE,5.0.0.RELEASE</p> </p> </details> <p></p>
non_process
ws medium detected in spring context release jar ws medium severity vulnerability vulnerable library spring context release jar spring context library home page a href path to dependency file fs agent pom xml path to vulnerable library root repository org springframework spring context release spring context release jar dependency hierarchy whitesource analysis via jar root library whitesource utilities jar spring web release jar x spring context release jar vulnerable library vulnerability details in spring framework versions release through release release through release and release through release are vulnerable to stack based buffer overflow which allows an authenticated attacker to crash the application when giving cronsequencegenerator a reversed range in the “minutes” or “hours” fields publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution org springframework spring context release release release release
0
7,148
10,289,323,736
IssuesEvent
2019-08-27 15:35:53
Holmeyoung/blog-comment
https://api.github.com/repos/Holmeyoung/blog-comment
opened
python 多线程与多进程 | 一只羊的碎碎念
/python-threading-multiprocessing/ Gitalk
https://www.holmeyoung.com/python-threading-multiprocessing/ 多进程进程池阻塞与非阻塞 异步进程池(非阻塞)123456789101112131415161718from multiprocessing import Pooldef test(i): print (i)if __name__=="__main__": pool = Pool(processes=10) for i in xrange(500): ''' For循环中执行步骤:
1.0
python 多线程与多进程 | 一只羊的碎碎念 - https://www.holmeyoung.com/python-threading-multiprocessing/ 多进程进程池阻塞与非阻塞 异步进程池(非阻塞)123456789101112131415161718from multiprocessing import Pooldef test(i): print (i)if __name__=="__main__": pool = Pool(processes=10) for i in xrange(500): ''' For循环中执行步骤:
process
python 多线程与多进程 一只羊的碎碎念 多进程进程池阻塞与非阻塞 异步进程池(非阻塞) multiprocessing import pooldef test i print i if name main pool pool processes for i in xrange for循环中执行步骤:
1
111,677
14,121,332,300
IssuesEvent
2020-11-09 01:39:42
rohmishra/personal_website
https://api.github.com/repos/rohmishra/personal_website
closed
General improvement and changes.
Design Needed enhancement help wanted no-issue-activity
**This report is tracking multiple changes to improve the design of website.** 1. ~Fix "My projects" text on dark mode chrome. #38~ ( Fixed in #43 ) 2. Change background design for designs pages 3. Improve the cards for projects. 4. Show more details for each project (pending design) 5. Prepare for transition to rmishra.me 6. Possibly disable contact me section till database is moved out of m-lab. 7. Add About me page #13 8. Add better CSS animations
1.0
General improvement and changes. - **This report is tracking multiple changes to improve the design of website.** 1. ~Fix "My projects" text on dark mode chrome. #38~ ( Fixed in #43 ) 2. Change background design for designs pages 3. Improve the cards for projects. 4. Show more details for each project (pending design) 5. Prepare for transition to rmishra.me 6. Possibly disable contact me section till database is moved out of m-lab. 7. Add About me page #13 8. Add better CSS animations
non_process
general improvement and changes this report is tracking multiple changes to improve the design of website fix my projects text on dark mode chrome fixed in change background design for designs pages improve the cards for projects show more details for each project pending design prepare for transition to rmishra me possibly disable contact me section till database is moved out of m lab add about me page add better css animations
0
303,827
26,230,700,791
IssuesEvent
2023-01-04 23:44:26
ImagingDataCommons/IDC-WebApp
https://api.github.com/repos/ImagingDataCommons/IDC-WebApp
closed
In test and prod tier, There is a warning message for BMI when a cohorts with BMI filters is saved.
bug testing needed testing passed
Note - I was unable to determine if this issue is actually a bug (We discussed this in our weekly meeting and i was suggested to open a ticket). Please feel free to delete the ticket if its not a bug. Summary - On the IDC webapp, When I select TCGA collection and BMI filter to create a cohort, I see the message that says Invalid operator seen for attribute 'bmi'; default of OR used instead. (Like in the image below) But I am unable to find where the OR is being used. It still shows AND in the filter definition and also in the filter URL, ![image](https://user-images.githubusercontent.com/85526819/197254272-1059c2e1-6cda-4cae-9010-42337d216fb1.png) To reproduce this issue in test, Select TCGA collections and related filter BMI (eg overweight) Save the cohort to see the warning on top of the page.
2.0
In test and prod tier, There is a warning message for BMI when a cohorts with BMI filters is saved. - Note - I was unable to determine if this issue is actually a bug (We discussed this in our weekly meeting and i was suggested to open a ticket). Please feel free to delete the ticket if its not a bug. Summary - On the IDC webapp, When I select TCGA collection and BMI filter to create a cohort, I see the message that says Invalid operator seen for attribute 'bmi'; default of OR used instead. (Like in the image below) But I am unable to find where the OR is being used. It still shows AND in the filter definition and also in the filter URL, ![image](https://user-images.githubusercontent.com/85526819/197254272-1059c2e1-6cda-4cae-9010-42337d216fb1.png) To reproduce this issue in test, Select TCGA collections and related filter BMI (eg overweight) Save the cohort to see the warning on top of the page.
non_process
in test and prod tier there is a warning message for bmi when a cohorts with bmi filters is saved note i was unable to determine if this issue is actually a bug we discussed this in our weekly meeting and i was suggested to open a ticket please feel free to delete the ticket if its not a bug summary on the idc webapp when i select tcga collection and bmi filter to create a cohort i see the message that says invalid operator seen for attribute bmi default of or used instead like in the image below but i am unable to find where the or is being used it still shows and in the filter definition and also in the filter url to reproduce this issue in test select tcga collections and related filter bmi eg overweight save the cohort to see the warning on top of the page
0
6,748
6,584,002,898
IssuesEvent
2017-09-13 08:35:08
vmware/docker-volume-vsphere
https://api.github.com/repos/vmware/docker-volume-vsphere
closed
Improve CI/CD log collection
component/ci-infrastructure kind/enhancement
- [ ] Dump relevant information for both successful and failed runs - [ ] Post logs to a location where developers can download it - [ ] Improve logging output in CI/CD
1.0
Improve CI/CD log collection - - [ ] Dump relevant information for both successful and failed runs - [ ] Post logs to a location where developers can download it - [ ] Improve logging output in CI/CD
non_process
improve ci cd log collection dump relevant information for both successful and failed runs post logs to a location where developers can download it improve logging output in ci cd
0
5,105
7,883,610,503
IssuesEvent
2018-06-27 06:08:31
ropensci/onboarding-meta
https://api.github.com/repos/ropensci/onboarding-meta
closed
Submitters not using reviews
process
What do we do if submitters don't want to make changes that reviewers/editors suggest? I think we have to be firm about this. If submitter gives good reasons for not making changes reviewers/editors suggest that's fine (and it should be item by item response too), but if there's a blanket statement that they just don't want to make the changes, I think we need to reject the submission. I think this scenario could arise from a submitter simply wanting to be under rOpenSci but not wanting to do the work.
1.0
Submitters not using reviews - What do we do if submitters don't want to make changes that reviewers/editors suggest? I think we have to be firm about this. If submitter gives good reasons for not making changes reviewers/editors suggest that's fine (and it should be item by item response too), but if there's a blanket statement that they just don't want to make the changes, I think we need to reject the submission. I think this scenario could arise from a submitter simply wanting to be under rOpenSci but not wanting to do the work.
process
submitters not using reviews what do we do if submitters don t want to make changes that reviewers editors suggest i think we have to be firm about this if submitter gives good reasons for not making changes reviewers editors suggest that s fine and it should be item by item response too but if there s a blanket statement that they just don t want to make the changes i think we need to reject the submission i think this scenario could arise from a submitter simply wanting to be under ropensci but not wanting to do the work
1
17,312
23,134,091,686
IssuesEvent
2022-07-28 13:01:41
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
opened
Cross compile qnx, Bazel does not produce shared libraries. Release 5.2.0 - $MONTH $YEAR
P1 type: process release team-OSS
**Description of the bug:** When I cross compile QNX, no dynamic library is generated. At this time, after I add `feature( name = "supports_dynamic_linker", enabled = True, )`, it prompts me that ld cannot be found. then i create a ld soft connection `ln aarch64-unknown-nto-qnx7.1.0-ld ld`, this problem is solved.i'm surprised that i specified ld in the tool_path like this `tool_path( name = "ld", path = QNX_HOST + "/usr/bin/aarch64-unknown-nto-qnx7.1.0-ld",),`, but it doesn't seem to work,How can I solve this problem gracefully。 **Which operating system are you running Bazel on?** Linux. **What is the output of bazel info release?** 5.2.0 **Have you found anything relevant by searching the web?** [When I cross compile the example , Bazel does not produce shared libraries](https://github.com/bazelbuild/bazel/issues/7637)
1.0
Cross compile qnx, Bazel does not produce shared libraries. Release 5.2.0 - $MONTH $YEAR - **Description of the bug:** When I cross compile QNX, no dynamic library is generated. At this time, after I add `feature( name = "supports_dynamic_linker", enabled = True, )`, it prompts me that ld cannot be found. then i create a ld soft connection `ln aarch64-unknown-nto-qnx7.1.0-ld ld`, this problem is solved.i'm surprised that i specified ld in the tool_path like this `tool_path( name = "ld", path = QNX_HOST + "/usr/bin/aarch64-unknown-nto-qnx7.1.0-ld",),`, but it doesn't seem to work,How can I solve this problem gracefully。 **Which operating system are you running Bazel on?** Linux. **What is the output of bazel info release?** 5.2.0 **Have you found anything relevant by searching the web?** [When I cross compile the example , Bazel does not produce shared libraries](https://github.com/bazelbuild/bazel/issues/7637)
process
cross compile qnx bazel does not produce shared libraries release month year description of the bug when i cross compile qnx no dynamic library is generated at this time after i add feature name supports dynamic linker enabled true , it prompts me that ld cannot be found then i create a ld soft connection ln unknown nto ld ld this problem is solved i m surprised that i specified ld in the tool path like this tool path name ld path qnx host usr bin unknown nto ld but it doesn t seem to work,how can i solve this problem gracefully。 which operating system are you running bazel on linux what is the output of bazel info release have you found anything relevant by searching the web
1
7,132
10,278,464,312
IssuesEvent
2019-08-25 14:39:06
nextmoov/nextmoov
https://api.github.com/repos/nextmoov/nextmoov
closed
Standard entrypoint: yarn start, yarn start:production
#Dev Tools & Processes
# Standard EntryPoint ```yarn start``` Run the project in *Development* ```yarn build``` Build the static assets to be served in production in /dist # Archived Discussion I would like to standardize the entrypoint for Front & Back project, for both Production & Development. I hereby propose that all our project (back & front) follow this convention: ```yarn start``` Run the project in *Development* ```yarn serve``` Run the project in *Production* (serve precompiled assets in /dist) #### Why not yarn start in Production? Create-React-App (and actually a lot of other framework) use ```yarn start``` as the entrypoint for Development — I propose that we don't override this behaviour
1.0
Standard entrypoint: yarn start, yarn start:production - # Standard EntryPoint ```yarn start``` Run the project in *Development* ```yarn build``` Build the static assets to be served in production in /dist # Archived Discussion I would like to standardize the entrypoint for Front & Back project, for both Production & Development. I hereby propose that all our project (back & front) follow this convention: ```yarn start``` Run the project in *Development* ```yarn serve``` Run the project in *Production* (serve precompiled assets in /dist) #### Why not yarn start in Production? Create-React-App (and actually a lot of other framework) use ```yarn start``` as the entrypoint for Development — I propose that we don't override this behaviour
process
standard entrypoint yarn start yarn start production standard entrypoint yarn start run the project in development yarn build build the static assets to be served in production in dist archived discussion i would like to standardize the entrypoint for front back project for both production development i hereby propose that all our project back front follow this convention yarn start run the project in development yarn serve run the project in production serve precompiled assets in dist why not yarn start in production create react app and actually a lot of other framework use yarn start as the entrypoint for development — i propose that we don t override this behaviour
1
54,708
3,071,123,078
IssuesEvent
2015-08-19 09:58:13
pavel-pimenov/flylinkdc-r5xx
https://api.github.com/repos/pavel-pimenov/flylinkdc-r5xx
closed
Некорректно работает перенос папки в очереди загрузки
bug Component-Logic imported Priority-Critical Usability
_From [rain.bipper@gmail.com](https://code.google.com/u/rain.bipper@gmail.com/) on November 06, 2009 12:26:45_ если в очереди загрузки выбрать "переместить" папку (подчеркиваю! именно папку, а не файлы), то после перемещения в конечной папке будут все файлы из исходной папки, а сама папка не создается. Например: в очереди есть папка1, которой есть N файлов. делаю переместить папка1 в папка2. в результате N файлов сыпется прямо в папка2, подпапка не создается! _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=49_
1.0
Некорректно работает перенос папки в очереди загрузки - _From [rain.bipper@gmail.com](https://code.google.com/u/rain.bipper@gmail.com/) on November 06, 2009 12:26:45_ если в очереди загрузки выбрать "переместить" папку (подчеркиваю! именно папку, а не файлы), то после перемещения в конечной папке будут все файлы из исходной папки, а сама папка не создается. Например: в очереди есть папка1, которой есть N файлов. делаю переместить папка1 в папка2. в результате N файлов сыпется прямо в папка2, подпапка не создается! _Original issue: http://code.google.com/p/flylinkdc/issues/detail?id=49_
non_process
некорректно работает перенос папки в очереди загрузки from on november если в очереди загрузки выбрать переместить папку подчеркиваю именно папку а не файлы то после перемещения в конечной папке будут все файлы из исходной папки а сама папка не создается например в очереди есть которой есть n файлов делаю переместить в в результате n файлов сыпется прямо в подпапка не создается original issue
0
12,638
15,016,625,529
IssuesEvent
2021-02-01 09:49:39
threefoldtech/js-sdk
https://api.github.com/repos/threefoldtech/js-sdk
closed
newly deployed solutions are not listed on 'marketplace' deployed solutions.
process_wontfix type_bug
just deployed taiga and its not listed on marketplace deployed solution <img width="1433" alt="Screenshot 2021-01-28 at 13 42 40" src="https://user-images.githubusercontent.com/43240801/106140257-d0a98700-616e-11eb-90dd-543a786022ea.png"> although it's shown on taiga/my workloads <img width="1404" alt="Screenshot 2021-01-28 at 13 42 34" src="https://user-images.githubusercontent.com/43240801/106140389-ffbff880-616e-11eb-8ce4-fd72aeabe6b1.png">
1.0
newly deployed solutions are not listed on 'marketplace' deployed solutions. - just deployed taiga and its not listed on marketplace deployed solution <img width="1433" alt="Screenshot 2021-01-28 at 13 42 40" src="https://user-images.githubusercontent.com/43240801/106140257-d0a98700-616e-11eb-90dd-543a786022ea.png"> although it's shown on taiga/my workloads <img width="1404" alt="Screenshot 2021-01-28 at 13 42 34" src="https://user-images.githubusercontent.com/43240801/106140389-ffbff880-616e-11eb-8ce4-fd72aeabe6b1.png">
process
newly deployed solutions are not listed on marketplace deployed solutions just deployed taiga and its not listed on marketplace deployed solution img width alt screenshot at src although it s shown on taiga my workloads img width alt screenshot at src
1
20,053
26,541,065,203
IssuesEvent
2023-01-19 19:18:31
Azure/azure-sdk-tools
https://api.github.com/repos/Azure/azure-sdk-tools
closed
APIView integration from Cadl PR
Epic APIView Central-EngSys Cadl WS: Process Tools & Automation
When a user creates a PR in the specs repo and is using Cadl, make sure there is an APIView created for it (as we currently do for SDK changes)
1.0
APIView integration from Cadl PR - When a user creates a PR in the specs repo and is using Cadl, make sure there is an APIView created for it (as we currently do for SDK changes)
process
apiview integration from cadl pr when a user creates a pr in the specs repo and is using cadl make sure there is an apiview created for it as we currently do for sdk changes
1
107,706
13,503,651,480
IssuesEvent
2020-09-13 14:30:52
rubyforgood/casa
https://api.github.com/repos/rubyforgood/casa
closed
Make edit profile accessible by dropdown menu in All CASA Admin dashboard
:cityscape: Multitenancy :globe_with_meridians: All CASA Admin :paintbrush: Design Good First Issue Help Wanted Priority: High Ruby For Good 🎃 Fall 2020
Part of epic #527 **What type of user is this for? [volunteer/admin/supervisor/all or All CASA Admin]** All CASA Admin **Where does/should this occur?** In the All CASA Admin dashboard (dependent on #628) **Description** All CASA Admin needs an entry point to their Edit Profile page. It should be accessible by clicking on their email in the upper right hand corner of their user dashboard. When they click, "Edit Profile" should appear as the only option in a drop down menu. Log in as any other user to see how this should look. <a href="https://github.com/rubyforgood/casa/blob/master/db/seeds.rb>Login instructions here.</a> **Screenshots** It should look like this: ![Screen Shot 2020-08-31 at 6 30 51 PM (2)](https://user-images.githubusercontent.com/62810851/91784656-45749700-ebb8-11ea-8dde-50e91d5f3016.png) Relates to issues: #627, #628, #630
1.0
Make edit profile accessible by dropdown menu in All CASA Admin dashboard - Part of epic #527 **What type of user is this for? [volunteer/admin/supervisor/all or All CASA Admin]** All CASA Admin **Where does/should this occur?** In the All CASA Admin dashboard (dependent on #628) **Description** All CASA Admin needs an entry point to their Edit Profile page. It should be accessible by clicking on their email in the upper right hand corner of their user dashboard. When they click, "Edit Profile" should appear as the only option in a drop down menu. Log in as any other user to see how this should look. <a href="https://github.com/rubyforgood/casa/blob/master/db/seeds.rb>Login instructions here.</a> **Screenshots** It should look like this: ![Screen Shot 2020-08-31 at 6 30 51 PM (2)](https://user-images.githubusercontent.com/62810851/91784656-45749700-ebb8-11ea-8dde-50e91d5f3016.png) Relates to issues: #627, #628, #630
non_process
make edit profile accessible by dropdown menu in all casa admin dashboard part of epic what type of user is this for all casa admin where does should this occur in the all casa admin dashboard dependent on description all casa admin needs an entry point to their edit profile page it should be accessible by clicking on their email in the upper right hand corner of their user dashboard when they click edit profile should appear as the only option in a drop down menu log in as any other user to see how this should look screenshots it should look like this relates to issues
0
14,772
18,049,104,580
IssuesEvent
2021-09-19 12:29:34
openstreetmap-polska/gugik2osm
https://api.github.com/repos/openstreetmap-polska/gugik2osm
closed
Budynki w budowie
data_processing
Aktualnie budynki w budowie są importowane jako zbudowane. Proponuję je oznaczać jako building=construction, ew. pomijać. ![image](https://user-images.githubusercontent.com/10471006/132632867-aa2f447e-7865-4e42-a977-c9a0128ef4f0.png) ![image](https://user-images.githubusercontent.com/10471006/132632969-38a05635-9b59-4729-9e0e-98617c809c70.png)
1.0
Budynki w budowie - Aktualnie budynki w budowie są importowane jako zbudowane. Proponuję je oznaczać jako building=construction, ew. pomijać. ![image](https://user-images.githubusercontent.com/10471006/132632867-aa2f447e-7865-4e42-a977-c9a0128ef4f0.png) ![image](https://user-images.githubusercontent.com/10471006/132632969-38a05635-9b59-4729-9e0e-98617c809c70.png)
process
budynki w budowie aktualnie budynki w budowie są importowane jako zbudowane proponuję je oznaczać jako building construction ew pomijać
1
84,822
16,563,155,407
IssuesEvent
2021-05-29 00:08:45
alefragnani/vscode-pascal-formatter
https://api.github.com/repos/alefragnani/vscode-pascal-formatter
closed
[FEATURE] - Support Workspace Trust API (limited)
enhancement vscode api adoption
<!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> Just like https://github.com/alefragnani/vscode-bookmarks/issues/430 The extension will have **limited** support, not allowing **indicate the formatter app path** on untrusted workspaces.
1.0
[FEATURE] - Support Workspace Trust API (limited) - <!-- Please search existing issues to avoid creating duplicates. --> <!-- Describe the feature you'd like. --> Just like https://github.com/alefragnani/vscode-bookmarks/issues/430 The extension will have **limited** support, not allowing **indicate the formatter app path** on untrusted workspaces.
non_process
support workspace trust api limited just like the extension will have limited support not allowing indicate the formatter app path on untrusted workspaces
0
93,613
3,906,621,301
IssuesEvent
2016-04-19 09:31:57
leoncastillejos/sonar
https://api.github.com/repos/leoncastillejos/sonar
closed
Sonar Web authentication and encryption
action:in_progress priority:high type:enhancement
Right now, Sonar web is cleartext, and does not ask for a password, therefore anyone is able to create, delete, view and modify machines and alerts. This is not ideal.
1.0
Sonar Web authentication and encryption - Right now, Sonar web is cleartext, and does not ask for a password, therefore anyone is able to create, delete, view and modify machines and alerts. This is not ideal.
non_process
sonar web authentication and encryption right now sonar web is cleartext and does not ask for a password therefore anyone is able to create delete view and modify machines and alerts this is not ideal
0
70
2,523,723,824
IssuesEvent
2015-01-20 13:06:53
sysown/proxysql-0.2
https://api.github.com/repos/sysown/proxysql-0.2
opened
Implement table stats_mysql_commands_counters to store counters about commands
ADMIN cxx_db cxx_pa development enhancement MYSQL PROTOCOL QUERY PROCESSOR STATISTICS
Table needs to be created in statsdb Optionally it can also store total execution time
1.0
Implement table stats_mysql_commands_counters to store counters about commands - Table needs to be created in statsdb Optionally it can also store total execution time
process
implement table stats mysql commands counters to store counters about commands table needs to be created in statsdb optionally it can also store total execution time
1
518,850
15,035,991,207
IssuesEvent
2021-02-02 14:46:04
StatCan/daaas
https://api.github.com/repos/StatCan/daaas
closed
Support Construction Starts project by investigating running fastai/pytorch jobs in distributed multi-GPU setup
component/kubeflow current-sprint kind/task priority/soon size/XL triage/support
Goals: * demonstrate running pytorch jobs in distributed multi-gpu setup on platform * demonstrate running fastai jobs in distributed multi-gpu setup on platform * advise construction starts project on how to achieve above multi-gpu setups to scale up their network training * build a tutorial/guide notebooks for distributed multi-gpu jobs
1.0
Support Construction Starts project by investigating running fastai/pytorch jobs in distributed multi-GPU setup - Goals: * demonstrate running pytorch jobs in distributed multi-gpu setup on platform * demonstrate running fastai jobs in distributed multi-gpu setup on platform * advise construction starts project on how to achieve above multi-gpu setups to scale up their network training * build a tutorial/guide notebooks for distributed multi-gpu jobs
non_process
support construction starts project by investigating running fastai pytorch jobs in distributed multi gpu setup goals demonstrate running pytorch jobs in distributed multi gpu setup on platform demonstrate running fastai jobs in distributed multi gpu setup on platform advise construction starts project on how to achieve above multi gpu setups to scale up their network training build a tutorial guide notebooks for distributed multi gpu jobs
0
14,258
9,221,057,571
IssuesEvent
2019-03-11 19:00:00
OriginProtocol/origin
https://api.github.com/repos/OriginProtocol/origin
opened
Shipping Information
dapp discussion html/css javascript p1 security transaction flow ui/ux
This is a placeholder for discussing the long-requested ability for a buyer to add a shipping address when making an offer. This hasn't been done yet because we obviously don't want to store unencrypted, sensitive information publicly. And the messaging infrastructure was not stable enough to rely on for this use.
True
Shipping Information - This is a placeholder for discussing the long-requested ability for a buyer to add a shipping address when making an offer. This hasn't been done yet because we obviously don't want to store unencrypted, sensitive information publicly. And the messaging infrastructure was not stable enough to rely on for this use.
non_process
shipping information this is a placeholder for discussing the long requested ability for a buyer to add a shipping address when making an offer this hasn t been done yet because we obviously don t want to store unencrypted sensitive information publicly and the messaging infrastructure was not stable enough to rely on for this use
0
199,284
22,693,267,196
IssuesEvent
2022-07-05 01:05:37
PGreaneyLYIT/easybuggy4django
https://api.github.com/repos/PGreaneyLYIT/easybuggy4django
opened
CVE-2022-34265 (High) detected in Django-2.0.3-py3-none-any.whl
security vulnerability
## CVE-2022-34265 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.0.3-py3-none-any.whl</b></p></summary> <p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl">https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl</a></p> <p>Path to dependency file: /tmp/ws-scm/easybuggy4django</p> <p>Path to vulnerable library: /tmp/ws-scm/easybuggy4django,/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Django-2.0.3-py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/PGreaneyLYIT/easybuggy4django/commit/5403b8fbcea4b699ce64d05146aeaa76d1062d89">5403b8fbcea4b699ce64d05146aeaa76d1062d89</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Django 3.2 before 3.2.14 and 4.0 before 4.0.6. The Trunc() and Extract() database functions are subject to SQL injection if untrusted data is used as a kind/lookup_name value. Applications that constrain the lookup name and kind choice to a known safe list are unaffected. <p>Publish Date: 2022-07-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34265>CVE-2022-34265</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.djangoproject.com/weblog/2022/jul/04/security-releases/">https://www.djangoproject.com/weblog/2022/jul/04/security-releases/</a></p> <p>Release Date: 2022-06-22</p> <p>Fix Resolution: Django - 3.2.14,4.0.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-34265 (High) detected in Django-2.0.3-py3-none-any.whl - ## CVE-2022-34265 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-2.0.3-py3-none-any.whl</b></p></summary> <p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl">https://files.pythonhosted.org/packages/3d/81/7e6cf5cb6f0f333946b5d3ee22e17c3c3f329d3bfeb86943a2a3cd861092/Django-2.0.3-py3-none-any.whl</a></p> <p>Path to dependency file: /tmp/ws-scm/easybuggy4django</p> <p>Path to vulnerable library: /tmp/ws-scm/easybuggy4django,/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Django-2.0.3-py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/PGreaneyLYIT/easybuggy4django/commit/5403b8fbcea4b699ce64d05146aeaa76d1062d89">5403b8fbcea4b699ce64d05146aeaa76d1062d89</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Django 3.2 before 3.2.14 and 4.0 before 4.0.6. The Trunc() and Extract() database functions are subject to SQL injection if untrusted data is used as a kind/lookup_name value. Applications that constrain the lookup name and kind choice to a known safe list are unaffected. <p>Publish Date: 2022-07-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34265>CVE-2022-34265</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.djangoproject.com/weblog/2022/jul/04/security-releases/">https://www.djangoproject.com/weblog/2022/jul/04/security-releases/</a></p> <p>Release Date: 2022-06-22</p> <p>Fix Resolution: Django - 3.2.14,4.0.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in django none any whl cve high severity vulnerability vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file tmp ws scm path to vulnerable library tmp ws scm requirements txt dependency hierarchy x django none any whl vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in django before and before the trunc and extract database functions are subject to sql injection if untrusted data is used as a kind lookup name value applications that constrain the lookup name and kind choice to a known safe list are unaffected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution django step up your open source security game with mend
0
664,690
22,285,284,120
IssuesEvent
2022-06-11 14:22:09
magento/magento2
https://api.github.com/repos/magento/magento2
closed
jQuery UI Slider and SelectMenu Mapping is not correct in requirejs-config.js
Issue: Confirmed Reproduced on 2.4.x Progress: PR in progress Priority: P2 Area: Framework Reported on 2.4.4
<!--- Thank you for contributing to Magento. To help us process this issue we recommend that you add the following information: - Summary of the issue, - Information on your environment, - Steps to reproduce, - Expected and actual results, Fields marked with (*) are required. Please don't remove the template. Please also have a look at our guidelines article before adding a new issue https://github.com/magento/magento2/wiki/Issue-reporting-guidelines --> ### Preconditions (*) <!--- Please provide as detailed information about your environment as possible. For example Magento version, tag, HEAD, PHP & MySQL version, etc.. --> On Magento 2.4.4 release, jQuery UI Slider and SelectMenu Mapping is not correct in vendor/magento/module-theme/view/base/requirejs-config.js. This causes an error on frontend as jQuery slider and selectmenu modules do not get loaded and we see an error saying "$j(..).slider is not a function". ### Steps to reproduce (*) <!--- It is important to provide a set of clear steps to reproduce this bug. If relevant please include code samples --> Open vendor/magento/module-theme/view/base/requirejs-config.js On line 42 you will notice the incorrect mapping - 'jquery-ui-modules/slider': 'jquery/ui-modules/widgets/selectmenu', ### Expected result (*) <!--- Tell us what should happen --> The correct mapping should be - 'jquery-ui-modules/selectmenu': 'jquery/ui-modules/widgets/selectmenu', 'jquery-ui-modules/slider': 'jquery/ui-modules/widgets/slider', ### Actual result (*) <!--- Tell us what happens instead --> This causes an error on frontend where jQuery slider is used and shows an error saying "$j(..).slider is not a function". <img width="1222" alt="Screenshot 2022-04-19 at 4 38 23 PM" src="https://user-images.githubusercontent.com/22843013/163991117-2a3bcc65-4215-4c43-81ea-991cd8d7e21e.png">
1.0
jQuery UI Slider and SelectMenu Mapping is not correct in requirejs-config.js - <!--- Thank you for contributing to Magento. To help us process this issue we recommend that you add the following information: - Summary of the issue, - Information on your environment, - Steps to reproduce, - Expected and actual results, Fields marked with (*) are required. Please don't remove the template. Please also have a look at our guidelines article before adding a new issue https://github.com/magento/magento2/wiki/Issue-reporting-guidelines --> ### Preconditions (*) <!--- Please provide as detailed information about your environment as possible. For example Magento version, tag, HEAD, PHP & MySQL version, etc.. --> On Magento 2.4.4 release, jQuery UI Slider and SelectMenu Mapping is not correct in vendor/magento/module-theme/view/base/requirejs-config.js. This causes an error on frontend as jQuery slider and selectmenu modules do not get loaded and we see an error saying "$j(..).slider is not a function". ### Steps to reproduce (*) <!--- It is important to provide a set of clear steps to reproduce this bug. If relevant please include code samples --> Open vendor/magento/module-theme/view/base/requirejs-config.js On line 42 you will notice the incorrect mapping - 'jquery-ui-modules/slider': 'jquery/ui-modules/widgets/selectmenu', ### Expected result (*) <!--- Tell us what should happen --> The correct mapping should be - 'jquery-ui-modules/selectmenu': 'jquery/ui-modules/widgets/selectmenu', 'jquery-ui-modules/slider': 'jquery/ui-modules/widgets/slider', ### Actual result (*) <!--- Tell us what happens instead --> This causes an error on frontend where jQuery slider is used and shows an error saying "$j(..).slider is not a function". <img width="1222" alt="Screenshot 2022-04-19 at 4 38 23 PM" src="https://user-images.githubusercontent.com/22843013/163991117-2a3bcc65-4215-4c43-81ea-991cd8d7e21e.png">
non_process
jquery ui slider and selectmenu mapping is not correct in requirejs config js thank you for contributing to magento to help us process this issue we recommend that you add the following information summary of the issue information on your environment steps to reproduce expected and actual results fields marked with are required please don t remove the template please also have a look at our guidelines article before adding a new issue preconditions please provide as detailed information about your environment as possible for example magento version tag head php mysql version etc on magento release jquery ui slider and selectmenu mapping is not correct in vendor magento module theme view base requirejs config js this causes an error on frontend as jquery slider and selectmenu modules do not get loaded and we see an error saying j slider is not a function steps to reproduce it is important to provide a set of clear steps to reproduce this bug if relevant please include code samples open vendor magento module theme view base requirejs config js on line you will notice the incorrect mapping jquery ui modules slider jquery ui modules widgets selectmenu expected result the correct mapping should be jquery ui modules selectmenu jquery ui modules widgets selectmenu jquery ui modules slider jquery ui modules widgets slider actual result this causes an error on frontend where jquery slider is used and shows an error saying j slider is not a function img width alt screenshot at pm src
0
2,055
2,603,975,613
IssuesEvent
2015-02-24 19:01:25
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳治疗湿尤需要多长时间
auto-migrated Priority-Medium Type-Defect
``` 沈阳治疗湿尤需要多长时间〓沈陽軍區政治部醫院性病〓TEL�� �024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治� ��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:17
1.0
沈阳治疗湿尤需要多长时间 - ``` 沈阳治疗湿尤需要多长时间〓沈陽軍區政治部醫院性病〓TEL�� �024-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治� ��。位于沈陽市沈河區二緯路32號。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:17
non_process
沈阳治疗湿尤需要多长时间 沈阳治疗湿尤需要多长时间〓沈陽軍區政治部醫院性病〓tel�� � 〓 , � ��。 。是一所與新中國同建立共� ��煌的歷史悠久、設備精良、技術權威、專家云集,是預防、 保健、醫療、科研康復為一體的綜合性醫院。是國家首批公�� �甲等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大� ��、東南大學等知名高等院校的教學醫院。曾被中國人民解放 軍空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立�� �體二等功。 original issue reported on code google com by gmail com on jun at
0
52,893
10,959,744,161
IssuesEvent
2019-11-27 12:06:05
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
[0.9.5.1] [Issue] [Graphics] Volume Icons on crewlist don't indicate volume.
Bug Code
Right now the new voice chat VOIP volume icon is a dot that extends into a bunch of curved lines to indicate volume. While it properly works in the lobby, only the dot shows up in the ingame crew list in the top left. Someone probably just forgot to port over the new system.
1.0
[0.9.5.1] [Issue] [Graphics] Volume Icons on crewlist don't indicate volume. - Right now the new voice chat VOIP volume icon is a dot that extends into a bunch of curved lines to indicate volume. While it properly works in the lobby, only the dot shows up in the ingame crew list in the top left. Someone probably just forgot to port over the new system.
non_process
volume icons on crewlist don t indicate volume right now the new voice chat voip volume icon is a dot that extends into a bunch of curved lines to indicate volume while it properly works in the lobby only the dot shows up in the ingame crew list in the top left someone probably just forgot to port over the new system
0
20,672
27,335,245,132
IssuesEvent
2023-02-26 05:18:32
python/cpython
https://api.github.com/repos/python/cpython
closed
Process and Thread resource recycling issue
type-bug expert-multiprocessing
# Summary <!-- 如果一个子进程为了提供服务而创建,子进程的父进程如果创建一个线程去轮询对应服务,则当前的进程默认退出机制会先释放子进程,导致线程轮询服务的线程raise Error --> For a child process with daemon=False creates the multiprocessing.Manager object, if it creates a thread for polling SyncManager-created Lock/Event objects, the default recycling mechanism will first release the grandchild process corresponding to the created multiprocessing.Manager object, which can interrupt the polling thread and raise Error. # Description <!-- 测试代码如下: --> The test code is as follows: ```python import time import queue import threading import multiprocessing class GlobalMgr(threading.Thread): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.mgr = multiprocessing.Manager() self.task_q = self.mgr.Queue() self.stop_event = threading.Event() def run(self): while not self.stop_event.is_set(): try: self.task_q.get(timeout=0.001) except queue.Empty as e: continue def subprocess(): mgr = GlobalMgr(daemon=True) mgr.start() time.sleep(1) if __name__ == "__main__": process = multiprocessing.Process(target=subprocess, daemon=False) process.start() process.join() ``` <!-- 上述测试代码会抛出如下异常 --> The above test code will throw the following exception: ```python Traceback (most recent call last): File "/usr/lib/python3.10/threading.py", line 1009, in _bootstrap_inner multiprocessing/process.py 318 source exit self.run() File "/home/hanjiangtao/workspace/program_learning/python_workspace/mp_resource_manager.py", line 18, in run self.task_q.get(timeout=0.001) File "<string>", line 2, in get File "/usr/lib/python3.10/multiprocessing/managers.py", line 833, in _callmethod raise convert_to_error(kind, result) multiprocessing.managers.RemoteError: --------------------------------------------------------------------------- Traceback (most recent call last): File "/usr/lib/python3.10/multiprocessing/managers.py", line 260, in serve_client self.id_to_local_proxy_obj[ident] KeyError: '7fcff61b4d00' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.10/multiprocessing/managers.py", line 262, in serve_client raise ke File "/usr/lib/python3.10/multiprocessing/managers.py", line 256, in serve_client obj, exposed, gettypeid = id_to_obj[ident] KeyError: '7fcff61b4d00' --------------------------------------------------------------------------- ``` <!-- 经过debug,发现上述异常抛出的根因是process子进程创建并start后,BaseProcess._bootstrap被调用,在程序资源释放也就是process对象退出时,走了BaseProcess._bootstrap中finally分支,导致提前调用util._exit_function,而util._exit_function方法会将BaseManger.start方法中注册的BaseManager._finalize_manager提前调用,导致thread还在运行时,SyncManager对象已经被释放,并抛出异常 --> After debugging, it was found that the main cause of the above exception was the invocation of BaseProcess._bootstrap after the subprocess was created and started.While releasing the program resources, that is, when the process object try to exit, the finally brach in BaseProcess._bootstrap will be taken. Which leds to calling util._exit_function in advance. The util._exit_function method will futher call the BaseManager._finalize_manager registered in the BaseManger.start method in advance, and this will trigger the exception during the runtime of thread for the accidental release of SyncManager object.</br> <!-- 但是,multiprocessing.util模块在被import的时候,已经将util._exit_function注册至atexit模块中,所以本身在子进程退出的时候,util._exit_function已经会被被动调用,那么我们是否可以认为在BaseProcess._bootstrap的finally分支的主动调用是非必需的? --> However, the multiprocessing.util module registers util._exit_function in the atexit module automatically when it is imported, which means that util._exit_function will be called naturally while subprocess exits. Consequently, can we consider the active invocation of BaseProcess._bootstrap's finally branch as unnecessary?</br> <!-- 当然,通过学习multiprocessing.util模块的代码,我们可以通过如下方案进行规避: --> Of course, after learning the multiprocessing.util module, we can avoid the issue by the following:</br> ```python import time import queue import threading import multiprocessing class GlobalMgr(threading.Thread): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.mgr = multiprocessing.Manager() self.task_q = self.mgr.Queue() self.stop_event = threading.Event() def run(self): while not self.stop_event.is_set(): try: self.task_q.get(timeout=0.001) except queue.Empty as e: continue def release(mgr): mgr.stop_event.set() mgr.join() def subprocess(): mgr = GlobalMgr(daemon=True) mgr.start() # Actively registers destructors for high-priority _exit_function calls multiprocessing.util.Finalize(None, release, args=(mgr,), exitpriority=100) time.sleep(1) if __name__ == "__main__": process = multiprocessing.Process(target=subprocess, daemon=False) process.start() process.join() ``` <!-- 但是multiprocessing.util.Finalize方法不是模块外部可见方法,所以我认为这只算是个规避的方法。 --> The multiprocessing.util.Finalize module is not an externally visible method, so it's only a hedge.' # Environment > Python 3.10.4</br> > Ubuntu x86_64</br>
1.0
Process and Thread resource recycling issue - # Summary <!-- 如果一个子进程为了提供服务而创建,子进程的父进程如果创建一个线程去轮询对应服务,则当前的进程默认退出机制会先释放子进程,导致线程轮询服务的线程raise Error --> For a child process with daemon=False creates the multiprocessing.Manager object, if it creates a thread for polling SyncManager-created Lock/Event objects, the default recycling mechanism will first release the grandchild process corresponding to the created multiprocessing.Manager object, which can interrupt the polling thread and raise Error. # Description <!-- 测试代码如下: --> The test code is as follows: ```python import time import queue import threading import multiprocessing class GlobalMgr(threading.Thread): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.mgr = multiprocessing.Manager() self.task_q = self.mgr.Queue() self.stop_event = threading.Event() def run(self): while not self.stop_event.is_set(): try: self.task_q.get(timeout=0.001) except queue.Empty as e: continue def subprocess(): mgr = GlobalMgr(daemon=True) mgr.start() time.sleep(1) if __name__ == "__main__": process = multiprocessing.Process(target=subprocess, daemon=False) process.start() process.join() ``` <!-- 上述测试代码会抛出如下异常 --> The above test code will throw the following exception: ```python Traceback (most recent call last): File "/usr/lib/python3.10/threading.py", line 1009, in _bootstrap_inner multiprocessing/process.py 318 source exit self.run() File "/home/hanjiangtao/workspace/program_learning/python_workspace/mp_resource_manager.py", line 18, in run self.task_q.get(timeout=0.001) File "<string>", line 2, in get File "/usr/lib/python3.10/multiprocessing/managers.py", line 833, in _callmethod raise convert_to_error(kind, result) multiprocessing.managers.RemoteError: --------------------------------------------------------------------------- Traceback (most recent call last): File "/usr/lib/python3.10/multiprocessing/managers.py", line 260, in serve_client self.id_to_local_proxy_obj[ident] KeyError: '7fcff61b4d00' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/lib/python3.10/multiprocessing/managers.py", line 262, in serve_client raise ke File "/usr/lib/python3.10/multiprocessing/managers.py", line 256, in serve_client obj, exposed, gettypeid = id_to_obj[ident] KeyError: '7fcff61b4d00' --------------------------------------------------------------------------- ``` <!-- 经过debug,发现上述异常抛出的根因是process子进程创建并start后,BaseProcess._bootstrap被调用,在程序资源释放也就是process对象退出时,走了BaseProcess._bootstrap中finally分支,导致提前调用util._exit_function,而util._exit_function方法会将BaseManger.start方法中注册的BaseManager._finalize_manager提前调用,导致thread还在运行时,SyncManager对象已经被释放,并抛出异常 --> After debugging, it was found that the main cause of the above exception was the invocation of BaseProcess._bootstrap after the subprocess was created and started.While releasing the program resources, that is, when the process object try to exit, the finally brach in BaseProcess._bootstrap will be taken. Which leds to calling util._exit_function in advance. The util._exit_function method will futher call the BaseManager._finalize_manager registered in the BaseManger.start method in advance, and this will trigger the exception during the runtime of thread for the accidental release of SyncManager object.</br> <!-- 但是,multiprocessing.util模块在被import的时候,已经将util._exit_function注册至atexit模块中,所以本身在子进程退出的时候,util._exit_function已经会被被动调用,那么我们是否可以认为在BaseProcess._bootstrap的finally分支的主动调用是非必需的? --> However, the multiprocessing.util module registers util._exit_function in the atexit module automatically when it is imported, which means that util._exit_function will be called naturally while subprocess exits. Consequently, can we consider the active invocation of BaseProcess._bootstrap's finally branch as unnecessary?</br> <!-- 当然,通过学习multiprocessing.util模块的代码,我们可以通过如下方案进行规避: --> Of course, after learning the multiprocessing.util module, we can avoid the issue by the following:</br> ```python import time import queue import threading import multiprocessing class GlobalMgr(threading.Thread): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.mgr = multiprocessing.Manager() self.task_q = self.mgr.Queue() self.stop_event = threading.Event() def run(self): while not self.stop_event.is_set(): try: self.task_q.get(timeout=0.001) except queue.Empty as e: continue def release(mgr): mgr.stop_event.set() mgr.join() def subprocess(): mgr = GlobalMgr(daemon=True) mgr.start() # Actively registers destructors for high-priority _exit_function calls multiprocessing.util.Finalize(None, release, args=(mgr,), exitpriority=100) time.sleep(1) if __name__ == "__main__": process = multiprocessing.Process(target=subprocess, daemon=False) process.start() process.join() ``` <!-- 但是multiprocessing.util.Finalize方法不是模块外部可见方法,所以我认为这只算是个规避的方法。 --> The multiprocessing.util.Finalize module is not an externally visible method, so it's only a hedge.' # Environment > Python 3.10.4</br> > Ubuntu x86_64</br>
process
process and thread resource recycling issue summary 如果一个子进程为了提供服务而创建,子进程的父进程如果创建一个线程去轮询对应服务,则当前的进程默认退出机制会先释放子进程,导致线程轮询服务的线程raise error for a child process with daemon false creates the multiprocessing manager object if it creates a thread for polling syncmanager created lock event objects the default recycling mechanism will first release the grandchild process corresponding to the created multiprocessing manager object which can interrupt the polling thread and raise error description 测试代码如下 the test code is as follows python import time import queue import threading import multiprocessing class globalmgr threading thread def init self args kwargs super init args kwargs self mgr multiprocessing manager self task q self mgr queue self stop event threading event def run self while not self stop event is set try self task q get timeout except queue empty as e continue def subprocess mgr globalmgr daemon true mgr start time sleep if name main process multiprocessing process target subprocess daemon false process start process join 上述测试代码会抛出如下异常 the above test code will throw the following exception python traceback most recent call last file usr lib threading py line in bootstrap inner multiprocessing process py source exit self run file home hanjiangtao workspace program learning python workspace mp resource manager py line in run self task q get timeout file line in get file usr lib multiprocessing managers py line in callmethod raise convert to error kind result multiprocessing managers remoteerror traceback most recent call last file usr lib multiprocessing managers py line in serve client self id to local proxy obj keyerror during handling of the above exception another exception occurred traceback most recent call last file usr lib multiprocessing managers py line in serve client raise ke file usr lib multiprocessing managers py line in serve client obj exposed gettypeid id to obj keyerror 经过debug,发现上述异常抛出的根因是process子进程创建并start后 baseprocess bootstrap被调用,在程序资源释放也就是process对象退出时,走了baseprocess bootstrap中finally分支,导致提前调用util exit function,而util exit function方法会将basemanger start方法中注册的basemanager finalize manager提前调用,导致thread还在运行时,syncmanager对象已经被释放,并抛出异常 after debugging it was found that the main cause of the above exception was the invocation of baseprocess bootstrap after the subprocess was created and started while releasing the program resources that is when the process object try to exit the finally brach in baseprocess bootstrap will be taken which leds to calling util exit function in advance the util exit function method will futher call the basemanager finalize manager registered in the basemanger start method in advance and this will trigger the exception during the runtime of thread for the accidental release of syncmanager object 但是,multiprocessing util模块在被import的时候,已经将util exit function注册至atexit模块中,所以本身在子进程退出的时候,util exit function已经会被被动调用,那么我们是否可以认为在baseprocess bootstrap的finally分支的主动调用是非必需的? however the multiprocessing util module registers util exit function in the atexit module automatically when it is imported which means that util exit function will be called naturally while subprocess exits consequently can we consider the active invocation of baseprocess bootstrap s finally branch as unnecessary 当然,通过学习multiprocessing util模块的代码,我们可以通过如下方案进行规避: of course after learning the multiprocessing util module we can avoid the issue by the following python import time import queue import threading import multiprocessing class globalmgr threading thread def init self args kwargs super init args kwargs self mgr multiprocessing manager self task q self mgr queue self stop event threading event def run self while not self stop event is set try self task q get timeout except queue empty as e continue def release mgr mgr stop event set mgr join def subprocess mgr globalmgr daemon true mgr start actively registers destructors for high priority exit function calls multiprocessing util finalize none release args mgr exitpriority time sleep if name main process multiprocessing process target subprocess daemon false process start process join 但是multiprocessing util finalize方法不是模块外部可见方法,所以我认为这只算是个规避的方法。 the multiprocessing util finalize module is not an externally visible method so it s only a hedge environment python ubuntu
1
7,039
10,197,375,067
IssuesEvent
2019-08-13 00:04:22
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
closed
False positive
whitelisting process
*@macuser666 commented on Aug 12, 2019, 12:33 PM UTC:* hi the following site is blocked [www.bcee.lu](http://www.bcee.lu) bcee.lu It's a legimate bank website of Banque et Caisse d'Épargne de l'État. *This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Ultimate.Hosts.Blacklist#519](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/519).*
1.0
False positive - *@macuser666 commented on Aug 12, 2019, 12:33 PM UTC:* hi the following site is blocked [www.bcee.lu](http://www.bcee.lu) bcee.lu It's a legimate bank website of Banque et Caisse d'Épargne de l'État. *This issue was moved by [funilrys](https://github.com/funilrys) from [mitchellkrogza/Ultimate.Hosts.Blacklist#519](https://github.com/mitchellkrogza/Ultimate.Hosts.Blacklist/issues/519).*
process
false positive commented on aug pm utc hi the following site is blocked bcee lu it s a legimate bank website of banque et caisse d épargne de l état this issue was moved by from
1
197,833
14,944,836,392
IssuesEvent
2021-01-26 02:28:28
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
test_cholesky_solve_batched_many_batches_cuda_complex128 has cuda illegal memory access
high priority module: crash module: linear algebra module: tests triaged
## 🐛 Bug test_cholesky_solve_batched_many_batches_cuda_complex128 has cuda illegal memory access. https://github.com/pytorch/pytorch/pull/47047 might be related. ## To Reproduce Steps to reproduce the behavior: ``` $ PYTORCH_TEST_WITH_SLOW=1 python test/test_linalg.py -v -k test_cholesky_solve_batched_many_batches_cuda_complex128 test_cholesky_solve_batched_many_batches_cuda_complex128 (__main__.TestLinalgCUDA) ... CUDA runtime error: an illegal memory access was encountered (700) in magma_zpotrf_batched at /home/xwang/Developer/magma-2.5.3/src/zpotrf_batched.cpp:234 CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:945 CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:946 CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:947 ERROR ====================================================================== ERROR: test_cholesky_solve_batched_many_batches_cuda_complex128 (__main__.TestLinalgCUDA) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_utils.py", line 864, in wrapper method(*args, **kwargs) File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 273, in instantiated_test result = test_fn(self, *args) File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_utils.py", line 542, in wrapper fn(*args, **kwargs) File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 545, in dep_fn return fn(slf, device, *args, **kwargs) File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 545, in dep_fn return fn(slf, device, *args, **kwargs) File "test/test_linalg.py", line 1906, in test_cholesky_solve_batched_many_batches b, A, L = self.cholesky_solve_test_helper(A_dims, b_dims, upper, device, dtype) File "test/test_linalg.py", line 1845, in cholesky_solve_test_helper L = torch.cholesky(A, upper=upper) RuntimeError: CUDA error: an illegal memory access was encountered Exception raised from magmaCholeskyBatched<c10::complex<double> > at /home/xwang/Developer/pytorch/aten/src/ATen/native/cuda/BatchLinearAlgebra.cu:653 (most recent call first): frame #0: c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >) + 0x68 (0x7efde8c50828 in /home/xwang/Developer/pytorch/torch/lib/libc10.so) frame #1: void at::native::magmaCholeskyBatched<c10::complex<double> >(magma_uplo_t, int, c10::complex<double>**, int, int*, int, at::native::MAGMAQueue const&) + 0x155 (0x7efde9fffbd5 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so) frame #2: at::native::_cholesky_helper_cuda(at::Tensor const&, bool) + 0x1709 (0x7efdea017d19 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so) frame #3: <unknown function> + 0x3336c13 (0x7efdebfdcc13 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so) frame #4: <unknown function> + 0x3336c84 (0x7efdebfdcc84 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so) frame #5: at::_cholesky_helper(at::Tensor const&, bool) + 0x116 (0x7efdfde476a6 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #6: at::native::cholesky(at::Tensor const&, bool) + 0xa9 (0x7efdfd8b2c99 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #7: <unknown function> + 0x1c388f3 (0x7efdfe0398f3 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #8: <unknown function> + 0x1c38964 (0x7efdfe039964 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #9: at::cholesky(at::Tensor const&, bool) + 0x116 (0x7efdfde46bc6 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #10: <unknown function> + 0x31ab9c4 (0x7efdff5ac9c4 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #11: <unknown function> + 0x31abcb4 (0x7efdff5accb4 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #12: at::Tensor::cholesky(bool) const + 0x116 (0x7efdfe1a7696 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #13: <unknown function> + 0x734e6d (0x7efe04b8de6d in /home/xwang/Developer/pytorch/torch/lib/libtorch_python.so) <omitting python frames> ---------------------------------------------------------------------- Ran 1 test in 2.616s FAILED (errors=1) ``` ## Expected behavior No fail ## Environment ``` Collecting environment information... PyTorch version: 1.8.0a0+533c837 Is debug build: False CUDA used to build PyTorch: 11.1 ROCM used to build PyTorch: N/A OS: Manjaro Linux (x86_64) GCC version: (GCC) 10.2.0 Clang version: Could not collect CMake version: version 3.18.4 Python version: 3.8 (64-bit runtime) Is CUDA available: True CUDA runtime version: Could not collect GPU models and configuration: GPU 0: GeForce RTX 2070 SUPER GPU 1: GeForce GTX 1070 Ti Nvidia driver version: 455.38 cuDNN version: Probably one of the following: /usr/lib/libcudnn.so.8.0.5 /usr/lib/libcudnn_adv_infer.so.8.0.5 /usr/lib/libcudnn_adv_train.so.8.0.5 /usr/lib/libcudnn_cnn_infer.so.8.0.5 /usr/lib/libcudnn_cnn_train.so.8.0.5 /usr/lib/libcudnn_ops_infer.so.8.0.5 /usr/lib/libcudnn_ops_train.so.8.0.5 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.19.4 [pip3] pytorch-ignite==0.4.2 [pip3] torch==1.8.0a0 [pip3] torchvision==0.9.0a0+74de51d [conda] Could not collect ``` ## Additional context Seems to be MAGMA-related. CC @ptrblck @mruberry cc @ezyang @gchanan @zou3519 @bdhirsh @jianyuh @nikitaved @pearu @mruberry @heitorschueroff @walterddr @VitalyFedyunin @IvanYashchuk
1.0
test_cholesky_solve_batched_many_batches_cuda_complex128 has cuda illegal memory access - ## 🐛 Bug test_cholesky_solve_batched_many_batches_cuda_complex128 has cuda illegal memory access. https://github.com/pytorch/pytorch/pull/47047 might be related. ## To Reproduce Steps to reproduce the behavior: ``` $ PYTORCH_TEST_WITH_SLOW=1 python test/test_linalg.py -v -k test_cholesky_solve_batched_many_batches_cuda_complex128 test_cholesky_solve_batched_many_batches_cuda_complex128 (__main__.TestLinalgCUDA) ... CUDA runtime error: an illegal memory access was encountered (700) in magma_zpotrf_batched at /home/xwang/Developer/magma-2.5.3/src/zpotrf_batched.cpp:234 CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:945 CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:946 CUDA runtime error: an illegal memory access was encountered (700) in magma_queue_destroy_internal at /home/xwang/Developer/magma-2.5.3/interface_cuda/interface.cpp:947 ERROR ====================================================================== ERROR: test_cholesky_solve_batched_many_batches_cuda_complex128 (__main__.TestLinalgCUDA) ---------------------------------------------------------------------- Traceback (most recent call last): File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_utils.py", line 864, in wrapper method(*args, **kwargs) File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 273, in instantiated_test result = test_fn(self, *args) File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_utils.py", line 542, in wrapper fn(*args, **kwargs) File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 545, in dep_fn return fn(slf, device, *args, **kwargs) File "/home/xwang/Developer/pytorch/torch/testing/_internal/common_device_type.py", line 545, in dep_fn return fn(slf, device, *args, **kwargs) File "test/test_linalg.py", line 1906, in test_cholesky_solve_batched_many_batches b, A, L = self.cholesky_solve_test_helper(A_dims, b_dims, upper, device, dtype) File "test/test_linalg.py", line 1845, in cholesky_solve_test_helper L = torch.cholesky(A, upper=upper) RuntimeError: CUDA error: an illegal memory access was encountered Exception raised from magmaCholeskyBatched<c10::complex<double> > at /home/xwang/Developer/pytorch/aten/src/ATen/native/cuda/BatchLinearAlgebra.cu:653 (most recent call first): frame #0: c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >) + 0x68 (0x7efde8c50828 in /home/xwang/Developer/pytorch/torch/lib/libc10.so) frame #1: void at::native::magmaCholeskyBatched<c10::complex<double> >(magma_uplo_t, int, c10::complex<double>**, int, int*, int, at::native::MAGMAQueue const&) + 0x155 (0x7efde9fffbd5 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so) frame #2: at::native::_cholesky_helper_cuda(at::Tensor const&, bool) + 0x1709 (0x7efdea017d19 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so) frame #3: <unknown function> + 0x3336c13 (0x7efdebfdcc13 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so) frame #4: <unknown function> + 0x3336c84 (0x7efdebfdcc84 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cuda.so) frame #5: at::_cholesky_helper(at::Tensor const&, bool) + 0x116 (0x7efdfde476a6 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #6: at::native::cholesky(at::Tensor const&, bool) + 0xa9 (0x7efdfd8b2c99 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #7: <unknown function> + 0x1c388f3 (0x7efdfe0398f3 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #8: <unknown function> + 0x1c38964 (0x7efdfe039964 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #9: at::cholesky(at::Tensor const&, bool) + 0x116 (0x7efdfde46bc6 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #10: <unknown function> + 0x31ab9c4 (0x7efdff5ac9c4 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #11: <unknown function> + 0x31abcb4 (0x7efdff5accb4 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #12: at::Tensor::cholesky(bool) const + 0x116 (0x7efdfe1a7696 in /home/xwang/Developer/pytorch/torch/lib/libtorch_cpu.so) frame #13: <unknown function> + 0x734e6d (0x7efe04b8de6d in /home/xwang/Developer/pytorch/torch/lib/libtorch_python.so) <omitting python frames> ---------------------------------------------------------------------- Ran 1 test in 2.616s FAILED (errors=1) ``` ## Expected behavior No fail ## Environment ``` Collecting environment information... PyTorch version: 1.8.0a0+533c837 Is debug build: False CUDA used to build PyTorch: 11.1 ROCM used to build PyTorch: N/A OS: Manjaro Linux (x86_64) GCC version: (GCC) 10.2.0 Clang version: Could not collect CMake version: version 3.18.4 Python version: 3.8 (64-bit runtime) Is CUDA available: True CUDA runtime version: Could not collect GPU models and configuration: GPU 0: GeForce RTX 2070 SUPER GPU 1: GeForce GTX 1070 Ti Nvidia driver version: 455.38 cuDNN version: Probably one of the following: /usr/lib/libcudnn.so.8.0.5 /usr/lib/libcudnn_adv_infer.so.8.0.5 /usr/lib/libcudnn_adv_train.so.8.0.5 /usr/lib/libcudnn_cnn_infer.so.8.0.5 /usr/lib/libcudnn_cnn_train.so.8.0.5 /usr/lib/libcudnn_ops_infer.so.8.0.5 /usr/lib/libcudnn_ops_train.so.8.0.5 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.19.4 [pip3] pytorch-ignite==0.4.2 [pip3] torch==1.8.0a0 [pip3] torchvision==0.9.0a0+74de51d [conda] Could not collect ``` ## Additional context Seems to be MAGMA-related. CC @ptrblck @mruberry cc @ezyang @gchanan @zou3519 @bdhirsh @jianyuh @nikitaved @pearu @mruberry @heitorschueroff @walterddr @VitalyFedyunin @IvanYashchuk
non_process
test cholesky solve batched many batches cuda has cuda illegal memory access 🐛 bug test cholesky solve batched many batches cuda has cuda illegal memory access might be related to reproduce steps to reproduce the behavior pytorch test with slow python test test linalg py v k test cholesky solve batched many batches cuda test cholesky solve batched many batches cuda main testlinalgcuda cuda runtime error an illegal memory access was encountered in magma zpotrf batched at home xwang developer magma src zpotrf batched cpp cuda runtime error an illegal memory access was encountered in magma queue destroy internal at home xwang developer magma interface cuda interface cpp cuda runtime error an illegal memory access was encountered in magma queue destroy internal at home xwang developer magma interface cuda interface cpp cuda runtime error an illegal memory access was encountered in magma queue destroy internal at home xwang developer magma interface cuda interface cpp error error test cholesky solve batched many batches cuda main testlinalgcuda traceback most recent call last file home xwang developer pytorch torch testing internal common utils py line in wrapper method args kwargs file home xwang developer pytorch torch testing internal common device type py line in instantiated test result test fn self args file home xwang developer pytorch torch testing internal common utils py line in wrapper fn args kwargs file home xwang developer pytorch torch testing internal common device type py line in dep fn return fn slf device args kwargs file home xwang developer pytorch torch testing internal common device type py line in dep fn return fn slf device args kwargs file test test linalg py line in test cholesky solve batched many batches b a l self cholesky solve test helper a dims b dims upper device dtype file test test linalg py line in cholesky solve test helper l torch cholesky a upper upper runtimeerror cuda error an illegal memory access was encountered exception raised from magmacholeskybatched at home xwang developer pytorch aten src aten native cuda batchlinearalgebra cu most recent call first frame error error sourcelocation std basic string std allocator in home xwang developer pytorch torch lib so frame void at native magmacholeskybatched magma uplo t int complex int int int at native magmaqueue const in home xwang developer pytorch torch lib libtorch cuda so frame at native cholesky helper cuda at tensor const bool in home xwang developer pytorch torch lib libtorch cuda so frame in home xwang developer pytorch torch lib libtorch cuda so frame in home xwang developer pytorch torch lib libtorch cuda so frame at cholesky helper at tensor const bool in home xwang developer pytorch torch lib libtorch cpu so frame at native cholesky at tensor const bool in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch cpu so frame at cholesky at tensor const bool in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch cpu so frame at tensor cholesky bool const in home xwang developer pytorch torch lib libtorch cpu so frame in home xwang developer pytorch torch lib libtorch python so ran test in failed errors expected behavior no fail environment collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os manjaro linux gcc version gcc clang version could not collect cmake version version python version bit runtime is cuda available true cuda runtime version could not collect gpu models and configuration gpu geforce rtx super gpu geforce gtx ti nvidia driver version cudnn version probably one of the following usr lib libcudnn so usr lib libcudnn adv infer so usr lib libcudnn adv train so usr lib libcudnn cnn infer so usr lib libcudnn cnn train so usr lib libcudnn ops infer so usr lib libcudnn ops train so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy pytorch ignite torch torchvision could not collect additional context seems to be magma related cc ptrblck mruberry cc ezyang gchanan bdhirsh jianyuh nikitaved pearu mruberry heitorschueroff walterddr vitalyfedyunin ivanyashchuk
0
163,811
12,745,623,589
IssuesEvent
2020-06-26 14:33:47
eclipse/openj9
https://api.github.com/repos/eclipse/openj9
closed
samplingObjectAllocation.soae001 failed, expected 1+ but got: 0
comp:gc test failure
https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/3 https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_x86-64_windows_Nightly_testList_0/10 cmdLineTester_jvmtitests_soae_3 ``` Testing: soae001 Test start time: 2020/06/05 04:04:16 Central Standard Time Running command: "C:/Users/jenkins/workspace/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/openjdkbinary/j2sdk-image\\bin\\java" -Xcompressedrefs -Xjit -Xgcpolicy:gencon -Xdump -agentlib:jvmtitest=test:soae001 -cp "C:/Users/jenkins/workspace/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/openjdk-tests///..//jvmtest\functional\cmdLineTests\jvmtitests\jvmtitest.jar" com.ibm.jvmti.tests.util.TestRunner Time spent starting: 15 milliseconds Time spent executing: 781 milliseconds Test result: FAILED Output from test: [OUT] *** Testing [1/1]: testDefaultInterval [OUT] Allocated a byte array with size 524288 [OUT] com.ibm.jvmti.tests.samplingObjectAllocation.soae001.check() failed, expected 1+ but got: 0 [OUT] *** Test took 15 milliseconds [OUT] FAILED [OUT] >> Success condition was not found: [Return code: 0] ``` https://github.com/eclipse/openj9/compare/589eaa5...5f0390e I expect this is caused by https://github.com/eclipse/openj9/pull/9745
1.0
samplingObjectAllocation.soae001 failed, expected 1+ but got: 0 - https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/3 https://ci.eclipse.org/openj9/job/Test_openjdk11_j9_sanity.functional_x86-64_windows_Nightly_testList_0/10 cmdLineTester_jvmtitests_soae_3 ``` Testing: soae001 Test start time: 2020/06/05 04:04:16 Central Standard Time Running command: "C:/Users/jenkins/workspace/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/openjdkbinary/j2sdk-image\\bin\\java" -Xcompressedrefs -Xjit -Xgcpolicy:gencon -Xdump -agentlib:jvmtitest=test:soae001 -cp "C:/Users/jenkins/workspace/Test_openjdk11_j9_sanity.functional_x86-64_windows_OpenJDK11_testList_0/openjdk-tests///..//jvmtest\functional\cmdLineTests\jvmtitests\jvmtitest.jar" com.ibm.jvmti.tests.util.TestRunner Time spent starting: 15 milliseconds Time spent executing: 781 milliseconds Test result: FAILED Output from test: [OUT] *** Testing [1/1]: testDefaultInterval [OUT] Allocated a byte array with size 524288 [OUT] com.ibm.jvmti.tests.samplingObjectAllocation.soae001.check() failed, expected 1+ but got: 0 [OUT] *** Test took 15 milliseconds [OUT] FAILED [OUT] >> Success condition was not found: [Return code: 0] ``` https://github.com/eclipse/openj9/compare/589eaa5...5f0390e I expect this is caused by https://github.com/eclipse/openj9/pull/9745
non_process
samplingobjectallocation failed expected but got cmdlinetester jvmtitests soae testing test start time central standard time running command c users jenkins workspace test sanity functional windows testlist openjdkbinary image bin java xcompressedrefs xjit xgcpolicy gencon xdump agentlib jvmtitest test cp c users jenkins workspace test sanity functional windows testlist openjdk tests jvmtest functional cmdlinetests jvmtitests jvmtitest jar com ibm jvmti tests util testrunner time spent starting milliseconds time spent executing milliseconds test result failed output from test testing testdefaultinterval allocated a byte array with size com ibm jvmti tests samplingobjectallocation check failed expected but got test took milliseconds failed success condition was not found i expect this is caused by
0
401,361
11,789,200,533
IssuesEvent
2020-03-17 16:45:42
aces/Loris
https://api.github.com/repos/aces/Loris
closed
[New Module] Electrophysiology browser
Feature Priority: High
List of bugs found in the current set up: - Task Name seems to be hardcoded to FaceHousCheck. Fix that so it takes that information from the physiological_parameter_file table (TaskName is one of the JSON file field). In jsx/components/eeg_session_panels.js and js/eeg_session_view.js
1.0
[New Module] Electrophysiology browser - List of bugs found in the current set up: - Task Name seems to be hardcoded to FaceHousCheck. Fix that so it takes that information from the physiological_parameter_file table (TaskName is one of the JSON file field). In jsx/components/eeg_session_panels.js and js/eeg_session_view.js
non_process
electrophysiology browser list of bugs found in the current set up task name seems to be hardcoded to facehouscheck fix that so it takes that information from the physiological parameter file table taskname is one of the json file field in jsx components eeg session panels js and js eeg session view js
0
17,102
22,622,771,145
IssuesEvent
2022-06-30 08:03:13
2i2c-org/infrastructure
https://api.github.com/repos/2i2c-org/infrastructure
closed
Define some first-line and second-line support processes
:label: team-process
### Background and proposal There are often cases where our support process is under-documented. For example, a few questions that people weren't sure how to answer: - How should we prioritize support requests? - What should we do if a request is not immediately "closeable"? Or if it requires ongoing follow-up work? - How can we communicate our inability to fulfill a request? - What kind of communication should we use throughout the support process? We should document some rough guidelines for these common questions, and also provide references to documentation about how to carry out first-line support. ### Implementation guide and constraints Another way to approach this is to ask "what are some common support situations, and what should we do in each situation?" We can draw from our experiences thus far to agree on some team practices to follow. ### Updates and ongoing work _Add items below as we learn more_ - [x] Do some research into support processes at other orgs (see refs below) - [x] #1155 - [x] #1154 ### Refs - [Distributed communication and incident response write-up](https://queue.acm.org/detail.cfm?id=3380779) - [Example support process from Sarah](https://github.com/2i2c-org/infrastructure/issues/1068#issuecomment-1063138772) - [Example support process from Chris](https://github.com/2i2c-org/infrastructure/issues/1068#issuecomment-1063516429) - [Wikimedia clinic duty](https://wikitech.wikimedia.org/wiki/SRE/Clinic_Duty#Responsibilities) - We have a semi-related issue to this here: https://github.com/2i2c-org/infrastructure/issues/1118
1.0
Define some first-line and second-line support processes - ### Background and proposal There are often cases where our support process is under-documented. For example, a few questions that people weren't sure how to answer: - How should we prioritize support requests? - What should we do if a request is not immediately "closeable"? Or if it requires ongoing follow-up work? - How can we communicate our inability to fulfill a request? - What kind of communication should we use throughout the support process? We should document some rough guidelines for these common questions, and also provide references to documentation about how to carry out first-line support. ### Implementation guide and constraints Another way to approach this is to ask "what are some common support situations, and what should we do in each situation?" We can draw from our experiences thus far to agree on some team practices to follow. ### Updates and ongoing work _Add items below as we learn more_ - [x] Do some research into support processes at other orgs (see refs below) - [x] #1155 - [x] #1154 ### Refs - [Distributed communication and incident response write-up](https://queue.acm.org/detail.cfm?id=3380779) - [Example support process from Sarah](https://github.com/2i2c-org/infrastructure/issues/1068#issuecomment-1063138772) - [Example support process from Chris](https://github.com/2i2c-org/infrastructure/issues/1068#issuecomment-1063516429) - [Wikimedia clinic duty](https://wikitech.wikimedia.org/wiki/SRE/Clinic_Duty#Responsibilities) - We have a semi-related issue to this here: https://github.com/2i2c-org/infrastructure/issues/1118
process
define some first line and second line support processes background and proposal there are often cases where our support process is under documented for example a few questions that people weren t sure how to answer how should we prioritize support requests what should we do if a request is not immediately closeable or if it requires ongoing follow up work how can we communicate our inability to fulfill a request what kind of communication should we use throughout the support process we should document some rough guidelines for these common questions and also provide references to documentation about how to carry out first line support implementation guide and constraints another way to approach this is to ask what are some common support situations and what should we do in each situation we can draw from our experiences thus far to agree on some team practices to follow updates and ongoing work add items below as we learn more do some research into support processes at other orgs see refs below refs we have a semi related issue to this here
1
136,184
18,722,443,415
IssuesEvent
2021-11-03 13:17:12
ioana-github-enterprise/simplebuild
https://api.github.com/repos/ioana-github-enterprise/simplebuild
opened
CVE-2019-16335 (High) detected in jackson-databind-2.7.9.jar
security vulnerability
## CVE-2019-16335 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: simplebuild/build.gradle</p> <p>Path to vulnerable library: le/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9/a4c0b14c7dd85bdf4d25da074e90a10fa4b9b88b/jackson-databind-2.7.9.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.7.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ioana-github-enterprise/simplebuild/commit/c99f55cc0f24d3137d4aa52ce1de8552a9d07579">c99f55cc0f24d3137d4aa52ce1de8552a9d07579</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariDataSource. This is a different vulnerability than CVE-2019-14540. <p>Publish Date: 2019-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16335>CVE-2019-16335</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/blob/master/release-notes/VERSION-2.x">https://github.com/FasterXML/jackson-databind/blob/master/release-notes/VERSION-2.x</a></p> <p>Release Date: 2020-10-20</p> <p>Fix Resolution: 2.9.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.9","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16335","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariDataSource. This is a different vulnerability than CVE-2019-14540.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16335","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-16335 (High) detected in jackson-databind-2.7.9.jar - ## CVE-2019-16335 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.7.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: simplebuild/build.gradle</p> <p>Path to vulnerable library: le/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.7.9/a4c0b14c7dd85bdf4d25da074e90a10fa4b9b88b/jackson-databind-2.7.9.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.7.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ioana-github-enterprise/simplebuild/commit/c99f55cc0f24d3137d4aa52ce1de8552a9d07579">c99f55cc0f24d3137d4aa52ce1de8552a9d07579</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariDataSource. This is a different vulnerability than CVE-2019-14540. <p>Publish Date: 2019-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16335>CVE-2019-16335</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/blob/master/release-notes/VERSION-2.x">https://github.com/FasterXML/jackson-databind/blob/master/release-notes/VERSION-2.x</a></p> <p>Release Date: 2020-10-20</p> <p>Fix Resolution: 2.9.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.7.9","packageFilePaths":["/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.7.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2019-16335","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariDataSource. This is a different vulnerability than CVE-2019-14540.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16335","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file simplebuild build gradle path to vulnerable library le caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikaridatasource this is a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikaridatasource this is a different vulnerability than cve vulnerabilityurl
0
14,433
17,483,498,373
IssuesEvent
2021-08-09 07:52:05
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[API Proposal]: System.Diagnostics.Process.WaitForExit() should return the Process to allow chaining of calls
api-suggestion area-System.Diagnostics.Process untriaged
### Background and motivation This would allow method calls to be chained and reduce the number of lines of code required for certain actions. ### API Proposal Before ```C# namespace System.Diagnostics { public class Process { public void WaitForExit(); } } ``` After: ```C# namespace System.Diagnostics { public class Process { public Process WaitForExit(); } } ``` ### API Usage Before: ```C# Process p = Process.Start(exeFilePath); p.WaitForExit(); if (p.ExitCode == 1) { // Handle the error somehow } ``` After: ```C# if (Process.Start(exeFilePath).WaitForExit().ExitCode == 1) { // Handle the error somehow } ``` ### Risks It currently returns void, so I'm not sure if this would cause any issues with backwards compatibility. Since if someone is already using the above 'Before' code it would continue working without any changes.
1.0
[API Proposal]: System.Diagnostics.Process.WaitForExit() should return the Process to allow chaining of calls - ### Background and motivation This would allow method calls to be chained and reduce the number of lines of code required for certain actions. ### API Proposal Before ```C# namespace System.Diagnostics { public class Process { public void WaitForExit(); } } ``` After: ```C# namespace System.Diagnostics { public class Process { public Process WaitForExit(); } } ``` ### API Usage Before: ```C# Process p = Process.Start(exeFilePath); p.WaitForExit(); if (p.ExitCode == 1) { // Handle the error somehow } ``` After: ```C# if (Process.Start(exeFilePath).WaitForExit().ExitCode == 1) { // Handle the error somehow } ``` ### Risks It currently returns void, so I'm not sure if this would cause any issues with backwards compatibility. Since if someone is already using the above 'Before' code it would continue working without any changes.
process
system diagnostics process waitforexit should return the process to allow chaining of calls background and motivation this would allow method calls to be chained and reduce the number of lines of code required for certain actions api proposal before c namespace system diagnostics public class process public void waitforexit after c namespace system diagnostics public class process public process waitforexit api usage before c process p process start exefilepath p waitforexit if p exitcode handle the error somehow after c if process start exefilepath waitforexit exitcode handle the error somehow risks it currently returns void so i m not sure if this would cause any issues with backwards compatibility since if someone is already using the above before code it would continue working without any changes
1
99,941
16,470,788,375
IssuesEvent
2021-05-23 11:19:40
axe-api/axe-api
https://api.github.com/repos/axe-api/axe-api
closed
Relationship queries
enhancement security
We need to add `with` feature to the **Queries**. ## Examples `api/users?with=posts{id|title|comments{content}}` ## Dependencies We should update the documentations.
True
Relationship queries - We need to add `with` feature to the **Queries**. ## Examples `api/users?with=posts{id|title|comments{content}}` ## Dependencies We should update the documentations.
non_process
relationship queries we need to add with feature to the queries examples api users with posts id title comments content dependencies we should update the documentations
0
518,712
15,033,172,642
IssuesEvent
2021-02-02 11:07:54
YangCatalog/backend
https://api.github.com/repos/YangCatalog/backend
closed
Reading log from .gz files
Priority: Critical bug
since we are compressing log files on daily basis we need to be able to read them back programmatically as well
1.0
Reading log from .gz files - since we are compressing log files on daily basis we need to be able to read them back programmatically as well
non_process
reading log from gz files since we are compressing log files on daily basis we need to be able to read them back programmatically as well
0
41,853
10,683,695,202
IssuesEvent
2019-10-22 08:54:49
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Regression in Firebird's SUBSTRING() implementation
C: DB: Firebird C: Functionality E: All Editions P: High R: Fixed T: Defect
We're currently generating bad SQL for `substring` in Firebird: ```sql select substring('abcde' from 1), substring('abcde' from 1 for2), substring('abcde' from 3), substring('abcde' from 3 for2) from RDB$DATABASE ``` The regression was introduced by this, or a related change: https://github.com/jOOQ/jOOQ/commit/b2d1119e24f33984c17af649ead532848f5c11ac
1.0
Regression in Firebird's SUBSTRING() implementation - We're currently generating bad SQL for `substring` in Firebird: ```sql select substring('abcde' from 1), substring('abcde' from 1 for2), substring('abcde' from 3), substring('abcde' from 3 for2) from RDB$DATABASE ``` The regression was introduced by this, or a related change: https://github.com/jOOQ/jOOQ/commit/b2d1119e24f33984c17af649ead532848f5c11ac
non_process
regression in firebird s substring implementation we re currently generating bad sql for substring in firebird sql select substring abcde from substring abcde from substring abcde from substring abcde from from rdb database the regression was introduced by this or a related change
0
760,717
26,653,997,110
IssuesEvent
2023-01-25 15:34:47
KingGizzard/Ballotbox
https://api.github.com/repos/KingGizzard/Ballotbox
closed
Build Agent 3
high priority
Agent 3 is the data request agent! You can see the dynamics of agent 3 in the [figma diagram here](https://www.figma.com/file/SkUdZjkenjeh5WQvJIe9vz/BallotBox?node-id=0%3A1&t=Gy4phXY2zIQ3Scjy-0) Agent 3 should be able to: - Request data from the Ballotbox.sol contract, and provide an email to be contacted through - Access IPFS data and save data locally ! - when the encryption methods for the [Oracle](https://github.com/KingGizzard/Ballotbox/issues/1) and [Agent 2](https://github.com/KingGizzard/Ballotbox/issues/4) are finished, then Agent 3 should be able to: - Decrypt IPFS data
1.0
Build Agent 3 - Agent 3 is the data request agent! You can see the dynamics of agent 3 in the [figma diagram here](https://www.figma.com/file/SkUdZjkenjeh5WQvJIe9vz/BallotBox?node-id=0%3A1&t=Gy4phXY2zIQ3Scjy-0) Agent 3 should be able to: - Request data from the Ballotbox.sol contract, and provide an email to be contacted through - Access IPFS data and save data locally ! - when the encryption methods for the [Oracle](https://github.com/KingGizzard/Ballotbox/issues/1) and [Agent 2](https://github.com/KingGizzard/Ballotbox/issues/4) are finished, then Agent 3 should be able to: - Decrypt IPFS data
non_process
build agent agent is the data request agent you can see the dynamics of agent in the agent should be able to request data from the ballotbox sol contract and provide an email to be contacted through access ipfs data and save data locally when the encryption methods for the and are finished then agent should be able to decrypt ipfs data
0
12,246
14,744,109,121
IssuesEvent
2021-01-07 14:51:23
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Holiday Fee Bug - Child accounts
anc-process anp-0.5 ant-bug ant-enhancement has attachment
In GitLab by @kdjstudios on Jan 2, 2020, 14:15 **Submitted by:** Kyle **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/10125583 **Server:** ALL **Client/Site:** NA **Account:** NA **Issue:** During a phone call with Leah we believe we found an issue with the holiday fee settings with parent child accounts. We setup a parent account (056-MX3177) with the Holiday fee enabled and a child account (056-M20004) with it disabled and a zero value. When we generated the draft invoice it was showing two line items for the holiday fee. It should only be charging the one as the child account has it disabled. ![image](/uploads/418ff31fe73a70cd919d1efe34f5777c/image.png)
1.0
Holiday Fee Bug - Child accounts - In GitLab by @kdjstudios on Jan 2, 2020, 14:15 **Submitted by:** Kyle **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/10125583 **Server:** ALL **Client/Site:** NA **Account:** NA **Issue:** During a phone call with Leah we believe we found an issue with the holiday fee settings with parent child accounts. We setup a parent account (056-MX3177) with the Holiday fee enabled and a child account (056-M20004) with it disabled and a zero value. When we generated the draft invoice it was showing two line items for the holiday fee. It should only be charging the one as the child account has it disabled. ![image](/uploads/418ff31fe73a70cd919d1efe34f5777c/image.png)
process
holiday fee bug child accounts in gitlab by kdjstudios on jan submitted by kyle helpdesk server all client site na account na issue during a phone call with leah we believe we found an issue with the holiday fee settings with parent child accounts we setup a parent account with the holiday fee enabled and a child account with it disabled and a zero value when we generated the draft invoice it was showing two line items for the holiday fee it should only be charging the one as the child account has it disabled uploads image png
1
23,672
2,660,221,062
IssuesEvent
2015-03-19 04:05:47
dartsim/dart
https://api.github.com/repos/dartsim/dart
closed
Bug in FreeJoint helper
Kind: Bug Priority: High
* File: dart/dynamics/FreeJoint.cpp * Functions compromised: convertToPositions (line 61) and converToTransform (line 70) * Problem description: I am generating a set of transformations with rotations around Z with angles in the range [0,2*PI], increasing at each step. I use these Transformations to set the joint positions of a freeJoint (using the same formula as used in convertToPositions). The joint positions I get are wrong for some angles. The joint positions correspond to the intended Transformation rotated 90 around Z. This happens usually after half of the range (either if I use [0,2PI] or [-PI,PI]. An example in which the bug happens is copied at the end of this issue. In this example I use: Input transform: ``` -0.815214 -0.57916 0 0.437796 -0.57916 0.815214 0 0.488175 0 0 -1 1.03 0 0 0 1 ``` Positions obtained using convertToPositions(): ``` -0.954925 -2.99295 -0 0.437796 0.488175 1.03 ``` Output transform obtained using convertToTransform() (should be the same as input transform): ``` -0.815214 0.57916 -5.39747e-16 0.437796 0.57916 0.815214 1.72211e-16 0.488175 5.39747e-16 -1.72211e-16 -1 1.03 6.95317e-310 1.38338e-322 6.92085e-310 2.12999e-317 ``` Notice the flipped signs in x an y columns. The problem described above is related to convertToPositions. I also noticed that in convertToTransform, there is a line missing: Tf.setIdentity(). Otherwise, the last row of Tf will be filled with garbage (as in the example above). * Final note: Please, if somebody can fix this quickly I would appreciate it. I spent most of today searching for the bug in my own code until I realized that the bug was (very possibly and unless I am mistaken) in dart so I would like to continue with my work once the bug is fixed. Thanks! PS.- Bug test file: ```cpp #include <dart/math/Geometry.h> #include <iostream> Eigen::Vector6d convertToPositions(const Eigen::Isometry3d& _tf); Eigen::Isometry3d convertToTransform(const Eigen::Vector6d& _positions); int main( int argc, char* argv[] ) { Eigen::Isometry3d Tf; Eigen::VectorXd q; Eigen::Isometry3d Tf_back; Tf.setIdentity(); Tf.translation() << 0.1, 0.2, 0.3; double dang = 10*M_PI/180.0; int n = (int) (2*M_PI /dang); Tf.matrix() << -0.815214, -0.57916, -0, 0.437796, -0.57916, 0.815214, 0, 0.488175, 0, 0, -1, 1.03, 0, 0, 0, 1; double det = Tf.linear().determinant(); std::cout << "Determinant of test TF: "<< det << std::endl; /* for( size_t i = 0; i < n; ++i ) { Eigen::Matrix3d rot; rot = Eigen::AngleAxisd( i*dang, Eigen::Vector3d(0,0,1) ); Tf.linear() = rot; */ // Get position with helper function q = convertToPositions( Tf ); // Get translation back with helper function Tf_back = convertToTransform( q ); //std::cout << "i: "<< i << std::endl; std::cout << "\n ** Tf original: \n"<< Tf.matrix() << std::endl; std::cout << "\n ** Positions from Tf: \n"<< q.transpose() << std::endl; std::cout << "\n ** Tf from q: \n" << Tf_back.matrix() << std::endl; std::cout << "\n\n" << std::endl; /* } */ } //============================================================================== Eigen::Vector6d convertToPositions(const Eigen::Isometry3d& _tf) { Eigen::Vector6d x; x.head<3>() = dart::math::logMap(_tf.linear()); x.tail<3>() = _tf.translation(); return x; } //============================================================================== Eigen::Isometry3d convertToTransform( const Eigen::Vector6d& _positions) { Eigen::Isometry3d tf; tf.linear() = dart::math::expMapRot(_positions.head<3>()); tf.translation() = _positions.tail<3>(); return tf; } ```
1.0
Bug in FreeJoint helper - * File: dart/dynamics/FreeJoint.cpp * Functions compromised: convertToPositions (line 61) and converToTransform (line 70) * Problem description: I am generating a set of transformations with rotations around Z with angles in the range [0,2*PI], increasing at each step. I use these Transformations to set the joint positions of a freeJoint (using the same formula as used in convertToPositions). The joint positions I get are wrong for some angles. The joint positions correspond to the intended Transformation rotated 90 around Z. This happens usually after half of the range (either if I use [0,2PI] or [-PI,PI]. An example in which the bug happens is copied at the end of this issue. In this example I use: Input transform: ``` -0.815214 -0.57916 0 0.437796 -0.57916 0.815214 0 0.488175 0 0 -1 1.03 0 0 0 1 ``` Positions obtained using convertToPositions(): ``` -0.954925 -2.99295 -0 0.437796 0.488175 1.03 ``` Output transform obtained using convertToTransform() (should be the same as input transform): ``` -0.815214 0.57916 -5.39747e-16 0.437796 0.57916 0.815214 1.72211e-16 0.488175 5.39747e-16 -1.72211e-16 -1 1.03 6.95317e-310 1.38338e-322 6.92085e-310 2.12999e-317 ``` Notice the flipped signs in x an y columns. The problem described above is related to convertToPositions. I also noticed that in convertToTransform, there is a line missing: Tf.setIdentity(). Otherwise, the last row of Tf will be filled with garbage (as in the example above). * Final note: Please, if somebody can fix this quickly I would appreciate it. I spent most of today searching for the bug in my own code until I realized that the bug was (very possibly and unless I am mistaken) in dart so I would like to continue with my work once the bug is fixed. Thanks! PS.- Bug test file: ```cpp #include <dart/math/Geometry.h> #include <iostream> Eigen::Vector6d convertToPositions(const Eigen::Isometry3d& _tf); Eigen::Isometry3d convertToTransform(const Eigen::Vector6d& _positions); int main( int argc, char* argv[] ) { Eigen::Isometry3d Tf; Eigen::VectorXd q; Eigen::Isometry3d Tf_back; Tf.setIdentity(); Tf.translation() << 0.1, 0.2, 0.3; double dang = 10*M_PI/180.0; int n = (int) (2*M_PI /dang); Tf.matrix() << -0.815214, -0.57916, -0, 0.437796, -0.57916, 0.815214, 0, 0.488175, 0, 0, -1, 1.03, 0, 0, 0, 1; double det = Tf.linear().determinant(); std::cout << "Determinant of test TF: "<< det << std::endl; /* for( size_t i = 0; i < n; ++i ) { Eigen::Matrix3d rot; rot = Eigen::AngleAxisd( i*dang, Eigen::Vector3d(0,0,1) ); Tf.linear() = rot; */ // Get position with helper function q = convertToPositions( Tf ); // Get translation back with helper function Tf_back = convertToTransform( q ); //std::cout << "i: "<< i << std::endl; std::cout << "\n ** Tf original: \n"<< Tf.matrix() << std::endl; std::cout << "\n ** Positions from Tf: \n"<< q.transpose() << std::endl; std::cout << "\n ** Tf from q: \n" << Tf_back.matrix() << std::endl; std::cout << "\n\n" << std::endl; /* } */ } //============================================================================== Eigen::Vector6d convertToPositions(const Eigen::Isometry3d& _tf) { Eigen::Vector6d x; x.head<3>() = dart::math::logMap(_tf.linear()); x.tail<3>() = _tf.translation(); return x; } //============================================================================== Eigen::Isometry3d convertToTransform( const Eigen::Vector6d& _positions) { Eigen::Isometry3d tf; tf.linear() = dart::math::expMapRot(_positions.head<3>()); tf.translation() = _positions.tail<3>(); return tf; } ```
non_process
bug in freejoint helper file dart dynamics freejoint cpp functions compromised converttopositions line and convertotransform line problem description i am generating a set of transformations with rotations around z with angles in the range increasing at each step i use these transformations to set the joint positions of a freejoint using the same formula as used in converttopositions the joint positions i get are wrong for some angles the joint positions correspond to the intended transformation rotated around z this happens usually after half of the range either if i use or an example in which the bug happens is copied at the end of this issue in this example i use input transform positions obtained using converttopositions output transform obtained using converttotransform should be the same as input transform notice the flipped signs in x an y columns the problem described above is related to converttopositions i also noticed that in converttotransform there is a line missing tf setidentity otherwise the last row of tf will be filled with garbage as in the example above final note please if somebody can fix this quickly i would appreciate it i spent most of today searching for the bug in my own code until i realized that the bug was very possibly and unless i am mistaken in dart so i would like to continue with my work once the bug is fixed thanks ps bug test file cpp include include eigen converttopositions const eigen tf eigen converttotransform const eigen positions int main int argc char argv eigen tf eigen vectorxd q eigen tf back tf setidentity tf translation double dang m pi int n int m pi dang tf matrix double det tf linear determinant std cout determinant of test tf det std endl for size t i i n i eigen rot rot eigen angleaxisd i dang eigen tf linear rot get position with helper function q converttopositions tf get translation back with helper function tf back converttotransform q std cout i i std endl std cout n tf original n tf matrix std endl std cout n positions from tf n q transpose std endl std cout n tf from q n tf back matrix std endl std cout n n std endl eigen converttopositions const eigen tf eigen x x head dart math logmap tf linear x tail tf translation return x eigen converttotransform const eigen positions eigen tf tf linear dart math expmaprot positions head tf translation positions tail return tf
0
16,500
21,481,344,694
IssuesEvent
2022-04-26 18:03:37
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
Iterate on intake form
intake-process-pct
# User story As a Platform team member, I need a way of submitting a request to the Content team, so that I can request Platform Content team support for Platform website pages. # Tasks - [ ] Revise intake form in YAML - [ ] Confirm that form submissions are appearing on documentation triage board # Acceptance criteria - Ticket is modified and available thru the VA.gov-team repo # Reference info - [Documentation triage board](https://github.com/department-of-veterans-affairs/va.gov-team/projects/16) - [Process steps for ad hoc documentaiton support requests (draft)](https://vfs.atlassian.net/wiki/spaces/~672346535/pages/2141487170/Process+steps+for+ad-hoc+documentation+support+requests+draft) - [Jill's question in #github_information](https://dsva.slack.com/archives/CU1E4CX9U/p1650908057955009)
1.0
Iterate on intake form - # User story As a Platform team member, I need a way of submitting a request to the Content team, so that I can request Platform Content team support for Platform website pages. # Tasks - [ ] Revise intake form in YAML - [ ] Confirm that form submissions are appearing on documentation triage board # Acceptance criteria - Ticket is modified and available thru the VA.gov-team repo # Reference info - [Documentation triage board](https://github.com/department-of-veterans-affairs/va.gov-team/projects/16) - [Process steps for ad hoc documentaiton support requests (draft)](https://vfs.atlassian.net/wiki/spaces/~672346535/pages/2141487170/Process+steps+for+ad-hoc+documentation+support+requests+draft) - [Jill's question in #github_information](https://dsva.slack.com/archives/CU1E4CX9U/p1650908057955009)
process
iterate on intake form user story as a platform team member i need a way of submitting a request to the content team so that i can request platform content team support for platform website pages tasks revise intake form in yaml confirm that form submissions are appearing on documentation triage board acceptance criteria ticket is modified and available thru the va gov team repo reference info
1
362,443
25,376,305,556
IssuesEvent
2022-11-21 14:21:26
redhat-plumbers-in-action/advanced-issue-labeler
https://api.github.com/repos/redhat-plumbers-in-action/advanced-issue-labeler
opened
Make it more visible in README which options don't work together :memo:
type: documentation
### Type of issue other ### Description _No response_ ### Describe the solution you'd like _No response_
1.0
Make it more visible in README which options don't work together :memo: - ### Type of issue other ### Description _No response_ ### Describe the solution you'd like _No response_
non_process
make it more visible in readme which options don t work together memo type of issue other description no response describe the solution you d like no response
0
68,199
17,191,958,982
IssuesEvent
2021-07-16 12:19:54
angular/angular-cli
https://api.github.com/repos/angular/angular-cli
closed
ng serve fails after some modification to the source code - module.buildInfo.jsonData TypeError: Cannot read property 'jsonData' of undefined
comp: devkit/build-angular freq1: low severity3: broken type: bug/fix
<!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 Oh hi there! 😄 To expedite issue processing please search open and closed issues before submitting a new one. Existing issues often contain information about workarounds, resolution, or progress updates. 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅--> # 🐞 Bug report ### Command (mark with an `x`) <!-- Can you pin-point the command or commands that are effected by this bug? --> <!-- ✍️edit: --> - [ ] new - [ ] build - [x] serve - [ ] test - [ ] e2e - [ ] generate - [ ] add - [ ] update - [ ] lint - [ ] extract-i18n - [ ] run - [ ] config - [ ] help - [ ] version - [ ] doc ### Is this a regression? <!-- Did this behavior use to work in the previous version? --> <!-- ✍️--> Yes, the previous version in which this bug was not present was: 12.1.1 ### Description <!-- ✍️--> Run ng serve like allways <pre> <code> Initial Chunk Files | Names | Size vendor.js | vendor | 17.10 MB main.js | main | 7.20 MB scripts.js | scripts | 1.01 MB styles.css | styles | 846.56 kB polyfills.js | polyfills | 406.35 kB runtime.js | runtime | 13.86 kB | Initial Total | 26.54 MB Lazy Chunk Files | Names | Size src_app_ceramicos-la-plata_administracion_administracion_module_ts.js | - | 4.83 MB default-src_app_ceramicos-la-plata_venta_venta_module_ts.js | - | 2.79 MB src_app_madeco_venta_venta_module_ts.js | - | 1.87 MB src_app_madeco_produccion_produccion_module_ts.js | - | 1.62 MB default-src_app_madeco_common-components_common-components_module_ts.js | - | 1.39 MB src_app_madeco_proveedores_proveedores_module_ts.js | - | 1.17 MB src_app_ventas_ventas_module_ts.js | - | 1.16 MB src_app_madeco_banco_bancos_module_ts.js | - | 917.62 kB src_app_ceramicos-la-plata_logistica_logistica_module_ts.js | - | 893.92 kB src_app_contabilidad_contabilidad_module_ts.js | - | 781.78 kB default-src_app_madeco_custom-directives_custom-directives_module_ts-src_app_main_service_con-9c34b3.js | - | 557.74 kB src_app_ceramicos-la-plata_banco_banco_module_ts.js | - | 482.17 kB src_app_ceramicos-la-plata_logistica-handhelds_logistica-handhelds_module_ts.js | - | 386.45 kB src_app_ceramicos-la-plata_planeamiento_planeamiento_module_ts.js | - | 385.68 kB src_app_ceramicos-la-plata_grupo_grupo_module_ts.js | - | 302.96 kB src_app_ceramicos-la-plata_facturacion_facturacion_module_ts.js | - | 277.80 kB src_app_shared-components_lector-qr-web-worker_qr-decoder_worker_ts.js | - | 253.25 kB src_app_ceramicos-la-plata_etiquetas_etiquetas_module_ts.js | - | 192.78 kB default-src_app_ceramicos-la-plata_venta_caja-ventas_caja-ventas_component_ts-src_app_ceramic-ecd1a6.js | - | 172.79 kB node_modules_firebase_messaging_dist_index_esm_js.js | - | 147.01 kB src_app_administracion_administracion_module_ts.js | - | 107.53 kB default-src_app_madeco_logistica_logistica_module_ts.js | - | 103.91 kB src_app_ceramicos-la-plata_tablero_tablero_module_ts.js | - | 55.45 kB default-src_app_ceramicos-la-plata_administracion_ingreso-egreso-motivo_ingreso-egreso-motivo-f547ef.js | - | 33.06 kB default-src_app_ceramicos-la-plata_logistica_local-storage_local-storage-clp-logistica-strategy_ts.js | - | 23.66 kB src_app_configuracion_configuracion_module_ts.js | - | 19.57 kB common.js | common | 9.92 kB </code> </pre> ## 🔬 Minimal Reproduction Modify some code then the error shows up after the rebuild of a lazy chunk ## 🔥 Exception or Error <pre> <code> Generating browser application bundles.../home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/json/JsonGenerator.js:119 let data = module.buildInfo.jsonData; ^ TypeError: Cannot read property 'jsonData' of undefined at JsonGenerator.getSize (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/json/JsonGenerator.js:119:31) at NormalModule.size (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModule.js:1224:43) at NormalModule.cleanupForCache (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModule.js:364:9) at NormalModuleFactory.cleanupForCache (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModuleFactory.js:670:11) at Compiler._cleanupLastNormalModuleFactory (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:383:34) at Compiler.createNormalModuleFactory (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1049:8) at Compiler.newCompilationParams (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1071:30) at Compiler.compile (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1082:23) at /home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Watching.js:188:19 at Hook.eval [as callAsync] (eval at create (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:22:1) </code></pre> ## 🌍 Your Environment <pre><code> <!-- run `ng version` and paste output below --> <!-- ✍️--> _ _ ____ _ ___ / \ _ __ __ _ _ _| | __ _ _ __ / ___| | |_ _| / △ \ | '_ \ / _` | | | | |/ _` | '__| | | | | | | / ___ \| | | | (_| | |_| | | (_| | | | |___| |___ | | /_/ \_\_| |_|\__, |\__,_|_|\__,_|_| \____|_____|___| |___/ Angular CLI: 12.1.2 Node: 14.17.0 Package Manager: npm 7.19.1 OS: linux x64 Angular: 12.1.2 ... animations, cli, common, compiler, compiler-cli, core, forms ... language-service, platform-browser, platform-browser-dynamic ... router, service-worker Package Version --------------------------------------------------------- @angular-devkit/architect 0.1201.2 @angular-devkit/build-angular 12.1.2 @angular-devkit/build-webpack 0.1201.2 @angular-devkit/core 12.1.2 @angular-devkit/schematics 12.1.2 @angular/fire 6.1.5 @schematics/angular 12.1.2 ng-packagr 12.1.2 rxjs 6.6.7 typescript 4.3.5 webpack 5.44.0 </code></pre> **ng build works fine**
1.0
ng serve fails after some modification to the source code - module.buildInfo.jsonData TypeError: Cannot read property 'jsonData' of undefined - <!--🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 Oh hi there! 😄 To expedite issue processing please search open and closed issues before submitting a new one. Existing issues often contain information about workarounds, resolution, or progress updates. 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅--> # 🐞 Bug report ### Command (mark with an `x`) <!-- Can you pin-point the command or commands that are effected by this bug? --> <!-- ✍️edit: --> - [ ] new - [ ] build - [x] serve - [ ] test - [ ] e2e - [ ] generate - [ ] add - [ ] update - [ ] lint - [ ] extract-i18n - [ ] run - [ ] config - [ ] help - [ ] version - [ ] doc ### Is this a regression? <!-- Did this behavior use to work in the previous version? --> <!-- ✍️--> Yes, the previous version in which this bug was not present was: 12.1.1 ### Description <!-- ✍️--> Run ng serve like allways <pre> <code> Initial Chunk Files | Names | Size vendor.js | vendor | 17.10 MB main.js | main | 7.20 MB scripts.js | scripts | 1.01 MB styles.css | styles | 846.56 kB polyfills.js | polyfills | 406.35 kB runtime.js | runtime | 13.86 kB | Initial Total | 26.54 MB Lazy Chunk Files | Names | Size src_app_ceramicos-la-plata_administracion_administracion_module_ts.js | - | 4.83 MB default-src_app_ceramicos-la-plata_venta_venta_module_ts.js | - | 2.79 MB src_app_madeco_venta_venta_module_ts.js | - | 1.87 MB src_app_madeco_produccion_produccion_module_ts.js | - | 1.62 MB default-src_app_madeco_common-components_common-components_module_ts.js | - | 1.39 MB src_app_madeco_proveedores_proveedores_module_ts.js | - | 1.17 MB src_app_ventas_ventas_module_ts.js | - | 1.16 MB src_app_madeco_banco_bancos_module_ts.js | - | 917.62 kB src_app_ceramicos-la-plata_logistica_logistica_module_ts.js | - | 893.92 kB src_app_contabilidad_contabilidad_module_ts.js | - | 781.78 kB default-src_app_madeco_custom-directives_custom-directives_module_ts-src_app_main_service_con-9c34b3.js | - | 557.74 kB src_app_ceramicos-la-plata_banco_banco_module_ts.js | - | 482.17 kB src_app_ceramicos-la-plata_logistica-handhelds_logistica-handhelds_module_ts.js | - | 386.45 kB src_app_ceramicos-la-plata_planeamiento_planeamiento_module_ts.js | - | 385.68 kB src_app_ceramicos-la-plata_grupo_grupo_module_ts.js | - | 302.96 kB src_app_ceramicos-la-plata_facturacion_facturacion_module_ts.js | - | 277.80 kB src_app_shared-components_lector-qr-web-worker_qr-decoder_worker_ts.js | - | 253.25 kB src_app_ceramicos-la-plata_etiquetas_etiquetas_module_ts.js | - | 192.78 kB default-src_app_ceramicos-la-plata_venta_caja-ventas_caja-ventas_component_ts-src_app_ceramic-ecd1a6.js | - | 172.79 kB node_modules_firebase_messaging_dist_index_esm_js.js | - | 147.01 kB src_app_administracion_administracion_module_ts.js | - | 107.53 kB default-src_app_madeco_logistica_logistica_module_ts.js | - | 103.91 kB src_app_ceramicos-la-plata_tablero_tablero_module_ts.js | - | 55.45 kB default-src_app_ceramicos-la-plata_administracion_ingreso-egreso-motivo_ingreso-egreso-motivo-f547ef.js | - | 33.06 kB default-src_app_ceramicos-la-plata_logistica_local-storage_local-storage-clp-logistica-strategy_ts.js | - | 23.66 kB src_app_configuracion_configuracion_module_ts.js | - | 19.57 kB common.js | common | 9.92 kB </code> </pre> ## 🔬 Minimal Reproduction Modify some code then the error shows up after the rebuild of a lazy chunk ## 🔥 Exception or Error <pre> <code> Generating browser application bundles.../home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/json/JsonGenerator.js:119 let data = module.buildInfo.jsonData; ^ TypeError: Cannot read property 'jsonData' of undefined at JsonGenerator.getSize (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/json/JsonGenerator.js:119:31) at NormalModule.size (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModule.js:1224:43) at NormalModule.cleanupForCache (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModule.js:364:9) at NormalModuleFactory.cleanupForCache (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/NormalModuleFactory.js:670:11) at Compiler._cleanupLastNormalModuleFactory (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:383:34) at Compiler.createNormalModuleFactory (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1049:8) at Compiler.newCompilationParams (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1071:30) at Compiler.compile (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Compiler.js:1082:23) at /home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/@angular-devkit/build-angular/node_modules/webpack/lib/Watching.js:188:19 at Hook.eval [as callAsync] (eval at create (/home/naxs/Trabajo/NaxsCloud/NaxsCloudWeb/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:22:1) </code></pre> ## 🌍 Your Environment <pre><code> <!-- run `ng version` and paste output below --> <!-- ✍️--> _ _ ____ _ ___ / \ _ __ __ _ _ _| | __ _ _ __ / ___| | |_ _| / △ \ | '_ \ / _` | | | | |/ _` | '__| | | | | | | / ___ \| | | | (_| | |_| | | (_| | | | |___| |___ | | /_/ \_\_| |_|\__, |\__,_|_|\__,_|_| \____|_____|___| |___/ Angular CLI: 12.1.2 Node: 14.17.0 Package Manager: npm 7.19.1 OS: linux x64 Angular: 12.1.2 ... animations, cli, common, compiler, compiler-cli, core, forms ... language-service, platform-browser, platform-browser-dynamic ... router, service-worker Package Version --------------------------------------------------------- @angular-devkit/architect 0.1201.2 @angular-devkit/build-angular 12.1.2 @angular-devkit/build-webpack 0.1201.2 @angular-devkit/core 12.1.2 @angular-devkit/schematics 12.1.2 @angular/fire 6.1.5 @schematics/angular 12.1.2 ng-packagr 12.1.2 rxjs 6.6.7 typescript 4.3.5 webpack 5.44.0 </code></pre> **ng build works fine**
non_process
ng serve fails after some modification to the source code module buildinfo jsondata typeerror cannot read property jsondata of undefined 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 oh hi there 😄 to expedite issue processing please search open and closed issues before submitting a new one existing issues often contain information about workarounds resolution or progress updates 🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅🔅 🐞 bug report command mark with an x new build serve test generate add update lint extract run config help version doc is this a regression yes the previous version in which this bug was not present was description run ng serve like allways initial chunk files names size vendor js vendor mb main js main mb scripts js scripts mb styles css styles kb polyfills js polyfills kb runtime js runtime kb initial total mb lazy chunk files names size src app ceramicos la plata administracion administracion module ts js mb default src app ceramicos la plata venta venta module ts js mb src app madeco venta venta module ts js mb src app madeco produccion produccion module ts js mb default src app madeco common components common components module ts js mb src app madeco proveedores proveedores module ts js mb src app ventas ventas module ts js mb src app madeco banco bancos module ts js kb src app ceramicos la plata logistica logistica module ts js kb src app contabilidad contabilidad module ts js kb default src app madeco custom directives custom directives module ts src app main service con js kb src app ceramicos la plata banco banco module ts js kb src app ceramicos la plata logistica handhelds logistica handhelds module ts js kb src app ceramicos la plata planeamiento planeamiento module ts js kb src app ceramicos la plata grupo grupo module ts js kb src app ceramicos la plata facturacion facturacion module ts js kb src app shared components lector qr web worker qr decoder worker ts js kb src app ceramicos la plata etiquetas etiquetas module ts js kb default src app ceramicos la plata venta caja ventas caja ventas component ts src app ceramic js kb node modules firebase messaging dist index esm js js kb src app administracion administracion module ts js kb default src app madeco logistica logistica module ts js kb src app ceramicos la plata tablero tablero module ts js kb default src app ceramicos la plata administracion ingreso egreso motivo ingreso egreso motivo js kb default src app ceramicos la plata logistica local storage local storage clp logistica strategy ts js kb src app configuracion configuracion module ts js kb common js common kb 🔬 minimal reproduction modify some code then the error shows up after the rebuild of a lazy chunk 🔥 exception or error generating browser application bundles home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib json jsongenerator js let data module buildinfo jsondata typeerror cannot read property jsondata of undefined at jsongenerator getsize home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib json jsongenerator js at normalmodule size home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib normalmodule js at normalmodule cleanupforcache home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib normalmodule js at normalmodulefactory cleanupforcache home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib normalmodulefactory js at compiler cleanuplastnormalmodulefactory home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib compiler js at compiler createnormalmodulefactory home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib compiler js at compiler newcompilationparams home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib compiler js at compiler compile home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib compiler js at home naxs trabajo naxscloud naxscloudweb node modules angular devkit build angular node modules webpack lib watching js at hook eval eval at create home naxs trabajo naxscloud naxscloudweb node modules tapable lib hookcodefactory js 🌍 your environment △ angular cli node package manager npm os linux angular animations cli common compiler compiler cli core forms language service platform browser platform browser dynamic router service worker package version angular devkit architect angular devkit build angular angular devkit build webpack angular devkit core angular devkit schematics angular fire schematics angular ng packagr rxjs typescript webpack ng build works fine
0
219,553
7,343,624,460
IssuesEvent
2018-03-07 12:02:53
trimstray/pvimport
https://api.github.com/repos/trimstray/pvimport
opened
Limiting disk usage during import.
Priority: Critical Status: In Progress Type: Maintenance
Next Release: **[testing](https://github.com/trimstray/pvimport/tree/testing)** Status: **In Progress** Limiting disk usage during import: - use **CFQ scheduler** - use **ionice** to change io scheduling class and priority
1.0
Limiting disk usage during import. - Next Release: **[testing](https://github.com/trimstray/pvimport/tree/testing)** Status: **In Progress** Limiting disk usage during import: - use **CFQ scheduler** - use **ionice** to change io scheduling class and priority
non_process
limiting disk usage during import next release status in progress limiting disk usage during import use cfq scheduler use ionice to change io scheduling class and priority
0
431,055
12,474,556,762
IssuesEvent
2020-05-29 09:52:05
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Don't show removed titles in lists
Priority: Medium Status: Fixed
![owner](https://user-images.githubusercontent.com/25908592/82090255-88882880-96f5-11ea-9ed5-7da10b05afb8.png) 1) It clutters up the list without any use. 2) In case of transfering a deed by mistake to one of the removed titles, an admin needs to reclaim it, as it's stuck otherwise. I'd have them removed from all such lists, actually.
1.0
Don't show removed titles in lists - ![owner](https://user-images.githubusercontent.com/25908592/82090255-88882880-96f5-11ea-9ed5-7da10b05afb8.png) 1) It clutters up the list without any use. 2) In case of transfering a deed by mistake to one of the removed titles, an admin needs to reclaim it, as it's stuck otherwise. I'd have them removed from all such lists, actually.
non_process
don t show removed titles in lists it clutters up the list without any use in case of transfering a deed by mistake to one of the removed titles an admin needs to reclaim it as it s stuck otherwise i d have them removed from all such lists actually
0
14,187
17,070,591,552
IssuesEvent
2021-07-07 12:54:05
opensrp/web
https://api.github.com/repos/opensrp/web
opened
[FHIR Road Map] Descriptive error messages for all forms
FHIR compatibility
## Issue Description add descriptive error messages to forms. Replace generic 'error occurred' messages in all forms with descriptive messages. ## Resources ### FHIR Resources to reference ## Please share any other relevant information about the change request --- Remember: 1. To alert the team in the [#opensrp-web](https://onaio.slack.com/archives/C01R82Y9N58) channel. 2. To follow up on delivery. You're in-charge.
True
[FHIR Road Map] Descriptive error messages for all forms - ## Issue Description add descriptive error messages to forms. Replace generic 'error occurred' messages in all forms with descriptive messages. ## Resources ### FHIR Resources to reference ## Please share any other relevant information about the change request --- Remember: 1. To alert the team in the [#opensrp-web](https://onaio.slack.com/archives/C01R82Y9N58) channel. 2. To follow up on delivery. You're in-charge.
non_process
descriptive error messages for all forms issue description add descriptive error messages to forms replace generic error occurred messages in all forms with descriptive messages resources fhir resources to reference please share any other relevant information about the change request remember to alert the team in the channel to follow up on delivery you re in charge
0
8,791
11,908,170,982
IssuesEvent
2020-03-31 00:10:51
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Change location of temporary executables for geoprocessing tools
Feature Request Processing
**Feature description.** <!-- A clear and concise description of what you want to happen. Ex. QGIS would rock even more if [...] --> I am not sure if I missed something somewhere, but it would be amazing to add the option to decide where the intermediate geoprocessing executables are stored and executed. We would like to use QGIS (with SAGA + GRASS + GDAL) at a university, and the fact that some intermediate executables are created inside the user's temporary folder means that we can't whitelist them. We end up having this type of error message: > "This program is blocked by group policy. For more information, contact your system administrator." In a GRASS example, the offending executable would be: > C:\Users\<username>\AppData\Local\Temp\processing_670dd4b4fc494c41a3761314d2eb317a\grassdata\grass_batch_job.cmd In the case of SAGA, it seems it tries to run a .bat file. We can't whitelist every user's entire `AppData\Local\Temp` folder, and the unpredictable naming of the folders does not allow to whitelist a more precise path. Could there be an option to define where all those intermediate executables will be going and executed from? For example, an extra option in Options > Processing > General **Additional context** Tested with QGIS 3.6.3 and 3.8.3 on Windows 10.
1.0
Change location of temporary executables for geoprocessing tools - **Feature description.** <!-- A clear and concise description of what you want to happen. Ex. QGIS would rock even more if [...] --> I am not sure if I missed something somewhere, but it would be amazing to add the option to decide where the intermediate geoprocessing executables are stored and executed. We would like to use QGIS (with SAGA + GRASS + GDAL) at a university, and the fact that some intermediate executables are created inside the user's temporary folder means that we can't whitelist them. We end up having this type of error message: > "This program is blocked by group policy. For more information, contact your system administrator." In a GRASS example, the offending executable would be: > C:\Users\<username>\AppData\Local\Temp\processing_670dd4b4fc494c41a3761314d2eb317a\grassdata\grass_batch_job.cmd In the case of SAGA, it seems it tries to run a .bat file. We can't whitelist every user's entire `AppData\Local\Temp` folder, and the unpredictable naming of the folders does not allow to whitelist a more precise path. Could there be an option to define where all those intermediate executables will be going and executed from? For example, an extra option in Options > Processing > General **Additional context** Tested with QGIS 3.6.3 and 3.8.3 on Windows 10.
process
change location of temporary executables for geoprocessing tools feature description i am not sure if i missed something somewhere but it would be amazing to add the option to decide where the intermediate geoprocessing executables are stored and executed we would like to use qgis with saga grass gdal at a university and the fact that some intermediate executables are created inside the user s temporary folder means that we can t whitelist them we end up having this type of error message this program is blocked by group policy for more information contact your system administrator in a grass example the offending executable would be c users appdata local temp processing grassdata grass batch job cmd in the case of saga it seems it tries to run a bat file we can t whitelist every user s entire appdata local temp folder and the unpredictable naming of the folders does not allow to whitelist a more precise path could there be an option to define where all those intermediate executables will be going and executed from for example an extra option in options processing general additional context tested with qgis and on windows
1
12,468
14,703,397,671
IssuesEvent
2021-01-04 14:59:29
fourMs/MGT-python
https://api.github.com/repos/fourMs/MGT-python
closed
Tkinter on Ubuntu
compatibility
Just discovered that installing the Tkinter package is slightly trickier on Ubuntu. In `musicalgestures`, in the cropping pre-processing step, when using manual cropping, I used tkinter to get the screensize and scale the image (_utils.py: `get_screen_resolution_scaled()`: *Gets the scaled screen resolution. Respects display scaling on high DPI displays.*), so it does not overflow the screen. On Ubuntu [there seems to be a problem](https://stackoverflow.com/questions/6084416/tkinter-module-not-found-on-ubuntu) with Tkinter by default. Using `sudo apt-get install python3-tk` solved it for me. But this means the installation process should include a similar step on Linux.
True
Tkinter on Ubuntu - Just discovered that installing the Tkinter package is slightly trickier on Ubuntu. In `musicalgestures`, in the cropping pre-processing step, when using manual cropping, I used tkinter to get the screensize and scale the image (_utils.py: `get_screen_resolution_scaled()`: *Gets the scaled screen resolution. Respects display scaling on high DPI displays.*), so it does not overflow the screen. On Ubuntu [there seems to be a problem](https://stackoverflow.com/questions/6084416/tkinter-module-not-found-on-ubuntu) with Tkinter by default. Using `sudo apt-get install python3-tk` solved it for me. But this means the installation process should include a similar step on Linux.
non_process
tkinter on ubuntu just discovered that installing the tkinter package is slightly trickier on ubuntu in musicalgestures in the cropping pre processing step when using manual cropping i used tkinter to get the screensize and scale the image utils py get screen resolution scaled gets the scaled screen resolution respects display scaling on high dpi displays so it does not overflow the screen on ubuntu with tkinter by default using sudo apt get install tk solved it for me but this means the installation process should include a similar step on linux
0
15,693
19,848,101,388
IssuesEvent
2022-01-21 09:12:44
ooi-data/CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_statistics_recovered
https://api.github.com/repos/ooi-data/CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_statistics_recovered
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:12:43.462049. ## Details Flow name: `CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_statistics_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T09:12:43.462049. ## Details Flow name: `CE04OSSM-SBD12-05-WAVSSA000-recovered_host-wavss_a_dcl_statistics_recovered` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 70, in __array__ return self.func(self.array) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/coding/variables.py", line 137, in _apply_mask data = np.asarray(data, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host wavss a dcl statistics recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray coding variables py line in array return self func self array file srv conda envs notebook lib site packages xarray coding variables py line in apply mask data np asarray data dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
725,419
24,961,611,792
IssuesEvent
2022-11-01 15:57:18
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
Can't add registration to MHR table
bug Priority1 Assets
Issue: 1. Logged in as staff, I want to add an existing registration (Active, Historical, Exempt) to my table. 2. When I search for an existing registration, example MHR "082086", I see a modal saying the registration exists 3. When I click "Add Registration to my table", I am served an error modal and am not able to add it to my table. ![Screen Shot 2022-10-03 at 3.27.46 PM.png](https://images.zenhubusercontent.com/5e1cd2cd5d67f0528e482520/4ce35b65-320b-415f-80bc-459a8f9c84d4)![Screen Shot 2022-10-03 at 3.27.51 PM.png](https://images.zenhubusercontent.com/5e1cd2cd5d67f0528e482520/676b2f22-da94-48b3-bf90-7130c6a5e44a)
1.0
Can't add registration to MHR table - Issue: 1. Logged in as staff, I want to add an existing registration (Active, Historical, Exempt) to my table. 2. When I search for an existing registration, example MHR "082086", I see a modal saying the registration exists 3. When I click "Add Registration to my table", I am served an error modal and am not able to add it to my table. ![Screen Shot 2022-10-03 at 3.27.46 PM.png](https://images.zenhubusercontent.com/5e1cd2cd5d67f0528e482520/4ce35b65-320b-415f-80bc-459a8f9c84d4)![Screen Shot 2022-10-03 at 3.27.51 PM.png](https://images.zenhubusercontent.com/5e1cd2cd5d67f0528e482520/676b2f22-da94-48b3-bf90-7130c6a5e44a)
non_process
can t add registration to mhr table issue logged in as staff i want to add an existing registration active historical exempt to my table when i search for an existing registration example mhr i see a modal saying the registration exists when i click add registration to my table i am served an error modal and am not able to add it to my table
0
80,957
15,609,560,455
IssuesEvent
2021-03-19 12:04:24
fac21/week2--Http-Project--AJMS
https://api.github.com/repos/fac21/week2--Http-Project--AJMS
closed
Tidy up commented out code
code review
There is a lot of commented out code, if you don't need it delete it 🧚‍♀️🧚‍♀️🧚‍♀️ it will make reading it easier for other people!
1.0
Tidy up commented out code - There is a lot of commented out code, if you don't need it delete it 🧚‍♀️🧚‍♀️🧚‍♀️ it will make reading it easier for other people!
non_process
tidy up commented out code there is a lot of commented out code if you don t need it delete it 🧚‍♀️🧚‍♀️🧚‍♀️ it will make reading it easier for other people
0
1,140
3,626,972,173
IssuesEvent
2016-02-10 05:13:16
worldspawn/mascis
https://api.github.com/repos/worldspawn/mascis
opened
Support Delete...Where
enhancement linq-expression-parser postgres-language-processor query-api t-sql-language-processor
- [ ] Linq Expression Parser - [ ] T-Sql Language Processor - [ ] Postgres Language Processor - [ ] Query Api
2.0
Support Delete...Where - - [ ] Linq Expression Parser - [ ] T-Sql Language Processor - [ ] Postgres Language Processor - [ ] Query Api
process
support delete where linq expression parser t sql language processor postgres language processor query api
1
11,823
7,719,337,037
IssuesEvent
2018-05-23 19:04:36
RoaringBitmap/CRoaring
https://api.github.com/repos/RoaringBitmap/CRoaring
opened
Add "roaring_bitmap_remove_many" function
help wanted performance
We would like to have a ``roaring_bitmap_remove_many`` function which allows you to remove from a bitmap all of the elements from an array. This can be done currently with repeated calls to ``roaring_bitmap_remove``. But it would be more convenient and efficient to have a dedicated function.
True
Add "roaring_bitmap_remove_many" function - We would like to have a ``roaring_bitmap_remove_many`` function which allows you to remove from a bitmap all of the elements from an array. This can be done currently with repeated calls to ``roaring_bitmap_remove``. But it would be more convenient and efficient to have a dedicated function.
non_process
add roaring bitmap remove many function we would like to have a roaring bitmap remove many function which allows you to remove from a bitmap all of the elements from an array this can be done currently with repeated calls to roaring bitmap remove but it would be more convenient and efficient to have a dedicated function
0
17,396
23,211,708,876
IssuesEvent
2022-08-02 10:43:18
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
opened
arduino/setup-protoc@v1 is not allowed to be used in apache/arrow-rs
bug development-process
**Describe the bug** <!-- A clear and concise description of what the bug is. --> https://github.com/apache/arrow-rs/actions/runs/2781430582 **To Reproduce** <!-- Steps to reproduce the behavior: --> **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> **Additional context** <!-- Add any other context about the problem here. -->
1.0
arduino/setup-protoc@v1 is not allowed to be used in apache/arrow-rs - **Describe the bug** <!-- A clear and concise description of what the bug is. --> https://github.com/apache/arrow-rs/actions/runs/2781430582 **To Reproduce** <!-- Steps to reproduce the behavior: --> **Expected behavior** <!-- A clear and concise description of what you expected to happen. --> **Additional context** <!-- Add any other context about the problem here. -->
process
arduino setup protoc is not allowed to be used in apache arrow rs describe the bug a clear and concise description of what the bug is to reproduce steps to reproduce the behavior expected behavior a clear and concise description of what you expected to happen additional context add any other context about the problem here
1
9,223
12,257,986,462
IssuesEvent
2020-05-06 14:29:01
MHRA/products
https://api.github.com/repos/MHRA/products
closed
Upload new PAR
EPIC - PARs process HIGH PRIORITY :arrow_double_up: SPIKE :hocho:
## User Want As a Medical Writer in the licensing team I would like to upload a new PAR so that it is available on products.mhra.gov.uk ## Acceptance Criteria ### Technical acceptance criteria - [ ] PAR pdf is in blob storage after upload - [ ] PAR metadata is attached - [ ] PAR is in the search index - [ ] blob/metadata/search index is handled by doc-index-updater ### Customer acceptance criteria - [ ] Medical writers can upload a file which will then be surfaced on products.mhra.gov.uk - [ ] The PAR can be linked to one or multiple products (which will have their own PL number) - [ ] PAR has product name - [ ] PAR has active substances, something about TMS list (unsure what this is, #no-acronyms) - [ ] PAR has one or more PLs / NR (unsure what this is) / THR (unsure what this is) ## Data - Potential impact **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
1.0
Upload new PAR - ## User Want As a Medical Writer in the licensing team I would like to upload a new PAR so that it is available on products.mhra.gov.uk ## Acceptance Criteria ### Technical acceptance criteria - [ ] PAR pdf is in blob storage after upload - [ ] PAR metadata is attached - [ ] PAR is in the search index - [ ] blob/metadata/search index is handled by doc-index-updater ### Customer acceptance criteria - [ ] Medical writers can upload a file which will then be surfaced on products.mhra.gov.uk - [ ] The PAR can be linked to one or multiple products (which will have their own PL number) - [ ] PAR has product name - [ ] PAR has active substances, something about TMS list (unsure what this is, #no-acronyms) - [ ] PAR has one or more PLs / NR (unsure what this is) / THR (unsure what this is) ## Data - Potential impact **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
process
upload new par user want as a medical writer in the licensing team i would like to upload a new par so that it is available on products mhra gov uk acceptance criteria technical acceptance criteria par pdf is in blob storage after upload par metadata is attached par is in the search index blob metadata search index is handled by doc index updater customer acceptance criteria medical writers can upload a file which will then be surfaced on products mhra gov uk the par can be linked to one or multiple products which will have their own pl number par has product name par has active substances something about tms list unsure what this is no acronyms par has one or more pls nr unsure what this is thr unsure what this is data potential impact size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
1
81,091
23,386,477,324
IssuesEvent
2022-08-11 14:09:01
reapit/foundations
https://api.github.com/repos/reapit/foundations
closed
App Builder making an entity editable launches the page creation wizard
feature front-end app-builder
**Background context or User story:** _Selecting editable from the customise panel on a table should launch a pre-populated page creation wizard. Should have entity type and page type table pre-selected_
1.0
App Builder making an entity editable launches the page creation wizard - **Background context or User story:** _Selecting editable from the customise panel on a table should launch a pre-populated page creation wizard. Should have entity type and page type table pre-selected_
non_process
app builder making an entity editable launches the page creation wizard background context or user story selecting editable from the customise panel on a table should launch a pre populated page creation wizard should have entity type and page type table pre selected
0
143,838
5,531,055,357
IssuesEvent
2017-03-21 05:46:13
knipferrc/plate
https://api.github.com/repos/knipferrc/plate
closed
Complete Plate Color Coordination
Priority: Medium Type: Feature
Attributes in place. Complete a random generation of colors for each unique plate.
1.0
Complete Plate Color Coordination - Attributes in place. Complete a random generation of colors for each unique plate.
non_process
complete plate color coordination attributes in place complete a random generation of colors for each unique plate
0
666
3,136,006,190
IssuesEvent
2015-09-10 17:49:45
PHPOffice/PHPWord
https://api.github.com/repos/PHPOffice/PHPWord
opened
Support arrays in TemplateProcessor.setValue method
Change Request Template Processor
To do: - arrays in `$subject` to handle headers, footers and main document part at once. - arrays in `$search` and in `$replacement` to avoid looping on client side. Requested by @OAFCROB (see https://github.com/PHPOffice/PHPWord/issues/513#issuecomment-136964062).
1.0
Support arrays in TemplateProcessor.setValue method - To do: - arrays in `$subject` to handle headers, footers and main document part at once. - arrays in `$search` and in `$replacement` to avoid looping on client side. Requested by @OAFCROB (see https://github.com/PHPOffice/PHPWord/issues/513#issuecomment-136964062).
process
support arrays in templateprocessor setvalue method to do arrays in subject to handle headers footers and main document part at once arrays in search and in replacement to avoid looping on client side requested by oafcrob see
1
11,203
13,957,703,535
IssuesEvent
2020-10-24 08:13:57
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
BE: Harvesting frequency
BE - Belgium Geoportal Harvesting process
Dear INSPIRE Geoportal Thematic viewer team, Dear Helpdesk; we have updated the harvesting frequencies of 2 catalogues over a week ago in the inspire registry (http://inspire-geoportal.ec.europa.eu/INSPIRERegistry/registers/services/allproposals ): * Geocatalogue de la r&eacute;gion Bruxelloise : from weekly to daily * Geopunt : from weekly to daily On the thematic geoportal the harvesting frequency has not yet been updated? Does this mean this has to be updated differently ? Thank you in advance Ouns Best regards, Ouns Kissiyar
1.0
BE: Harvesting frequency - Dear INSPIRE Geoportal Thematic viewer team, Dear Helpdesk; we have updated the harvesting frequencies of 2 catalogues over a week ago in the inspire registry (http://inspire-geoportal.ec.europa.eu/INSPIRERegistry/registers/services/allproposals ): * Geocatalogue de la r&eacute;gion Bruxelloise : from weekly to daily * Geopunt : from weekly to daily On the thematic geoportal the harvesting frequency has not yet been updated? Does this mean this has to be updated differently ? Thank you in advance Ouns Best regards, Ouns Kissiyar
process
be harvesting frequency dear inspire geoportal thematic viewer team dear helpdesk we have updated the harvesting frequencies of catalogues over a week ago in the inspire registry geocatalogue de la r eacute gion bruxelloise from weekly to daily geopunt from weekly to daily on the thematic geoportal the harvesting frequency has not yet been updated does this mean this has to be updated differently thank you in advance ouns best regards ouns kissiyar
1
15,595
19,721,759,324
IssuesEvent
2022-01-13 15:59:49
eunseo2/JAVA
https://api.github.com/repos/eunseo2/JAVA
opened
Iterating
process
> Task :Iterating.main() 1 2 3 total = 6 ```java public static List<User> userList(){ List<User> userList = new ArrayList<>(); userList.add(new User(1,"user1","user1@gmail.com")); userList.add(new User(2,"user2","user2@gmail.com")); userList.add(new User(3,"user3","user3@gmail.com")); return userList; } ``` user1의 Id 값 + user2의 Id 값 + user3의 Id 값 = 6
1.0
Iterating - > Task :Iterating.main() 1 2 3 total = 6 ```java public static List<User> userList(){ List<User> userList = new ArrayList<>(); userList.add(new User(1,"user1","user1@gmail.com")); userList.add(new User(2,"user2","user2@gmail.com")); userList.add(new User(3,"user3","user3@gmail.com")); return userList; } ``` user1의 Id 값 + user2의 Id 값 + user3의 Id 값 = 6
process
iterating task iterating main total java public static list userlist list userlist new arraylist userlist add new user gmail com userlist add new user gmail com userlist add new user gmail com return userlist id 값 id 값 id 값
1
17,229
22,917,655,674
IssuesEvent
2022-07-17 07:34:07
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[SQL connector] Users using value.format should receive validationException, and they are not
compute/data-processing
Community user reported that when using `value.format` they get a NoSuchElementException. This should not occur as when users specify `value.format`, they should get a ValidationException.
1.0
[SQL connector] Users using value.format should receive validationException, and they are not - Community user reported that when using `value.format` they get a NoSuchElementException. This should not occur as when users specify `value.format`, they should get a ValidationException.
process
users using value format should receive validationexception and they are not community user reported that when using value format they get a nosuchelementexception this should not occur as when users specify value format they should get a validationexception
1
424,505
29,117,249,548
IssuesEvent
2023-05-17 02:48:54
pilcrowOnPaper/lucia
https://api.github.com/repos/pilcrowOnPaper/lucia
closed
[Docs-Bug]: Drizzle ORM Sqlite incorrect
documentation bug
### Describe the bug In the Drizzle ORM documentation page, the SQLite section has the same code as the mysql section, and refers to functions (i.e. varchar) that are not available in the sqlite adapter (text is used instead). Current Documentation ```typescript import { sqliteTable, varchar, integer } from "drizzle-orm/sqlite-core"; const user = mysqlTable("auth_user", { id: varchar("id", { length: 15 // change this when using custom user ids }).primaryKey() // other user attributes }); const session = mysqlTable("auth_session", { id: varchar("id", { length: 128 }).primaryKey(), userId: varchar("user_id", { length: 15 }) .notNull() .references(() => user.id), activeExpires: integer("active_expires").notNull(), idleExpires: integer("idle_expires").notNull() }); const key = mysqlTable("auth_key", { id: varchar("id", { length: 255 }).primaryKey(), userId: varchar("user_id", { length: 15 }) .notNull() .references(() => user.id), primaryKey: integer().notNull(), hashedPassword: varchar("hashed_password", { length: 255 }) }); ```
1.0
[Docs-Bug]: Drizzle ORM Sqlite incorrect - ### Describe the bug In the Drizzle ORM documentation page, the SQLite section has the same code as the mysql section, and refers to functions (i.e. varchar) that are not available in the sqlite adapter (text is used instead). Current Documentation ```typescript import { sqliteTable, varchar, integer } from "drizzle-orm/sqlite-core"; const user = mysqlTable("auth_user", { id: varchar("id", { length: 15 // change this when using custom user ids }).primaryKey() // other user attributes }); const session = mysqlTable("auth_session", { id: varchar("id", { length: 128 }).primaryKey(), userId: varchar("user_id", { length: 15 }) .notNull() .references(() => user.id), activeExpires: integer("active_expires").notNull(), idleExpires: integer("idle_expires").notNull() }); const key = mysqlTable("auth_key", { id: varchar("id", { length: 255 }).primaryKey(), userId: varchar("user_id", { length: 15 }) .notNull() .references(() => user.id), primaryKey: integer().notNull(), hashedPassword: varchar("hashed_password", { length: 255 }) }); ```
non_process
drizzle orm sqlite incorrect describe the bug in the drizzle orm documentation page the sqlite section has the same code as the mysql section and refers to functions i e varchar that are not available in the sqlite adapter text is used instead current documentation typescript import sqlitetable varchar integer from drizzle orm sqlite core const user mysqltable auth user id varchar id length change this when using custom user ids primarykey other user attributes const session mysqltable auth session id varchar id length primarykey userid varchar user id length notnull references user id activeexpires integer active expires notnull idleexpires integer idle expires notnull const key mysqltable auth key id varchar id length primarykey userid varchar user id length notnull references user id primarykey integer notnull hashedpassword varchar hashed password length
0
5,537
8,392,233,582
IssuesEvent
2018-10-09 17:02:12
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Do not get goaccess to work with my log file
log-processing log/date/time format question
Hi, I went through all kind of issues, but I am not able to get the default vcombined to work THis is an example of a line: testma.mantagifts.com:443 ip-217-105-39-28.ip.prioritytelecom.net - mantasalesapp [04/Oct/2018:07:53:48 +0200] "POST /sales_app/backups/backup.php HTTP/1.1" 200 427 "-" "Appcelerator Titanium/7.4.0 (iPad/12.0; iOS; en_US;)" This is my command: > goaccess /var/log/apache2/access_ssl_testma.mantagifts.com.log --log-format=VCOMBINED .. Token 'ip-217-105-39-28.ip.prioritytelecom.net' doesn't match specifier '%h' .. Format Errors - Verify your log/date/time format I have tried multiple thinkgs (custom log, time etc),but nothing seem to work. Can you help me with this? Thanks
1.0
Do not get goaccess to work with my log file - Hi, I went through all kind of issues, but I am not able to get the default vcombined to work THis is an example of a line: testma.mantagifts.com:443 ip-217-105-39-28.ip.prioritytelecom.net - mantasalesapp [04/Oct/2018:07:53:48 +0200] "POST /sales_app/backups/backup.php HTTP/1.1" 200 427 "-" "Appcelerator Titanium/7.4.0 (iPad/12.0; iOS; en_US;)" This is my command: > goaccess /var/log/apache2/access_ssl_testma.mantagifts.com.log --log-format=VCOMBINED .. Token 'ip-217-105-39-28.ip.prioritytelecom.net' doesn't match specifier '%h' .. Format Errors - Verify your log/date/time format I have tried multiple thinkgs (custom log, time etc),but nothing seem to work. Can you help me with this? Thanks
process
do not get goaccess to work with my log file hi i went through all kind of issues but i am not able to get the default vcombined to work this is an example of a line testma mantagifts com ip ip prioritytelecom net mantasalesapp post sales app backups backup php http appcelerator titanium ipad ios en us this is my command goaccess var log access ssl testma mantagifts com log log format vcombined token ip ip prioritytelecom net doesn t match specifier h format errors verify your log date time format i have tried multiple thinkgs custom log time etc but nothing seem to work can you help me with this thanks
1
2,427
5,203,079,979
IssuesEvent
2017-01-24 11:37:26
rogerthat-platform/rogerthat-android-client
https://api.github.com/repos/rogerthat-platform/rogerthat-android-client
closed
Speed up database initialization on fresh installs
process_duplicate type_feature
Update script from version 0 to version 73 + support for upgrading from 73 to 74, if 74 exists.
1.0
Speed up database initialization on fresh installs - Update script from version 0 to version 73 + support for upgrading from 73 to 74, if 74 exists.
process
speed up database initialization on fresh installs update script from version to version support for upgrading from to if exists
1
6,306
9,309,369,444
IssuesEvent
2019-03-25 16:20:06
googleapis/google-cloud-cpp
https://api.github.com/repos/googleapis/google-cloud-cpp
closed
Test driver scripts may miss errors.
type: process
The driver scripts for tests and examples use expressions like this: ```sh local foo="blah-$(date +%s)" ``` These are a problem because errors in the shell expansion `$(...)` are not reported in this case. We can use: ```sh local foo foo="blah-$(date +%s)" ``` but those do not work with `local -r`. We can replace the "date as a source of random numbers" with `${RANDOM}` in some cases. Regardless, we should search for such expressions and fix them appropriately. Thanks to @dopiera for pointing this out.
1.0
Test driver scripts may miss errors. - The driver scripts for tests and examples use expressions like this: ```sh local foo="blah-$(date +%s)" ``` These are a problem because errors in the shell expansion `$(...)` are not reported in this case. We can use: ```sh local foo foo="blah-$(date +%s)" ``` but those do not work with `local -r`. We can replace the "date as a source of random numbers" with `${RANDOM}` in some cases. Regardless, we should search for such expressions and fix them appropriately. Thanks to @dopiera for pointing this out.
process
test driver scripts may miss errors the driver scripts for tests and examples use expressions like this sh local foo blah date s these are a problem because errors in the shell expansion are not reported in this case we can use sh local foo foo blah date s but those do not work with local r we can replace the date as a source of random numbers with random in some cases regardless we should search for such expressions and fix them appropriately thanks to dopiera for pointing this out
1
1,674
4,312,642,471
IssuesEvent
2016-07-22 06:56:57
Jumpscale/jscockpit
https://api.github.com/repos/Jumpscale/jscockpit
reopened
Install a BP to deploy a CS which is already deployed doesn't update its properties
process_wontfix type_bug
**Steps** - Create a CS in du-conv-2 with name "checkdeleteissue". Its ID is 1647 ![screenshot from 2016-07-21 13-39-43](https://cloud.githubusercontent.com/assets/17981089/17021490/fbf61df4-4f50-11e6-83dc-b3c1d51dcb21.png) - Create a new Repo in a cockpit and enter this bp which will install a CS with "checkdeleteissue" (same CS in the same env) ``` g8client__dubai: g8.url: 'du-conv-2.demo.greenitglobe.com' g8.login: 'gig' g8.password: 'KrOe6gE9K5nCQdmretfXnj' g8.account: 'gig' vdc__checkdeleteissue: g8.client.name: 'dubai' maxMemoryCapacity: 2 maxVDiskCapacity: 10 maxCPUCapacity: 2 maxNASCapacity: 20 maxArchiveCapacity: 20 maxNetworkOptTransfer: 5 maxNetworkPeerTransfer: 15 maxNumPublicIP: 1 ``` - Execute BP - Init Repo - Install ![screenshot from 2016-07-21 13-43-44](https://cloud.githubusercontent.com/assets/17981089/17021573/8ad63c70-4f51-11e6-9358-caaa72da7506.png) - Check the Repo instance ![screenshot from 2016-07-21 13-44-36](https://cloud.githubusercontent.com/assets/17981089/17022291/d6524b0e-4f55-11e6-9df9-16e88e7401fd.png) **Expected result** Cockpit should check if this CS is already deployed and if yes, It must produce an error message.
1.0
Install a BP to deploy a CS which is already deployed doesn't update its properties - **Steps** - Create a CS in du-conv-2 with name "checkdeleteissue". Its ID is 1647 ![screenshot from 2016-07-21 13-39-43](https://cloud.githubusercontent.com/assets/17981089/17021490/fbf61df4-4f50-11e6-83dc-b3c1d51dcb21.png) - Create a new Repo in a cockpit and enter this bp which will install a CS with "checkdeleteissue" (same CS in the same env) ``` g8client__dubai: g8.url: 'du-conv-2.demo.greenitglobe.com' g8.login: 'gig' g8.password: 'KrOe6gE9K5nCQdmretfXnj' g8.account: 'gig' vdc__checkdeleteissue: g8.client.name: 'dubai' maxMemoryCapacity: 2 maxVDiskCapacity: 10 maxCPUCapacity: 2 maxNASCapacity: 20 maxArchiveCapacity: 20 maxNetworkOptTransfer: 5 maxNetworkPeerTransfer: 15 maxNumPublicIP: 1 ``` - Execute BP - Init Repo - Install ![screenshot from 2016-07-21 13-43-44](https://cloud.githubusercontent.com/assets/17981089/17021573/8ad63c70-4f51-11e6-9358-caaa72da7506.png) - Check the Repo instance ![screenshot from 2016-07-21 13-44-36](https://cloud.githubusercontent.com/assets/17981089/17022291/d6524b0e-4f55-11e6-9df9-16e88e7401fd.png) **Expected result** Cockpit should check if this CS is already deployed and if yes, It must produce an error message.
process
install a bp to deploy a cs which is already deployed doesn t update its properties steps create a cs in du conv with name checkdeleteissue its id is create a new repo in a cockpit and enter this bp which will install a cs with checkdeleteissue same cs in the same env dubai url du conv demo greenitglobe com login gig password account gig vdc checkdeleteissue client name dubai maxmemorycapacity maxvdiskcapacity maxcpucapacity maxnascapacity maxarchivecapacity maxnetworkopttransfer maxnetworkpeertransfer maxnumpublicip execute bp init repo install check the repo instance expected result cockpit should check if this cs is already deployed and if yes it must produce an error message
1
236,193
18,074,622,066
IssuesEvent
2021-09-21 08:29:10
jupyterhub/binderhub
https://api.github.com/repos/jupyterhub/binderhub
closed
Going from z2jh to z2bh
enhancement documentation
Related to #611 and https://github.com/jupyterhub/zero-to-jupyterhub-k8s/issues/643 @aculich and I discussed working through the tutorials in the following order: 1. littlest jupyterhub on GCP 2. zero to jupyterhub 3. zero to binderhub @betatim and @matthew-brett @LotteWeerts started working through this yesterday and we have some ideas on how to document this. It would be useful to document the best way to transition from each tutorial so you don't destroy your session each time.
1.0
Going from z2jh to z2bh - Related to #611 and https://github.com/jupyterhub/zero-to-jupyterhub-k8s/issues/643 @aculich and I discussed working through the tutorials in the following order: 1. littlest jupyterhub on GCP 2. zero to jupyterhub 3. zero to binderhub @betatim and @matthew-brett @LotteWeerts started working through this yesterday and we have some ideas on how to document this. It would be useful to document the best way to transition from each tutorial so you don't destroy your session each time.
non_process
going from to related to and aculich and i discussed working through the tutorials in the following order littlest jupyterhub on gcp zero to jupyterhub zero to binderhub betatim and matthew brett lotteweerts started working through this yesterday and we have some ideas on how to document this it would be useful to document the best way to transition from each tutorial so you don t destroy your session each time
0
302,559
26,150,595,593
IssuesEvent
2022-12-30 13:00:13
nupac/nupac
https://api.github.com/repos/nupac/nupac
opened
Add tests with `--long` flag
good first issue priority: low tests
### Describe the issue There are no tests using the `--long` flag, which means we don't test if it raises some errors. There's no need for robust parsing, just test if each command works just fine with the flag — basically `ASSERT_NO_THROW` ### Argumentation We don't test if the code actually works with this flag ### Screenshots and other helpful media _No response_ ### Additional context _No response_
1.0
Add tests with `--long` flag - ### Describe the issue There are no tests using the `--long` flag, which means we don't test if it raises some errors. There's no need for robust parsing, just test if each command works just fine with the flag — basically `ASSERT_NO_THROW` ### Argumentation We don't test if the code actually works with this flag ### Screenshots and other helpful media _No response_ ### Additional context _No response_
non_process
add tests with long flag describe the issue there are no tests using the long flag which means we don t test if it raises some errors there s no need for robust parsing just test if each command works just fine with the flag — basically assert no throw argumentation we don t test if the code actually works with this flag screenshots and other helpful media no response additional context no response
0
3,336
6,470,674,941
IssuesEvent
2017-08-17 09:37:15
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
Add ability to search for a vdisk based on the name and vpool
process_duplicate
Currently we only have a limited number of ways to search for vdisks. An awesome, additional way would be to be able to search with the name and vpool of a vdisk. Now we have to loop over all vdisks and compare the name to find the matching disk.
1.0
Add ability to search for a vdisk based on the name and vpool - Currently we only have a limited number of ways to search for vdisks. An awesome, additional way would be to be able to search with the name and vpool of a vdisk. Now we have to loop over all vdisks and compare the name to find the matching disk.
process
add ability to search for a vdisk based on the name and vpool currently we only have a limited number of ways to search for vdisks an awesome additional way would be to be able to search with the name and vpool of a vdisk now we have to loop over all vdisks and compare the name to find the matching disk
1
3,734
6,733,143,621
IssuesEvent
2017-10-18 13:58:51
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
closed
Warranty Assessment Species Analysis - View
enhancement process workflow report
Create a report to display the species analysis for a certain warranty period
1.0
Warranty Assessment Species Analysis - View - Create a report to display the species analysis for a certain warranty period
process
warranty assessment species analysis view create a report to display the species analysis for a certain warranty period
1
927
3,389,328,879
IssuesEvent
2015-11-30 00:37:42
MaretEngineering/MROV
https://api.github.com/repos/MaretEngineering/MROV
opened
Be able to turn off Serial by changing one constant
Necessary Addition Processing
All the lines that refer to serial should be inside if statements that refer to that constant.
1.0
Be able to turn off Serial by changing one constant - All the lines that refer to serial should be inside if statements that refer to that constant.
process
be able to turn off serial by changing one constant all the lines that refer to serial should be inside if statements that refer to that constant
1
117,941
15,193,346,814
IssuesEvent
2021-02-16 00:27:29
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Limit output resolution/size from darkroom
feature: redesign no-issue-activity priority: low scope: UI
From https://redmine.darktable.org/issues/12351. When processing noisy images (e.g. ISO 6400), I do not export with full pixel count but limit to a reasonable size that hides the noise a bit. The same holds for slightly unsharp images. That way, the maximum usable image size, e.g. maximum print size, is given by the image resolution and limited such that the technical flaws are not visible for the people viewing the picture. Unfortunately, the output size/resolution is determined in lighttable mode and the image itself does not carry this information. When I export an old picture again, I maybe forget that it was a blurry shot and therefore it is sent out at e.g. full resolution. Furthermore, when processing a set of images, one always has to export one by one to account for the blurry/noisy shots. It would therefore be great if one could limit the resolution from within darkroom such that is cannot be overridden by the export module (smaller export would still be possible). A new module “limit output size” would be a reasonable way to implement this feature. Within the module, the output size limit could be selected as either (maybe a dropdown to choose among these) * absolute pixel numbers (input of two numbers, max. horizontal and vertical pixel count; dependent on the aspect ratio, one of the numbers wins, as in the export module), * megapixel count (input one number, e.g. 10M, and the closest smaller size that would match the aspect ratio is used), or * factor of the original size (e.g. “2” for downscaling to half the size; factor 2 would therefore mean quarter of the pixel count; or, instead, the factor of the pixel count to avoid the necessity of sqrt(2) input; or maybe both options). The darkroom view would of course follow the limitation to have a reasonable preview. The module itself would also show the actual output size limit of the image. I thought a bit more about that idea, and came to the conclusion that there is an easy part and a difficult one. The easy part would be to implement a factor of the original size. This could, in the UI, just be a text entry or a slider. A widget that computes and shows the resulting image size would be convenient. The benefit is that this does not interfere with changes in the crop. It would only require one additional entry in the xmp/data base, which is the scaling factor as a float number. The other two ways to specify the scaling have the issue that they fail for strong crops, where you may even go below the specified output size. Therefore, for the other two options, the UI would have to reflect what it refers to, the original image size or the crop (a check box named “referred to the crop size” which is initially not checked, and an “upscaling allowed” check box. The first, easy, solution would already help a lot, so the implementation of this feature would not have to go the entire way at a time. Slightly related: https://redmine.darktable.org/issues/12234
1.0
Limit output resolution/size from darkroom - From https://redmine.darktable.org/issues/12351. When processing noisy images (e.g. ISO 6400), I do not export with full pixel count but limit to a reasonable size that hides the noise a bit. The same holds for slightly unsharp images. That way, the maximum usable image size, e.g. maximum print size, is given by the image resolution and limited such that the technical flaws are not visible for the people viewing the picture. Unfortunately, the output size/resolution is determined in lighttable mode and the image itself does not carry this information. When I export an old picture again, I maybe forget that it was a blurry shot and therefore it is sent out at e.g. full resolution. Furthermore, when processing a set of images, one always has to export one by one to account for the blurry/noisy shots. It would therefore be great if one could limit the resolution from within darkroom such that is cannot be overridden by the export module (smaller export would still be possible). A new module “limit output size” would be a reasonable way to implement this feature. Within the module, the output size limit could be selected as either (maybe a dropdown to choose among these) * absolute pixel numbers (input of two numbers, max. horizontal and vertical pixel count; dependent on the aspect ratio, one of the numbers wins, as in the export module), * megapixel count (input one number, e.g. 10M, and the closest smaller size that would match the aspect ratio is used), or * factor of the original size (e.g. “2” for downscaling to half the size; factor 2 would therefore mean quarter of the pixel count; or, instead, the factor of the pixel count to avoid the necessity of sqrt(2) input; or maybe both options). The darkroom view would of course follow the limitation to have a reasonable preview. The module itself would also show the actual output size limit of the image. I thought a bit more about that idea, and came to the conclusion that there is an easy part and a difficult one. The easy part would be to implement a factor of the original size. This could, in the UI, just be a text entry or a slider. A widget that computes and shows the resulting image size would be convenient. The benefit is that this does not interfere with changes in the crop. It would only require one additional entry in the xmp/data base, which is the scaling factor as a float number. The other two ways to specify the scaling have the issue that they fail for strong crops, where you may even go below the specified output size. Therefore, for the other two options, the UI would have to reflect what it refers to, the original image size or the crop (a check box named “referred to the crop size” which is initially not checked, and an “upscaling allowed” check box. The first, easy, solution would already help a lot, so the implementation of this feature would not have to go the entire way at a time. Slightly related: https://redmine.darktable.org/issues/12234
non_process
limit output resolution size from darkroom from when processing noisy images e g iso i do not export with full pixel count but limit to a reasonable size that hides the noise a bit the same holds for slightly unsharp images that way the maximum usable image size e g maximum print size is given by the image resolution and limited such that the technical flaws are not visible for the people viewing the picture unfortunately the output size resolution is determined in lighttable mode and the image itself does not carry this information when i export an old picture again i maybe forget that it was a blurry shot and therefore it is sent out at e g full resolution furthermore when processing a set of images one always has to export one by one to account for the blurry noisy shots it would therefore be great if one could limit the resolution from within darkroom such that is cannot be overridden by the export module smaller export would still be possible a new module “limit output size” would be a reasonable way to implement this feature within the module the output size limit could be selected as either maybe a dropdown to choose among these absolute pixel numbers input of two numbers max horizontal and vertical pixel count dependent on the aspect ratio one of the numbers wins as in the export module megapixel count input one number e g and the closest smaller size that would match the aspect ratio is used or factor of the original size e g “ ” for downscaling to half the size factor would therefore mean quarter of the pixel count or instead the factor of the pixel count to avoid the necessity of sqrt input or maybe both options the darkroom view would of course follow the limitation to have a reasonable preview the module itself would also show the actual output size limit of the image i thought a bit more about that idea and came to the conclusion that there is an easy part and a difficult one the easy part would be to implement a factor of the original size this could in the ui just be a text entry or a slider a widget that computes and shows the resulting image size would be convenient the benefit is that this does not interfere with changes in the crop it would only require one additional entry in the xmp data base which is the scaling factor as a float number the other two ways to specify the scaling have the issue that they fail for strong crops where you may even go below the specified output size therefore for the other two options the ui would have to reflect what it refers to the original image size or the crop a check box named “referred to the crop size” which is initially not checked and an “upscaling allowed” check box the first easy solution would already help a lot so the implementation of this feature would not have to go the entire way at a time slightly related
0
43,319
9,417,819,382
IssuesEvent
2019-04-10 17:40:34
grpc/grpc-java
https://api.github.com/repos/grpc/grpc-java
closed
Releases can be uploaded without signing being successful
code health
We configure signing to be required, but it isn't working for some reason.
1.0
Releases can be uploaded without signing being successful - We configure signing to be required, but it isn't working for some reason.
non_process
releases can be uploaded without signing being successful we configure signing to be required but it isn t working for some reason
0
15,230
19,101,357,993
IssuesEvent
2021-11-29 23:05:26
ooi-data/CE09OSSM-MFD37-03-DOSTAD000-telemetered-dosta_abcdjm_ctdbp_dcl_instrument
https://api.github.com/repos/ooi-data/CE09OSSM-MFD37-03-DOSTAD000-telemetered-dosta_abcdjm_ctdbp_dcl_instrument
closed
🛑 Processing failed: ResponseParserError
process
## Overview `ResponseParserError` found in `processing_task` task during run ended on 2021-05-05T16:35:55.842847. ## Details Flow name: `CE09OSSM-MFD37-03-DOSTAD000-telemetered-dosta_abcdjm_ctdbp_dcl_instrument` Task name: `processing_task` Error type: `ResponseParserError` Error message: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 452, in _parse_xml_string_to_dom root = parser.close() File "<string>", line None xml.etree.ElementTree.ParseError: no element found: line 2, column 0 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/share/miniconda/envs/harvester/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 71, in processing_task File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 311, in finalize_zarr source_store.fs.delete(source_store.root, recursive=True) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1151, in delete return self.rm(path, recursive=recursive, maxdepth=maxdepth) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 72, in wrapper return sync(self.loop, func, *args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 53, in sync raise result[0] File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 20, in _runner result[0] = await coro File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1510, in _rm await asyncio.gather( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1498, in _bulk_delete await self._call_s3("delete_objects", kwargs, Bucket=bucket, Delete=delete_keys) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 252, in _call_s3 raise translate_boto_error(err) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 233, in _call_s3 out = await method(**additional_kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 140, in _make_api_call http, parsed_response = await self._make_request( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 160, in _make_request return await self._endpoint.make_request(operation_model, request_dict) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 101, in _send_request success_response, exception = await self._get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 120, in _get_response success_response, exception = await self._do_get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 180, in _do_get_response parsed_response = parser.parse( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 245, in parse parsed = self._do_parse(response, shape) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 809, in _do_parse self._add_modeled_parse(response, shape, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 818, in _add_modeled_parse self._parse_payload(response, shape, member_shapes, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 858, in _parse_payload original_parsed = self._initial_body_parse(response['body']) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 944, in _initial_body_parse return self._parse_xml_string_to_dom(xml_string) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 454, in _parse_xml_string_to_dom raise ResponseParserError( botocore.parsers.ResponseParserError: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' ``` </details>
1.0
🛑 Processing failed: ResponseParserError - ## Overview `ResponseParserError` found in `processing_task` task during run ended on 2021-05-05T16:35:55.842847. ## Details Flow name: `CE09OSSM-MFD37-03-DOSTAD000-telemetered-dosta_abcdjm_ctdbp_dcl_instrument` Task name: `processing_task` Error type: `ResponseParserError` Error message: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 452, in _parse_xml_string_to_dom root = parser.close() File "<string>", line None xml.etree.ElementTree.ParseError: no element found: line 2, column 0 During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/share/miniconda/envs/harvester/lib/python3.8/site-packages/ooi_harvester/processor/pipeline.py", line 71, in processing_task File "/srv/conda/envs/notebook/lib/python3.8/site-packages/ooi_harvester/processor/__init__.py", line 311, in finalize_zarr source_store.fs.delete(source_store.root, recursive=True) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/spec.py", line 1151, in delete return self.rm(path, recursive=recursive, maxdepth=maxdepth) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 72, in wrapper return sync(self.loop, func, *args, **kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 53, in sync raise result[0] File "/srv/conda/envs/notebook/lib/python3.8/site-packages/fsspec/asyn.py", line 20, in _runner result[0] = await coro File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1510, in _rm await asyncio.gather( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 1498, in _bulk_delete await self._call_s3("delete_objects", kwargs, Bucket=bucket, Delete=delete_keys) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 252, in _call_s3 raise translate_boto_error(err) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/s3fs/core.py", line 233, in _call_s3 out = await method(**additional_kwargs) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 140, in _make_api_call http, parsed_response = await self._make_request( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/client.py", line 160, in _make_request return await self._endpoint.make_request(operation_model, request_dict) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 101, in _send_request success_response, exception = await self._get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 120, in _get_response success_response, exception = await self._do_get_response( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/aiobotocore/endpoint.py", line 180, in _do_get_response parsed_response = parser.parse( File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 245, in parse parsed = self._do_parse(response, shape) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 809, in _do_parse self._add_modeled_parse(response, shape, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 818, in _add_modeled_parse self._parse_payload(response, shape, member_shapes, final_parsed) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 858, in _parse_payload original_parsed = self._initial_body_parse(response['body']) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 944, in _initial_body_parse return self._parse_xml_string_to_dom(xml_string) File "/srv/conda/envs/notebook/lib/python3.8/site-packages/botocore/parsers.py", line 454, in _parse_xml_string_to_dom raise ResponseParserError( botocore.parsers.ResponseParserError: Unable to parse response (no element found: line 2, column 0), invalid XML received. Further retries may succeed: b'<?xml version="1.0" encoding="UTF-8"?>\n' ``` </details>
process
🛑 processing failed responseparsererror overview responseparsererror found in processing task task during run ended on details flow name telemetered dosta abcdjm ctdbp dcl instrument task name processing task error type responseparsererror error message unable to parse response no element found line column invalid xml received further retries may succeed b n traceback traceback most recent call last file srv conda envs notebook lib site packages botocore parsers py line in parse xml string to dom root parser close file line none xml etree elementtree parseerror no element found line column during handling of the above exception another exception occurred traceback most recent call last file usr share miniconda envs harvester lib site packages ooi harvester processor pipeline py line in processing task file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize zarr source store fs delete source store root recursive true file srv conda envs notebook lib site packages fsspec spec py line in delete return self rm path recursive recursive maxdepth maxdepth file srv conda envs notebook lib site packages fsspec asyn py line in wrapper return sync self loop func args kwargs file srv conda envs notebook lib site packages fsspec asyn py line in sync raise result file srv conda envs notebook lib site packages fsspec asyn py line in runner result await coro file srv conda envs notebook lib site packages core py line in rm await asyncio gather file srv conda envs notebook lib site packages core py line in bulk delete await self call delete objects kwargs bucket bucket delete delete keys file srv conda envs notebook lib site packages core py line in call raise translate boto error err file srv conda envs notebook lib site packages core py line in call out await method additional kwargs file srv conda envs notebook lib site packages aiobotocore client py line in make api call http parsed response await self make request file srv conda envs notebook lib site packages aiobotocore client py line in make request return await self endpoint make request operation model request dict file srv conda envs notebook lib site packages aiobotocore endpoint py line in send request success response exception await self get response file srv conda envs notebook lib site packages aiobotocore endpoint py line in get response success response exception await self do get response file srv conda envs notebook lib site packages aiobotocore endpoint py line in do get response parsed response parser parse file srv conda envs notebook lib site packages botocore parsers py line in parse parsed self do parse response shape file srv conda envs notebook lib site packages botocore parsers py line in do parse self add modeled parse response shape final parsed file srv conda envs notebook lib site packages botocore parsers py line in add modeled parse self parse payload response shape member shapes final parsed file srv conda envs notebook lib site packages botocore parsers py line in parse payload original parsed self initial body parse response file srv conda envs notebook lib site packages botocore parsers py line in initial body parse return self parse xml string to dom xml string file srv conda envs notebook lib site packages botocore parsers py line in parse xml string to dom raise responseparsererror botocore parsers responseparsererror unable to parse response no element found line column invalid xml received further retries may succeed b n
1
212,285
16,437,903,573
IssuesEvent
2021-05-20 11:22:29
ethersphere/bee
https://api.github.com/repos/ethersphere/bee
closed
TestNeighborhoodDepth flakes
flaky-test issue
``` === RUN TestNeighborhoodDepth kademlia_test.go:142: timed out waiting for depth. want 14 got 15 --- FAIL: TestNeighborhoodDepth (4.31s) ``` Probably related to balanced kademlia changes
1.0
TestNeighborhoodDepth flakes - ``` === RUN TestNeighborhoodDepth kademlia_test.go:142: timed out waiting for depth. want 14 got 15 --- FAIL: TestNeighborhoodDepth (4.31s) ``` Probably related to balanced kademlia changes
non_process
testneighborhooddepth flakes run testneighborhooddepth kademlia test go timed out waiting for depth want got fail testneighborhooddepth probably related to balanced kademlia changes
0
62,533
17,023,941,337
IssuesEvent
2021-07-03 04:41:07
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
OSM API v0.6 Reopening closed note bug
Component: api Priority: minor Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 9.56am, Wednesday, 12th December 2018]** when using the api `https://api.openstreetmap.org/api/0.6/notes/#id/reopen?text=Comment`, if the text contains `=` error code 401 will be thrown. while the comment api `https://api.openstreetmap.org/api/0.6/notes/#id/comment?text=ThisIsANoteComment` does not have such restrictions
1.0
OSM API v0.6 Reopening closed note bug - **[Submitted to the original trac issue database at 9.56am, Wednesday, 12th December 2018]** when using the api `https://api.openstreetmap.org/api/0.6/notes/#id/reopen?text=Comment`, if the text contains `=` error code 401 will be thrown. while the comment api `https://api.openstreetmap.org/api/0.6/notes/#id/comment?text=ThisIsANoteComment` does not have such restrictions
non_process
osm api reopening closed note bug when using the api if the text contains error code will be thrown while the comment api does not have such restrictions
0