Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
34,586
7,846,098,987
IssuesEvent
2018-06-19 14:38:59
eamodio/vscode-gitlens
https://api.github.com/repos/eamodio/vscode-gitlens
closed
Is it possible to not have the code jump on opening files due to GitLens
question upstream/vscode
<!-- If you are encountering an issue that says `See output channel for more details`, please enable output channel logging by setting `"gitlens.outputLevel": "verbose"` in your settings.json. This will enable logging to the GitLens channel in the Output pane. Once enabled, please attempt to reproduce the issue (if possible) and attach the relevant log lines from the GitLens channel. --> - GitLens Version: 8.3 - VSCode Version: 1.24.0 - OS Version: 10.11.6 Steps to Reproduce: 1. Open any new file. 2. Wait ~ 1 second. The text will jump down to make space for the GitLens commit overlay. I find it extremely annoying how when I open any file the text jumps down a bit to make space for this overlay: ![2018-06-15 at 01 14](https://user-images.githubusercontent.com/6391776/41442821-71225a06-7039-11e8-80d2-b2959f1c47cc.png) Is it possible to either turn it off completely or make it so that when I open a file, the text is already moved so that the text doesn't shift to make space for that overlay. Thank you.
1.0
Is it possible to not have the code jump on opening files due to GitLens - <!-- If you are encountering an issue that says `See output channel for more details`, please enable output channel logging by setting `"gitlens.outputLevel": "verbose"` in your settings.json. This will enable logging to the GitLens channel in the Output pane. Once enabled, please attempt to reproduce the issue (if possible) and attach the relevant log lines from the GitLens channel. --> - GitLens Version: 8.3 - VSCode Version: 1.24.0 - OS Version: 10.11.6 Steps to Reproduce: 1. Open any new file. 2. Wait ~ 1 second. The text will jump down to make space for the GitLens commit overlay. I find it extremely annoying how when I open any file the text jumps down a bit to make space for this overlay: ![2018-06-15 at 01 14](https://user-images.githubusercontent.com/6391776/41442821-71225a06-7039-11e8-80d2-b2959f1c47cc.png) Is it possible to either turn it off completely or make it so that when I open a file, the text is already moved so that the text doesn't shift to make space for that overlay. Thank you.
non_defect
is it possible to not have the code jump on opening files due to gitlens if you are encountering an issue that says see output channel for more details please enable output channel logging by setting gitlens outputlevel verbose in your settings json this will enable logging to the gitlens channel in the output pane once enabled please attempt to reproduce the issue if possible and attach the relevant log lines from the gitlens channel gitlens version vscode version os version steps to reproduce open any new file wait second the text will jump down to make space for the gitlens commit overlay i find it extremely annoying how when i open any file the text jumps down a bit to make space for this overlay is it possible to either turn it off completely or make it so that when i open a file the text is already moved so that the text doesn t shift to make space for that overlay thank you
0
70,008
22,783,136,111
IssuesEvent
2022-07-08 23:02:47
Clever-ISA/Clever-ISA
https://api.github.com/repos/Clever-ISA/Clever-ISA
closed
All Jumps should not sync instructions
X-main I-defect S-blocked-on-maintainer V-1.0
It is currently impossible to have a single page mapping that is both writable and executable. Therefore under most circumstances, it is impossible for the program to modify the instruction stream being executed without a call to a supervisor. Further, branches syncing instruction stream nearly negates the ability for branch prediction and prefetching branches (except special `fast` jumps that explicitly do not sync).
1.0
All Jumps should not sync instructions - It is currently impossible to have a single page mapping that is both writable and executable. Therefore under most circumstances, it is impossible for the program to modify the instruction stream being executed without a call to a supervisor. Further, branches syncing instruction stream nearly negates the ability for branch prediction and prefetching branches (except special `fast` jumps that explicitly do not sync).
defect
all jumps should not sync instructions it is currently impossible to have a single page mapping that is both writable and executable therefore under most circumstances it is impossible for the program to modify the instruction stream being executed without a call to a supervisor further branches syncing instruction stream nearly negates the ability for branch prediction and prefetching branches except special fast jumps that explicitly do not sync
1
50,239
10,469,369,471
IssuesEvent
2019-09-22 20:12:14
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
closed
[RuyJIT/ARM] Generated code quaility
area-CodeGen question
As I remember there was no any activities to check whether RuyJIT generates well optimized code. So how we can do it and what is the references to check with? The easiest way I see is look to generated code and search not efficient places. For test programs could be used: 1. Benchmarks/micro-benchmarks, but they are quite far from real user scenarios 1. Real applications (we can analyze only hot methods, to reduce amount of work). @jkotas could you recommend some other ways and metrics? I'm looking approaches: 1. How to search problem places 1. How to find ways to fix them 1. How to estimate gain (it's nice to make optimizations but they should be valuable for real scenarios=\ )
1.0
[RuyJIT/ARM] Generated code quaility - As I remember there was no any activities to check whether RuyJIT generates well optimized code. So how we can do it and what is the references to check with? The easiest way I see is look to generated code and search not efficient places. For test programs could be used: 1. Benchmarks/micro-benchmarks, but they are quite far from real user scenarios 1. Real applications (we can analyze only hot methods, to reduce amount of work). @jkotas could you recommend some other ways and metrics? I'm looking approaches: 1. How to search problem places 1. How to find ways to fix them 1. How to estimate gain (it's nice to make optimizations but they should be valuable for real scenarios=\ )
non_defect
generated code quaility as i remember there was no any activities to check whether ruyjit generates well optimized code so how we can do it and what is the references to check with the easiest way i see is look to generated code and search not efficient places for test programs could be used benchmarks micro benchmarks but they are quite far from real user scenarios real applications we can analyze only hot methods to reduce amount of work jkotas could you recommend some other ways and metrics i m looking approaches how to search problem places how to find ways to fix them how to estimate gain it s nice to make optimizations but they should be valuable for real scenarios
0
45,187
11,599,228,626
IssuesEvent
2020-02-25 01:32:36
openenclave/openenclave
https://api.github.com/repos/openenclave/openenclave
closed
Transition to using Terraform with Jenkins for CI/CD
build ci/cd triaged
This issue is for tracking improvements to our CI/CD infrastructure to using Terraform to provision environments with SGX-enabled nodes for both Linux and Windows. This will encompass bare-metal testing and supporting additional OS's seamlessly We should have at least one machine scale set for each distribution/OS/SGX configuration (kabyLake etc). These nodes should be connected to Jenkins and all CI/CD processes should run in containers, with necessary exceptions. We should automate the maintenance of the SGX driver (and security updates on the nodes too). Enable scaling rules to the CI/CD resources to handle high demand.
1.0
Transition to using Terraform with Jenkins for CI/CD - This issue is for tracking improvements to our CI/CD infrastructure to using Terraform to provision environments with SGX-enabled nodes for both Linux and Windows. This will encompass bare-metal testing and supporting additional OS's seamlessly We should have at least one machine scale set for each distribution/OS/SGX configuration (kabyLake etc). These nodes should be connected to Jenkins and all CI/CD processes should run in containers, with necessary exceptions. We should automate the maintenance of the SGX driver (and security updates on the nodes too). Enable scaling rules to the CI/CD resources to handle high demand.
non_defect
transition to using terraform with jenkins for ci cd this issue is for tracking improvements to our ci cd infrastructure to using terraform to provision environments with sgx enabled nodes for both linux and windows this will encompass bare metal testing and supporting additional os s seamlessly we should have at least one machine scale set for each distribution os sgx configuration kabylake etc these nodes should be connected to jenkins and all ci cd processes should run in containers with necessary exceptions we should automate the maintenance of the sgx driver and security updates on the nodes too enable scaling rules to the ci cd resources to handle high demand
0
22,377
3,642,191,605
IssuesEvent
2016-02-14 05:23:47
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
Non-Admin Execution can erroneously result in VirtualStore contents
C: Client - Setup P: Undetermined T: Defect
**Reported by JacobKlein on 25 Jul 43370817 13:34 UTC** Somehow, my BOINC installation resulted in a !VirtualStore directory being created. A later installation, which fixed BOINC to reference the proper !ProgramData BOINC directory, still did not quite work correctly, because of the existence of the !VirtualStore data. We should try to:[Find how the !VirtualStore directory got created, and prevent it if possible[[BR]([BR]]-)]- Cleanup any existing BOINC !VirtualStore directories, if possible, since their mere existence can cause problems. The issue I had was that the GPUGrid.net project would use some files from C:\!ProgramData\BOINC\slots\0 ... and some files from !VirtualStore\!ProgramData\BOINC\slots\0 ... and the end effect was that tasks would complete immediately, be marked successful, and I would be granted credit. Not good. Migrated-From: http://boinc.berkeley.edu/trac/ticket/1247
1.0
Non-Admin Execution can erroneously result in VirtualStore contents - **Reported by JacobKlein on 25 Jul 43370817 13:34 UTC** Somehow, my BOINC installation resulted in a !VirtualStore directory being created. A later installation, which fixed BOINC to reference the proper !ProgramData BOINC directory, still did not quite work correctly, because of the existence of the !VirtualStore data. We should try to:[Find how the !VirtualStore directory got created, and prevent it if possible[[BR]([BR]]-)]- Cleanup any existing BOINC !VirtualStore directories, if possible, since their mere existence can cause problems. The issue I had was that the GPUGrid.net project would use some files from C:\!ProgramData\BOINC\slots\0 ... and some files from !VirtualStore\!ProgramData\BOINC\slots\0 ... and the end effect was that tasks would complete immediately, be marked successful, and I would be granted credit. Not good. Migrated-From: http://boinc.berkeley.edu/trac/ticket/1247
defect
non admin execution can erroneously result in virtualstore contents reported by jacobklein on jul utc somehow my boinc installation resulted in a virtualstore directory being created a later installation which fixed boinc to reference the proper programdata boinc directory still did not quite work correctly because of the existence of the virtualstore data we should try to cleanup any existing boinc virtualstore directories if possible since their mere existence can cause problems the issue i had was that the gpugrid net project would use some files from c programdata boinc slots and some files from virtualstore programdata boinc slots and the end effect was that tasks would complete immediately be marked successful and i would be granted credit not good migrated from
1
101,292
31,019,034,478
IssuesEvent
2023-08-10 02:39:01
netsampler/goflow2
https://api.github.com/repos/netsampler/goflow2
closed
Redhat OS Error `GLIBC_2.32' not found (required by goflow2)
build
I installed goflow2-1.3.4-1.x86_64.rpm in my redhat machine, when i execute it throws error as below: $goflow2 -h goflow2: /lib64/libc.so.6: version `GLIBC_2.32' not found (required by goflow2) goflow2: /lib64/libc.so.6: version `GLIBC_2.34' not found (required by goflow2) Can anyone help?
1.0
Redhat OS Error `GLIBC_2.32' not found (required by goflow2) - I installed goflow2-1.3.4-1.x86_64.rpm in my redhat machine, when i execute it throws error as below: $goflow2 -h goflow2: /lib64/libc.so.6: version `GLIBC_2.32' not found (required by goflow2) goflow2: /lib64/libc.so.6: version `GLIBC_2.34' not found (required by goflow2) Can anyone help?
non_defect
redhat os error glibc not found required by i installed rpm in my redhat machine when i execute it throws error as below h libc so version glibc not found required by libc so version glibc not found required by can anyone help
0
15,629
2,866,663,383
IssuesEvent
2015-06-05 08:25:01
Guake/guake
https://api.github.com/repos/Guake/guake
closed
Guake fails to start due to a GlobalHotkey related C call
Priority:High Type: Defect
This is the traceback: Traceback (most recent call last): File "/usr/lib/python2.7/runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name) File "/usr/lib/python2.7/runpy.py", line 72, in _run_code exec code in run_globals File "/usr/lib/python2.7/site-packages/guake/main.py", line 235, in <module> exec_main() File "/usr/lib/python2.7/site-packages/guake/main.py", line 231, in exec_main if not main(): File "/usr/lib/python2.7/site-packages/guake/main.py", line 147, in main instance = Guake() File "/usr/lib/python2.7/site-packages/guake/guake_app.py", line 287, in __init__ self.hotkeys = guake.globalhotkeys.GlobalHotkey() SystemError: NULL result without error in PyObject_Call This started happening since today. I tried the usual reboot-and-hope-things-work scenario, but didn't work. For the record, I am using Arch Linux.
1.0
Guake fails to start due to a GlobalHotkey related C call - This is the traceback: Traceback (most recent call last): File "/usr/lib/python2.7/runpy.py", line 162, in _run_module_as_main "__main__", fname, loader, pkg_name) File "/usr/lib/python2.7/runpy.py", line 72, in _run_code exec code in run_globals File "/usr/lib/python2.7/site-packages/guake/main.py", line 235, in <module> exec_main() File "/usr/lib/python2.7/site-packages/guake/main.py", line 231, in exec_main if not main(): File "/usr/lib/python2.7/site-packages/guake/main.py", line 147, in main instance = Guake() File "/usr/lib/python2.7/site-packages/guake/guake_app.py", line 287, in __init__ self.hotkeys = guake.globalhotkeys.GlobalHotkey() SystemError: NULL result without error in PyObject_Call This started happening since today. I tried the usual reboot-and-hope-things-work scenario, but didn't work. For the record, I am using Arch Linux.
defect
guake fails to start due to a globalhotkey related c call this is the traceback traceback most recent call last file usr lib runpy py line in run module as main main fname loader pkg name file usr lib runpy py line in run code exec code in run globals file usr lib site packages guake main py line in exec main file usr lib site packages guake main py line in exec main if not main file usr lib site packages guake main py line in main instance guake file usr lib site packages guake guake app py line in init self hotkeys guake globalhotkeys globalhotkey systemerror null result without error in pyobject call this started happening since today i tried the usual reboot and hope things work scenario but didn t work for the record i am using arch linux
1
271,241
29,368,818,927
IssuesEvent
2023-05-29 01:04:39
Hi-Fi/remotesikulilibrary
https://api.github.com/repos/Hi-Fi/remotesikulilibrary
opened
CVE-2023-26464 (High) detected in log4j-1.2.16.jar
Mend: dependency security vulnerability
## CVE-2023-26464 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.16.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/log4j/log4j/1.2.16/log4j-1.2.16.jar</p> <p> Dependency Hierarchy: - jrobotremoteserver-3.0.jar (Root Library) - :x: **log4j-1.2.16.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> ** UNSUPPORTED WHEN ASSIGNED ** When using the Chainsaw or SocketAppender components with Log4j 1.x on JRE less than 1.7, an attacker that manages to cause a logging entry involving a specially-crafted (ie, deeply nested) hashmap or hashtable (depending on which logging component is in use) to be processed could exhaust the available memory in the virtual machine and achieve Denial of Service when the object is deserialized. This issue affects Apache Log4j before 2. Affected users are recommended to update to Log4j 2.x. NOTE: This vulnerability only affects products that are no longer supported by the maintainer. <p>Publish Date: 2023-03-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26464>CVE-2023-26464</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-vp98-w2p3-mv35">https://github.com/advisories/GHSA-vp98-w2p3-mv35</a></p> <p>Release Date: 2023-03-10</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-26464 (High) detected in log4j-1.2.16.jar - ## CVE-2023-26464 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.16.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/log4j/log4j/1.2.16/log4j-1.2.16.jar</p> <p> Dependency Hierarchy: - jrobotremoteserver-3.0.jar (Root Library) - :x: **log4j-1.2.16.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> ** UNSUPPORTED WHEN ASSIGNED ** When using the Chainsaw or SocketAppender components with Log4j 1.x on JRE less than 1.7, an attacker that manages to cause a logging entry involving a specially-crafted (ie, deeply nested) hashmap or hashtable (depending on which logging component is in use) to be processed could exhaust the available memory in the virtual machine and achieve Denial of Service when the object is deserialized. This issue affects Apache Log4j before 2. Affected users are recommended to update to Log4j 2.x. NOTE: This vulnerability only affects products that are no longer supported by the maintainer. <p>Publish Date: 2023-03-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-26464>CVE-2023-26464</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-vp98-w2p3-mv35">https://github.com/advisories/GHSA-vp98-w2p3-mv35</a></p> <p>Release Date: 2023-03-10</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in jar cve high severity vulnerability vulnerable library jar apache path to dependency file pom xml path to vulnerable library root repository jar dependency hierarchy jrobotremoteserver jar root library x jar vulnerable library vulnerability details unsupported when assigned when using the chainsaw or socketappender components with x on jre less than an attacker that manages to cause a logging entry involving a specially crafted ie deeply nested hashmap or hashtable depending on which logging component is in use to be processed could exhaust the available memory in the virtual machine and achieve denial of service when the object is deserialized this issue affects apache before affected users are recommended to update to x note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core step up your open source security game with mend
0
5,601
2,783,739,024
IssuesEvent
2015-05-07 02:56:49
OData/odata.net
https://api.github.com/repos/OData/odata.net
closed
Uri parser can't work for short data type (Edm.Int16) for function parameter
3 - Tested bug fix ready ODataLib
Supposed I have a function as: ``` xml <Function Name="ShortFunction" IsBound="true"> <Parameter Name="bindingParameter" Type="Collection(NS.Customer)" /> <Parameter Name="number" Type="Edm.Int16" Nullable="false" /> <ReturnType Type="Edm.String" Unicode="false" /> </Function> ``` Then, I call this function as: ``` c# ~/odata/Customers/Default.ShortFunction(number=1) ``` It fails as ``` c# "Expression of type 'Edm.Int32' cannot be converted to type 'Edm.Int16'." ``` The call stack is: ``` c# at Microsoft.OData.Core.UriParser.Parsers.MetadataBindingUtils.ConvertToTypeIfNeeded(SingleValueNode source, IEdmTypeReference targetTypeReference) at Microsoft.OData.Core.UriParser.Parsers.FunctionCallBinder.BindSegmentParameters(ODataUriParserConfiguration configuration, IEdmOperation functionOrOpertion, ICollection`1 segmentParameterTokens) at Microsoft.OData.Core.UriParser.Parsers.ODataPathParser.TryBindingParametersAndMatchingOperation(String identifier, String parenthesisExpression, IEdmType bindingType, ODataUriParserConfiguration configuration, ICollection`1& boundParameters, IEdmOperation& matchingOperation) at Microsoft.OData.Core.UriParser.Parsers.ODataPathParser.TryCreateSegmentForOperation(ODataPathSegment previousSegment, String identifier, String parenthesisExpression) at Microsoft.OData.Core.UriParser.Parsers.ODataPathParser.CreateNextSegment(String text) at Microsoft.OData.Core.UriParser.Parsers.ODataPathParser.ParsePath(ICollection`1 segments) at Microsoft.OData.Core.UriParser.Parsers.ODataPathFactory.BindPath(ICollection`1 segments, ODataUriParserConfiguration configuration) at Microsoft.OData.Core.UriParser.ODataUriParser.ParsePathImplementation() at Microsoft.OData.Core.UriParser.ODataUriParser.Initialize() at Microsoft.OData.Core.UriParser.ODataUriParser.ParsePath() at System.Web.OData.Routing.DefaultODataPathHandler.Parse(IEdmModel model, String serviceRoot, String odataPath, ODataUriResolverSetttings resolverSettings, Boolean enableUriTemplateParsing) in d:\github\WebApi\OData\src\System.Web.OData\OData\Routing\DefaultODataPathHandler.cs:line 137 at System.Web.OData.Routing.DefaultODataPathHandler.Parse(IEdmModel model, String serviceRoot, String odataPath) in d:\github\WebApi\OData\src\System.Web.OData\OData\Routing\DefaultODataPathHandler.cs:line 55 at System.Web.OData.Routing.ODataPathRouteConstraint.Match(HttpRequestMessage request, IHttpRoute route, String parameterName, IDictionary`2 values, HttpRouteDirection routeDirection) in d:\github\WebApi\OData\src\System.Web.OData\OData\Routing\ODataPathRouteConstraint.cs:line 171 ``` I am using Web API 2.2 for OData V4. Thanks.
1.0
Uri parser can't work for short data type (Edm.Int16) for function parameter - Supposed I have a function as: ``` xml <Function Name="ShortFunction" IsBound="true"> <Parameter Name="bindingParameter" Type="Collection(NS.Customer)" /> <Parameter Name="number" Type="Edm.Int16" Nullable="false" /> <ReturnType Type="Edm.String" Unicode="false" /> </Function> ``` Then, I call this function as: ``` c# ~/odata/Customers/Default.ShortFunction(number=1) ``` It fails as ``` c# "Expression of type 'Edm.Int32' cannot be converted to type 'Edm.Int16'." ``` The call stack is: ``` c# at Microsoft.OData.Core.UriParser.Parsers.MetadataBindingUtils.ConvertToTypeIfNeeded(SingleValueNode source, IEdmTypeReference targetTypeReference) at Microsoft.OData.Core.UriParser.Parsers.FunctionCallBinder.BindSegmentParameters(ODataUriParserConfiguration configuration, IEdmOperation functionOrOpertion, ICollection`1 segmentParameterTokens) at Microsoft.OData.Core.UriParser.Parsers.ODataPathParser.TryBindingParametersAndMatchingOperation(String identifier, String parenthesisExpression, IEdmType bindingType, ODataUriParserConfiguration configuration, ICollection`1& boundParameters, IEdmOperation& matchingOperation) at Microsoft.OData.Core.UriParser.Parsers.ODataPathParser.TryCreateSegmentForOperation(ODataPathSegment previousSegment, String identifier, String parenthesisExpression) at Microsoft.OData.Core.UriParser.Parsers.ODataPathParser.CreateNextSegment(String text) at Microsoft.OData.Core.UriParser.Parsers.ODataPathParser.ParsePath(ICollection`1 segments) at Microsoft.OData.Core.UriParser.Parsers.ODataPathFactory.BindPath(ICollection`1 segments, ODataUriParserConfiguration configuration) at Microsoft.OData.Core.UriParser.ODataUriParser.ParsePathImplementation() at Microsoft.OData.Core.UriParser.ODataUriParser.Initialize() at Microsoft.OData.Core.UriParser.ODataUriParser.ParsePath() at System.Web.OData.Routing.DefaultODataPathHandler.Parse(IEdmModel model, String serviceRoot, String odataPath, ODataUriResolverSetttings resolverSettings, Boolean enableUriTemplateParsing) in d:\github\WebApi\OData\src\System.Web.OData\OData\Routing\DefaultODataPathHandler.cs:line 137 at System.Web.OData.Routing.DefaultODataPathHandler.Parse(IEdmModel model, String serviceRoot, String odataPath) in d:\github\WebApi\OData\src\System.Web.OData\OData\Routing\DefaultODataPathHandler.cs:line 55 at System.Web.OData.Routing.ODataPathRouteConstraint.Match(HttpRequestMessage request, IHttpRoute route, String parameterName, IDictionary`2 values, HttpRouteDirection routeDirection) in d:\github\WebApi\OData\src\System.Web.OData\OData\Routing\ODataPathRouteConstraint.cs:line 171 ``` I am using Web API 2.2 for OData V4. Thanks.
non_defect
uri parser can t work for short data type edm for function parameter supposed i have a function as xml then i call this function as c odata customers default shortfunction number it fails as c expression of type edm cannot be converted to type edm the call stack is c at microsoft odata core uriparser parsers metadatabindingutils converttotypeifneeded singlevaluenode source iedmtypereference targettypereference at microsoft odata core uriparser parsers functioncallbinder bindsegmentparameters odatauriparserconfiguration configuration iedmoperation functionoropertion icollection segmentparametertokens at microsoft odata core uriparser parsers odatapathparser trybindingparametersandmatchingoperation string identifier string parenthesisexpression iedmtype bindingtype odatauriparserconfiguration configuration icollection boundparameters iedmoperation matchingoperation at microsoft odata core uriparser parsers odatapathparser trycreatesegmentforoperation odatapathsegment previoussegment string identifier string parenthesisexpression at microsoft odata core uriparser parsers odatapathparser createnextsegment string text at microsoft odata core uriparser parsers odatapathparser parsepath icollection segments at microsoft odata core uriparser parsers odatapathfactory bindpath icollection segments odatauriparserconfiguration configuration at microsoft odata core uriparser odatauriparser parsepathimplementation at microsoft odata core uriparser odatauriparser initialize at microsoft odata core uriparser odatauriparser parsepath at system web odata routing defaultodatapathhandler parse iedmmodel model string serviceroot string odatapath odatauriresolversetttings resolversettings boolean enableuritemplateparsing in d github webapi odata src system web odata odata routing defaultodatapathhandler cs line at system web odata routing defaultodatapathhandler parse iedmmodel model string serviceroot string odatapath in d github webapi odata src system web odata odata routing defaultodatapathhandler cs line at system web odata routing odatapathrouteconstraint match httprequestmessage request ihttproute route string parametername idictionary values httproutedirection routedirection in d github webapi odata src system web odata odata routing odatapathrouteconstraint cs line i am using web api for odata thanks
0
282,853
24,500,092,942
IssuesEvent
2022-10-10 12:06:22
wpfoodmanager/wp-food-manager
https://api.github.com/repos/wpfoodmanager/wp-food-manager
closed
Reset button need add on the dashboard page
In Testing
Reset button need add on the dashboard page. ![image](https://user-images.githubusercontent.com/75515088/194309428-f60f9bd4-f279-4551-b4f3-ecb318b8dcb4.png)
1.0
Reset button need add on the dashboard page - Reset button need add on the dashboard page. ![image](https://user-images.githubusercontent.com/75515088/194309428-f60f9bd4-f279-4551-b4f3-ecb318b8dcb4.png)
non_defect
reset button need add on the dashboard page reset button need add on the dashboard page
0
80,348
3,561,016,322
IssuesEvent
2016-01-23 14:07:42
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
closed
Missing constants from cv2 python interface
affected: 2.4 auto-transferred bug category: python bindings priority: normal
Transferred from http://code.opencv.org/issues/3181 ``` || Kyle Schmitt on 2013-07-28 17:16 || Priority: Normal || Affected: 2.4.0 - 2.4.6 || Category: python bindings || Tracker: Bug || Difficulty: || PR: || Platform: x64 / Linux ``` Missing constants from cv2 python interface ----------- ``` The video related constants (and possibly more), are missing from the cv2 python interface. According to the docs: http://docs.opencv.org/modules/highgui/doc/reading_and_writing_images_and_video.html#videocapture-get The capture properties such as CV_CAP_PROP_FRAME_WIDTH should be available in the cv2 module, but they are not. They are only available through cv, or cv2.cv. A snippet from a bpython session: >>> import cv2 >>> import cv >>> cv2.VideoCapture(0).get(cv2.CV_CAP_PROP_FRAME_WIDTH) Traceback (most recent call last): File "<input>", line 1, in <module> AttributeError: 'module' object has no attribute 'CV_CAP_PROP_FRAME_WIDTH' >>> cv2.VideoCapture(0).get(cv2.cv.CV_CAP_PROP_FRAME_WIDTH) 176.0 >>> cv2.VideoCapture(0).get(cv.CV_CAP_PROP_FRAME_WIDTH) 176.0 >>> ``` History ------- ##### Victor Kocheganov on 2013-08-05 10:51 ``` Hello Kyle Schmitt! Thank you for reporting the issue and detail description! Unfortunately our human resources are highly limited and it would be much appreciated if you have time to investigate it by yourself and provide a fix to community (please see http://www.code.opencv.org/projects/opencv/wiki/How_to_contribute for details)! Thank you in advance, Victor Kocheganov - Target version set to 2.4.7 - Assignee set to Vadim Pisarevsky - Status changed from New to Open - Category set to python bindings ``` ##### Victor Kocheganov on 2013-08-08 06:51 ``` - Assignee changed from Vadim Pisarevsky to Andrey Pavlenko ``` ##### abid rahman on 2013-09-24 02:20 ``` It is already available in master branch. <pre> cv2.CAP_PROP_FRAME_COUNT </pre> ``` ##### Alexander Smorkalov on 2013-11-27 10:28 ``` Andrey Pavlenko, you've made some fixes related to wrappers generation for VideoCapture. Is this problem exist in 2.4 right now? - Affected version changed from 2.4.0 - 2.4.5 to 2.4.0 - 2.4.6 ``` ##### Alexander Smorkalov on 2013-11-28 05:57 ``` - Target version changed from 2.4.7 to 2.4.8 ``` ##### Alexander Smorkalov on 2013-12-30 10:38 ``` - Target version changed from 2.4.8 to 2.4.9 ``` ##### Alexander Smorkalov on 2014-04-30 19:05 ``` - Target version changed from 2.4.9 to 2.4.10 ```
1.0
Missing constants from cv2 python interface - Transferred from http://code.opencv.org/issues/3181 ``` || Kyle Schmitt on 2013-07-28 17:16 || Priority: Normal || Affected: 2.4.0 - 2.4.6 || Category: python bindings || Tracker: Bug || Difficulty: || PR: || Platform: x64 / Linux ``` Missing constants from cv2 python interface ----------- ``` The video related constants (and possibly more), are missing from the cv2 python interface. According to the docs: http://docs.opencv.org/modules/highgui/doc/reading_and_writing_images_and_video.html#videocapture-get The capture properties such as CV_CAP_PROP_FRAME_WIDTH should be available in the cv2 module, but they are not. They are only available through cv, or cv2.cv. A snippet from a bpython session: >>> import cv2 >>> import cv >>> cv2.VideoCapture(0).get(cv2.CV_CAP_PROP_FRAME_WIDTH) Traceback (most recent call last): File "<input>", line 1, in <module> AttributeError: 'module' object has no attribute 'CV_CAP_PROP_FRAME_WIDTH' >>> cv2.VideoCapture(0).get(cv2.cv.CV_CAP_PROP_FRAME_WIDTH) 176.0 >>> cv2.VideoCapture(0).get(cv.CV_CAP_PROP_FRAME_WIDTH) 176.0 >>> ``` History ------- ##### Victor Kocheganov on 2013-08-05 10:51 ``` Hello Kyle Schmitt! Thank you for reporting the issue and detail description! Unfortunately our human resources are highly limited and it would be much appreciated if you have time to investigate it by yourself and provide a fix to community (please see http://www.code.opencv.org/projects/opencv/wiki/How_to_contribute for details)! Thank you in advance, Victor Kocheganov - Target version set to 2.4.7 - Assignee set to Vadim Pisarevsky - Status changed from New to Open - Category set to python bindings ``` ##### Victor Kocheganov on 2013-08-08 06:51 ``` - Assignee changed from Vadim Pisarevsky to Andrey Pavlenko ``` ##### abid rahman on 2013-09-24 02:20 ``` It is already available in master branch. <pre> cv2.CAP_PROP_FRAME_COUNT </pre> ``` ##### Alexander Smorkalov on 2013-11-27 10:28 ``` Andrey Pavlenko, you've made some fixes related to wrappers generation for VideoCapture. Is this problem exist in 2.4 right now? - Affected version changed from 2.4.0 - 2.4.5 to 2.4.0 - 2.4.6 ``` ##### Alexander Smorkalov on 2013-11-28 05:57 ``` - Target version changed from 2.4.7 to 2.4.8 ``` ##### Alexander Smorkalov on 2013-12-30 10:38 ``` - Target version changed from 2.4.8 to 2.4.9 ``` ##### Alexander Smorkalov on 2014-04-30 19:05 ``` - Target version changed from 2.4.9 to 2.4.10 ```
non_defect
missing constants from python interface transferred from kyle schmitt on priority normal affected category python bindings tracker bug difficulty pr platform linux missing constants from python interface the video related constants and possibly more are missing from the python interface according to the docs the capture properties such as cv cap prop frame width should be available in the module but they are not they are only available through cv or cv a snippet from a bpython session import import cv videocapture get cv cap prop frame width traceback most recent call last file line in attributeerror module object has no attribute cv cap prop frame width videocapture get cv cv cap prop frame width videocapture get cv cv cap prop frame width history victor kocheganov on hello kyle schmitt thank you for reporting the issue and detail description unfortunately our human resources are highly limited and it would be much appreciated if you have time to investigate it by yourself and provide a fix to community please see for details thank you in advance victor kocheganov target version set to assignee set to vadim pisarevsky status changed from new to open category set to python bindings victor kocheganov on assignee changed from vadim pisarevsky to andrey pavlenko abid rahman on it is already available in master branch cap prop frame count alexander smorkalov on andrey pavlenko you ve made some fixes related to wrappers generation for videocapture is this problem exist in right now affected version changed from to alexander smorkalov on target version changed from to alexander smorkalov on target version changed from to alexander smorkalov on target version changed from to
0
45,016
12,520,276,966
IssuesEvent
2020-06-03 15:37:38
ka65359/kai-kong-music-lib
https://api.github.com/repos/ka65359/kai-kong-music-lib
opened
Add song clears filter string
Defect Sev 3
### Description The filter is cleared when a new song is added. We should persist it. #### Steps to reproduce 1. 2.
1.0
Add song clears filter string - ### Description The filter is cleared when a new song is added. We should persist it. #### Steps to reproduce 1. 2.
defect
add song clears filter string description the filter is cleared when a new song is added we should persist it steps to reproduce
1
123,834
12,219,637,787
IssuesEvent
2020-05-01 22:18:22
textileio/js-threads
https://api.github.com/repos/textileio/js-threads
closed
Update to use _id instead of ID in instances
documentation enhancement
**Describe the bug** On the Go side, we've now switched to using _id for the identifier on instances. This is great as it will feel more mongodb-like, and is pretty standard in JS. Since we don't rely on Go to validate our instances here, we'll need to make sure we implement this the same way that we do in Go. This ticket therefore is linked with #57, which is also what we use by default in Go now.
1.0
Update to use _id instead of ID in instances - **Describe the bug** On the Go side, we've now switched to using _id for the identifier on instances. This is great as it will feel more mongodb-like, and is pretty standard in JS. Since we don't rely on Go to validate our instances here, we'll need to make sure we implement this the same way that we do in Go. This ticket therefore is linked with #57, which is also what we use by default in Go now.
non_defect
update to use id instead of id in instances describe the bug on the go side we ve now switched to using id for the identifier on instances this is great as it will feel more mongodb like and is pretty standard in js since we don t rely on go to validate our instances here we ll need to make sure we implement this the same way that we do in go this ticket therefore is linked with which is also what we use by default in go now
0
40,394
2,868,917,014
IssuesEvent
2015-06-05 21:56:52
dart-lang/dart_style
https://api.github.com/repos/dart-lang/dart_style
closed
Mis-formatting of strings
AssumedStale bug Priority-Medium
<a href="https://github.com/stevemessick"><img src="https://avatars.githubusercontent.com/u/8518285?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [stevemessick](https://github.com/stevemessick)** _Originally opened as dart-lang/sdk#18206_ ---- An edge case for the formatter. Format the code in the screenshot. It concatenates both of the strings onto one line, making the line length too long, and also doesn't combine them. It seems to be sensitive to the exact lengths of the interpolated variable names, and I also need to have it inside the if statement. &nbsp;- Alan Knight //////////////////////////////////////////////////////////////////////////////////// Editor: 1.4.0.edge_034945 (2014-04-10) OS: Mac OS X - x86_64 (10.9.2) JVM: 1.6.0_65 # projects: 4 # open dart files: 25 auto-run pub: false localhost resolves to: 127.0.0.1 mem max/total/free: 1983 / 620 / 307 MB thread count: 34 index: 1554905 relationships in 177578 keys in 1445 sources SDK installed: true Dartium installed: true ______ **Attachment:** [screenshot.png](https://storage.googleapis.com/google-code-attachments/dart/issue-18206/comment-0/screenshot.png) (144.83 KB)
1.0
Mis-formatting of strings - <a href="https://github.com/stevemessick"><img src="https://avatars.githubusercontent.com/u/8518285?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [stevemessick](https://github.com/stevemessick)** _Originally opened as dart-lang/sdk#18206_ ---- An edge case for the formatter. Format the code in the screenshot. It concatenates both of the strings onto one line, making the line length too long, and also doesn't combine them. It seems to be sensitive to the exact lengths of the interpolated variable names, and I also need to have it inside the if statement. &nbsp;- Alan Knight //////////////////////////////////////////////////////////////////////////////////// Editor: 1.4.0.edge_034945 (2014-04-10) OS: Mac OS X - x86_64 (10.9.2) JVM: 1.6.0_65 # projects: 4 # open dart files: 25 auto-run pub: false localhost resolves to: 127.0.0.1 mem max/total/free: 1983 / 620 / 307 MB thread count: 34 index: 1554905 relationships in 177578 keys in 1445 sources SDK installed: true Dartium installed: true ______ **Attachment:** [screenshot.png](https://storage.googleapis.com/google-code-attachments/dart/issue-18206/comment-0/screenshot.png) (144.83 KB)
non_defect
mis formatting of strings issue by originally opened as dart lang sdk an edge case for the formatter format the code in the screenshot it concatenates both of the strings onto one line making the line length too long and also doesn t combine them it seems to be sensitive to the exact lengths of the interpolated variable names and i also need to have it inside the if statement nbsp alan knight editor edge os mac os x jvm projects open dart files auto run pub false localhost resolves to mem max total free mb thread count index relationships in keys in sources sdk installed true dartium installed true attachment kb
0
64,064
18,161,462,126
IssuesEvent
2021-09-27 10:05:19
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Beta dot animation breaking fonts subpixel antialiasing on Windows in Chrome
T-Defect S-Minor A-User-Menu A11y O-Occasional
It seems that the animation `mx_Beta_bluePulse` used in class `mx_BetaDot` is making an enormous layout trashing and is a second source of fonts not being correctly subpixel antialiased on Windows. The main source so far is the background-blur on left panel, which is being taken care of here: https://github.com/matrix-org/matrix-react-sdk/pull/6262 To reproduce proper behavior comment out class `.mx_BetaDot` in `res/css/views/beta/_BetaCard.scss` and remove all `backdrop-filter` references from our css codebase. The other way to verify if rendering is proper is to look for layer borders checkbox which is accessible via Chrome Devtools cog menu (Rendering). Screenshots below: Valid fonts rendering (properly antialiased) will have this kind of layer borders: ![CleanShot 2021-08-16 at 11 58 40](https://user-images.githubusercontent.com/3636685/129546467-c1f2c807-173b-4555-a61e-d8e96023c59a.png) Invalid fonts rendering layer borders (it's super easy to just look at the bottom-right corner with icons): ![CleanShot 2021-08-16 at 11 59 40](https://user-images.githubusercontent.com/3636685/129546494-70f4fcfe-752b-492a-8e69-56ab3f82f979.png) The fonts rendering is irreproducible on anywhere except Windows and Chrome/Chromium, yet the layers part should be easily reproducible on any system using Chrome. Related: https://github.com/matrix-org/matrix-react-sdk/pull/6262 (gets rid of background-blur and replaces it with canvas, which doesn't trash the layout) Related: https://github.com/vector-im/element-web/issues/15594
1.0
Beta dot animation breaking fonts subpixel antialiasing on Windows in Chrome - It seems that the animation `mx_Beta_bluePulse` used in class `mx_BetaDot` is making an enormous layout trashing and is a second source of fonts not being correctly subpixel antialiased on Windows. The main source so far is the background-blur on left panel, which is being taken care of here: https://github.com/matrix-org/matrix-react-sdk/pull/6262 To reproduce proper behavior comment out class `.mx_BetaDot` in `res/css/views/beta/_BetaCard.scss` and remove all `backdrop-filter` references from our css codebase. The other way to verify if rendering is proper is to look for layer borders checkbox which is accessible via Chrome Devtools cog menu (Rendering). Screenshots below: Valid fonts rendering (properly antialiased) will have this kind of layer borders: ![CleanShot 2021-08-16 at 11 58 40](https://user-images.githubusercontent.com/3636685/129546467-c1f2c807-173b-4555-a61e-d8e96023c59a.png) Invalid fonts rendering layer borders (it's super easy to just look at the bottom-right corner with icons): ![CleanShot 2021-08-16 at 11 59 40](https://user-images.githubusercontent.com/3636685/129546494-70f4fcfe-752b-492a-8e69-56ab3f82f979.png) The fonts rendering is irreproducible on anywhere except Windows and Chrome/Chromium, yet the layers part should be easily reproducible on any system using Chrome. Related: https://github.com/matrix-org/matrix-react-sdk/pull/6262 (gets rid of background-blur and replaces it with canvas, which doesn't trash the layout) Related: https://github.com/vector-im/element-web/issues/15594
defect
beta dot animation breaking fonts subpixel antialiasing on windows in chrome it seems that the animation mx beta bluepulse used in class mx betadot is making an enormous layout trashing and is a second source of fonts not being correctly subpixel antialiased on windows the main source so far is the background blur on left panel which is being taken care of here to reproduce proper behavior comment out class mx betadot in res css views beta betacard scss and remove all backdrop filter references from our css codebase the other way to verify if rendering is proper is to look for layer borders checkbox which is accessible via chrome devtools cog menu rendering screenshots below valid fonts rendering properly antialiased will have this kind of layer borders invalid fonts rendering layer borders it s super easy to just look at the bottom right corner with icons the fonts rendering is irreproducible on anywhere except windows and chrome chromium yet the layers part should be easily reproducible on any system using chrome related gets rid of background blur and replaces it with canvas which doesn t trash the layout related
1
752,256
26,278,065,913
IssuesEvent
2023-01-07 01:53:23
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
reopened
[Talent] Shadowy Apparitions
Class: Priest Pathfinding Talent Priority: Medium Status: Needs Confirmation Bug Report from Discord
Mitosis OP — 12/11/2022 10:31 AM Pathing is janky as hell, and proc rate is significantly lower than what it should be. In particular while moving you do not seem to get any bonus chance to summon where it should be 5x as likely
1.0
[Talent] Shadowy Apparitions - Mitosis OP — 12/11/2022 10:31 AM Pathing is janky as hell, and proc rate is significantly lower than what it should be. In particular while moving you do not seem to get any bonus chance to summon where it should be 5x as likely
non_defect
shadowy apparitions mitosis op — am pathing is janky as hell and proc rate is significantly lower than what it should be in particular while moving you do not seem to get any bonus chance to summon where it should be as likely
0
26
2,492,544,954
IssuesEvent
2015-01-05 00:55:37
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
Redis class not exists
Defect
Cake 2.5.8 - if `Redis` class doesn't exists, this error will be thrown: ```php PHP Fatal error: Uncaught exception 'CacheException' with message 'Cache engine _cake_core_ is not properly configured.' in /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/Cache/Cache.php:181 Stack trace: #0 /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/Cache/Cache.php(151): Cache::_buildEngine('_cake_core_') #1 /var/www/cake258/App/Config/core.php(282): Cache::config('_cake_core_', Array) #2 /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/Core/Configure.php(72): include('/var/www/cake258/...') #3 /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/bootstrap.php(175): Configure::bootstrap(true) #4 /var/www/cake258/App/webroot/index.php(85): include('/var/www/cake258/...') #5 {main} thrown in /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/Cache/Cache.php on line 181 ``` Should this part be tested a bit before? https://github.com/cakephp/cakephp/blob/master/lib/Cake/Cache/Engine/RedisEngine.php#L57
1.0
Redis class not exists - Cake 2.5.8 - if `Redis` class doesn't exists, this error will be thrown: ```php PHP Fatal error: Uncaught exception 'CacheException' with message 'Cache engine _cake_core_ is not properly configured.' in /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/Cache/Cache.php:181 Stack trace: #0 /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/Cache/Cache.php(151): Cache::_buildEngine('_cake_core_') #1 /var/www/cake258/App/Config/core.php(282): Cache::config('_cake_core_', Array) #2 /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/Core/Configure.php(72): include('/var/www/cake258/...') #3 /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/bootstrap.php(175): Configure::bootstrap(true) #4 /var/www/cake258/App/webroot/index.php(85): include('/var/www/cake258/...') #5 {main} thrown in /var/www/cake258/Vendor/cakephp/cakephp/lib/Cake/Cache/Cache.php on line 181 ``` Should this part be tested a bit before? https://github.com/cakephp/cakephp/blob/master/lib/Cake/Cache/Engine/RedisEngine.php#L57
defect
redis class not exists cake if redis class doesn t exists this error will be thrown php php fatal error uncaught exception cacheexception with message cache engine cake core is not properly configured in var www vendor cakephp cakephp lib cake cache cache php stack trace var www vendor cakephp cakephp lib cake cache cache php cache buildengine cake core var www app config core php cache config cake core array var www vendor cakephp cakephp lib cake core configure php include var www var www vendor cakephp cakephp lib cake bootstrap php configure bootstrap true var www app webroot index php include var www main thrown in var www vendor cakephp cakephp lib cake cache cache php on line should this part be tested a bit before
1
41,041
10,271,844,454
IssuesEvent
2019-08-23 15:00:36
openanthem/nimbus-core
https://api.github.com/repos/openanthem/nimbus-core
closed
Data entered using Chrome's autofill feature is not getting saved when submitting the form.
Defect Open
# Issue Details **Type of Issue** (check one with "X") ``` [X] Bug Report => Please search GitHub for a similar issue or PR before submitting [ ] Feature Request => Please ensure feature is not already in progress [ ] Support Request => Please do not submit support requests here, instead see: https://discourse.oss.antheminc.com/ ``` ## Current Behavior <!-- When submitting the form using autofill feature of chrome, data is not getting persisted in the database. --> ## Expected Behavior <!-- Data filled in using the autofill should also be persisted in the database. --> ## How to Reproduce the Issue ### Steps to Reproduce <!-- 1) Fill some fields using the autofill feature, for eg: address so city, zip and state are prepopulated. 2) Hit submit button 3) Data will not be persisted for city, zip and state (data that was prepopulated) in the DB --> ### Code Snippet <!-- Please add any code that is necessary to reproduce the issue. If the current behavior is recreated better with an example, please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via https://plnkr.co or similar. --> # Environment Details * **Nimbus Version:** <!-- 1.3.x --> * **Browser:** <!-- Please list all browsers where this could be reproduced. -->
1.0
Data entered using Chrome's autofill feature is not getting saved when submitting the form. - # Issue Details **Type of Issue** (check one with "X") ``` [X] Bug Report => Please search GitHub for a similar issue or PR before submitting [ ] Feature Request => Please ensure feature is not already in progress [ ] Support Request => Please do not submit support requests here, instead see: https://discourse.oss.antheminc.com/ ``` ## Current Behavior <!-- When submitting the form using autofill feature of chrome, data is not getting persisted in the database. --> ## Expected Behavior <!-- Data filled in using the autofill should also be persisted in the database. --> ## How to Reproduce the Issue ### Steps to Reproduce <!-- 1) Fill some fields using the autofill feature, for eg: address so city, zip and state are prepopulated. 2) Hit submit button 3) Data will not be persisted for city, zip and state (data that was prepopulated) in the DB --> ### Code Snippet <!-- Please add any code that is necessary to reproduce the issue. If the current behavior is recreated better with an example, please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via https://plnkr.co or similar. --> # Environment Details * **Nimbus Version:** <!-- 1.3.x --> * **Browser:** <!-- Please list all browsers where this could be reproduced. -->
defect
data entered using chrome s autofill feature is not getting saved when submitting the form issue details type of issue check one with x bug report please search github for a similar issue or pr before submitting feature request please ensure feature is not already in progress support request please do not submit support requests here instead see current behavior when submitting the form using autofill feature of chrome data is not getting persisted in the database expected behavior data filled in using the autofill should also be persisted in the database how to reproduce the issue steps to reproduce fill some fields using the autofill feature for eg address so city zip and state are prepopulated hit submit button data will not be persisted for city zip and state data that was prepopulated in the db code snippet please add any code that is necessary to reproduce the issue if the current behavior is recreated better with an example please provide the steps to reproduce and if possible a minimal demo of the problem via or similar environment details nimbus version x browser please list all browsers where this could be reproduced
1
46,782
13,055,975,343
IssuesEvent
2020-07-30 03:16:55
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[DOMLauncher] Emulate SLC bit-packing (Trac #1864)
Incomplete Migration Migrated from Trac combo simulation defect
Migrated from https://code.icecube.wisc.edu/ticket/1864 ```json { "status": "closed", "changetime": "2019-02-13T14:13:24", "description": "In real data the SLC charge stamps are packed into 9 bits, dropping the LSB from each if bit 10 in the peak sample is high. DOMLauncher should do the same. The payload format is documented here:\n\nhttps://docushare.icecube.wisc.edu/dsweb/Get/Document-20568\n\nSee also: #1863", "reporter": "jvansanten", "cc": "", "resolution": "fixed", "_ts": "1550067204154158", "component": "combo simulation", "summary": "[DOMLauncher] Emulate SLC bit-packing", "priority": "normal", "keywords": "", "time": "2016-09-30T14:58:12", "milestone": "", "owner": "cweaver", "type": "defect" } ```
1.0
[DOMLauncher] Emulate SLC bit-packing (Trac #1864) - Migrated from https://code.icecube.wisc.edu/ticket/1864 ```json { "status": "closed", "changetime": "2019-02-13T14:13:24", "description": "In real data the SLC charge stamps are packed into 9 bits, dropping the LSB from each if bit 10 in the peak sample is high. DOMLauncher should do the same. The payload format is documented here:\n\nhttps://docushare.icecube.wisc.edu/dsweb/Get/Document-20568\n\nSee also: #1863", "reporter": "jvansanten", "cc": "", "resolution": "fixed", "_ts": "1550067204154158", "component": "combo simulation", "summary": "[DOMLauncher] Emulate SLC bit-packing", "priority": "normal", "keywords": "", "time": "2016-09-30T14:58:12", "milestone": "", "owner": "cweaver", "type": "defect" } ```
defect
emulate slc bit packing trac migrated from json status closed changetime description in real data the slc charge stamps are packed into bits dropping the lsb from each if bit in the peak sample is high domlauncher should do the same the payload format is documented here n n also reporter jvansanten cc resolution fixed ts component combo simulation summary emulate slc bit packing priority normal keywords time milestone owner cweaver type defect
1
17,500
3,010,324,598
IssuesEvent
2015-07-28 12:37:00
oozcitak/imagelistview
https://api.github.com/repos/oozcitak/imagelistview
closed
Please move to Github
auto-migrated Priority-Medium Type-Defect
``` To save the project from disappearing when Google Code will be shut down, I suggest that you create a GitHub repository by clicking the "Export to GitHub" button. I could do quite well by myself, but I feel that you as the original author should do that :-) ``` Original issue reported on code.google.com by `uwe.k...@gmail.com` on 16 Jul 2015 at 7:29
1.0
Please move to Github - ``` To save the project from disappearing when Google Code will be shut down, I suggest that you create a GitHub repository by clicking the "Export to GitHub" button. I could do quite well by myself, but I feel that you as the original author should do that :-) ``` Original issue reported on code.google.com by `uwe.k...@gmail.com` on 16 Jul 2015 at 7:29
defect
please move to github to save the project from disappearing when google code will be shut down i suggest that you create a github repository by clicking the export to github button i could do quite well by myself but i feel that you as the original author should do that original issue reported on code google com by uwe k gmail com on jul at
1
115,233
24,736,344,913
IssuesEvent
2022-10-20 22:24:14
bnreplah/verademo
https://api.github.com/repos/bnreplah/verademo
opened
Improper Neutralization of Script-Related HTML Tags in a Web Page (Basic XSS) [VID:80:WEB-INF/views/profile.jsp:160]
VeracodeFlaw: Medium Veracode Pipeline Scan
**Filename:** WEB-INF/views/profile.jsp **Line:** 160 **CWE:** 80 (Improper Neutralization of Script-Related HTML Tags in a Web Page (Basic XSS)) <span>This call to javax.servlet.jsp.JspWriter.print() contains a cross-site scripting (XSS) flaw. The application populates the HTTP response with untrusted input, allowing an attacker to embed malicious content, such as Javascript code, which will be executed in the context of the victim's browser. XSS vulnerabilities are commonly exploited to steal or manipulate cookies, modify presentation of content, and compromise confidential information, with new attack vectors being discovered on a regular basis. The first argument to print() contains tainted data from the variable heckler.getUsername(). The tainted data originated from an earlier call to java.sql.PreparedStatement.executeQuery. The tainted data is directed into an output stream returned by javax.servlet.jsp.JspWriter.</span> <span>Use contextual escaping on all untrusted data before using it to construct any portion of an HTTP response. The escaping method should be chosen based on the specific use case of the untrusted data, otherwise it may not protect fully against the attack. For example, if the data is being written to the body of an HTML page, use HTML entity escaping; if the data is being written to an attribute, use attribute escaping; etc. Both the OWASP Java Encoder library and the Microsoft AntiXSS library provide contextual escaping methods. For more details on contextual escaping, see https://github.com/OWASP/CheatSheetSeries/blob/master/cheatsheets/Cross_Site_Scripting_Prevention_Cheat_Sheet.md. In addition, as a best practice, always validate untrusted input to ensure that it conforms to the expected format, using centralized data validation routines when possible.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/79.html">CWE</a> <a href="https://owasp.org/www-community/attacks/xss/">OWASP</a> <a href="https://docs.veracode.com/r/review_cleansers">Supported Cleansers</a></span>
2.0
Improper Neutralization of Script-Related HTML Tags in a Web Page (Basic XSS) [VID:80:WEB-INF/views/profile.jsp:160] - **Filename:** WEB-INF/views/profile.jsp **Line:** 160 **CWE:** 80 (Improper Neutralization of Script-Related HTML Tags in a Web Page (Basic XSS)) <span>This call to javax.servlet.jsp.JspWriter.print() contains a cross-site scripting (XSS) flaw. The application populates the HTTP response with untrusted input, allowing an attacker to embed malicious content, such as Javascript code, which will be executed in the context of the victim's browser. XSS vulnerabilities are commonly exploited to steal or manipulate cookies, modify presentation of content, and compromise confidential information, with new attack vectors being discovered on a regular basis. The first argument to print() contains tainted data from the variable heckler.getUsername(). The tainted data originated from an earlier call to java.sql.PreparedStatement.executeQuery. The tainted data is directed into an output stream returned by javax.servlet.jsp.JspWriter.</span> <span>Use contextual escaping on all untrusted data before using it to construct any portion of an HTTP response. The escaping method should be chosen based on the specific use case of the untrusted data, otherwise it may not protect fully against the attack. For example, if the data is being written to the body of an HTML page, use HTML entity escaping; if the data is being written to an attribute, use attribute escaping; etc. Both the OWASP Java Encoder library and the Microsoft AntiXSS library provide contextual escaping methods. For more details on contextual escaping, see https://github.com/OWASP/CheatSheetSeries/blob/master/cheatsheets/Cross_Site_Scripting_Prevention_Cheat_Sheet.md. In addition, as a best practice, always validate untrusted input to ensure that it conforms to the expected format, using centralized data validation routines when possible.</span> <span>References: <a href="https://cwe.mitre.org/data/definitions/79.html">CWE</a> <a href="https://owasp.org/www-community/attacks/xss/">OWASP</a> <a href="https://docs.veracode.com/r/review_cleansers">Supported Cleansers</a></span>
non_defect
improper neutralization of script related html tags in a web page basic xss filename web inf views profile jsp line cwe improper neutralization of script related html tags in a web page basic xss this call to javax servlet jsp jspwriter print contains a cross site scripting xss flaw the application populates the http response with untrusted input allowing an attacker to embed malicious content such as javascript code which will be executed in the context of the victim s browser xss vulnerabilities are commonly exploited to steal or manipulate cookies modify presentation of content and compromise confidential information with new attack vectors being discovered on a regular basis the first argument to print contains tainted data from the variable heckler getusername the tainted data originated from an earlier call to java sql preparedstatement executequery the tainted data is directed into an output stream returned by javax servlet jsp jspwriter use contextual escaping on all untrusted data before using it to construct any portion of an http response the escaping method should be chosen based on the specific use case of the untrusted data otherwise it may not protect fully against the attack for example if the data is being written to the body of an html page use html entity escaping if the data is being written to an attribute use attribute escaping etc both the owasp java encoder library and the microsoft antixss library provide contextual escaping methods for more details on contextual escaping see in addition as a best practice always validate untrusted input to ensure that it conforms to the expected format using centralized data validation routines when possible references
0
144,246
11,599,246,911
IssuesEvent
2020-02-25 01:36:22
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Manual test run on OS X for 1.4.x - Release
OS/macOS QA/Yes release-notes/exclude tests
## Per release specialty tests ### Installer - [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window ### Data(Upgrade from previous release) - [ ] Make sure that data from the last version appears in the new version OK - [ ] With data from the last version, verify that - [ ] bookmarks on the bookmark toolbar and bookmark folders can be opened - [ ] cookies are preserved - [ ] installed extensions are retained and work correctly - [ ] opened tabs can be reloaded - [ ] stored passwords are preserved - [ ] unpinned tabs can be pinned ## Extensions/Plugins tests - [x] Verify one item from Brave Update server is installable (Example: Ad-block DAT file on fresh extension) - [x] Verify one item from Google Update server is installable (Example: Extensions from CWS) - [x] Verify PDFJS, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled - [x] Verify magnet links and .torrent files loads Torrent viewer page and able to download torrent ### CWS - [x] Verify installing ABP from CWS shows warning message `NOT A RECOMMENDED BRAVE EXTENSION!` but still allows to install the extension - [x] Verify installing LastPass from CWS doesn't show any warning message ### PDF - [x] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf - [x] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf ### Widevine - [x] Verify `Widevine Notification` is shown when you visit Netflix for the first time - [x] Test that you can stream on Netflix on a fresh profile after installing Widevine ### Bravery settings - [x] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/ - [x] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/ - [x] Verify that toggling `Ads and trackers blocked` works as expected - [x] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items - [x] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly - [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked ### Fingerprint Tests - [x] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00` - [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on - [x] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/ - [x] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on ### Rewards - [x] Verify wallet is auto created after enabling rewards - [ ] Verify account balance shows correct BAT and USD value - [ ] Verify you are able to restore a wallet - [x] Verify wallet address matches the QR code that is generated under `Add funds` - [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel - [ ] Verify adding funds via any of the currencies flows into wallet after specified amount of time - [ ] Verify adding funds to an existing wallet with amount, adjusts the BAT value appropriately - [ ] Verify monthly budget shows correct BAT and USD value - [ ] Verify you are able to exclude a publisher from the auto-contribute table by clicking on the `x` in auto-contribute table and popup list of sites - [ ] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel - [ ] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %) - [ ] Verify when you click on `Send a tip`, the custom tip banner displays - [ ] Verify you are able to make one-time tip and they display in tips panel - [ ] Verify you are able to make recurring tip and they display in tips panel - [ ] Verify you can tip a verified publisher - [ ] Verify you can tip a verified YouTube creator - [ ] Verify tip panel shows a verified checkmark for a verified publisher/verified YouTube creator - [ ] Verify tip panel shows a message about unverified publisher - [ ] Verify BR panel shows message about an unverified publisher - [ ] Verify you are able to perform a contribution - [ ] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators - [ ] Verify that disabling Rewards and enabling it again does not lose state - [ ] Verify that disabling auto-contribute and enabling it again does not lose state - [ ] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings - [ ] Upgrade from older version - [ ] Verify the wallet balance is retained and wallet backup code isn't corrupted - [ ] Verify auto-contribute list is not lost after upgrade - [ ] Verify tips list is not lost after upgrade - [ ] Verify wallet panel transactions list is not lost after upgrade ### Ads Upgrade Tests: - [x] Install 0.62.51 and enable Rewards (Ads are not available on this version). Update on `test` channel to the hotfix version. Verify Ads are off by default, should get a BAT logo notification to alert you that Ads are available. - [x] Install 0.64.77 and enable Rewards. Ads are on by default. View an Ad. Update on `test` channel to the hotfix version. Verify Ads are still on after update, Ads panel information was not lost after upgrade, no BAT logo notification. - [x] Install 0.64.77 and enable Rewards. Disable Ads. Update on `test` channel to the hotfix version. Verify Ads are still off after update, no BAT logo notification. - [x] Install 1.3.118 and enable Rewards. Ads are on by default. View an ad. Update on `test` channel to the hotfix version. Verify Ads are still on after update, Ads panel information was not lost after upgrade, no BAT logo notification. - [x] install 1.3.118 and enable Rewards. Disable Ads. Update on `test` channel to the hotfix version. Verify Ads are still off after update, no BAT logo notification. ### Tor Tabs - [x] Visit https://check.torproject.org in a Tor window, ensure its shows success message for using a Tor exit node - [x] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads - [x] Visit https://protonirockerxow.onion/ in a Tor window, ensure login page is shown - [x] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown ### Session storage - [x] Temporarily move away your browser profile and test that a new profile is created when browser is launched - macOS - `~/Library/Application\ Support/BraveSoftware/` - Windows - `%userprofile%\appdata\Local\BraveSoftware\` - Linux(Ubuntu) - `~/.config/BraveSoftware/` - [x] Test that windows and tabs restore when closed, including active tab - [x] Ensure that the tabs in the above session are being lazy loaded when the session is restored ## Update tests - [x] Verify visiting `brave://settings/help` triggers update check - [x] Verify once update is downloaded, prompts to `Relaunch` to install update ## Chromium upgrade tests - [x] Verify `brave://gpu` on Brave and `chrome://gpu` on Chrome are similar for the same Chromium version on both browsers #### Adblock - [x] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/` #### Components - [x] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components ## Crypto Wallets - [x] ensure that you can create a new wallet without any issues - [x] ensure that you can restore a previous CW wallet without any issues - [x] ensure that you can restore a previous MM wallet without any issues - [x] ensure that you can create a transaction (sending crypto) with a CW wallet - [x] ensure that you can create a transaction (sending crypto) using a restored MM wallet
1.0
Manual test run on OS X for 1.4.x - Release - ## Per release specialty tests ### Installer - [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window ### Data(Upgrade from previous release) - [ ] Make sure that data from the last version appears in the new version OK - [ ] With data from the last version, verify that - [ ] bookmarks on the bookmark toolbar and bookmark folders can be opened - [ ] cookies are preserved - [ ] installed extensions are retained and work correctly - [ ] opened tabs can be reloaded - [ ] stored passwords are preserved - [ ] unpinned tabs can be pinned ## Extensions/Plugins tests - [x] Verify one item from Brave Update server is installable (Example: Ad-block DAT file on fresh extension) - [x] Verify one item from Google Update server is installable (Example: Extensions from CWS) - [x] Verify PDFJS, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled - [x] Verify magnet links and .torrent files loads Torrent viewer page and able to download torrent ### CWS - [x] Verify installing ABP from CWS shows warning message `NOT A RECOMMENDED BRAVE EXTENSION!` but still allows to install the extension - [x] Verify installing LastPass from CWS doesn't show any warning message ### PDF - [x] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf - [x] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf ### Widevine - [x] Verify `Widevine Notification` is shown when you visit Netflix for the first time - [x] Test that you can stream on Netflix on a fresh profile after installing Widevine ### Bravery settings - [x] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/ - [x] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/ - [x] Verify that toggling `Ads and trackers blocked` works as expected - [x] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items - [x] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly - [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked ### Fingerprint Tests - [x] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00` - [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on - [x] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/ - [x] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on ### Rewards - [x] Verify wallet is auto created after enabling rewards - [ ] Verify account balance shows correct BAT and USD value - [ ] Verify you are able to restore a wallet - [x] Verify wallet address matches the QR code that is generated under `Add funds` - [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel - [ ] Verify adding funds via any of the currencies flows into wallet after specified amount of time - [ ] Verify adding funds to an existing wallet with amount, adjusts the BAT value appropriately - [ ] Verify monthly budget shows correct BAT and USD value - [ ] Verify you are able to exclude a publisher from the auto-contribute table by clicking on the `x` in auto-contribute table and popup list of sites - [ ] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel - [ ] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %) - [ ] Verify when you click on `Send a tip`, the custom tip banner displays - [ ] Verify you are able to make one-time tip and they display in tips panel - [ ] Verify you are able to make recurring tip and they display in tips panel - [ ] Verify you can tip a verified publisher - [ ] Verify you can tip a verified YouTube creator - [ ] Verify tip panel shows a verified checkmark for a verified publisher/verified YouTube creator - [ ] Verify tip panel shows a message about unverified publisher - [ ] Verify BR panel shows message about an unverified publisher - [ ] Verify you are able to perform a contribution - [ ] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators - [ ] Verify that disabling Rewards and enabling it again does not lose state - [ ] Verify that disabling auto-contribute and enabling it again does not lose state - [ ] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings - [ ] Upgrade from older version - [ ] Verify the wallet balance is retained and wallet backup code isn't corrupted - [ ] Verify auto-contribute list is not lost after upgrade - [ ] Verify tips list is not lost after upgrade - [ ] Verify wallet panel transactions list is not lost after upgrade ### Ads Upgrade Tests: - [x] Install 0.62.51 and enable Rewards (Ads are not available on this version). Update on `test` channel to the hotfix version. Verify Ads are off by default, should get a BAT logo notification to alert you that Ads are available. - [x] Install 0.64.77 and enable Rewards. Ads are on by default. View an Ad. Update on `test` channel to the hotfix version. Verify Ads are still on after update, Ads panel information was not lost after upgrade, no BAT logo notification. - [x] Install 0.64.77 and enable Rewards. Disable Ads. Update on `test` channel to the hotfix version. Verify Ads are still off after update, no BAT logo notification. - [x] Install 1.3.118 and enable Rewards. Ads are on by default. View an ad. Update on `test` channel to the hotfix version. Verify Ads are still on after update, Ads panel information was not lost after upgrade, no BAT logo notification. - [x] install 1.3.118 and enable Rewards. Disable Ads. Update on `test` channel to the hotfix version. Verify Ads are still off after update, no BAT logo notification. ### Tor Tabs - [x] Visit https://check.torproject.org in a Tor window, ensure its shows success message for using a Tor exit node - [x] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads - [x] Visit https://protonirockerxow.onion/ in a Tor window, ensure login page is shown - [x] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown ### Session storage - [x] Temporarily move away your browser profile and test that a new profile is created when browser is launched - macOS - `~/Library/Application\ Support/BraveSoftware/` - Windows - `%userprofile%\appdata\Local\BraveSoftware\` - Linux(Ubuntu) - `~/.config/BraveSoftware/` - [x] Test that windows and tabs restore when closed, including active tab - [x] Ensure that the tabs in the above session are being lazy loaded when the session is restored ## Update tests - [x] Verify visiting `brave://settings/help` triggers update check - [x] Verify once update is downloaded, prompts to `Relaunch` to install update ## Chromium upgrade tests - [x] Verify `brave://gpu` on Brave and `chrome://gpu` on Chrome are similar for the same Chromium version on both browsers #### Adblock - [x] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/` #### Components - [x] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components ## Crypto Wallets - [x] ensure that you can create a new wallet without any issues - [x] ensure that you can restore a previous CW wallet without any issues - [x] ensure that you can restore a previous MM wallet without any issues - [x] ensure that you can create a transaction (sending crypto) with a CW wallet - [x] ensure that you can create a transaction (sending crypto) using a restored MM wallet
non_defect
manual test run on os x for x release per release specialty tests installer check signature if os run spctl assess verbose applications brave browser beta app and make sure it returns accepted if windows right click on the brave installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window data upgrade from previous release make sure that data from the last version appears in the new version ok with data from the last version verify that bookmarks on the bookmark toolbar and bookmark folders can be opened cookies are preserved installed extensions are retained and work correctly opened tabs can be reloaded stored passwords are preserved unpinned tabs can be pinned extensions plugins tests verify one item from brave update server is installable example ad block dat file on fresh extension verify one item from google update server is installable example extensions from cws verify pdfjs torrent viewer extensions are installed automatically on fresh profile and cannot be disabled verify magnet links and torrent files loads torrent viewer page and able to download torrent cws verify installing abp from cws shows warning message not a recommended brave extension but still allows to install the extension verify installing lastpass from cws doesn t show any warning message pdf test that pdf is loaded over https at test that pdf is loaded over http at widevine verify widevine notification is shown when you visit netflix for the first time test that you can stream on netflix on a fresh profile after installing widevine bravery settings verify that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to verify that toggling ads and trackers blocked works as expected visit verify that safe browsing via our proxy works for all the listed items visit and then turn on script blocking page should not load allow it from the script blocking ui in the url bar and it should load the page correctly test that party storage results are blank at when party cookies are blocked and not blank when party cookies are unblocked fingerprint tests visit ensure blocked items are listed in shields result window should show got canvas fingerprint and got webgl fingerprint test that audio fingerprint is blocked at only when block all fingerprinting protection is on test that brave browser isn t detected on test that doesn t leak ip address when block all fingerprinting protection is on rewards verify wallet is auto created after enabling rewards verify account balance shows correct bat and usd value verify you are able to restore a wallet verify wallet address matches the qr code that is generated under add funds verify actions taken claiming grant tipping auto contribute display in wallet panel verify adding funds via any of the currencies flows into wallet after specified amount of time verify adding funds to an existing wallet with amount adjusts the bat value appropriately verify monthly budget shows correct bat and usd value verify you are able to exclude a publisher from the auto contribute table by clicking on the x in auto contribute table and popup list of sites verify you are able to exclude a publisher by using the toggle on the rewards panel verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify when you click on send a tip the custom tip banner displays verify you are able to make one time tip and they display in tips panel verify you are able to make recurring tip and they display in tips panel verify you can tip a verified publisher verify you can tip a verified youtube creator verify tip panel shows a verified checkmark for a verified publisher verified youtube creator verify tip panel shows a message about unverified publisher verify br panel shows message about an unverified publisher verify you are able to perform a contribution verify if you disable auto contribute you are still able to tip regular sites and youtube creators verify that disabling rewards and enabling it again does not lose state verify that disabling auto contribute and enabling it again does not lose state adjust min visit time in settings visit some sites and youtube channels to verify they are added to the table after the specified settings upgrade from older version verify the wallet balance is retained and wallet backup code isn t corrupted verify auto contribute list is not lost after upgrade verify tips list is not lost after upgrade verify wallet panel transactions list is not lost after upgrade ads upgrade tests install and enable rewards ads are not available on this version update on test channel to the hotfix version verify ads are off by default should get a bat logo notification to alert you that ads are available install and enable rewards ads are on by default view an ad update on test channel to the hotfix version verify ads are still on after update ads panel information was not lost after upgrade no bat logo notification install and enable rewards disable ads update on test channel to the hotfix version verify ads are still off after update no bat logo notification install and enable rewards ads are on by default view an ad update on test channel to the hotfix version verify ads are still on after update ads panel information was not lost after upgrade no bat logo notification install and enable rewards disable ads update on test channel to the hotfix version verify ads are still off after update no bat logo notification tor tabs visit in a tor window ensure its shows success message for using a tor exit node visit in a tor window note down exit node ip address do a hard refresh ctrl shift r cmd shift r ensure exit ip changes after page reloads visit in a tor window ensure login page is shown visit in a tor window ensure location isn t shown session storage temporarily move away your browser profile and test that a new profile is created when browser is launched macos library application support bravesoftware windows userprofile appdata local bravesoftware linux ubuntu config bravesoftware test that windows and tabs restore when closed including active tab ensure that the tabs in the above session are being lazy loaded when the session is restored update tests verify visiting brave settings help triggers update check verify once update is downloaded prompts to relaunch to install update chromium upgrade tests verify brave gpu on brave and chrome gpu on chrome are similar for the same chromium version on both browsers adblock verify referrer blocking works properly for tld visit and verify adblock works properly similar to components delete adblock folder from browser profile and restart browser visit brave components and verify brave ad block updater downloads and update the component repeat for all brave components crypto wallets ensure that you can create a new wallet without any issues ensure that you can restore a previous cw wallet without any issues ensure that you can restore a previous mm wallet without any issues ensure that you can create a transaction sending crypto with a cw wallet ensure that you can create a transaction sending crypto using a restored mm wallet
0
48,258
13,067,582,872
IssuesEvent
2020-07-31 00:55:39
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
Assertion "`bin_edges[q+1]-bin_edges[q] > 0' failed." triggered in MillipedeDOMCacheMap.cxx:247 (Trac #2092)
Migrated from Trac combo reconstruction defect
Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion: /home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed. Unfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered? Migrated from https://code.icecube.wisc.edu/ticket/2092 ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "description": "Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion:\n\n/home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed.\n\nUnfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered?", "reporter": "claudio.kopper", "cc": "austin.schneider", "resolution": "insufficient resources", "_ts": "1550067323910946", "component": "combo reconstruction", "summary": "Assertion \"`bin_edges[q+1]-bin_edges[q] > 0' failed.\" triggered in MillipedeDOMCacheMap.cxx:247", "priority": "normal", "keywords": "", "time": "2017-09-15T16:30:44", "milestone": "", "owner": "jvansanten", "type": "defect" } ```
1.0
Assertion "`bin_edges[q+1]-bin_edges[q] > 0' failed." triggered in MillipedeDOMCacheMap.cxx:247 (Trac #2092) - Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion: /home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed. Unfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered? Migrated from https://code.icecube.wisc.edu/ticket/2092 ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "description": "Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion:\n\n/home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed.\n\nUnfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered?", "reporter": "claudio.kopper", "cc": "austin.schneider", "resolution": "insufficient resources", "_ts": "1550067323910946", "component": "combo reconstruction", "summary": "Assertion \"`bin_edges[q+1]-bin_edges[q] > 0' failed.\" triggered in MillipedeDOMCacheMap.cxx:247", "priority": "normal", "keywords": "", "time": "2017-09-15T16:30:44", "milestone": "", "owner": "jvansanten", "type": "defect" } ```
defect
assertion bin edges bin edges failed triggered in millipededomcachemap cxx trac austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion home aschneider combo source millipede private millipede millipededomcachemap cxx int rebin bayesian blocks int const double const double bool double double double double assertion bin edges bin edges failed unfortunately we do not have a minimal example right now i guess distribute needs better reporting or we would need to add logging but maybe it is obvious why this assertion could be triggered migrated from json status closed changetime description austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion n n home aschneider combo source millipede private millipede millipededomcachemap cxx int rebin bayesian blocks int const double const double bool double double double double assertion bin edges bin edges failed n nunfortunately we do not have a minimal example right now i guess distribute needs better reporting or we would need to add logging but maybe it is obvious why this assertion could be triggered reporter claudio kopper cc austin schneider resolution insufficient resources ts component combo reconstruction summary assertion bin edges bin edges failed triggered in millipededomcachemap cxx priority normal keywords time milestone owner jvansanten type defect
1
10,361
2,622,147,426
IssuesEvent
2015-03-04 00:04:42
byzhang/i7z
https://api.github.com/repos/byzhang/i7z
closed
mandir should be prefixed with $(prefix)
auto-migrated Priority-Medium Type-Defect
``` commit 95 adds mandir to the Makefile, but missing $(prefix). Please add this to allow non "/" installations. ``` Original issue reported on code.google.com by `ju.lec...@googlemail.com` on 24 Jul 2012 at 6:27
1.0
mandir should be prefixed with $(prefix) - ``` commit 95 adds mandir to the Makefile, but missing $(prefix). Please add this to allow non "/" installations. ``` Original issue reported on code.google.com by `ju.lec...@googlemail.com` on 24 Jul 2012 at 6:27
defect
mandir should be prefixed with prefix commit adds mandir to the makefile but missing prefix please add this to allow non installations original issue reported on code google com by ju lec googlemail com on jul at
1
35,034
7,887,543,491
IssuesEvent
2018-06-27 18:49:55
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Hackers have infaltrated com_contact through XSS and [POST:jform]
No Code Attached Yet
### Steps to reproduce the issue Upgraded to 3.8.10. As soon as I did Hackers from Germany hit com_contact with XSS attacks and [POST:jform] and send multiple emails through this component. I don't even use this component on my site. ### Expected result No hacking and XSS attack of com_contact ### Actual result emails are sent to me using this component 5 emails every 5 minutes ### System information (as much as possible) PHP 7.0.30 Joomla 3.8.10 ### Additional comments At this time I have installed Security Check Pro and it is stripping the XSS attack and stopping this. Won't upgrade any more sites until this is rectified.
1.0
Hackers have infaltrated com_contact through XSS and [POST:jform] - ### Steps to reproduce the issue Upgraded to 3.8.10. As soon as I did Hackers from Germany hit com_contact with XSS attacks and [POST:jform] and send multiple emails through this component. I don't even use this component on my site. ### Expected result No hacking and XSS attack of com_contact ### Actual result emails are sent to me using this component 5 emails every 5 minutes ### System information (as much as possible) PHP 7.0.30 Joomla 3.8.10 ### Additional comments At this time I have installed Security Check Pro and it is stripping the XSS attack and stopping this. Won't upgrade any more sites until this is rectified.
non_defect
hackers have infaltrated com contact through xss and steps to reproduce the issue upgraded to as soon as i did hackers from germany hit com contact with xss attacks and and send multiple emails through this component i don t even use this component on my site expected result no hacking and xss attack of com contact actual result emails are sent to me using this component emails every minutes system information as much as possible php joomla additional comments at this time i have installed security check pro and it is stripping the xss attack and stopping this won t upgrade any more sites until this is rectified
0
55,910
14,789,851,447
IssuesEvent
2021-01-12 11:09:21
hazelcast/hazelcast-jet
https://api.github.com/repos/hazelcast/hazelcast-jet
closed
Not requiring columns not required for JSON format
defect sql
This command should fail, but it succeeds: ```sql CREATE MAPPING latest_trades TYPE IMap OPTIONS ( 'keyFormat' = 'java', 'keyJavaClass' = 'java.lang.String', 'valueFormat' = 'json' ); ``` For JSON we can't resolve the columns. If the `keyFormat` was also `json`, the command will fail with "The resolved field list is empty".
1.0
Not requiring columns not required for JSON format - This command should fail, but it succeeds: ```sql CREATE MAPPING latest_trades TYPE IMap OPTIONS ( 'keyFormat' = 'java', 'keyJavaClass' = 'java.lang.String', 'valueFormat' = 'json' ); ``` For JSON we can't resolve the columns. If the `keyFormat` was also `json`, the command will fail with "The resolved field list is empty".
defect
not requiring columns not required for json format this command should fail but it succeeds sql create mapping latest trades type imap options keyformat java keyjavaclass java lang string valueformat json for json we can t resolve the columns if the keyformat was also json the command will fail with the resolved field list is empty
1
82,066
31,896,357,492
IssuesEvent
2023-09-18 02:20:42
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Text-based user avatars: abnormal appearance.
T-Defect
### Steps to reproduce n/a ### Outcome #### What did you expect? n/a #### What happened instead? As shown in the figure, some avatar is stretched vertically. **But limited to text-based avatars, image-based avatars are normal.** ![{4FA45B7F-E917-7CF7-3654-9863EA8C3114}](https://github.com/vector-im/element-web/assets/48104960/f3eab991-f086-41ca-a75e-67436faa2e30) ### Operating system x64 Windows 11 (10.0.22621) ### Application version Element Desktop 1.11.43 ### How did you install the app? https://packages.element.io/desktop/update/win32/x64 ### Homeserver Synapse 1.92.1 ### Will you send logs? No
1.0
Text-based user avatars: abnormal appearance. - ### Steps to reproduce n/a ### Outcome #### What did you expect? n/a #### What happened instead? As shown in the figure, some avatar is stretched vertically. **But limited to text-based avatars, image-based avatars are normal.** ![{4FA45B7F-E917-7CF7-3654-9863EA8C3114}](https://github.com/vector-im/element-web/assets/48104960/f3eab991-f086-41ca-a75e-67436faa2e30) ### Operating system x64 Windows 11 (10.0.22621) ### Application version Element Desktop 1.11.43 ### How did you install the app? https://packages.element.io/desktop/update/win32/x64 ### Homeserver Synapse 1.92.1 ### Will you send logs? No
defect
text based user avatars abnormal appearance steps to reproduce n a outcome what did you expect n a what happened instead as shown in the figure some avatar is stretched vertically but limited to text based avatars image based avatars are normal operating system windows application version element desktop how did you install the app homeserver synapse will you send logs no
1
264,867
20,036,511,708
IssuesEvent
2022-02-02 12:30:00
LForchini/Multiverse-Web-Store
https://api.github.com/repos/LForchini/Multiverse-Web-Store
closed
Test Endpoints using postman
documentation enhancement
ENDPOINTS: * `GET /cart/` * `POST /cart/add` * `PATCH /cart/` * `DELETE /cart/:productId` * `GET /products/` * `GET /products/:productId` * `GET /users/` * `GET /users/:userId` * `DELETE /users/:userId` * `PATCH /users/:userId` * `POST /users/login` * `POST /users/logout` * `POST /users/`
1.0
Test Endpoints using postman - ENDPOINTS: * `GET /cart/` * `POST /cart/add` * `PATCH /cart/` * `DELETE /cart/:productId` * `GET /products/` * `GET /products/:productId` * `GET /users/` * `GET /users/:userId` * `DELETE /users/:userId` * `PATCH /users/:userId` * `POST /users/login` * `POST /users/logout` * `POST /users/`
non_defect
test endpoints using postman endpoints get cart post cart add patch cart delete cart productid get products get products productid get users get users userid delete users userid patch users userid post users login post users logout post users
0
43,548
11,746,386,492
IssuesEvent
2020-03-12 11:37:37
primefaces/primereact
https://api.github.com/repos/primefaces/primereact
closed
DataTable scrollHeight doesn't calculate correctly with frozen columns and header groups
defect
### There is no guarantee in receiving an immediate response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeReact PRO Support* where support is provided within 4 business hours **I'm submitting a ...** (check one with "x") ``` [X] bug report [ ] feature request [ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57 ``` **Codesandbox Case (Bug Reports)** Please fork the codesandbox below and create a case demonstrating your bug report. Issues without a codesandbox have much less possibility to be reviewed. It's a convoluted example, but this is essentially what we have in our project. A table with the first 2 columns fixed and header groups to highlight different areas of the report table https://codesandbox.io/s/primereact-test-vzsde **Current behavior** <!-- Describe how the bug manifests. --> When viewing a table with frozen columns and header groups the frozen table gets calculated to one height and the unfrozen part to a completely different height. **Expected behavior** <!-- Describe what the behavior would be without the bug. --> Both sides of the table should be the same height regardless of percentage or fixed height or whether they are frozen or have colum header grops. **Minimal reproduction of the problem with instructions** <!-- If the current behavior is a bug or you can illustrate your feature request better with an example, please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via https://codesandbox.io or similar (you can use this template as a starting point: https://codesandbox.io/s/qjx332qq4). --> https://codesandbox.io/s/primereact-test-vzsde has everything demonstrated, but the following steps were followed: * create a datatable with some data (make sure there are a large number of columns) * mark the first few columns as frozen * create fromColumnHeaderGroup and a headerGroup (not sure if multiple rows make a difference, but our project requires multiple rows **Please tell us about your environment:** <!-- Operating system, IDE, package manager, HTTP server, ... --> Windows 10, Visual Code, NPM, Node * **React version:** <!-- Check whether this is still an issue in the most recent React version --> 16.12.0 * **PrimeReact version:** <!-- Check whether this is still an issue in the most recent PrimeReact version --> 3.4.0 * **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] <!-- All browsers where this could be reproduced --> assume all - tried it on Chrome 80, Firefox 72, IE11, Edge 82 * **Language:** [all | TypeScript X.X | ES6/7 | ES5] Typescript
1.0
DataTable scrollHeight doesn't calculate correctly with frozen columns and header groups - ### There is no guarantee in receiving an immediate response in GitHub Issue Tracker, If you'd like to secure our response, you may consider *PrimeReact PRO Support* where support is provided within 4 business hours **I'm submitting a ...** (check one with "x") ``` [X] bug report [ ] feature request [ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57 ``` **Codesandbox Case (Bug Reports)** Please fork the codesandbox below and create a case demonstrating your bug report. Issues without a codesandbox have much less possibility to be reviewed. It's a convoluted example, but this is essentially what we have in our project. A table with the first 2 columns fixed and header groups to highlight different areas of the report table https://codesandbox.io/s/primereact-test-vzsde **Current behavior** <!-- Describe how the bug manifests. --> When viewing a table with frozen columns and header groups the frozen table gets calculated to one height and the unfrozen part to a completely different height. **Expected behavior** <!-- Describe what the behavior would be without the bug. --> Both sides of the table should be the same height regardless of percentage or fixed height or whether they are frozen or have colum header grops. **Minimal reproduction of the problem with instructions** <!-- If the current behavior is a bug or you can illustrate your feature request better with an example, please provide the *STEPS TO REPRODUCE* and if possible a *MINIMAL DEMO* of the problem via https://codesandbox.io or similar (you can use this template as a starting point: https://codesandbox.io/s/qjx332qq4). --> https://codesandbox.io/s/primereact-test-vzsde has everything demonstrated, but the following steps were followed: * create a datatable with some data (make sure there are a large number of columns) * mark the first few columns as frozen * create fromColumnHeaderGroup and a headerGroup (not sure if multiple rows make a difference, but our project requires multiple rows **Please tell us about your environment:** <!-- Operating system, IDE, package manager, HTTP server, ... --> Windows 10, Visual Code, NPM, Node * **React version:** <!-- Check whether this is still an issue in the most recent React version --> 16.12.0 * **PrimeReact version:** <!-- Check whether this is still an issue in the most recent PrimeReact version --> 3.4.0 * **Browser:** [all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView ] <!-- All browsers where this could be reproduced --> assume all - tried it on Chrome 80, Firefox 72, IE11, Edge 82 * **Language:** [all | TypeScript X.X | ES6/7 | ES5] Typescript
defect
datatable scrollheight doesn t calculate correctly with frozen columns and header groups there is no guarantee in receiving an immediate response in github issue tracker if you d like to secure our response you may consider primereact pro support where support is provided within business hours i m submitting a check one with x bug report feature request support request please do not submit support request here instead see codesandbox case bug reports please fork the codesandbox below and create a case demonstrating your bug report issues without a codesandbox have much less possibility to be reviewed it s a convoluted example but this is essentially what we have in our project a table with the first columns fixed and header groups to highlight different areas of the report table current behavior when viewing a table with frozen columns and header groups the frozen table gets calculated to one height and the unfrozen part to a completely different height expected behavior both sides of the table should be the same height regardless of percentage or fixed height or whether they are frozen or have colum header grops minimal reproduction of the problem with instructions if the current behavior is a bug or you can illustrate your feature request better with an example please provide the steps to reproduce and if possible a minimal demo of the problem via or similar you can use this template as a starting point has everything demonstrated but the following steps were followed create a datatable with some data make sure there are a large number of columns mark the first few columns as frozen create fromcolumnheadergroup and a headergroup not sure if multiple rows make a difference but our project requires multiple rows please tell us about your environment windows visual code npm node react version primereact version browser assume all tried it on chrome firefox edge language typescript
1
23,466
3,829,676,953
IssuesEvent
2016-03-31 11:47:57
pearu/pylibnidaqmx
https://api.github.com/repos/pearu/pylibnidaqmx
closed
Importing Analog/digital input/output task does not work
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Installation pylibnidaqmx from Source 2. 3. What is the expected output? What do you see instead? Expected output: the import should be successful. Instead I get the following message: from nidaqmx import AnalogInputTask ------------------------------------------------------------ Traceback (most recent call last): File "<ipython console>", line 1, in <module> ImportError: cannot import name AnalogInputTask What version of the product are you using? On what operating system? NI6008; Pythonxy bundle, pylibnidaqmx-0.2.0 Please provide any additional information below. import nidaqmx works. But fails to import classes in "libnidaqmx.py" file. Thank you, Regards, Kousik ``` Original issue reported on code.google.com by `sarathyk...@gmail.com` on 27 Jul 2011 at 2:09
1.0
Importing Analog/digital input/output task does not work - ``` What steps will reproduce the problem? 1. Installation pylibnidaqmx from Source 2. 3. What is the expected output? What do you see instead? Expected output: the import should be successful. Instead I get the following message: from nidaqmx import AnalogInputTask ------------------------------------------------------------ Traceback (most recent call last): File "<ipython console>", line 1, in <module> ImportError: cannot import name AnalogInputTask What version of the product are you using? On what operating system? NI6008; Pythonxy bundle, pylibnidaqmx-0.2.0 Please provide any additional information below. import nidaqmx works. But fails to import classes in "libnidaqmx.py" file. Thank you, Regards, Kousik ``` Original issue reported on code.google.com by `sarathyk...@gmail.com` on 27 Jul 2011 at 2:09
defect
importing analog digital input output task does not work what steps will reproduce the problem installation pylibnidaqmx from source what is the expected output what do you see instead expected output the import should be successful instead i get the following message from nidaqmx import analoginputtask traceback most recent call last file line in importerror cannot import name analoginputtask what version of the product are you using on what operating system pythonxy bundle pylibnidaqmx please provide any additional information below import nidaqmx works but fails to import classes in libnidaqmx py file thank you regards kousik original issue reported on code google com by sarathyk gmail com on jul at
1
18,988
3,117,750,427
IssuesEvent
2015-09-04 05:03:57
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
opened
Emit/output error when using += operator with Delegates
defect ready
A severe output error happens whenever an Delegate is bound using the += operator. The bug was reported here [Delegate Compilation Bug](http://forums.bridge.net/forum/bridge-net-pro/bugs/614) For instance: - broken: `test.Delegate += Callback;` - works: `test.Delegate = test.Delegate + Callback;`
1.0
Emit/output error when using += operator with Delegates - A severe output error happens whenever an Delegate is bound using the += operator. The bug was reported here [Delegate Compilation Bug](http://forums.bridge.net/forum/bridge-net-pro/bugs/614) For instance: - broken: `test.Delegate += Callback;` - works: `test.Delegate = test.Delegate + Callback;`
defect
emit output error when using operator with delegates a severe output error happens whenever an delegate is bound using the operator the bug was reported here for instance broken test delegate callback works test delegate test delegate callback
1
193,671
15,384,306,511
IssuesEvent
2021-03-03 04:26:04
ProfKleberSouza/projeto-pratico-brunosamuelfernandogleydiston
https://api.github.com/repos/ProfKleberSouza/projeto-pratico-brunosamuelfernandogleydiston
closed
Revisor da documentação
documentation
Responsável por revisar o conteúdo em Contexto e Especificações do Projeto.
1.0
Revisor da documentação - Responsável por revisar o conteúdo em Contexto e Especificações do Projeto.
non_defect
revisor da documentação responsável por revisar o conteúdo em contexto e especificações do projeto
0
18,314
3,041,795,907
IssuesEvent
2015-08-08 00:44:38
supertuxkart/stk-code
https://api.github.com/repos/supertuxkart/stk-code
closed
Moved far back when rescuing in Green Valley
P3: normal T: defect
Driving the first part of Green Valley, the rescuing works as expected, putting the kart a little back. Rescuing after the invisible checkline however, always moves the kart back to said checkline. How to easily reproduce: drive Green Valley as normal, and rescue when coming to the second or the last bridge. There might be more tracks affected by this, but Green Valley is the one I'm certain of at the moment.
1.0
Moved far back when rescuing in Green Valley - Driving the first part of Green Valley, the rescuing works as expected, putting the kart a little back. Rescuing after the invisible checkline however, always moves the kart back to said checkline. How to easily reproduce: drive Green Valley as normal, and rescue when coming to the second or the last bridge. There might be more tracks affected by this, but Green Valley is the one I'm certain of at the moment.
defect
moved far back when rescuing in green valley driving the first part of green valley the rescuing works as expected putting the kart a little back rescuing after the invisible checkline however always moves the kart back to said checkline how to easily reproduce drive green valley as normal and rescue when coming to the second or the last bridge there might be more tracks affected by this but green valley is the one i m certain of at the moment
1
467,989
13,459,672,801
IssuesEvent
2020-09-09 12:36:34
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
registration.ap.gov.in - site is not usable
browser-firefox engine-gecko priority-normal
<!-- @browser: Firefox 81.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:81.0) Gecko/20100101 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/57970 --> **URL**: http://registration.ap.gov.in/APCARDECClient/FetchHelpDetails **Browser / Version**: Firefox 81.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: website not working <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/59605230-5a7b-4f18-be0e-b85c91460e6d.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200908191057</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/9/5ba8daa3-4f34-4223-b966-774109c5f313) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
registration.ap.gov.in - site is not usable - <!-- @browser: Firefox 81.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:81.0) Gecko/20100101 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/57970 --> **URL**: http://registration.ap.gov.in/APCARDECClient/FetchHelpDetails **Browser / Version**: Firefox 81.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: website not working <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/9/59605230-5a7b-4f18-be0e-b85c91460e6d.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200908191057</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/9/5ba8daa3-4f34-4223-b966-774109c5f313) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_defect
registration ap gov in site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce website not working view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
75,878
26,115,650,084
IssuesEvent
2022-12-28 05:34:39
jMonkeyEngine/jmonkeyengine
https://api.github.com/repos/jMonkeyEngine/jmonkeyengine
closed
A multithreading issue with GltfLoader
defect
I use multithreading to load gltf models, when loaded, all model materials are messed up. After I debug GltfLoader, I found out that there may be threading issue in the GltfLoader. The variable "defaultMaterialAdapters" cannot be "static". This variable will be shared when all threads load materials from the model. A simple way to resolve this issue is to remove the "static". Hope this helps. I use jme version: 3.5.2-stable ![20220505173134](https://user-images.githubusercontent.com/93635496/166900456-62948438-1335-4563-99c0-0cdada11880f.jpg) ![resolve](https://user-images.githubusercontent.com/93635496/166900714-56e25c54-caf5-4ccd-a1ed-106435b68acc.jpg) I use this way to resolve my problem. ![err0](https://user-images.githubusercontent.com/93635496/166901053-864bf782-4d8d-4a99-bc30-da51fa42adfa.jpg) ![err1](https://user-images.githubusercontent.com/93635496/166901072-d6d99275-fb64-488c-a350-b8b85237ff03.jpg) The wrong materials before ![correct](https://user-images.githubusercontent.com/93635496/166901114-9cb068e3-c7a4-42c4-a25b-31d50eeae9db.jpg) The correct materials after
1.0
A multithreading issue with GltfLoader - I use multithreading to load gltf models, when loaded, all model materials are messed up. After I debug GltfLoader, I found out that there may be threading issue in the GltfLoader. The variable "defaultMaterialAdapters" cannot be "static". This variable will be shared when all threads load materials from the model. A simple way to resolve this issue is to remove the "static". Hope this helps. I use jme version: 3.5.2-stable ![20220505173134](https://user-images.githubusercontent.com/93635496/166900456-62948438-1335-4563-99c0-0cdada11880f.jpg) ![resolve](https://user-images.githubusercontent.com/93635496/166900714-56e25c54-caf5-4ccd-a1ed-106435b68acc.jpg) I use this way to resolve my problem. ![err0](https://user-images.githubusercontent.com/93635496/166901053-864bf782-4d8d-4a99-bc30-da51fa42adfa.jpg) ![err1](https://user-images.githubusercontent.com/93635496/166901072-d6d99275-fb64-488c-a350-b8b85237ff03.jpg) The wrong materials before ![correct](https://user-images.githubusercontent.com/93635496/166901114-9cb068e3-c7a4-42c4-a25b-31d50eeae9db.jpg) The correct materials after
defect
a multithreading issue with gltfloader i use multithreading to load gltf models when loaded all model materials are messed up after i debug gltfloader i found out that there may be threading issue in the gltfloader the variable defaultmaterialadapters cannot be static this variable will be shared when all threads load materials from the model a simple way to resolve this issue is to remove the static hope this helps i use jme version stable i use this way to resolve my problem the wrong materials before the correct materials after
1
74,896
25,395,662,730
IssuesEvent
2022-11-22 08:28:47
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
WriteJdbcP does not rollback the transaction if SQLException is thrown [HZ-1744]
Type: Defect Source: Internal to-jira Team: Platform
**Describe the bug** WriteJdbcP is calling connection.commit() but does not rollback the transaction upon SQLException. As a result of this Hikari is showing these logs *"{} - Executed rollback on connection {} due to dirty commit state on close()"* **Expected behavior** `connection.rollback() shold be called when SQLException is caught` **To Reproduce** Steps to reproduce the behavior: Configure an IMap with MapStore. Try to put() the same item two times. Hikari will show the logs mentioned
1.0
WriteJdbcP does not rollback the transaction if SQLException is thrown [HZ-1744] - **Describe the bug** WriteJdbcP is calling connection.commit() but does not rollback the transaction upon SQLException. As a result of this Hikari is showing these logs *"{} - Executed rollback on connection {} due to dirty commit state on close()"* **Expected behavior** `connection.rollback() shold be called when SQLException is caught` **To Reproduce** Steps to reproduce the behavior: Configure an IMap with MapStore. Try to put() the same item two times. Hikari will show the logs mentioned
defect
writejdbcp does not rollback the transaction if sqlexception is thrown describe the bug writejdbcp is calling connection commit but does not rollback the transaction upon sqlexception as a result of this hikari is showing these logs executed rollback on connection due to dirty commit state on close expected behavior connection rollback shold be called when sqlexception is caught to reproduce steps to reproduce the behavior configure an imap with mapstore try to put the same item two times hikari will show the logs mentioned
1
445,543
31,239,894,489
IssuesEvent
2023-08-20 18:33:40
burmilla/os
https://api.github.com/repos/burmilla/os
closed
Add to releases a format supported by Digital Ocean
documentation wontfix
https://www.digitalocean.com/docs/images/custom-images/#image-requirements > Image Requirements > >Images you upload to DigitalOcean must meet the following requirements: > > Operating system. Images must have a Unix-like OS. > > File format. Images must be in one of the following file formats: > Raw (.img) with an MBR or GPT partition table > qcow2 > VHDX > VDI > VMDK > > Size. Images must be 100 GB or less when uncompressed, including the filesystem. > > Filesystem. Images must support the ext3 or ext4 filesystems. > > cloud-init. Images must have cloud-init 0.7.7 or higher, cloudbase-init, coreos-cloudinit, ignition, or bsd-cloudinit installed and configured correctly. If your image's default cloud-init configuration lists the NoCloud datasource before the ConfigDrive datasource, Droplets created from your image will not function properly. Click here to display detailed cloud-init instructions. > > SSH configuration. Images must have sshd installed and configured to run on boot. If your image does not have sshd set up, you will not have SSH access to Droplets created from that image unless you recover access using the Droplet console. > > You can also upload a custom image that meets the above criteria as a compressed gzip or bzip2 file.
1.0
Add to releases a format supported by Digital Ocean - https://www.digitalocean.com/docs/images/custom-images/#image-requirements > Image Requirements > >Images you upload to DigitalOcean must meet the following requirements: > > Operating system. Images must have a Unix-like OS. > > File format. Images must be in one of the following file formats: > Raw (.img) with an MBR or GPT partition table > qcow2 > VHDX > VDI > VMDK > > Size. Images must be 100 GB or less when uncompressed, including the filesystem. > > Filesystem. Images must support the ext3 or ext4 filesystems. > > cloud-init. Images must have cloud-init 0.7.7 or higher, cloudbase-init, coreos-cloudinit, ignition, or bsd-cloudinit installed and configured correctly. If your image's default cloud-init configuration lists the NoCloud datasource before the ConfigDrive datasource, Droplets created from your image will not function properly. Click here to display detailed cloud-init instructions. > > SSH configuration. Images must have sshd installed and configured to run on boot. If your image does not have sshd set up, you will not have SSH access to Droplets created from that image unless you recover access using the Droplet console. > > You can also upload a custom image that meets the above criteria as a compressed gzip or bzip2 file.
non_defect
add to releases a format supported by digital ocean image requirements images you upload to digitalocean must meet the following requirements operating system images must have a unix like os file format images must be in one of the following file formats raw img with an mbr or gpt partition table vhdx vdi vmdk size images must be gb or less when uncompressed including the filesystem filesystem images must support the or filesystems cloud init images must have cloud init or higher cloudbase init coreos cloudinit ignition or bsd cloudinit installed and configured correctly if your image s default cloud init configuration lists the nocloud datasource before the configdrive datasource droplets created from your image will not function properly click here to display detailed cloud init instructions ssh configuration images must have sshd installed and configured to run on boot if your image does not have sshd set up you will not have ssh access to droplets created from that image unless you recover access using the droplet console you can also upload a custom image that meets the above criteria as a compressed gzip or file
0
49,879
13,187,284,402
IssuesEvent
2020-08-13 02:55:39
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Typo in py2-v3.0.1 setup.sh (Trac #2213)
Incomplete Migration Migrated from Trac cvmfs defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2213">https://code.icecube.wisc.edu/ticket/2213</a>, reported by icecube and owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "description": "Running eval `/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/setup.sh` fails on some hosts (such as submit-1, and my personal macbook) with the following error:\n\n/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/setup.sh: line 358: [: argument expected\n\nLooks to be a typo in this line of the setup script.\n", "reporter": "icecube", "cc": "", "resolution": "fixed", "_ts": "1550067323910946", "component": "cvmfs", "summary": "Typo in py2-v3.0.1 setup.sh", "priority": "normal", "keywords": "", "time": "2018-11-30T08:51:21", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
1.0
Typo in py2-v3.0.1 setup.sh (Trac #2213) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/2213">https://code.icecube.wisc.edu/ticket/2213</a>, reported by icecube and owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "description": "Running eval `/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/setup.sh` fails on some hosts (such as submit-1, and my personal macbook) with the following error:\n\n/cvmfs/icecube.opensciencegrid.org/py2-v3.0.1/setup.sh: line 358: [: argument expected\n\nLooks to be a typo in this line of the setup script.\n", "reporter": "icecube", "cc": "", "resolution": "fixed", "_ts": "1550067323910946", "component": "cvmfs", "summary": "Typo in py2-v3.0.1 setup.sh", "priority": "normal", "keywords": "", "time": "2018-11-30T08:51:21", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
defect
typo in setup sh trac migrated from json status closed changetime description running eval cvmfs icecube opensciencegrid org setup sh fails on some hosts such as submit and my personal macbook with the following error n n cvmfs icecube opensciencegrid org setup sh line argument expected n nlooks to be a typo in this line of the setup script n reporter icecube cc resolution fixed ts component cvmfs summary typo in setup sh priority normal keywords time milestone owner david schultz type defect
1
48,869
13,184,761,814
IssuesEvent
2020-08-12 20:02:47
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'> (Trac #348)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/348 , reported by mase and owned by nega_</summary> <p> ```json { "status": "closed", "changetime": "2019-01-11T20:08:32", "description": "I can't see the revision log for romeo trunk\n\n\n==== How to Reproduce ====\n\nWhile doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.\n\n''(please provide additional details here)''\n\n\nRequest parameters:\n{{{\n{'path': u'/projects/romeo/trunk'}\n}}}\n\n\nUser Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`\n\n==== System Information ====\n\n|| '''Trac''' || `0.11.1` ||\n|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||\n|| '''setuptools''' || `0.6` ||\n|| '''psycopg2''' || `2.2.1` ||\n|| '''Genshi''' || `0.6` ||\n|| '''mod_python''' || `3.3.1` ||\n|| '''Pygments''' || `0.10` ||\n|| '''Subversion''' || `1.5.1 (r32289)` ||\n|| '''jQuery:''' || `1.2.6` ||\n\n==== Python Traceback ====\n{{{\nTraceback (most recent call last):\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 423, in _dispatch_request\n dispatcher.dispatch(req)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 219, in dispatch\n data, content_type)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 715, in render_template\n return stream.render(method, doctype=doctype)\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 183, in render\n return encode(generator, method=method, encoding=encoding, out=out)\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 57, in encode\n return _encode(''.join(list(iterator)))\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 339, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 826, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 670, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 771, in __call__\n for kind, data, pos in chain(stream, [(None, None, None)]):\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 586, in __call__\n for ev in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 767, in _strip_accesskeys\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 756, in _generate\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 327, in _match\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 565, in _flatten\n result = _eval_expr(data, ctxt, vars)\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 277, in _eval_expr\n retval = expr.evaluate(ctxt)\n File \"/usr/lib/pymodules/python2.5/genshi/template/eval.py\", line 178, in evaluate\n return eval(self.code, _globals, {'__data__': data})\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html\", line 134, in <Expression u'dateinfo(change.date)'>\n <td class=\"date\" py:content=\"dateinfo(change.date)\" />\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py\", line 230, in dateinfo\n pretty_timedelta(date),\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 71, in pretty_timedelta\n time1 = to_datetime(time1)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 51, in to_datetime\n type(t))\nTypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>\n\n}}}\n ", "reporter": "mase", "cc": "", "resolution": "fixed", "_ts": "1547237312321270", "component": "combo simulation", "summary": "TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>", "priority": "normal", "keywords": "", "time": "2012-01-30T03:41:32", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'> (Trac #348) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/348 , reported by mase and owned by nega_</summary> <p> ```json { "status": "closed", "changetime": "2019-01-11T20:08:32", "description": "I can't see the revision log for romeo trunk\n\n\n==== How to Reproduce ====\n\nWhile doing a GET operation on `/log/projects/romeo/trunk`, Trac issued an internal error.\n\n''(please provide additional details here)''\n\n\nRequest parameters:\n{{{\n{'path': u'/projects/romeo/trunk'}\n}}}\n\n\nUser Agent was: `Mozilla/5.0 (X11; U; Linux i686; ja; rv:1.9.2.24) Gecko/20111107 Ubuntu/10.10 (maverick) Firefox/3.6.24`\n\n==== System Information ====\n\n|| '''Trac''' || `0.11.1` ||\n|| '''Python''' || `2.5.5 (r255:77872, Nov 28 2010, 19:14:39) ` [[br]] `[GCC 4.4.5]` ||\n|| '''setuptools''' || `0.6` ||\n|| '''psycopg2''' || `2.2.1` ||\n|| '''Genshi''' || `0.6` ||\n|| '''mod_python''' || `3.3.1` ||\n|| '''Pygments''' || `0.10` ||\n|| '''Subversion''' || `1.5.1 (r32289)` ||\n|| '''jQuery:''' || `1.2.6` ||\n\n==== Python Traceback ====\n{{{\nTraceback (most recent call last):\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 423, in _dispatch_request\n dispatcher.dispatch(req)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/main.py\", line 219, in dispatch\n data, content_type)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 715, in render_template\n return stream.render(method, doctype=doctype)\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 183, in render\n return encode(generator, method=method, encoding=encoding, out=out)\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 57, in encode\n return _encode(''.join(list(iterator)))\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 339, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 826, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 670, in __call__\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 771, in __call__\n for kind, data, pos in chain(stream, [(None, None, None)]):\n File \"/usr/lib/pymodules/python2.5/genshi/output.py\", line 586, in __call__\n for ev in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 767, in _strip_accesskeys\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/web/chrome.py\", line 756, in _generate\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 378, in _match\n ctxt, start=idx + 1, **vars):\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 327, in _match\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 545, in _flatten\n for kind, data, pos in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/core.py\", line 288, in _ensure\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/path.py\", line 588, in _generate\n subevent = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 605, in _include\n for event in stream:\n File \"/usr/lib/pymodules/python2.5/genshi/template/markup.py\", line 316, in _strip\n event = next()\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 565, in _flatten\n result = _eval_expr(data, ctxt, vars)\n File \"/usr/lib/pymodules/python2.5/genshi/template/base.py\", line 277, in _eval_expr\n retval = expr.evaluate(ctxt)\n File \"/usr/lib/pymodules/python2.5/genshi/template/eval.py\", line 178, in evaluate\n return eval(self.code, _globals, {'__data__': data})\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/versioncontrol/templates/revisionlog.html\", line 134, in <Expression u'dateinfo(change.date)'>\n <td class=\"date\" py:content=\"dateinfo(change.date)\" />\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/timeline/web_ui.py\", line 230, in dateinfo\n pretty_timedelta(date),\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 71, in pretty_timedelta\n time1 = to_datetime(time1)\n File \"/opt/trac/local/lib/python2.5/site-packages/Trac-0.11.1-py2.5.egg/trac/util/datefmt.py\", line 51, in to_datetime\n type(t))\nTypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>\n\n}}}\n ", "reporter": "mase", "cc": "", "resolution": "fixed", "_ts": "1547237312321270", "component": "combo simulation", "summary": "TypeError: expecting datetime, int, long, float, or None; got <class 'genshi.template.eval.Undefined'>", "priority": "normal", "keywords": "", "time": "2012-01-30T03:41:32", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
typeerror expecting datetime int long float or none got trac migrated from reported by mase and owned by nega json status closed changetime description i can t see the revision log for romeo trunk n n n how to reproduce n nwhile doing a get operation on log projects romeo trunk trac issued an internal error n n please provide additional details here n n nrequest parameters n n path u projects romeo trunk n n n nuser agent was mozilla u linux ja rv gecko ubuntu maverick firefox n n system information n n trac n python nov n setuptools n n genshi n mod python n pygments n subversion n jquery n n python traceback n ntraceback most recent call last n file opt trac local lib site packages trac egg trac web main py line in dispatch request n dispatcher dispatch req n file opt trac local lib site packages trac egg trac web main py line in dispatch n data content type n file opt trac local lib site packages trac egg trac web chrome py line in render template n return stream render method doctype doctype n file usr lib pymodules genshi core py line in render n return encode generator method method encoding encoding out out n file usr lib pymodules genshi output py line in encode n return encode join list iterator n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in stream n file usr lib pymodules genshi output py line in call n for kind data pos in chain stream n file usr lib pymodules genshi output py line in call n for ev in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file opt trac local lib site packages trac egg trac web chrome py line in strip accesskeys n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file opt trac local lib site packages trac egg trac web chrome py line in generate n for kind data pos in stream n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in match n ctxt start idx vars n file usr lib pymodules genshi template markup py line in match n ctxt start idx vars n file usr lib pymodules genshi template markup py line in match n for event in stream n file usr lib pymodules genshi template base py line in flatten n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi path py line in generate n subevent next n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in strip n event next n file usr lib pymodules genshi template base py line in flatten n for kind data pos in stream n file usr lib pymodules genshi core py line in ensure n for event in stream n file usr lib pymodules genshi path py line in generate n subevent next n file usr lib pymodules genshi template base py line in include n for event in stream n file usr lib pymodules genshi template markup py line in strip n event next n file usr lib pymodules genshi template base py line in flatten n result eval expr data ctxt vars n file usr lib pymodules genshi template base py line in eval expr n retval expr evaluate ctxt n file usr lib pymodules genshi template eval py line in evaluate n return eval self code globals data data n file opt trac local lib site packages trac egg trac versioncontrol templates revisionlog html line in n n file opt trac local lib site packages trac egg trac timeline web ui py line in dateinfo n pretty timedelta date n file opt trac local lib site packages trac egg trac util datefmt py line in pretty timedelta n to datetime n file opt trac local lib site packages trac egg trac util datefmt py line in to datetime n type t ntypeerror expecting datetime int long float or none got n n n reporter mase cc resolution fixed ts component combo simulation summary typeerror expecting datetime int long float or none got priority normal keywords time milestone owner nega type defect
1
225,693
24,881,138,897
IssuesEvent
2022-10-28 01:16:51
scumdestroy/docs
https://api.github.com/repos/scumdestroy/docs
opened
Twisted-22.4.0-py3-none-any.whl: 1 vulnerabilities (highest severity is: 5.4)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Twisted-22.4.0-py3-none-any.whl</b></p></summary> <p>An asynchronous networking framework written in Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/db/99/38622ff95bb740bcc991f548eb46295bba62fcb6e907db1987c4d92edd09/Twisted-22.4.0-py3-none-any.whl">https://files.pythonhosted.org/packages/db/99/38622ff95bb740bcc991f548eb46295bba62fcb6e907db1987c4d92edd09/Twisted-22.4.0-py3-none-any.whl</a></p> <p>Path to dependency file: /ci/requirements.txt</p> <p>Path to vulnerable library: /ci/requirements.txt,/ci/requirements.txt</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (Twisted version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-39348](https://www.mend.io/vulnerability-database/CVE-2022-39348) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.4 | Twisted-22.4.0-py3-none-any.whl | Direct | twisted - 19.2.1,18.4.0;Twisted - 22.10.0rc1 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-39348</summary> ### Vulnerable Library - <b>Twisted-22.4.0-py3-none-any.whl</b></p> <p>An asynchronous networking framework written in Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/db/99/38622ff95bb740bcc991f548eb46295bba62fcb6e907db1987c4d92edd09/Twisted-22.4.0-py3-none-any.whl">https://files.pythonhosted.org/packages/db/99/38622ff95bb740bcc991f548eb46295bba62fcb6e907db1987c4d92edd09/Twisted-22.4.0-py3-none-any.whl</a></p> <p>Path to dependency file: /ci/requirements.txt</p> <p>Path to vulnerable library: /ci/requirements.txt,/ci/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Twisted-22.4.0-py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> <p></p> ### Vulnerability Details <p> Twisted is an event-based framework for internet applications. Started with version 0.9.4, when the host header does not match a configured host `twisted.web.vhost.NameVirtualHost` will return a `NoResource` resource which renders the Host header unescaped into the 404 response allowing HTML and script injection. In practice this should be very difficult to exploit as being able to modify the Host header of a normal HTTP request implies that one is already in a privileged position. This issue was fixed in version 22.10.0rc1. There are no known workarounds. <p>Publish Date: 2022-10-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-39348>CVE-2022-39348</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.4</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-39348">https://nvd.nist.gov/vuln/detail/CVE-2022-39348</a></p> <p>Release Date: 2022-10-26</p> <p>Fix Resolution: twisted - 19.2.1,18.4.0;Twisted - 22.10.0rc1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
Twisted-22.4.0-py3-none-any.whl: 1 vulnerabilities (highest severity is: 5.4) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Twisted-22.4.0-py3-none-any.whl</b></p></summary> <p>An asynchronous networking framework written in Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/db/99/38622ff95bb740bcc991f548eb46295bba62fcb6e907db1987c4d92edd09/Twisted-22.4.0-py3-none-any.whl">https://files.pythonhosted.org/packages/db/99/38622ff95bb740bcc991f548eb46295bba62fcb6e907db1987c4d92edd09/Twisted-22.4.0-py3-none-any.whl</a></p> <p>Path to dependency file: /ci/requirements.txt</p> <p>Path to vulnerable library: /ci/requirements.txt,/ci/requirements.txt</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (Twisted version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-39348](https://www.mend.io/vulnerability-database/CVE-2022-39348) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.4 | Twisted-22.4.0-py3-none-any.whl | Direct | twisted - 19.2.1,18.4.0;Twisted - 22.10.0rc1 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-39348</summary> ### Vulnerable Library - <b>Twisted-22.4.0-py3-none-any.whl</b></p> <p>An asynchronous networking framework written in Python</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/db/99/38622ff95bb740bcc991f548eb46295bba62fcb6e907db1987c4d92edd09/Twisted-22.4.0-py3-none-any.whl">https://files.pythonhosted.org/packages/db/99/38622ff95bb740bcc991f548eb46295bba62fcb6e907db1987c4d92edd09/Twisted-22.4.0-py3-none-any.whl</a></p> <p>Path to dependency file: /ci/requirements.txt</p> <p>Path to vulnerable library: /ci/requirements.txt,/ci/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Twisted-22.4.0-py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> <p></p> ### Vulnerability Details <p> Twisted is an event-based framework for internet applications. Started with version 0.9.4, when the host header does not match a configured host `twisted.web.vhost.NameVirtualHost` will return a `NoResource` resource which renders the Host header unescaped into the 404 response allowing HTML and script injection. In practice this should be very difficult to exploit as being able to modify the Host header of a normal HTTP request implies that one is already in a privileged position. This issue was fixed in version 22.10.0rc1. There are no known workarounds. <p>Publish Date: 2022-10-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-39348>CVE-2022-39348</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.4</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-39348">https://nvd.nist.gov/vuln/detail/CVE-2022-39348</a></p> <p>Release Date: 2022-10-26</p> <p>Fix Resolution: twisted - 19.2.1,18.4.0;Twisted - 22.10.0rc1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_defect
twisted none any whl vulnerabilities highest severity is vulnerable library twisted none any whl an asynchronous networking framework written in python library home page a href path to dependency file ci requirements txt path to vulnerable library ci requirements txt ci requirements txt vulnerabilities cve severity cvss dependency type fixed in twisted version remediation available medium twisted none any whl direct twisted twisted details cve vulnerable library twisted none any whl an asynchronous networking framework written in python library home page a href path to dependency file ci requirements txt path to vulnerable library ci requirements txt ci requirements txt dependency hierarchy x twisted none any whl vulnerable library found in base branch develop vulnerability details twisted is an event based framework for internet applications started with version when the host header does not match a configured host twisted web vhost namevirtualhost will return a noresource resource which renders the host header unescaped into the response allowing html and script injection in practice this should be very difficult to exploit as being able to modify the host header of a normal http request implies that one is already in a privileged position this issue was fixed in version there are no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution twisted twisted step up your open source security game with mend
0
826,611
31,682,174,403
IssuesEvent
2023-09-08 01:27:56
Team-Ampersand/DUS
https://api.github.com/repos/Team-Ampersand/DUS
closed
(#80) DotoriCalendar에 onDaySelected parameter 추가
✨ Type: Feature 0️⃣ Priority: Critical
# Describe - 날짜를 클릭했을 때 실행되는 onDaySelected parameter 추가 ## etc - 기타사항을 입력하세요.
1.0
(#80) DotoriCalendar에 onDaySelected parameter 추가 - # Describe - 날짜를 클릭했을 때 실행되는 onDaySelected parameter 추가 ## etc - 기타사항을 입력하세요.
non_defect
dotoricalendar에 ondayselected parameter 추가 describe 날짜를 클릭했을 때 실행되는 ondayselected parameter 추가 etc 기타사항을 입력하세요
0
21,288
3,482,064,590
IssuesEvent
2015-12-29 20:29:48
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
[TypeScript] Optional parameter becomes mandatory
defect
```C# using Bridge; namespace Demo { public class App { public DoSomething(int i = 5) { } } } ``` Generated d.ts: ```JavaScript export interface App { doSomething(i: Number): void; } ``` Expected d.ts: ```JavaScript export interface App { doSomething(i?: Number): void; } ```
1.0
[TypeScript] Optional parameter becomes mandatory - ```C# using Bridge; namespace Demo { public class App { public DoSomething(int i = 5) { } } } ``` Generated d.ts: ```JavaScript export interface App { doSomething(i: Number): void; } ``` Expected d.ts: ```JavaScript export interface App { doSomething(i?: Number): void; } ```
defect
optional parameter becomes mandatory c using bridge namespace demo public class app public dosomething int i generated d ts javascript export interface app dosomething i number void expected d ts javascript export interface app dosomething i number void
1
31,399
6,518,370,739
IssuesEvent
2017-08-28 07:44:50
OlafvdSpek/ctemplate
https://api.github.com/repos/OlafvdSpek/ctemplate
closed
Linker errors while attempting to compile on SmartOS (Solaris 11) with gcc 4.7.3 (also tried 4.9
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Check out sources from trunk (also tried tagged release 2.3) 2. Attempt to build and link using FLAGS="-m64 -std=c++11" CFLAGS="$FLAGS" CXXFLAGS="$CFLAGS" LDFLAGS="$FLAGS -Lsrc/solaris" \ ./configure --prefix=/usr/local/ctemplate Initially tried without -Lsrc/solaris, but this does not seem to be relevant on Solaris 11 in any case. Both lead to same error. 3. Build using make What is the expected output? What do you see instead? All link attempts fail with errors similar to the following: CXXLD libctemplate.la Undefined first referenced symbol in file std::_Rb_tree<unsigned long, std::pair<unsigned long const, ctemplate::TemplateString>, std::_Select1st<std::pair<unsigned long const, ctemplate::TemplateString> >, std::less<unsigned long>, std::allocator<std::pair<unsigned long const, ctemplate::TemplateString> > >::_Rb_tree_impl<std::less<unsigned long>, true>::_Rb_tree_impl() .libs/libctemplate.so.3.0.0.exp ctemplate::IndentedWriter::~IndentedWriter() .libs/libctemplate.so.3.0.0.exp std::__detail::_Hash_code_base<ctemplate::TemplateCache::RefcountedTemplate*, std::pair<ctemplate::TemplateCache::RefcountedTemplate* const, int>, std::_Select1st<std::pair<ctemplate::TemplateCache::RefcountedTemplate* const, int> >, ctemplate::TemplateCache::RefTplPtrHash, std::__detail::_Mod_range_hashing, std::__detail::_Default_ranged_hash, true>::_Hash_code_base(std::_Select1st<std::pair<ctemplate::TemplateCache::Refco untedTemplate* const, int> > const&, ctemplate::TemplateCache::RefTplPtrHash const&, std::__detail::_Mod_range_hashing const&, std::__detail::_Default_ranged_hash const&) .libs/libctemplate.so.3.0.0.exp std::pair<std::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::vector<ctemplate::TemplateDictionary*, ctemplate::ArenaAllocator<ctemplate::TemplateDictionary*, ctemplate::UnsafeArena> > const*>::pair<std::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::vector<ctemplate::TemplateDictionary*, ctemplate::ArenaAllocator<ctemplate::TemplateDictionary*, ctemplate::UnsafeArena> > const*, void>(std::basic_string<char, std::char_traits<char>, std::allocator<char> >&&, std::vector<ctemplate::TemplateDictionary*, ctemplate::ArenaAllocator<ctemplate::TemplateDictionary*, ctemplate::UnsafeArena> > const*&&) .libs/libctemplate.so.3.0.0.exp -- and a lot more similar errors What version of the product are you using? On what operating system? Latest sources from trunk. Joyent SmartOS (SunOS 5.11 joyent_20140501T225642Z) Please provide any additional information below. ``` Original issue reported on code.google.com by `rakesh.v...@gmail.com` on 9 Feb 2015 at 12:48
1.0
Linker errors while attempting to compile on SmartOS (Solaris 11) with gcc 4.7.3 (also tried 4.9 - ``` What steps will reproduce the problem? 1. Check out sources from trunk (also tried tagged release 2.3) 2. Attempt to build and link using FLAGS="-m64 -std=c++11" CFLAGS="$FLAGS" CXXFLAGS="$CFLAGS" LDFLAGS="$FLAGS -Lsrc/solaris" \ ./configure --prefix=/usr/local/ctemplate Initially tried without -Lsrc/solaris, but this does not seem to be relevant on Solaris 11 in any case. Both lead to same error. 3. Build using make What is the expected output? What do you see instead? All link attempts fail with errors similar to the following: CXXLD libctemplate.la Undefined first referenced symbol in file std::_Rb_tree<unsigned long, std::pair<unsigned long const, ctemplate::TemplateString>, std::_Select1st<std::pair<unsigned long const, ctemplate::TemplateString> >, std::less<unsigned long>, std::allocator<std::pair<unsigned long const, ctemplate::TemplateString> > >::_Rb_tree_impl<std::less<unsigned long>, true>::_Rb_tree_impl() .libs/libctemplate.so.3.0.0.exp ctemplate::IndentedWriter::~IndentedWriter() .libs/libctemplate.so.3.0.0.exp std::__detail::_Hash_code_base<ctemplate::TemplateCache::RefcountedTemplate*, std::pair<ctemplate::TemplateCache::RefcountedTemplate* const, int>, std::_Select1st<std::pair<ctemplate::TemplateCache::RefcountedTemplate* const, int> >, ctemplate::TemplateCache::RefTplPtrHash, std::__detail::_Mod_range_hashing, std::__detail::_Default_ranged_hash, true>::_Hash_code_base(std::_Select1st<std::pair<ctemplate::TemplateCache::Refco untedTemplate* const, int> > const&, ctemplate::TemplateCache::RefTplPtrHash const&, std::__detail::_Mod_range_hashing const&, std::__detail::_Default_ranged_hash const&) .libs/libctemplate.so.3.0.0.exp std::pair<std::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::vector<ctemplate::TemplateDictionary*, ctemplate::ArenaAllocator<ctemplate::TemplateDictionary*, ctemplate::UnsafeArena> > const*>::pair<std::basic_string<char, std::char_traits<char>, std::allocator<char> >, std::vector<ctemplate::TemplateDictionary*, ctemplate::ArenaAllocator<ctemplate::TemplateDictionary*, ctemplate::UnsafeArena> > const*, void>(std::basic_string<char, std::char_traits<char>, std::allocator<char> >&&, std::vector<ctemplate::TemplateDictionary*, ctemplate::ArenaAllocator<ctemplate::TemplateDictionary*, ctemplate::UnsafeArena> > const*&&) .libs/libctemplate.so.3.0.0.exp -- and a lot more similar errors What version of the product are you using? On what operating system? Latest sources from trunk. Joyent SmartOS (SunOS 5.11 joyent_20140501T225642Z) Please provide any additional information below. ``` Original issue reported on code.google.com by `rakesh.v...@gmail.com` on 9 Feb 2015 at 12:48
defect
linker errors while attempting to compile on smartos solaris with gcc also tried what steps will reproduce the problem check out sources from trunk also tried tagged release attempt to build and link using flags std c cflags flags cxxflags cflags ldflags flags lsrc solaris configure prefix usr local ctemplate initially tried without lsrc solaris but this does not seem to be relevant on solaris in any case both lead to same error build using make what is the expected output what do you see instead all link attempts fail with errors similar to the following cxxld libctemplate la undefined first referenced symbol in file std rb tree unsigned long std pair unsigned long const ctemplate templatestring std std pair unsigned long const ctemplate templatestring std less std allocator rb tree impl true rb tree impl libs libctemplate so exp ctemplate indentedwriter indentedwriter libs libctemplate so exp std detail hash code base ctemplate templatecache refcountedtemplate std pair std std pair ctemplate templatecache refcountedtemplate const int ctemplate templatecache reftplptrhash std detail mod range hashing std detail default ranged hash true hash code base std std pair ctemplate templatecache refco untedtemplate const int const ctemplate templatecache reftplptrhash const std detail mod range hashing const std detail default ranged hash const libs libctemplate so exp std pair std allocator std vector ctemplate templatedictionary ctemplate arenaallocator ctemplate templatedictionary ctemplate unsafearena const pair std basic string char std char traits std allocator std vector ctemplate templatedictionary ctemplate arenaallocator ctemplate templatedictionary ctemplate unsafearena const void std basic string char std char traits std allocator std vector ctemplate templatedictionary ctemplate arenaallocator ctemplate templatedictionary ctemplate unsafearena const libs libctemplate so exp and a lot more similar errors what version of the product are you using on what operating system latest sources from trunk joyent smartos sunos joyent please provide any additional information below original issue reported on code google com by rakesh v gmail com on feb at
1
358,378
10,618,527,202
IssuesEvent
2019-10-13 05:25:38
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
opened
/blog/_posts/2017-01-00-Fission-Serverless-Functions-As-Service-For-Kubernetes.md
lang/zh priority/P1 sync/new version/1.16 welcome
Source File: [/blog/_posts/2017-01-00-Fission-Serverless-Functions-As-Service-For-Kubernetes.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/blog/_posts/2017-01-00-Fission-Serverless-Functions-As-Service-For-Kubernetes.md)
1.0
/blog/_posts/2017-01-00-Fission-Serverless-Functions-As-Service-For-Kubernetes.md - Source File: [/blog/_posts/2017-01-00-Fission-Serverless-Functions-As-Service-For-Kubernetes.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/blog/_posts/2017-01-00-Fission-Serverless-Functions-As-Service-For-Kubernetes.md)
non_defect
blog posts fission serverless functions as service for kubernetes md source file
0
19,063
3,130,941,391
IssuesEvent
2015-09-09 12:27:20
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
SecureSocket.initialize fails if certificate database path is non-ASCII
Area-Library Library-IO Priority-Low Triaged Type-Defect
Due to the Mozilla bug &quot;NSS does not allow specifying a profile path containing characters outside of the system native codepage&quot;, the Dart SecureSocket class can't be initialized with a certificate database with a non-ASCII path. See https://bugzilla.mozilla.org/show_bug.cgi?id=336829 Since there doesn't seem to be a fix unless it is fixed upstream, I will mark the priority low on this issue. Users have encountered this issue when running pub, from an installation directory with a non-ASCII path. See issue https://code.google.com/p/dart/issues/detail?id=15364
1.0
SecureSocket.initialize fails if certificate database path is non-ASCII - Due to the Mozilla bug &quot;NSS does not allow specifying a profile path containing characters outside of the system native codepage&quot;, the Dart SecureSocket class can't be initialized with a certificate database with a non-ASCII path. See https://bugzilla.mozilla.org/show_bug.cgi?id=336829 Since there doesn't seem to be a fix unless it is fixed upstream, I will mark the priority low on this issue. Users have encountered this issue when running pub, from an installation directory with a non-ASCII path. See issue https://code.google.com/p/dart/issues/detail?id=15364
defect
securesocket initialize fails if certificate database path is non ascii due to the mozilla bug quot nss does not allow specifying a profile path containing characters outside of the system native codepage quot the dart securesocket class can t be initialized with a certificate database with a non ascii path see since there doesn t seem to be a fix unless it is fixed upstream i will mark the priority low on this issue users have encountered this issue when running pub from an installation directory with a non ascii path see issue
1
366,397
10,820,315,742
IssuesEvent
2019-11-08 16:08:26
workcraft/workcraft
https://api.github.com/repos/workcraft/workcraft
closed
Verification for presence of consensus cubes in Circuit implementation
enhancement priority:high status:confirmed tag:backend:mpsat tag:model:circuit
We have a gate with output `O` computing function `F` that has a binate input `Bin`. This function `F` is implemented as a _sum-of-products_ cover, perhaps with some redundant cubes. Witness of property violation for the _positive phase_: * `O` is enabled, and * `Bin` is enabled, and * `~dF/dBin`, and * for each cube `C` in the cover: `~(C & ~dC/dBin)`. A similar witness needs to be derived for the _negative phase_.
1.0
Verification for presence of consensus cubes in Circuit implementation - We have a gate with output `O` computing function `F` that has a binate input `Bin`. This function `F` is implemented as a _sum-of-products_ cover, perhaps with some redundant cubes. Witness of property violation for the _positive phase_: * `O` is enabled, and * `Bin` is enabled, and * `~dF/dBin`, and * for each cube `C` in the cover: `~(C & ~dC/dBin)`. A similar witness needs to be derived for the _negative phase_.
non_defect
verification for presence of consensus cubes in circuit implementation we have a gate with output o computing function f that has a binate input bin this function f is implemented as a sum of products cover perhaps with some redundant cubes witness of property violation for the positive phase o is enabled and bin is enabled and df dbin and for each cube c in the cover c dc dbin a similar witness needs to be derived for the negative phase
0
217,682
16,860,454,281
IssuesEvent
2021-06-21 12:24:59
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[Security Solution] Long text name for Policy is not truncated under the Administration tab.
OLM Sprint QA:Ready for Testing Team: SecuritySolution Team:Onboarding and Lifecycle Mgt bug impact:low v7.14.0
**Describe the bug:** Long text name for Policy is not truncated under the Administration tab. **Build Details:** ``` Platform: Staging Version: 7.11.0-SNAPSHOT Commit : 1a6d7308e58787bf0448f65c039739b40d4a6e3d Build: 37074 Artifact Page: https://artifacts-api.elastic.co/v1/search/7.11.0-SNAPSHOT ``` **Browser Details** All **Preconditions** 1. Cloud environment on staging should exist. 2. Agent should be deployed with a Security Endpoint integration enabled 3. Provide a long name to the policy without any space. **Steps to Reproduce** 1. Navigate to Kibana URL on Browser. 2. Click on "Administration" under Security from the left navigation bar. 3. Click on the policy name and observe the long name is not truncated on UI. **Test data** N/A **Impacted Test case(s)** N/A **Actual Result** Long text name for Policy is not truncated under the Administration tab. **Expected Result** Long text name for Policy should be truncated under the Administration tab. **What's Working** N/A **What's not Working** N/A **Screenshots** ![Integration name](https://user-images.githubusercontent.com/60252716/101769417-f02c1800-3b0c-11eb-8d93-f0f0baa62733.png) **Logs** N/A
1.0
[Security Solution] Long text name for Policy is not truncated under the Administration tab. - **Describe the bug:** Long text name for Policy is not truncated under the Administration tab. **Build Details:** ``` Platform: Staging Version: 7.11.0-SNAPSHOT Commit : 1a6d7308e58787bf0448f65c039739b40d4a6e3d Build: 37074 Artifact Page: https://artifacts-api.elastic.co/v1/search/7.11.0-SNAPSHOT ``` **Browser Details** All **Preconditions** 1. Cloud environment on staging should exist. 2. Agent should be deployed with a Security Endpoint integration enabled 3. Provide a long name to the policy without any space. **Steps to Reproduce** 1. Navigate to Kibana URL on Browser. 2. Click on "Administration" under Security from the left navigation bar. 3. Click on the policy name and observe the long name is not truncated on UI. **Test data** N/A **Impacted Test case(s)** N/A **Actual Result** Long text name for Policy is not truncated under the Administration tab. **Expected Result** Long text name for Policy should be truncated under the Administration tab. **What's Working** N/A **What's not Working** N/A **Screenshots** ![Integration name](https://user-images.githubusercontent.com/60252716/101769417-f02c1800-3b0c-11eb-8d93-f0f0baa62733.png) **Logs** N/A
non_defect
long text name for policy is not truncated under the administration tab describe the bug long text name for policy is not truncated under the administration tab build details platform staging version snapshot commit build artifact page browser details all preconditions cloud environment on staging should exist agent should be deployed with a security endpoint integration enabled provide a long name to the policy without any space steps to reproduce navigate to kibana url on browser click on administration under security from the left navigation bar click on the policy name and observe the long name is not truncated on ui test data n a impacted test case s n a actual result long text name for policy is not truncated under the administration tab expected result long text name for policy should be truncated under the administration tab what s working n a what s not working n a screenshots logs n a
0
75,919
26,149,562,210
IssuesEvent
2022-12-30 11:22:57
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Double clicking „start video call“ adds two jitsi widgets
T-Defect S-Minor Help Wanted A-Jitsi O-Occasional good first issue
### Steps to reproduce 1. Have a room with some members 2. Double click camera to start a video call (some people are likely to double-click on everything) ### Outcome #### What did you expect? One call starting #### What happened instead? Two calls starting ![multiple_jits_doublei](https://user-images.githubusercontent.com/6216686/183071182-313dbad1-2bff-4de4-a8bd-8e37ced5673e.gif) ### Operating system Ubuntu 22.04 LTS ### Browser information Firefox 103.0.1 ### URL for webapp https://develop.element.io/ ### Application version Version von Element: 3e3f279469ad-react-b61cc4850b0f-js-3907d1c28f87 Version von Olm: 3.2.12 ### Homeserver _No response_ ### Will you send logs? No
1.0
Double clicking „start video call“ adds two jitsi widgets - ### Steps to reproduce 1. Have a room with some members 2. Double click camera to start a video call (some people are likely to double-click on everything) ### Outcome #### What did you expect? One call starting #### What happened instead? Two calls starting ![multiple_jits_doublei](https://user-images.githubusercontent.com/6216686/183071182-313dbad1-2bff-4de4-a8bd-8e37ced5673e.gif) ### Operating system Ubuntu 22.04 LTS ### Browser information Firefox 103.0.1 ### URL for webapp https://develop.element.io/ ### Application version Version von Element: 3e3f279469ad-react-b61cc4850b0f-js-3907d1c28f87 Version von Olm: 3.2.12 ### Homeserver _No response_ ### Will you send logs? No
defect
double clicking „start video call“ adds two jitsi widgets steps to reproduce have a room with some members double click camera to start a video call some people are likely to double click on everything outcome what did you expect one call starting what happened instead two calls starting operating system ubuntu lts browser information firefox url for webapp application version version von element react js version von olm homeserver no response will you send logs no
1
106,546
13,308,180,651
IssuesEvent
2020-08-26 00:11:28
phetsims/collision-lab
https://api.github.com/repos/phetsims/collision-lab
closed
Some pixel polishes
design:polish
Discussed some pixel polishes with @ariel-phet on zoom today. - [x] Change the font for the velocity/momentum readouts on top of the balls and COM to 14. - [x] Change the vertical spacing between the momentum vectors in the momenta diagram to 0.5 for 1D screens. - [x] Remove the furthest zoom level (all the way zoomed out) for the momenta diagram (all screens). - [x] Change the default to one more level zoomed in (all screens). - [x] Add one more level of zoom-in (all screens).
1.0
Some pixel polishes - Discussed some pixel polishes with @ariel-phet on zoom today. - [x] Change the font for the velocity/momentum readouts on top of the balls and COM to 14. - [x] Change the vertical spacing between the momentum vectors in the momenta diagram to 0.5 for 1D screens. - [x] Remove the furthest zoom level (all the way zoomed out) for the momenta diagram (all screens). - [x] Change the default to one more level zoomed in (all screens). - [x] Add one more level of zoom-in (all screens).
non_defect
some pixel polishes discussed some pixel polishes with ariel phet on zoom today change the font for the velocity momentum readouts on top of the balls and com to change the vertical spacing between the momentum vectors in the momenta diagram to for screens remove the furthest zoom level all the way zoomed out for the momenta diagram all screens change the default to one more level zoomed in all screens add one more level of zoom in all screens
0
23,192
3,775,435,193
IssuesEvent
2016-03-17 13:34:21
mgm3746/garbagecat
https://api.github.com/repos/mgm3746/garbagecat
closed
Handle JDK 1.7 ParNew logging
auto-migrated Priority-Medium Type-Defect
``` OracleJDK 1.7 update 55 2014-07-08T13:25:56.259+0100: 196.082: [GC2014-07-08T13:25:56.259+0100: 196.082: [ParNew: 471871K->35576K(471872K), 0.2695773 secs] 586079K->162457K(3093312K), 0.2699776 secs] [Times: user=0.48 sys=0.00, real=0.27 secs] ``` Original issue reported on code.google.com by `mgm...@gmail.com` on 10 Jul 2014 at 3:03
1.0
Handle JDK 1.7 ParNew logging - ``` OracleJDK 1.7 update 55 2014-07-08T13:25:56.259+0100: 196.082: [GC2014-07-08T13:25:56.259+0100: 196.082: [ParNew: 471871K->35576K(471872K), 0.2695773 secs] 586079K->162457K(3093312K), 0.2699776 secs] [Times: user=0.48 sys=0.00, real=0.27 secs] ``` Original issue reported on code.google.com by `mgm...@gmail.com` on 10 Jul 2014 at 3:03
defect
handle jdk parnew logging oraclejdk update secs times user sys real secs original issue reported on code google com by mgm gmail com on jul at
1
81,142
30,727,865,922
IssuesEvent
2023-07-27 21:20:54
vector-im/element-x-android
https://api.github.com/repos/vector-im/element-x-android
closed
Deadlock when opening a room.
S-Major T-Defect O-Occasional A-Timeline
We recently fixed several deadlocks (application in infinite loading state), but we can still repro one from time to time. 1. open the room list 2. open a room 3. go back to the room list 4. open the same room again 5. Repeat 3 and 4 until the timeline does not display messages but a loading wheel forever. We can repro on large rooms (10k+ members), as well as on DMs (2 members) Note: after a deadlock, repeat 3 and 4 again and the messages can be displayed. Looks like sometimes the method `addTimelineListener` from the Uniffi layer never returns.
1.0
Deadlock when opening a room. - We recently fixed several deadlocks (application in infinite loading state), but we can still repro one from time to time. 1. open the room list 2. open a room 3. go back to the room list 4. open the same room again 5. Repeat 3 and 4 until the timeline does not display messages but a loading wheel forever. We can repro on large rooms (10k+ members), as well as on DMs (2 members) Note: after a deadlock, repeat 3 and 4 again and the messages can be displayed. Looks like sometimes the method `addTimelineListener` from the Uniffi layer never returns.
defect
deadlock when opening a room we recently fixed several deadlocks application in infinite loading state but we can still repro one from time to time open the room list open a room go back to the room list open the same room again repeat and until the timeline does not display messages but a loading wheel forever we can repro on large rooms members as well as on dms members note after a deadlock repeat and again and the messages can be displayed looks like sometimes the method addtimelinelistener from the uniffi layer never returns
1
37,515
8,410,719,055
IssuesEvent
2018-10-12 11:36:49
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
closed
Phylanx fails with latest hpx
category: CMake type: defect
## Expected Behavior Expect Phylanx to compile with latest HPX ## Actual Behavior Phylanx Fails compilation with the following error listed below. I can confirm that it works up to the following commit dbbb94662ba0818cc5d6120f52e4a7b330075709 ( Merge pull request #3476 from STEllAR-GROUP/hide_credentials ). It looks to me the changes that were made to CMAKE recently is causing the issue. The buildbot link for the failing build is here: http://ktau.nic.uoregon.edu:8020/#/builders/8/builds/215 ``` [ 12%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/assert_condition.cpp.o /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/patterns.cpp: In function ‘std::__cxx11::string phylanx::execution_tree::find_help(const string&)’: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/patterns.cpp:68:30: error: comparison between signed and unsigned integer expressions [-Werror=sign-compare] for(int i=0;i<p.size();i++) { ~^~~~~~~~~ /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp: In instantiation of ‘std::vector<phylanx::ast::expression> phylanx::execution_tree::compiler::compiler::extract_lambda_arguments(const std::pair<_FIter, _FIter>&, const phylanx::ast::tagged&) [with Iterator = std::_Rb_tree_iterator<std::pair<const std::__cxx11::basic_string<char>, phylanx::ast::expression> >]’: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp:314:62: required from here /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp:201:41: error: comparison between signed and unsigned integer expressions [-Werror=sign-compare] if (count != 0 && count != size-1 && ~~~~~~^~~~~~~~~ /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp: In instantiation of ‘std::vector<phylanx::ast::expression> phylanx::execution_tree::compiler::compiler::extract_define_arguments(const std::pair<_FIter, _FIter>&, const phylanx::ast::tagged&) [with Iterator = std::_Rb_tree_iterator<std::pair<const std::__cxx11::basic_string<char>, phylanx::ast::expression> >]’: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp:336:62: required from here /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp:160:41: error: comparison between signed and unsigned integer expressions [-Werror=sign-compare] if (count != 0 && count != size-1 && ~~~~~~^~~~~~~~~ [ 12%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/base_primitive.cpp.o [ 12%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/call_function.cpp.o [ 12%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/console_output.cpp.o In file included from /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/ast/parser/expression.cpp:8:0: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/phylanx/ast/parser/expression_def.hpp: In instantiation of ‘phylanx::ast::parser::expression_base<Iterator>::expression_base(phylanx::ast::parser::error_handler<Iterator>&) [with Iterator = __gnu_cxx::__normal_iterator<const char*, std::__cxx11::basic_string<char> >]’: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/phylanx/ast/parser/expression_def.hpp:26:62: required from ‘phylanx::ast::parser::expression<Iterator>::expression(phylanx::ast::parser::error_handler<Iterator>&) [with Iterator = __gnu_cxx::__normal_iterator<const char*, std::__cxx11::basic_string<char> >]’ /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/ast/parser/expression.cpp:13:39: required from here /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/phylanx/ast/parser/expression_def.hpp:126:28: error: suggest parentheses around comparison in operand of ‘|’ [-Werror=parentheses] | '(' > expr > ')' ~~~~~~~~~~~^~~~~ [ 14%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/debug_output.cpp.o [ 14%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/define_variable.cpp.o [ 14%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/enable_tracing.cpp.o [ 14%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/format_string.cpp.o [ 17%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/function.cpp.o [ 17%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/lambda.cpp.o [ 17%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/node_data_helpers.cpp.o [ 17%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/primitive_component.cpp.o [ 19%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/primitive_component_base.cpp.o cc1plus: some warnings being treated as errors make[2]: *** [src/CMakeFiles/phylanx_component.dir/execution_tree/patterns.cpp.o] Error 1 make[2]: *** Waiting for unfinished jobs.... cc1plus: some warnings being treated as errors make[2]: *** [src/CMakeFiles/phylanx_component.dir/execution_tree/compiler/compiler.cpp.o] Error 1 /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/primitives/base_primitive.cpp:64:21: warning: type qualifiers ignored on function return type [-Wignored-qualifiers] static char const* const get_primitive_argument_type_name(std::size_t index) ^~~~~ cc1plus: some warnings being treated as errors make[2]: *** [src/CMakeFiles/phylanx_component.dir/ast/parser/expression.cpp.o] Error 1 make[1]: *** [src/CMakeFiles/phylanx_component.dir/all] Error 2 make: *** [all] Error 2 program finished with exit code 2 elapsedTime=80.792626 ``` ## Steps to Reproduce the Problem Build Phylanx against latest hpx ## Specifications - HPX Version: Master - Platform (compiler, OS): Ubuntu 18.04 / gcc 7.3
1.0
Phylanx fails with latest hpx - ## Expected Behavior Expect Phylanx to compile with latest HPX ## Actual Behavior Phylanx Fails compilation with the following error listed below. I can confirm that it works up to the following commit dbbb94662ba0818cc5d6120f52e4a7b330075709 ( Merge pull request #3476 from STEllAR-GROUP/hide_credentials ). It looks to me the changes that were made to CMAKE recently is causing the issue. The buildbot link for the failing build is here: http://ktau.nic.uoregon.edu:8020/#/builders/8/builds/215 ``` [ 12%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/assert_condition.cpp.o /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/patterns.cpp: In function ‘std::__cxx11::string phylanx::execution_tree::find_help(const string&)’: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/patterns.cpp:68:30: error: comparison between signed and unsigned integer expressions [-Werror=sign-compare] for(int i=0;i<p.size();i++) { ~^~~~~~~~~ /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp: In instantiation of ‘std::vector<phylanx::ast::expression> phylanx::execution_tree::compiler::compiler::extract_lambda_arguments(const std::pair<_FIter, _FIter>&, const phylanx::ast::tagged&) [with Iterator = std::_Rb_tree_iterator<std::pair<const std::__cxx11::basic_string<char>, phylanx::ast::expression> >]’: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp:314:62: required from here /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp:201:41: error: comparison between signed and unsigned integer expressions [-Werror=sign-compare] if (count != 0 && count != size-1 && ~~~~~~^~~~~~~~~ /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp: In instantiation of ‘std::vector<phylanx::ast::expression> phylanx::execution_tree::compiler::compiler::extract_define_arguments(const std::pair<_FIter, _FIter>&, const phylanx::ast::tagged&) [with Iterator = std::_Rb_tree_iterator<std::pair<const std::__cxx11::basic_string<char>, phylanx::ast::expression> >]’: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp:336:62: required from here /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/compiler/compiler.cpp:160:41: error: comparison between signed and unsigned integer expressions [-Werror=sign-compare] if (count != 0 && count != size-1 && ~~~~~~^~~~~~~~~ [ 12%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/base_primitive.cpp.o [ 12%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/call_function.cpp.o [ 12%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/console_output.cpp.o In file included from /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/ast/parser/expression.cpp:8:0: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/phylanx/ast/parser/expression_def.hpp: In instantiation of ‘phylanx::ast::parser::expression_base<Iterator>::expression_base(phylanx::ast::parser::error_handler<Iterator>&) [with Iterator = __gnu_cxx::__normal_iterator<const char*, std::__cxx11::basic_string<char> >]’: /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/phylanx/ast/parser/expression_def.hpp:26:62: required from ‘phylanx::ast::parser::expression<Iterator>::expression(phylanx::ast::parser::error_handler<Iterator>&) [with Iterator = __gnu_cxx::__normal_iterator<const char*, std::__cxx11::basic_string<char> >]’ /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/ast/parser/expression.cpp:13:39: required from here /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/phylanx/ast/parser/expression_def.hpp:126:28: error: suggest parentheses around comparison in operand of ‘|’ [-Werror=parentheses] | '(' > expr > ')' ~~~~~~~~~~~^~~~~ [ 14%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/debug_output.cpp.o [ 14%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/define_variable.cpp.o [ 14%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/enable_tracing.cpp.o [ 14%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/format_string.cpp.o [ 17%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/function.cpp.o [ 17%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/lambda.cpp.o [ 17%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/node_data_helpers.cpp.o [ 17%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/primitive_component.cpp.o [ 19%] Building CXX object src/CMakeFiles/phylanx_component.dir/execution_tree/primitives/primitive_component_base.cpp.o cc1plus: some warnings being treated as errors make[2]: *** [src/CMakeFiles/phylanx_component.dir/execution_tree/patterns.cpp.o] Error 1 make[2]: *** Waiting for unfinished jobs.... cc1plus: some warnings being treated as errors make[2]: *** [src/CMakeFiles/phylanx_component.dir/execution_tree/compiler/compiler.cpp.o] Error 1 /var/lib/buildbot/slaves/phylanx/x86_64-gcc7-release/build/src/execution_tree/primitives/base_primitive.cpp:64:21: warning: type qualifiers ignored on function return type [-Wignored-qualifiers] static char const* const get_primitive_argument_type_name(std::size_t index) ^~~~~ cc1plus: some warnings being treated as errors make[2]: *** [src/CMakeFiles/phylanx_component.dir/ast/parser/expression.cpp.o] Error 1 make[1]: *** [src/CMakeFiles/phylanx_component.dir/all] Error 2 make: *** [all] Error 2 program finished with exit code 2 elapsedTime=80.792626 ``` ## Steps to Reproduce the Problem Build Phylanx against latest hpx ## Specifications - HPX Version: Master - Platform (compiler, OS): Ubuntu 18.04 / gcc 7.3
defect
phylanx fails with latest hpx expected behavior expect phylanx to compile with latest hpx actual behavior phylanx fails compilation with the following error listed below i can confirm that it works up to the following commit merge pull request from stellar group hide credentials it looks to me the changes that were made to cmake recently is causing the issue the buildbot link for the failing build is here building cxx object src cmakefiles phylanx component dir execution tree primitives assert condition cpp o var lib buildbot slaves phylanx release build src execution tree patterns cpp in function ‘std string phylanx execution tree find help const string ’ var lib buildbot slaves phylanx release build src execution tree patterns cpp error comparison between signed and unsigned integer expressions for int i i p size i var lib buildbot slaves phylanx release build src execution tree compiler compiler cpp in instantiation of ‘std vector phylanx execution tree compiler compiler extract lambda arguments const std pair const phylanx ast tagged ’ var lib buildbot slaves phylanx release build src execution tree compiler compiler cpp required from here var lib buildbot slaves phylanx release build src execution tree compiler compiler cpp error comparison between signed and unsigned integer expressions if count count size var lib buildbot slaves phylanx release build src execution tree compiler compiler cpp in instantiation of ‘std vector phylanx execution tree compiler compiler extract define arguments const std pair const phylanx ast tagged ’ var lib buildbot slaves phylanx release build src execution tree compiler compiler cpp required from here var lib buildbot slaves phylanx release build src execution tree compiler compiler cpp error comparison between signed and unsigned integer expressions if count count size building cxx object src cmakefiles phylanx component dir execution tree primitives base primitive cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives call function cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives console output cpp o in file included from var lib buildbot slaves phylanx release build src ast parser expression cpp var lib buildbot slaves phylanx release build phylanx ast parser expression def hpp in instantiation of ‘phylanx ast parser expression base expression base phylanx ast parser error handler ’ var lib buildbot slaves phylanx release build phylanx ast parser expression def hpp required from ‘phylanx ast parser expression expression phylanx ast parser error handler ’ var lib buildbot slaves phylanx release build src ast parser expression cpp required from here var lib buildbot slaves phylanx release build phylanx ast parser expression def hpp error suggest parentheses around comparison in operand of ‘ ’ expr building cxx object src cmakefiles phylanx component dir execution tree primitives debug output cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives define variable cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives enable tracing cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives format string cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives function cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives lambda cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives node data helpers cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives primitive component cpp o building cxx object src cmakefiles phylanx component dir execution tree primitives primitive component base cpp o some warnings being treated as errors make error make waiting for unfinished jobs some warnings being treated as errors make error var lib buildbot slaves phylanx release build src execution tree primitives base primitive cpp warning type qualifiers ignored on function return type static char const const get primitive argument type name std size t index some warnings being treated as errors make error make error make error program finished with exit code elapsedtime steps to reproduce the problem build phylanx against latest hpx specifications hpx version master platform compiler os ubuntu gcc
1
122,775
12,159,648,877
IssuesEvent
2020-04-26 10:00:47
kiwicom/schemathesis
https://api.github.com/repos/kiwicom/schemathesis
closed
Add code examples
Priority: Low Type: Documentation
It will be extremely helpful to have some code examples , so it will be easier to start using Schemathesis. - Straightforward test. the whole api, without filtration or app insights - Filters for specific endpoints - Tests with insights about the app data (id of some existing DB object for GET / PUT calls) - Using `subtesthack` to unlock pytest fixtures per hypothesis test
1.0
Add code examples - It will be extremely helpful to have some code examples , so it will be easier to start using Schemathesis. - Straightforward test. the whole api, without filtration or app insights - Filters for specific endpoints - Tests with insights about the app data (id of some existing DB object for GET / PUT calls) - Using `subtesthack` to unlock pytest fixtures per hypothesis test
non_defect
add code examples it will be extremely helpful to have some code examples so it will be easier to start using schemathesis straightforward test the whole api without filtration or app insights filters for specific endpoints tests with insights about the app data id of some existing db object for get put calls using subtesthack to unlock pytest fixtures per hypothesis test
0
50,267
13,187,410,224
IssuesEvent
2020-08-13 03:19:37
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
clsim doesn't jive with OpenCL Framework (Trac #415)
Migrated from Trac combo simulation defect
http://builds.icecube.wisc.edu/builders/quick_simulation_osx_server_10.6.7/builds/506/steps/compile/logs/stdio <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/415 , reported by nega and owned by claudio.kopper_</summary> <p> ```json { "status": "closed", "changetime": "2012-06-12T22:32:22", "description": "http://builds.icecube.wisc.edu/builders/quick_simulation_osx_server_10.6.7/builds/506/steps/compile/logs/stdio\n\n", "reporter": "nega", "cc": "jvs", "resolution": "invalid", "_ts": "1339540342000000", "component": "combo simulation", "summary": "clsim doesn't jive with OpenCL Framework", "priority": "normal", "keywords": "clsim", "time": "2012-06-12T22:08:39", "milestone": "", "owner": "claudio.kopper", "type": "defect" } ``` </p> </details>
1.0
clsim doesn't jive with OpenCL Framework (Trac #415) - http://builds.icecube.wisc.edu/builders/quick_simulation_osx_server_10.6.7/builds/506/steps/compile/logs/stdio <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/415 , reported by nega and owned by claudio.kopper_</summary> <p> ```json { "status": "closed", "changetime": "2012-06-12T22:32:22", "description": "http://builds.icecube.wisc.edu/builders/quick_simulation_osx_server_10.6.7/builds/506/steps/compile/logs/stdio\n\n", "reporter": "nega", "cc": "jvs", "resolution": "invalid", "_ts": "1339540342000000", "component": "combo simulation", "summary": "clsim doesn't jive with OpenCL Framework", "priority": "normal", "keywords": "clsim", "time": "2012-06-12T22:08:39", "milestone": "", "owner": "claudio.kopper", "type": "defect" } ``` </p> </details>
defect
clsim doesn t jive with opencl framework trac migrated from reported by nega and owned by claudio kopper json status closed changetime description reporter nega cc jvs resolution invalid ts component combo simulation summary clsim doesn t jive with opencl framework priority normal keywords clsim time milestone owner claudio kopper type defect
1
31,035
2,731,183,698
IssuesEvent
2015-04-16 18:51:33
million12/M12.Foundation
https://api.github.com/repos/million12/M12.Foundation
closed
Raw edit mode: inline CSS styling should be ignored
bug improvement low-priority
Via Inspector user can provide custom inline CSS for majority of M12.Foundation components. When in raw edit mode, these styles should be ignored / not set.
1.0
Raw edit mode: inline CSS styling should be ignored - Via Inspector user can provide custom inline CSS for majority of M12.Foundation components. When in raw edit mode, these styles should be ignored / not set.
non_defect
raw edit mode inline css styling should be ignored via inspector user can provide custom inline css for majority of foundation components when in raw edit mode these styles should be ignored not set
0
137,373
18,752,697,375
IssuesEvent
2021-11-05 05:50:56
madhans23/linux-4.15
https://api.github.com/repos/madhans23/linux-4.15
opened
CVE-2020-12464 (Medium) detected in linuxv4.15
security vulnerability
## CVE-2020-12464 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv4.15</b></p></summary> <p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/brodo/linux.git>https://git.kernel.org/pub/scm/linux/kernel/git/brodo/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.15/commit/d96ee498864d1a0b6222cfb17d64ca8196014940">d96ee498864d1a0b6222cfb17d64ca8196014940</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> usb_sg_cancel in drivers/usb/core/message.c in the Linux kernel before 5.6.8 has a use-after-free because a transfer occurs without a reference, aka CID-056ad39ee925. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-12464>CVE-2020-12464</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-12464">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-12464</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: v5.7-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-12464 (Medium) detected in linuxv4.15 - ## CVE-2020-12464 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv4.15</b></p></summary> <p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/brodo/linux.git>https://git.kernel.org/pub/scm/linux/kernel/git/brodo/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.15/commit/d96ee498864d1a0b6222cfb17d64ca8196014940">d96ee498864d1a0b6222cfb17d64ca8196014940</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> usb_sg_cancel in drivers/usb/core/message.c in the Linux kernel before 5.6.8 has a use-after-free because a transfer occurs without a reference, aka CID-056ad39ee925. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-12464>CVE-2020-12464</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-12464">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-12464</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: v5.7-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in cve medium severity vulnerability vulnerable library library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details usb sg cancel in drivers usb core message c in the linux kernel before has a use after free because a transfer occurs without a reference aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
40,991
10,257,772,637
IssuesEvent
2019-08-21 20:56:19
mercury-hpc/mercury
https://api.github.com/repos/mercury-hpc/mercury
closed
Default Spack install fails
defect task
**Describe the bug** Fixing #299 with [this commit](https://github.com/mercury-hpc/mercury/commit/05831f8db8c2916628d03cb0caee339134a2997c) seems to have caused a side effect when installing with Spack. **To Reproduce** The default version of Mercury in the Spack package is 1.0.1. So if installing everything fresh with Spack, `spack install mercury`, you get `cmake@3.15.1` and `mercury@1.0.1` which then reproduces the issue in #299: ``` [prompt]$ spack install mercury ==> Installing mercury ⋮ ==> Executing phase: 'cmake' ⋮ >> 30 CMake Error at src/util/CMakeLists.txt:38 (CHECK_SYMBOL_EXISTS): 31 Unknown CMake command "CHECK_SYMBOL_EXISTS". 32 33 34 -- Configuring incomplete, errors occurred! ``` After troubleshooting and finding the issue, was able to use workarounds of either using the master branch: ``` spack install mercury@master ``` or an older version of cmake ``` spack install mercury ^cmake@3.14.5 ``` **Expected behavior** One of two things, 1. `spack install mercury` will default to a working combination (newer version, master branch, or older cmake) 2. `spack install mercury@1.0.1 ^cmake@3.15.1` will provide a proper error **Platform (please complete the following information):** - System description - Compiler version: [gcc 4.9.3] - Plugin and protocol used [ofi, sm] - Dependency version [spack defaults: libfabric 1.8.0, bmi, boost 1.70.0] **Potential solutions** Add a `conflicts` or `depends_on` to the Mercury Spack package so users know what is wrong if trying to install with the latest cmake and default version. 1. Add a `conflicts`: ``` conflicts('^cmake@3.15.0:', when='@:1.0.1') ``` results in: ``` [prompt]$ spack spec mercury@1.0.1 ^cmake@3.15.1 Input spec -------------------------------- mercury Concretized -------------------------------- ==> Error: Conflicts in concretized spec ⋮ 1. "^cmake@3.15.0:" conflicts with "mercury@:1.0.1" ``` or 2. Add a `depends_on`: ``` depends_on(cmake@:3.14.999', when='@:1.0.1', type='build') ``` results in: ``` [prompt]$ spack spec mercury@1.0.1 ^cmake@3.15.1 Input spec -------------------------------- mercury Concretized -------------------------------- ==> Error: An unsatisfiable version constraint has been detected for spec: ⋮ mercury requires cmake version 2.8.12.2:3.14.999, but spec asked for 3.15.1 ``` Other combinations (i.e., `spack install mercury@master`) succeed.
1.0
Default Spack install fails - **Describe the bug** Fixing #299 with [this commit](https://github.com/mercury-hpc/mercury/commit/05831f8db8c2916628d03cb0caee339134a2997c) seems to have caused a side effect when installing with Spack. **To Reproduce** The default version of Mercury in the Spack package is 1.0.1. So if installing everything fresh with Spack, `spack install mercury`, you get `cmake@3.15.1` and `mercury@1.0.1` which then reproduces the issue in #299: ``` [prompt]$ spack install mercury ==> Installing mercury ⋮ ==> Executing phase: 'cmake' ⋮ >> 30 CMake Error at src/util/CMakeLists.txt:38 (CHECK_SYMBOL_EXISTS): 31 Unknown CMake command "CHECK_SYMBOL_EXISTS". 32 33 34 -- Configuring incomplete, errors occurred! ``` After troubleshooting and finding the issue, was able to use workarounds of either using the master branch: ``` spack install mercury@master ``` or an older version of cmake ``` spack install mercury ^cmake@3.14.5 ``` **Expected behavior** One of two things, 1. `spack install mercury` will default to a working combination (newer version, master branch, or older cmake) 2. `spack install mercury@1.0.1 ^cmake@3.15.1` will provide a proper error **Platform (please complete the following information):** - System description - Compiler version: [gcc 4.9.3] - Plugin and protocol used [ofi, sm] - Dependency version [spack defaults: libfabric 1.8.0, bmi, boost 1.70.0] **Potential solutions** Add a `conflicts` or `depends_on` to the Mercury Spack package so users know what is wrong if trying to install with the latest cmake and default version. 1. Add a `conflicts`: ``` conflicts('^cmake@3.15.0:', when='@:1.0.1') ``` results in: ``` [prompt]$ spack spec mercury@1.0.1 ^cmake@3.15.1 Input spec -------------------------------- mercury Concretized -------------------------------- ==> Error: Conflicts in concretized spec ⋮ 1. "^cmake@3.15.0:" conflicts with "mercury@:1.0.1" ``` or 2. Add a `depends_on`: ``` depends_on(cmake@:3.14.999', when='@:1.0.1', type='build') ``` results in: ``` [prompt]$ spack spec mercury@1.0.1 ^cmake@3.15.1 Input spec -------------------------------- mercury Concretized -------------------------------- ==> Error: An unsatisfiable version constraint has been detected for spec: ⋮ mercury requires cmake version 2.8.12.2:3.14.999, but spec asked for 3.15.1 ``` Other combinations (i.e., `spack install mercury@master`) succeed.
defect
default spack install fails describe the bug fixing with seems to have caused a side effect when installing with spack to reproduce the default version of mercury in the spack package is so if installing everything fresh with spack spack install mercury you get cmake and mercury which then reproduces the issue in spack install mercury installing mercury ⋮ executing phase cmake ⋮ cmake error at src util cmakelists txt check symbol exists unknown cmake command check symbol exists configuring incomplete errors occurred after troubleshooting and finding the issue was able to use workarounds of either using the master branch spack install mercury master or an older version of cmake spack install mercury cmake expected behavior one of two things spack install mercury will default to a working combination newer version master branch or older cmake spack install mercury cmake will provide a proper error platform please complete the following information system description compiler version plugin and protocol used dependency version potential solutions add a conflicts or depends on to the mercury spack package so users know what is wrong if trying to install with the latest cmake and default version add a conflicts conflicts cmake when results in spack spec mercury cmake input spec mercury concretized error conflicts in concretized spec ⋮ cmake conflicts with mercury or add a depends on depends on cmake when type build results in spack spec mercury cmake input spec mercury concretized error an unsatisfiable version constraint has been detected for spec ⋮ mercury requires cmake version but spec asked for other combinations i e spack install mercury master succeed
1
25,361
4,306,653,915
IssuesEvent
2016-07-21 04:47:46
kraigs-android/kraigsandroid
https://api.github.com/repos/kraigs-android/kraigsandroid
closed
translation
auto-migrated Priority-Medium Type-Defect
``` Russian translation provided in Issue 48 contains several mistakes - semantic, misprints and just not very good interpretation (in my opinion). So I decided to correct them. Please find attached my version of translation. And here there is the description of main errors I've corrected: ------ 1. "default_settings": provided Russian text said "Restore standard settings", which is absolutely wrong 2. "alarm_fade": ru text said "Range of volume changing", I changed that to "Volume increasing". Unfortunately there is no direct short translation for the word "fade"... 3. "scheduled_days": ru text said "Days of week", changed that to "Ring on these days". 4. "snooze": again there is no conventional translation, but the provided one used informal word, which sounds too rude in my opinion. I changed that to a literary word 5. "debug_mode_summary": ru text said "Enabling debugging variants", changed that to be closer to the original meaning. 6. "time_out_title": provided Russian text said "Alarm time has finished", I changed that to "Limit of ringing duration" ------ Also I fixed one misprint in the original English file. Patch file is attached. ``` Original issue reported on code.google.com by `av.n...@gmail.com` on 29 Jun 2011 at 6:37 Attachments: * [strings-ru.xml](https://storage.googleapis.com/google-code-attachments/kraigsandroid/issue-63/comment-0/strings-ru.xml) * [en-strings.patch](https://storage.googleapis.com/google-code-attachments/kraigsandroid/issue-63/comment-0/en-strings.patch)
1.0
translation - ``` Russian translation provided in Issue 48 contains several mistakes - semantic, misprints and just not very good interpretation (in my opinion). So I decided to correct them. Please find attached my version of translation. And here there is the description of main errors I've corrected: ------ 1. "default_settings": provided Russian text said "Restore standard settings", which is absolutely wrong 2. "alarm_fade": ru text said "Range of volume changing", I changed that to "Volume increasing". Unfortunately there is no direct short translation for the word "fade"... 3. "scheduled_days": ru text said "Days of week", changed that to "Ring on these days". 4. "snooze": again there is no conventional translation, but the provided one used informal word, which sounds too rude in my opinion. I changed that to a literary word 5. "debug_mode_summary": ru text said "Enabling debugging variants", changed that to be closer to the original meaning. 6. "time_out_title": provided Russian text said "Alarm time has finished", I changed that to "Limit of ringing duration" ------ Also I fixed one misprint in the original English file. Patch file is attached. ``` Original issue reported on code.google.com by `av.n...@gmail.com` on 29 Jun 2011 at 6:37 Attachments: * [strings-ru.xml](https://storage.googleapis.com/google-code-attachments/kraigsandroid/issue-63/comment-0/strings-ru.xml) * [en-strings.patch](https://storage.googleapis.com/google-code-attachments/kraigsandroid/issue-63/comment-0/en-strings.patch)
defect
translation russian translation provided in issue contains several mistakes semantic misprints and just not very good interpretation in my opinion so i decided to correct them please find attached my version of translation and here there is the description of main errors i ve corrected default settings provided russian text said restore standard settings which is absolutely wrong alarm fade ru text said range of volume changing i changed that to volume increasing unfortunately there is no direct short translation for the word fade scheduled days ru text said days of week changed that to ring on these days snooze again there is no conventional translation but the provided one used informal word which sounds too rude in my opinion i changed that to a literary word debug mode summary ru text said enabling debugging variants changed that to be closer to the original meaning time out title provided russian text said alarm time has finished i changed that to limit of ringing duration also i fixed one misprint in the original english file patch file is attached original issue reported on code google com by av n gmail com on jun at attachments
1
51,638
12,753,998,109
IssuesEvent
2020-06-28 02:18:28
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Rustbuild should be able to test libstd without building a compiler
A-rustbuild A-testsuite C-enhancement
Currently, any attempt to build or test a part of libstd forces rustc and rustdoc to be built, even though the downloaded bootstrap compiler already contains rustc and rustdoc executables. What's worse is that the only way to prevent the compiler build is by using `--keep-stage`, which has numerous issues and can even break the build directory in certain cases. Rustbuild should be taught to use the the downloaded toolchain to build and test parts of libstd. It should either do so by default, or at least when `--stage=0` is passed.
1.0
Rustbuild should be able to test libstd without building a compiler - Currently, any attempt to build or test a part of libstd forces rustc and rustdoc to be built, even though the downloaded bootstrap compiler already contains rustc and rustdoc executables. What's worse is that the only way to prevent the compiler build is by using `--keep-stage`, which has numerous issues and can even break the build directory in certain cases. Rustbuild should be taught to use the the downloaded toolchain to build and test parts of libstd. It should either do so by default, or at least when `--stage=0` is passed.
non_defect
rustbuild should be able to test libstd without building a compiler currently any attempt to build or test a part of libstd forces rustc and rustdoc to be built even though the downloaded bootstrap compiler already contains rustc and rustdoc executables what s worse is that the only way to prevent the compiler build is by using keep stage which has numerous issues and can even break the build directory in certain cases rustbuild should be taught to use the the downloaded toolchain to build and test parts of libstd it should either do so by default or at least when stage is passed
0
68,814
21,914,725,701
IssuesEvent
2022-05-21 16:30:32
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
ZTS reproducibly panics on `zfs_rename_014_neg` on ppc64el
Type: Defect
### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | 11 Kernel Version | 5.15.41 Architecture | ppc64el OpenZFS Version | 2cd0f98f4 ### Describe the problem you're observing On trying to put the BLAKE3 PR through its paces, I hit a kernel panic on `zfs_rename_014_neg`. Confused, I tried running it against vanilla git, and lo, I had two panics. ### Describe how to reproduce the problem `scripts/zfs-tests.sh -T zfs_rename` on a ppc64el system, AFAICT. ### Include any warning/errors/backtraces from the system logs ``` [ 913.019952] synth uevent: /devices/vio: failed to send uevent [ 913.020021] vio vio: uevent: failed to send synthetic uevent [ 913.341239] synth uevent: /devices/vio: failed to send uevent [ 913.341316] vio vio: uevent: failed to send synthetic uevent [ 913.668152] synth uevent: /devices/vio: failed to send uevent [ 913.668200] vio vio: uevent: failed to send synthetic uevent [ 914.001152] synth uevent: /devices/vio: failed to send uevent [ 914.001231] vio vio: uevent: failed to send synthetic uevent [ 926.728484] synth uevent: /devices/vio: failed to send uevent [ 926.728594] vio vio: uevent: failed to send synthetic uevent [ 927.736509] synth uevent: /devices/vio: failed to send uevent [ 927.736576] vio vio: uevent: failed to send synthetic uevent [ 958.949429] Kernel panic - not syncing: corrupted stack end detected inside scheduler [ 958.949489] CPU: 1 PID: 134834 Comm: txg_sync Kdump: loaded Tainted: P OE 5.15.41-pristine #1 [ 958.949535] Call Trace: [ 958.949548] [c000000071a8ac30] [c00000000078dbd0] dump_stack_lvl+0x74/0xa8 (unreliable) [ 958.949592] [c000000071a8ac70] [c00000000013ada8] panic+0x154/0x3dc [ 958.949623] [c000000071a8ad00] [c000000000c81a3c] __schedule+0xb9c/0xba0 [ 958.949656] [c000000071a8add0] [c000000000c81c84] __cond_resched+0x64/0x90 [ 958.949690] [c000000071a8ae00] [c000000000c853e8] down_read+0x28/0x110 [ 958.949721] [c000000071a8ae30] [c008000009b63ff8] dnode_hold_impl+0x120/0x14e0 [zfs] [ 958.949808] [c000000071a8af00] [c008000009b3b680] dmu_bonus_hold+0x58/0xe0 [zfs] [ 958.949889] [c000000071a8af50] [c008000009b7e3f8] dsl_dataset_hold_obj+0x60/0xae0 [zfs] [ 958.949971] [c000000071a8b0d0] [c008000009b7ecb8] dsl_dataset_hold_obj+0x920/0xae0 [zfs] [ 958.950054] [c000000071a8b250] [c008000009b47b90] dmu_objset_find_dp_impl+0x158/0x510 [zfs] [ 958.950136] [c000000071a8b310] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950218] [c000000071a8b3d0] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950299] [c000000071a8b490] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950382] [c000000071a8b550] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950464] [c000000071a8b610] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950553] [c000000071a8b6d0] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950641] [c000000071a8b790] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950728] [c000000071a8b850] [c008000009b4b1d8] dmu_objset_find_dp+0x110/0x2d0 [zfs] [ 958.950815] [c000000071a8b940] [c008000009b94bc0] dsl_dir_rename_check+0x1e8/0x710 [zfs] [ 958.950902] [c000000071a8b9f0] [c008000009ba92b0] dsl_sync_task_sync+0xb8/0x1c0 [zfs] [ 958.950991] [c000000071a8ba30] [c008000009b97578] dsl_pool_sync+0x520/0x6a0 [zfs] [ 958.951079] [c000000071a8bb20] [c008000009bd55f4] spa_sync+0x62c/0x11f0 [zfs] [ 958.951166] [c000000071a8bc90] [c008000009c001bc] txg_sync_thread+0x2b4/0x450 [zfs] [ 958.951249] [c000000071a8bd60] [c0080000014ac2d0] thread_generic_wrapper+0x98/0xd0 [spl] [ 958.951293] [c000000071a8bda0] [c0000000001749e0] kthread+0x180/0x190 [ 958.951328] [c000000071a8be10] [c00000000000cf64] ret_from_kernel_thread+0x5c/0x64 [ 958.951382] Sending IPI to other CPUs [ 958.952928] IPI complete [ 958.955298] kexec: Starting switchover sequence. ```
1.0
ZTS reproducibly panics on `zfs_rename_014_neg` on ppc64el - ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | 11 Kernel Version | 5.15.41 Architecture | ppc64el OpenZFS Version | 2cd0f98f4 ### Describe the problem you're observing On trying to put the BLAKE3 PR through its paces, I hit a kernel panic on `zfs_rename_014_neg`. Confused, I tried running it against vanilla git, and lo, I had two panics. ### Describe how to reproduce the problem `scripts/zfs-tests.sh -T zfs_rename` on a ppc64el system, AFAICT. ### Include any warning/errors/backtraces from the system logs ``` [ 913.019952] synth uevent: /devices/vio: failed to send uevent [ 913.020021] vio vio: uevent: failed to send synthetic uevent [ 913.341239] synth uevent: /devices/vio: failed to send uevent [ 913.341316] vio vio: uevent: failed to send synthetic uevent [ 913.668152] synth uevent: /devices/vio: failed to send uevent [ 913.668200] vio vio: uevent: failed to send synthetic uevent [ 914.001152] synth uevent: /devices/vio: failed to send uevent [ 914.001231] vio vio: uevent: failed to send synthetic uevent [ 926.728484] synth uevent: /devices/vio: failed to send uevent [ 926.728594] vio vio: uevent: failed to send synthetic uevent [ 927.736509] synth uevent: /devices/vio: failed to send uevent [ 927.736576] vio vio: uevent: failed to send synthetic uevent [ 958.949429] Kernel panic - not syncing: corrupted stack end detected inside scheduler [ 958.949489] CPU: 1 PID: 134834 Comm: txg_sync Kdump: loaded Tainted: P OE 5.15.41-pristine #1 [ 958.949535] Call Trace: [ 958.949548] [c000000071a8ac30] [c00000000078dbd0] dump_stack_lvl+0x74/0xa8 (unreliable) [ 958.949592] [c000000071a8ac70] [c00000000013ada8] panic+0x154/0x3dc [ 958.949623] [c000000071a8ad00] [c000000000c81a3c] __schedule+0xb9c/0xba0 [ 958.949656] [c000000071a8add0] [c000000000c81c84] __cond_resched+0x64/0x90 [ 958.949690] [c000000071a8ae00] [c000000000c853e8] down_read+0x28/0x110 [ 958.949721] [c000000071a8ae30] [c008000009b63ff8] dnode_hold_impl+0x120/0x14e0 [zfs] [ 958.949808] [c000000071a8af00] [c008000009b3b680] dmu_bonus_hold+0x58/0xe0 [zfs] [ 958.949889] [c000000071a8af50] [c008000009b7e3f8] dsl_dataset_hold_obj+0x60/0xae0 [zfs] [ 958.949971] [c000000071a8b0d0] [c008000009b7ecb8] dsl_dataset_hold_obj+0x920/0xae0 [zfs] [ 958.950054] [c000000071a8b250] [c008000009b47b90] dmu_objset_find_dp_impl+0x158/0x510 [zfs] [ 958.950136] [c000000071a8b310] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950218] [c000000071a8b3d0] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950299] [c000000071a8b490] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950382] [c000000071a8b550] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950464] [c000000071a8b610] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950553] [c000000071a8b6d0] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950641] [c000000071a8b790] [c008000009b47de8] dmu_objset_find_dp_impl+0x3b0/0x510 [zfs] [ 958.950728] [c000000071a8b850] [c008000009b4b1d8] dmu_objset_find_dp+0x110/0x2d0 [zfs] [ 958.950815] [c000000071a8b940] [c008000009b94bc0] dsl_dir_rename_check+0x1e8/0x710 [zfs] [ 958.950902] [c000000071a8b9f0] [c008000009ba92b0] dsl_sync_task_sync+0xb8/0x1c0 [zfs] [ 958.950991] [c000000071a8ba30] [c008000009b97578] dsl_pool_sync+0x520/0x6a0 [zfs] [ 958.951079] [c000000071a8bb20] [c008000009bd55f4] spa_sync+0x62c/0x11f0 [zfs] [ 958.951166] [c000000071a8bc90] [c008000009c001bc] txg_sync_thread+0x2b4/0x450 [zfs] [ 958.951249] [c000000071a8bd60] [c0080000014ac2d0] thread_generic_wrapper+0x98/0xd0 [spl] [ 958.951293] [c000000071a8bda0] [c0000000001749e0] kthread+0x180/0x190 [ 958.951328] [c000000071a8be10] [c00000000000cf64] ret_from_kernel_thread+0x5c/0x64 [ 958.951382] Sending IPI to other CPUs [ 958.952928] IPI complete [ 958.955298] kexec: Starting switchover sequence. ```
defect
zts reproducibly panics on zfs rename neg on system information type version name distribution name debian distribution version kernel version architecture openzfs version describe the problem you re observing on trying to put the pr through its paces i hit a kernel panic on zfs rename neg confused i tried running it against vanilla git and lo i had two panics describe how to reproduce the problem scripts zfs tests sh t zfs rename on a system afaict include any warning errors backtraces from the system logs synth uevent devices vio failed to send uevent vio vio uevent failed to send synthetic uevent synth uevent devices vio failed to send uevent vio vio uevent failed to send synthetic uevent synth uevent devices vio failed to send uevent vio vio uevent failed to send synthetic uevent synth uevent devices vio failed to send uevent vio vio uevent failed to send synthetic uevent synth uevent devices vio failed to send uevent vio vio uevent failed to send synthetic uevent synth uevent devices vio failed to send uevent vio vio uevent failed to send synthetic uevent kernel panic not syncing corrupted stack end detected inside scheduler cpu pid comm txg sync kdump loaded tainted p oe pristine call trace dump stack lvl unreliable panic schedule cond resched down read dnode hold impl dmu bonus hold dsl dataset hold obj dsl dataset hold obj dmu objset find dp impl dmu objset find dp impl dmu objset find dp impl dmu objset find dp impl dmu objset find dp impl dmu objset find dp impl dmu objset find dp impl dmu objset find dp impl dmu objset find dp dsl dir rename check dsl sync task sync dsl pool sync spa sync txg sync thread thread generic wrapper kthread ret from kernel thread sending ipi to other cpus ipi complete kexec starting switchover sequence
1
66,545
20,267,675,643
IssuesEvent
2022-02-15 13:36:35
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
New search experience can't find metaspaces
T-Defect S-Minor A-Spaces O-Occasional Z-Community-Testing A-Spotlight-Search
> I tried searching for People metapace, but it didn't come up.
1.0
New search experience can't find metaspaces - > I tried searching for People metapace, but it didn't come up.
defect
new search experience can t find metaspaces i tried searching for people metapace but it didn t come up
1
498,308
14,405,504,205
IssuesEvent
2020-12-03 18:48:27
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
opened
POST FormData Failure With cy.request
internal-priority
### Description: Cypress currently does not allow for formData to be correctly sent. GitHub Issue Created April 2018: [https://github.com/cypress-io/cypress/issues/1647](https://github.com/cypress-io/cypress/issues/1647|smart-link) Case 1: **User/UUID**: Elena Franko, Sr QA - [42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea ](https://app.redash.io/cypressio/dashboard/org-info?p_USER_ID_UUID_NAME_EMAIL_GITHUB_ID=42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea&p_ORG_ID_UUID_NAME=42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea)([Impersonate](https://dashboard.cypress.io/admin/impersonate/22361)) **Org/UUID**: New10 - [7ee9a890-d981-404a-ae03-a3580e9d4916](https://app.redash.io/cypressio/dashboard/org-info?p_USER_ID_UUID_NAME_EMAIL_GITHUB_ID=42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea&p_ORG_ID_UUID_NAME=42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea) **Plan:** Custom - Paid Plan **Email** : [HelpScout Conversation](https://secure.helpscout.net/conversation/1197425776/6877?folderId=3841065) ### Context: * New10 has had some escalations recently and are up for a renewal so we are offering hands-on assistance. * Elena is trying to send formData with a PDF using `cy.request` but her implementation is not working. She has tried other workarounds highlighted in [this issue](https://github.com/cypress-io/cypress/issues/1647) but none have helped. ﹍Initial Email:﹍ {quote}My name is Elena and I work as Quality Engineer at New10, Amsterdam. Hope you can help me with one question. I need to upload a PDF file as a FormData to one of our back-end services. Here is my attempt to do so: {quote} let formData = new FormData() let pdfFile = new jsPDF() pdfFile.text( `Dummy PDF file for testing, contract:${contract.contractType}, applicationId: ${applicationId}`, 20, 20 ) let generatedPdfFile = pdfFile.output('blob') formData.append( `dummy_applicationId_${applicationId}.pdf`, generatedPdfFile, `dummy_applicationId_${applicationId}.pdf` ) cy.request({ method: 'POST', url: `${DOCUMENT_MANAGEMENT_SERVICE_URL}/documents?identityId=${this.identityId}&companyId=${this.companyId}&applicationId=${this.applicationId}&documentType=${contract.contractType}&vaultLevel=EXTERNAL_VAULT_LEVEL&fileName=${contract.contractType}.pdf`, headers: { Accept: 'application/json', 'Content-Type': `multipart/form-data; boundary=${formData._boundary}`, Authorization: `Bearer ${this.idToken}`, }, body: formData, form: true, }) {quote} This call is successful if I use, for instance, axios to send a request, so it should be correct. Here is what I get if I send it with cy.request. The request we sent was: {quote} `Method: POST URL: https://service-name/documents?identityId=identy&companyId=company&applicationId=id&documentType=PLEDGE&vaultLevel=EXTERNAL_VAULT_LEVEL&fileName=PLEDGE.pdf Headers: { "Connection": "keep-alive", "Content-Type": "application/x-www-form-urlencoded", "Authorization": "Bearer tockenRemovedForThisEmail", "user-agent": "Cypress", "accept": "application/json", "cookie": "CloudFront-Signature=signature CloudFront-Policy=someValue; CloudFront-Key-Pair-Id=someValue", "accept-encoding": "gzip, deflate", "content-length": 0 } Body: ----------------------------------------------------------- ` {quote}The response we got was: {quote} `Status: 500 - Server Error Headers: { "date": "Wed, 17 Jun 2020 11:47:07 GMT", "content-type": "application/json; charset=utf-8", "content-length": "193", "connection": "keep-alive", "vary": "Origin", "access-control-allow-origin": "false" } Body: { "statusCode": 500, "code": "FST_ERR_BAD_STATUS_CODE", "error": "Internal Server Error", "message": "FST_ERR_BAD_STATUS_CODE: Called reply with an invalid status code: FST_ERR_CTP_INVALID_MEDIA_TYPE" } ` {quote}I changes some values in the log above for a security reason. To me it seems like an empty body. Might be related to this open issue: https://github.com/cypress-io/cypress/issues/1647 Could you please let me know if I made a mistake while creating a request? {quote} ### Ask: I have given her three solutions to try in our conversation before this escalation and before we schedule a call. Now we are close to resolving whether it is a limitation of Cypress, a bug, or an implementation issue.
1.0
POST FormData Failure With cy.request - ### Description: Cypress currently does not allow for formData to be correctly sent. GitHub Issue Created April 2018: [https://github.com/cypress-io/cypress/issues/1647](https://github.com/cypress-io/cypress/issues/1647|smart-link) Case 1: **User/UUID**: Elena Franko, Sr QA - [42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea ](https://app.redash.io/cypressio/dashboard/org-info?p_USER_ID_UUID_NAME_EMAIL_GITHUB_ID=42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea&p_ORG_ID_UUID_NAME=42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea)([Impersonate](https://dashboard.cypress.io/admin/impersonate/22361)) **Org/UUID**: New10 - [7ee9a890-d981-404a-ae03-a3580e9d4916](https://app.redash.io/cypressio/dashboard/org-info?p_USER_ID_UUID_NAME_EMAIL_GITHUB_ID=42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea&p_ORG_ID_UUID_NAME=42b5e8d3-514c-4aa7-b4e8-c8bcb8e919ea) **Plan:** Custom - Paid Plan **Email** : [HelpScout Conversation](https://secure.helpscout.net/conversation/1197425776/6877?folderId=3841065) ### Context: * New10 has had some escalations recently and are up for a renewal so we are offering hands-on assistance. * Elena is trying to send formData with a PDF using `cy.request` but her implementation is not working. She has tried other workarounds highlighted in [this issue](https://github.com/cypress-io/cypress/issues/1647) but none have helped. ﹍Initial Email:﹍ {quote}My name is Elena and I work as Quality Engineer at New10, Amsterdam. Hope you can help me with one question. I need to upload a PDF file as a FormData to one of our back-end services. Here is my attempt to do so: {quote} let formData = new FormData() let pdfFile = new jsPDF() pdfFile.text( `Dummy PDF file for testing, contract:${contract.contractType}, applicationId: ${applicationId}`, 20, 20 ) let generatedPdfFile = pdfFile.output('blob') formData.append( `dummy_applicationId_${applicationId}.pdf`, generatedPdfFile, `dummy_applicationId_${applicationId}.pdf` ) cy.request({ method: 'POST', url: `${DOCUMENT_MANAGEMENT_SERVICE_URL}/documents?identityId=${this.identityId}&companyId=${this.companyId}&applicationId=${this.applicationId}&documentType=${contract.contractType}&vaultLevel=EXTERNAL_VAULT_LEVEL&fileName=${contract.contractType}.pdf`, headers: { Accept: 'application/json', 'Content-Type': `multipart/form-data; boundary=${formData._boundary}`, Authorization: `Bearer ${this.idToken}`, }, body: formData, form: true, }) {quote} This call is successful if I use, for instance, axios to send a request, so it should be correct. Here is what I get if I send it with cy.request. The request we sent was: {quote} `Method: POST URL: https://service-name/documents?identityId=identy&companyId=company&applicationId=id&documentType=PLEDGE&vaultLevel=EXTERNAL_VAULT_LEVEL&fileName=PLEDGE.pdf Headers: { "Connection": "keep-alive", "Content-Type": "application/x-www-form-urlencoded", "Authorization": "Bearer tockenRemovedForThisEmail", "user-agent": "Cypress", "accept": "application/json", "cookie": "CloudFront-Signature=signature CloudFront-Policy=someValue; CloudFront-Key-Pair-Id=someValue", "accept-encoding": "gzip, deflate", "content-length": 0 } Body: ----------------------------------------------------------- ` {quote}The response we got was: {quote} `Status: 500 - Server Error Headers: { "date": "Wed, 17 Jun 2020 11:47:07 GMT", "content-type": "application/json; charset=utf-8", "content-length": "193", "connection": "keep-alive", "vary": "Origin", "access-control-allow-origin": "false" } Body: { "statusCode": 500, "code": "FST_ERR_BAD_STATUS_CODE", "error": "Internal Server Error", "message": "FST_ERR_BAD_STATUS_CODE: Called reply with an invalid status code: FST_ERR_CTP_INVALID_MEDIA_TYPE" } ` {quote}I changes some values in the log above for a security reason. To me it seems like an empty body. Might be related to this open issue: https://github.com/cypress-io/cypress/issues/1647 Could you please let me know if I made a mistake while creating a request? {quote} ### Ask: I have given her three solutions to try in our conversation before this escalation and before we schedule a call. Now we are close to resolving whether it is a limitation of Cypress, a bug, or an implementation issue.
non_defect
post formdata failure with cy request description cypress currently does not allow for formdata to be correctly sent github issue created april case user uuid elena franko sr qa org uuid plan custom paid plan email context has had some escalations recently and are up for a renewal so we are offering hands on assistance elena is trying to send formdata with a pdf using cy request but her implementation is not working she has tried other workarounds highlighted in but none have helped ﹍initial email ﹍ quote my name is elena and i work as quality engineer at amsterdam hope you can help me with one question i need to upload a pdf file as a formdata to one of our back end services here is my attempt to do so quote let formdata new formdata let pdffile new jspdf pdffile text dummy pdf file for testing contract contract contracttype applicationid applicationid let generatedpdffile pdffile output blob formdata append dummy applicationid applicationid pdf generatedpdffile dummy applicationid applicationid pdf cy request method post url document management service url documents identityid this identityid companyid this companyid applicationid this applicationid documenttype contract contracttype vaultlevel external vault level filename contract contracttype pdf headers accept application json content type multipart form data boundary formdata boundary authorization bearer this idtoken body formdata form true quote this call is successful if i use for instance axios to send a request so it should be correct here is what i get if i send it with cy request the request we sent was quote method post url headers connection keep alive content type application x www form urlencoded authorization bearer tockenremovedforthisemail user agent cypress accept application json cookie cloudfront signature signature cloudfront policy somevalue cloudfront key pair id somevalue accept encoding gzip deflate content length body quote the response we got was quote status server error headers date wed jun gmt content type application json charset utf content length connection keep alive vary origin access control allow origin false body statuscode code fst err bad status code error internal server error message fst err bad status code called reply with an invalid status code fst err ctp invalid media type quote i changes some values in the log above for a security reason to me it seems like an empty body might be related to this open issue could you please let me know if i made a mistake while creating a request quote ask i have given her three solutions to try in our conversation before this escalation and before we schedule a call now we are close to resolving whether it is a limitation of cypress a bug or an implementation issue
0
22,773
2,650,920,825
IssuesEvent
2015-03-16 06:45:56
grepper/tovid
https://api.github.com/repos/grepper/tovid
opened
Source video rotation
enhancement imported Priority-Low
_From [just...@gmail.com](https://code.google.com/u/118426383622680449865/) on September 26, 2007 08:07:46_ Modern photocameras automatically detect the orientation of photo (portrait or landscape) and embed this information in produced JPGs. Unfortunately QucikTime .movs do not support this feature so if you rotate your camera during filming you end up with rotated video file. It's easy to watch this on PC (with mplayer -vf rotate=1), but not so easy to convert to VCD/DVD - as you cannot simply add this filter to mencode options, cause tovid is expecting different resulting video resolution and fails. _Original issue: http://code.google.com/p/tovid/issues/detail?id=12_
1.0
Source video rotation - _From [just...@gmail.com](https://code.google.com/u/118426383622680449865/) on September 26, 2007 08:07:46_ Modern photocameras automatically detect the orientation of photo (portrait or landscape) and embed this information in produced JPGs. Unfortunately QucikTime .movs do not support this feature so if you rotate your camera during filming you end up with rotated video file. It's easy to watch this on PC (with mplayer -vf rotate=1), but not so easy to convert to VCD/DVD - as you cannot simply add this filter to mencode options, cause tovid is expecting different resulting video resolution and fails. _Original issue: http://code.google.com/p/tovid/issues/detail?id=12_
non_defect
source video rotation from on september modern photocameras automatically detect the orientation of photo portrait or landscape and embed this information in produced jpgs unfortunately quciktime movs do not support this feature so if you rotate your camera during filming you end up with rotated video file it s easy to watch this on pc with mplayer vf rotate but not so easy to convert to vcd dvd as you cannot simply add this filter to mencode options cause tovid is expecting different resulting video resolution and fails original issue
0
20,770
11,502,221,028
IssuesEvent
2020-02-12 18:41:12
aws/aws-sdk-ruby
https://api.github.com/repos/aws/aws-sdk-ruby
closed
Aws::CognitoIdentityProvider::Client initiate_auth exposes user password in logs
service-api
Please fill out the sections below to help us address your issue ### Issue description In a rails 6 application, I'm using Cognito to manage user authentication. I recently noticed that calls' logs are automatically added to the logs but there is an issue - Cognito initiate_auth method with `auth_flow: "USER_PASSWORD_AUTH"` is exposing the password provided by the user: ``` [Aws::CognitoIdentityProvider::Client 200 1.012806 0 retries] initiate_auth(client_id:"[FILTERED]",auth_flow:"USER_PASSWORD_AUTH",auth_parameters:{"USERNAME"=>"test@gmail.com","PASSWORD"=>"Test1234!"}) ``` ### Gem name ('aws-sdk', 'aws-sdk-resources' or service gems like 'aws-sdk-s3') and its version aws-sdk-core 3.67.0 aws-sdk-cognitoidentityprovider 1.25.0 ### Version of Ruby, OS environment ruby 2.6.3 Mac OS 10.13.6 ### Code snippets / steps to reproduce my code: ``` COGNITO_CLIENT = Aws::CognitoIdentityProvider::Client.new COGNITO_CLIENT.initiate_auth( client_id: ENV['AWS_COGNITO_CLIENT_ID'], auth_flow: 'USER_PASSWORD_AUTH', auth_parameters: { 'USERNAME' => username, 'PASSWORD' => password } ) ``` The issue is related to the filter methods in https://github.com/aws/aws-sdk-ruby/blob/master/gems/aws-sdk-core/lib/aws-sdk-core/log/param_filter.rb ``` def filter_hash(values) filtered = {} values.each_pair do |key, value| filtered[key] = @filters.include?(key) ? '[FILTERED]' : filter(value) end filtered end ``` During parsing `:auth_parameters` hash keys of this hash are actually string written in the capital letters, so they do not match symbols form `SENSITIVE` constant. My suggestion is to include `'PASSWORD'` or `:auth_parameters` in `SENSITIVE` constant or to add some parsing to the key in this method to make it more universal like: ``` def filter_hash(values) filtered = {} values.each_pair do |key, value| key = key.downcase.to_sym filtered[key] = @filters.include?(key) ? '[FILTERED]' : filter(value) end filtered end ``` This change produces log like: ``` [Aws::CognitoIdentityProvider::Client 200 0.836941 0 retries] initiate_auth(client_id:"[FILTERED]",auth_flow:"USER_PASSWORD_AUTH",auth_parameters:{username:"[FILTERED]",password:"[FILTERED]"}) ```
1.0
Aws::CognitoIdentityProvider::Client initiate_auth exposes user password in logs - Please fill out the sections below to help us address your issue ### Issue description In a rails 6 application, I'm using Cognito to manage user authentication. I recently noticed that calls' logs are automatically added to the logs but there is an issue - Cognito initiate_auth method with `auth_flow: "USER_PASSWORD_AUTH"` is exposing the password provided by the user: ``` [Aws::CognitoIdentityProvider::Client 200 1.012806 0 retries] initiate_auth(client_id:"[FILTERED]",auth_flow:"USER_PASSWORD_AUTH",auth_parameters:{"USERNAME"=>"test@gmail.com","PASSWORD"=>"Test1234!"}) ``` ### Gem name ('aws-sdk', 'aws-sdk-resources' or service gems like 'aws-sdk-s3') and its version aws-sdk-core 3.67.0 aws-sdk-cognitoidentityprovider 1.25.0 ### Version of Ruby, OS environment ruby 2.6.3 Mac OS 10.13.6 ### Code snippets / steps to reproduce my code: ``` COGNITO_CLIENT = Aws::CognitoIdentityProvider::Client.new COGNITO_CLIENT.initiate_auth( client_id: ENV['AWS_COGNITO_CLIENT_ID'], auth_flow: 'USER_PASSWORD_AUTH', auth_parameters: { 'USERNAME' => username, 'PASSWORD' => password } ) ``` The issue is related to the filter methods in https://github.com/aws/aws-sdk-ruby/blob/master/gems/aws-sdk-core/lib/aws-sdk-core/log/param_filter.rb ``` def filter_hash(values) filtered = {} values.each_pair do |key, value| filtered[key] = @filters.include?(key) ? '[FILTERED]' : filter(value) end filtered end ``` During parsing `:auth_parameters` hash keys of this hash are actually string written in the capital letters, so they do not match symbols form `SENSITIVE` constant. My suggestion is to include `'PASSWORD'` or `:auth_parameters` in `SENSITIVE` constant or to add some parsing to the key in this method to make it more universal like: ``` def filter_hash(values) filtered = {} values.each_pair do |key, value| key = key.downcase.to_sym filtered[key] = @filters.include?(key) ? '[FILTERED]' : filter(value) end filtered end ``` This change produces log like: ``` [Aws::CognitoIdentityProvider::Client 200 0.836941 0 retries] initiate_auth(client_id:"[FILTERED]",auth_flow:"USER_PASSWORD_AUTH",auth_parameters:{username:"[FILTERED]",password:"[FILTERED]"}) ```
non_defect
aws cognitoidentityprovider client initiate auth exposes user password in logs please fill out the sections below to help us address your issue issue description in a rails application i m using cognito to manage user authentication i recently noticed that calls logs are automatically added to the logs but there is an issue cognito initiate auth method with auth flow user password auth is exposing the password provided by the user initiate auth client id auth flow user password auth auth parameters username test gmail com password gem name aws sdk aws sdk resources or service gems like aws sdk and its version aws sdk core aws sdk cognitoidentityprovider version of ruby os environment ruby mac os code snippets steps to reproduce my code cognito client aws cognitoidentityprovider client new cognito client initiate auth client id env auth flow user password auth auth parameters username username password password the issue is related to the filter methods in def filter hash values filtered values each pair do key value filtered filters include key filter value end filtered end during parsing auth parameters hash keys of this hash are actually string written in the capital letters so they do not match symbols form sensitive constant my suggestion is to include password or auth parameters in sensitive constant or to add some parsing to the key in this method to make it more universal like def filter hash values filtered values each pair do key value key key downcase to sym filtered filters include key filter value end filtered end this change produces log like initiate auth client id auth flow user password auth auth parameters username password
0
37,253
8,309,738,540
IssuesEvent
2018-09-24 08:07:20
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
SelectCheckboxMenu: disabled items are shown as selected labels if multiple=true
defect
## 1) Environment - PrimeFaces version: 6.2 - Does it work on the newest released PrimeFaces version? Version? No, 6.2 - Does it work on the newest sources in GitHub? (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source) No - Application server + version: Wildfly 11, Jetty - Affected browsers: FF, Chrome, ## 2) Expected behavior Having a SelectCheckboxMenu with disabled SelectItems and multible=true configured. If I click the first checkbox for selecting all items I expect that only the labels of the not disabled items are shown. ## 3) Actual behavior All items are shown. ## 4) Steps to reproduce Just alter the SelectCheckboxMenu example from the showcase to have disabled SelectItems and multible=true configured. ## 5) Sample XHTML ``` <p:selectCheckboxMenu id="multiple" value="#{checkboxView.selectedCities}" label="Cities" multiple="true" filter="true" filterMatchMode="startsWith" panelStyle="width:250px"> <f:selectItems value="#{checkboxView.cities}" /> </p:selectCheckboxMenu> ``` ## 6) Sample bean ``` @ManagedBean @ViewScoped public class CheckboxView { private String[] selectedCities; private List<SelectItem> cities; @PostConstruct public void init() { cities = new ArrayList<>(); add(cities, "Miami"); add(cities, "London"); add(cities, "Paris"); add(cities, "Istanbul"); add(cities, "Berlin"); add(cities, "Barcelona"); add(cities, "Rome"); add(cities, "Brasilia"); add(cities, "Amsterdam"); } private void add(List<SelectItem> list, String label) { list.add(new SelectItem(label, label, null, (((list.size()) %2) == 1))); } public String[] getSelectedCities() { return selectedCities; } public void setSelectedCities(String[] selectedCities) { this.selectedCities = selectedCities; } public List<SelectItem> getCities() { return cities; } } ``` Screenshot: ![selectcheckboxmenumultiple](https://user-images.githubusercontent.com/11403755/45802368-e6f9cb00-bcb6-11e8-89e1-69f4a92d99bc.png)
1.0
SelectCheckboxMenu: disabled items are shown as selected labels if multiple=true - ## 1) Environment - PrimeFaces version: 6.2 - Does it work on the newest released PrimeFaces version? Version? No, 6.2 - Does it work on the newest sources in GitHub? (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source) No - Application server + version: Wildfly 11, Jetty - Affected browsers: FF, Chrome, ## 2) Expected behavior Having a SelectCheckboxMenu with disabled SelectItems and multible=true configured. If I click the first checkbox for selecting all items I expect that only the labels of the not disabled items are shown. ## 3) Actual behavior All items are shown. ## 4) Steps to reproduce Just alter the SelectCheckboxMenu example from the showcase to have disabled SelectItems and multible=true configured. ## 5) Sample XHTML ``` <p:selectCheckboxMenu id="multiple" value="#{checkboxView.selectedCities}" label="Cities" multiple="true" filter="true" filterMatchMode="startsWith" panelStyle="width:250px"> <f:selectItems value="#{checkboxView.cities}" /> </p:selectCheckboxMenu> ``` ## 6) Sample bean ``` @ManagedBean @ViewScoped public class CheckboxView { private String[] selectedCities; private List<SelectItem> cities; @PostConstruct public void init() { cities = new ArrayList<>(); add(cities, "Miami"); add(cities, "London"); add(cities, "Paris"); add(cities, "Istanbul"); add(cities, "Berlin"); add(cities, "Barcelona"); add(cities, "Rome"); add(cities, "Brasilia"); add(cities, "Amsterdam"); } private void add(List<SelectItem> list, String label) { list.add(new SelectItem(label, label, null, (((list.size()) %2) == 1))); } public String[] getSelectedCities() { return selectedCities; } public void setSelectedCities(String[] selectedCities) { this.selectedCities = selectedCities; } public List<SelectItem> getCities() { return cities; } } ``` Screenshot: ![selectcheckboxmenumultiple](https://user-images.githubusercontent.com/11403755/45802368-e6f9cb00-bcb6-11e8-89e1-69f4a92d99bc.png)
defect
selectcheckboxmenu disabled items are shown as selected labels if multiple true environment primefaces version does it work on the newest released primefaces version version no does it work on the newest sources in github build by source no application server version wildfly jetty affected browsers ff chrome expected behavior having a selectcheckboxmenu with disabled selectitems and multible true configured if i click the first checkbox for selecting all items i expect that only the labels of the not disabled items are shown actual behavior all items are shown steps to reproduce just alter the selectcheckboxmenu example from the showcase to have disabled selectitems and multible true configured sample xhtml p selectcheckboxmenu id multiple value checkboxview selectedcities label cities multiple true filter true filtermatchmode startswith panelstyle width sample bean managedbean viewscoped public class checkboxview private string selectedcities private list cities postconstruct public void init cities new arraylist add cities miami add cities london add cities paris add cities istanbul add cities berlin add cities barcelona add cities rome add cities brasilia add cities amsterdam private void add list list string label list add new selectitem label label null list size public string getselectedcities return selectedcities public void setselectedcities string selectedcities this selectedcities selectedcities public list getcities return cities screenshot
1
26,118
4,593,613,302
IssuesEvent
2016-09-21 02:02:04
afisher1/GridLAB-D
https://api.github.com/repos/afisher1/GridLAB-D
closed
#41 Implement variables using current naming convention and units,
defect
Not all object classes use the current naming convention of 'description[unit]'. ,
1.0
#41 Implement variables using current naming convention and units, - Not all object classes use the current naming convention of 'description[unit]'. ,
defect
implement variables using current naming convention and units not all object classes use the current naming convention of description
1
55,045
14,146,723,306
IssuesEvent
2020-11-10 19:39:51
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Schedule: Locales not fully respected
defect
All European locales have first day of the week Monday [[link](http://chartsbin.com/view/41671)]. But e.g. Norwegian is stariting by Sunday (søn.) in `p:schedule` component, which is wrong. All, except EN schould be Mon: | PF version | EN | DE | SV | NO | |:------------|:-----| :-----|:-----|:-----| | 8.0 | Sun | Mon | Mon | **Sun** | | 8.0.2 | Sun | Mon | Mon | **Sun** | | 8.0.5 | Sun | Mon | Mon | **Sun** | | snapshot* | Sun | **Sun** | **Sun** | **Sun** | (*) primefaces-master-6_1-gf9d6444-4838.jar * FullCalendar demo is working correctly: https://fullcalendar.io/docs/locale-demo. E.g if I choose "nb" locale (nb - Norwegian Bokmål, nn - Norwegian Nynorsk), then week starts by monday (man.). * Setting of `p:schedule weekNumberCalculation` to "ISO" does not help, even if FullCalendar is saying "If weekNumberCalculation is set to 'ISO', this option defaults to 1 (Monday)". https://fullcalendar.io/docs/firstDay **XHTML** ```html <p:schedule id="myCalendarId" view="agendaWeek" rightHeaderTemplate="false" value="#{scheduleController.eventModel}" locale="#{scheduleController.locale}" timeZone="#{scheduleController.timeZone}" clientTimeZone="#{scheduleController.clientTimeZone}" weekNumberCalculation="local" showWeekends="true" allDaySlot="false" aspectRatio="1.2" timeFormat="HH:mm"> </p:schedule> ``` **ScheduleController.java** ```java private static final String clientTimeZone = "local"; private static final Locale LOCALE_EN_US = new Locale("en", "US"); private static final Locale LOCALE_DE_DE = new Locale("de", "DE"); private static final Locale LOCALE_SV_SE = new Locale("sv", "SE"); private static final Locale LOCALE_NO_NO = new Locale("no", "NO"); public static final ZoneId ZONE_ID_UTC = ZoneId.of("UTC"); public static final ZoneId ZONE_ID_BERLIN = ZoneId.of("Europe/Berlin"); public static final ZoneId ZONE_ID_STOCKHOLM = ZoneId.of("Europe/Stockholm"); public static final ZoneId ZONE_ID_OSLO = ZoneId.of("Europe/Oslo"); ```
1.0
Schedule: Locales not fully respected - All European locales have first day of the week Monday [[link](http://chartsbin.com/view/41671)]. But e.g. Norwegian is stariting by Sunday (søn.) in `p:schedule` component, which is wrong. All, except EN schould be Mon: | PF version | EN | DE | SV | NO | |:------------|:-----| :-----|:-----|:-----| | 8.0 | Sun | Mon | Mon | **Sun** | | 8.0.2 | Sun | Mon | Mon | **Sun** | | 8.0.5 | Sun | Mon | Mon | **Sun** | | snapshot* | Sun | **Sun** | **Sun** | **Sun** | (*) primefaces-master-6_1-gf9d6444-4838.jar * FullCalendar demo is working correctly: https://fullcalendar.io/docs/locale-demo. E.g if I choose "nb" locale (nb - Norwegian Bokmål, nn - Norwegian Nynorsk), then week starts by monday (man.). * Setting of `p:schedule weekNumberCalculation` to "ISO" does not help, even if FullCalendar is saying "If weekNumberCalculation is set to 'ISO', this option defaults to 1 (Monday)". https://fullcalendar.io/docs/firstDay **XHTML** ```html <p:schedule id="myCalendarId" view="agendaWeek" rightHeaderTemplate="false" value="#{scheduleController.eventModel}" locale="#{scheduleController.locale}" timeZone="#{scheduleController.timeZone}" clientTimeZone="#{scheduleController.clientTimeZone}" weekNumberCalculation="local" showWeekends="true" allDaySlot="false" aspectRatio="1.2" timeFormat="HH:mm"> </p:schedule> ``` **ScheduleController.java** ```java private static final String clientTimeZone = "local"; private static final Locale LOCALE_EN_US = new Locale("en", "US"); private static final Locale LOCALE_DE_DE = new Locale("de", "DE"); private static final Locale LOCALE_SV_SE = new Locale("sv", "SE"); private static final Locale LOCALE_NO_NO = new Locale("no", "NO"); public static final ZoneId ZONE_ID_UTC = ZoneId.of("UTC"); public static final ZoneId ZONE_ID_BERLIN = ZoneId.of("Europe/Berlin"); public static final ZoneId ZONE_ID_STOCKHOLM = ZoneId.of("Europe/Stockholm"); public static final ZoneId ZONE_ID_OSLO = ZoneId.of("Europe/Oslo"); ```
defect
schedule locales not fully respected all european locales have first day of the week monday but e g norwegian is stariting by sunday søn in p schedule component which is wrong all except en schould be mon pf version en de sv no sun mon mon sun sun mon mon sun sun mon mon sun snapshot sun sun sun sun primefaces master jar fullcalendar demo is working correctly e g if i choose nb locale nb norwegian bokmål nn norwegian nynorsk then week starts by monday man setting of p schedule weeknumbercalculation to iso does not help even if fullcalendar is saying if weeknumbercalculation is set to iso this option defaults to monday xhtml html p schedule id mycalendarid view agendaweek rightheadertemplate false value schedulecontroller eventmodel locale schedulecontroller locale timezone schedulecontroller timezone clienttimezone schedulecontroller clienttimezone weeknumbercalculation local showweekends true alldayslot false aspectratio timeformat hh mm schedulecontroller java java private static final string clienttimezone local private static final locale locale en us new locale en us private static final locale locale de de new locale de de private static final locale locale sv se new locale sv se private static final locale locale no no new locale no no public static final zoneid zone id utc zoneid of utc public static final zoneid zone id berlin zoneid of europe berlin public static final zoneid zone id stockholm zoneid of europe stockholm public static final zoneid zone id oslo zoneid of europe oslo
1
87,936
15,790,359,327
IssuesEvent
2021-04-02 01:14:37
SmartBear/ready-mqtt-plugin
https://api.github.com/repos/SmartBear/ready-mqtt-plugin
opened
WS-2019-0509 (Medium) detected in bcprov-jdk15on-1.64.jar
security vulnerability
## WS-2019-0509 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.64.jar</b></p></summary> <p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for Java 5 to Java 11.</p> <p>Library home page: <a href="https://www.bouncycastle.org/java.html">https://www.bouncycastle.org/java.html</a></p> <p>Path to dependency file: ready-mqtt-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.64/bcprov-jdk15on-1.64.jar</p> <p> Dependency Hierarchy: - bcpkix-jdk15on-1.64.jar (Root Library) - :x: **bcprov-jdk15on-1.64.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Bouncy Castle through 1.68 is vulnerable to Denial of Service (DoS). The Dump.class file utilizes the FileInputStream object when reading from user-provided files, but doesn't close these streams properly. <p>Publish Date: 2019-12-24 <p>URL: <a href=https://github.com/bcgit/bc-java/issues/634>WS-2019-0509</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.bouncycastle","packageName":"bcprov-jdk15on","packageVersion":"1.64","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.bouncycastle:bcpkix-jdk15on:1.64;org.bouncycastle:bcprov-jdk15on:1.64","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0509","vulnerabilityDetails":"Bouncy Castle through 1.68 is vulnerable to Denial of Service (DoS). The Dump.class file utilizes the FileInputStream object when reading from user-provided files, but doesn\u0027t close these streams properly.","vulnerabilityUrl":"https://github.com/bcgit/bc-java/issues/634","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0509 (Medium) detected in bcprov-jdk15on-1.64.jar - ## WS-2019-0509 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.64.jar</b></p></summary> <p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for Java 5 to Java 11.</p> <p>Library home page: <a href="https://www.bouncycastle.org/java.html">https://www.bouncycastle.org/java.html</a></p> <p>Path to dependency file: ready-mqtt-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.64/bcprov-jdk15on-1.64.jar</p> <p> Dependency Hierarchy: - bcpkix-jdk15on-1.64.jar (Root Library) - :x: **bcprov-jdk15on-1.64.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Bouncy Castle through 1.68 is vulnerable to Denial of Service (DoS). The Dump.class file utilizes the FileInputStream object when reading from user-provided files, but doesn't close these streams properly. <p>Publish Date: 2019-12-24 <p>URL: <a href=https://github.com/bcgit/bc-java/issues/634>WS-2019-0509</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.bouncycastle","packageName":"bcprov-jdk15on","packageVersion":"1.64","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.bouncycastle:bcpkix-jdk15on:1.64;org.bouncycastle:bcprov-jdk15on:1.64","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2019-0509","vulnerabilityDetails":"Bouncy Castle through 1.68 is vulnerable to Denial of Service (DoS). The Dump.class file utilizes the FileInputStream object when reading from user-provided files, but doesn\u0027t close these streams properly.","vulnerabilityUrl":"https://github.com/bcgit/bc-java/issues/634","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
ws medium detected in bcprov jar ws medium severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for java to java library home page a href path to dependency file ready mqtt plugin pom xml path to vulnerable library home wss scanner repository org bouncycastle bcprov bcprov jar dependency hierarchy bcpkix jar root library x bcprov jar vulnerable library found in base branch master vulnerability details bouncy castle through is vulnerable to denial of service dos the dump class file utilizes the fileinputstream object when reading from user provided files but doesn t close these streams properly publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org bouncycastle bcpkix org bouncycastle bcprov isminimumfixversionavailable false basebranches vulnerabilityidentifier ws vulnerabilitydetails bouncy castle through is vulnerable to denial of service dos the dump class file utilizes the fileinputstream object when reading from user provided files but doesn close these streams properly vulnerabilityurl
0
213,957
7,261,822,772
IssuesEvent
2018-02-19 00:37:32
leo-project/leofs
https://api.github.com/repos/leo-project/leofs
reopened
[all] Handle supervisor restart properly
Priority-HIGH survey v1.4
As reported on https://github.com/leo-project/leofs/issues/966#issuecomment-358582625, the restarts of leo_mq_sup or leo_object_storage_sup might cause the supervisor process tree to be broken (part of worker processes failed to restart) the below is the crash.log when that happened. ```erlang 2018-01-10 20:23:50 =SUPERVISOR REPORT==== Supervisor: {local,leo_storage_sup} Context: child_terminated Reason: shutdown Offender: [{pid,<0.111.0>},{id,leo_object_storage_sup},{mfargs,{leo_object_storage_sup,start_link,[[{64,"/mnt/avs1/bodies"},{64,"/mnt/avs2/bodies"},{64,"/mnt/avs3/bodies"},{64,"/mnt/avs4/bodies"}],leo_storage_msg_collector]}},{restart_type,permanent},{shutdown,2000},{child_type,supervisor}] 2018-01-10 20:23:50 =ERROR REPORT==== {module,"leo_backend_db_eleveldb"},{function,"open/1"},{line,96},{body,{db_open,"IO error: lock /mnt/avs1/bodies/metadata/63/LOCK: already held by process"}} 2018-01-10 20:23:50 =CRASH REPORT==== crasher: initial call: leo_backend_db_server:init/1 pid: <0.17246.1505> registered_name: [] exception exit: {{db_open,"IO error: lock /mnt/avs1/bodies/metadata/63/LOCK: already held by process"},[{gen_server,init_it,6,[{file,"gen_server.erl"},{line,344}]},{proc_lib,init_p_do_apply,3,[{file,"proc_lib.erl"},{line,247}]}]} ancestors: [leo_backend_db_sup,leo_object_storage_sup,leo_storage_sup,<0.46.0>] messages: [] links: [<0.17234.1505>] dictionary: [] trap_exit: false status: running heap_size: 987 stack_size: 27 reductions: 371 neighbours: 2018-01-10 20:23:50 =ERROR REPORT==== {module,"leo_backend_db_eleveldb"},{function,"open/1"},{line,96},{body,{db_open,"IO error: lock /mnt/avs2/bodies/metadata/10063/LOCK: already held by process"}} 2018-01-10 20:23:50 =CRASH REPORT==== crasher: initial call: leo_backend_db_server:init/1 pid: <0.17237.1505> registered_name: [] exception exit: {{db_open,"IO error: lock /mnt/avs2/bodies/metadata/10063/LOCK: already held by process"},[{gen_server,init_it,6,[{file,"gen_server.erl"},{line,344}]},{proc_lib,init_p_do_apply,3,[{file,"proc_lib.erl"},{line,247}]}]} ancestors: [leo_backend_db_sup,leo_object_storage_sup,leo_storage_sup,<0.46.0>] messages: [] links: [<0.17234.1505>] dictionary: [] trap_exit: false status: running heap_size: 987 stack_size: 27 reductions: 347 neighbours: ``` Just in case, we have to test/check if all supervisor restarts work as expected.
1.0
[all] Handle supervisor restart properly - As reported on https://github.com/leo-project/leofs/issues/966#issuecomment-358582625, the restarts of leo_mq_sup or leo_object_storage_sup might cause the supervisor process tree to be broken (part of worker processes failed to restart) the below is the crash.log when that happened. ```erlang 2018-01-10 20:23:50 =SUPERVISOR REPORT==== Supervisor: {local,leo_storage_sup} Context: child_terminated Reason: shutdown Offender: [{pid,<0.111.0>},{id,leo_object_storage_sup},{mfargs,{leo_object_storage_sup,start_link,[[{64,"/mnt/avs1/bodies"},{64,"/mnt/avs2/bodies"},{64,"/mnt/avs3/bodies"},{64,"/mnt/avs4/bodies"}],leo_storage_msg_collector]}},{restart_type,permanent},{shutdown,2000},{child_type,supervisor}] 2018-01-10 20:23:50 =ERROR REPORT==== {module,"leo_backend_db_eleveldb"},{function,"open/1"},{line,96},{body,{db_open,"IO error: lock /mnt/avs1/bodies/metadata/63/LOCK: already held by process"}} 2018-01-10 20:23:50 =CRASH REPORT==== crasher: initial call: leo_backend_db_server:init/1 pid: <0.17246.1505> registered_name: [] exception exit: {{db_open,"IO error: lock /mnt/avs1/bodies/metadata/63/LOCK: already held by process"},[{gen_server,init_it,6,[{file,"gen_server.erl"},{line,344}]},{proc_lib,init_p_do_apply,3,[{file,"proc_lib.erl"},{line,247}]}]} ancestors: [leo_backend_db_sup,leo_object_storage_sup,leo_storage_sup,<0.46.0>] messages: [] links: [<0.17234.1505>] dictionary: [] trap_exit: false status: running heap_size: 987 stack_size: 27 reductions: 371 neighbours: 2018-01-10 20:23:50 =ERROR REPORT==== {module,"leo_backend_db_eleveldb"},{function,"open/1"},{line,96},{body,{db_open,"IO error: lock /mnt/avs2/bodies/metadata/10063/LOCK: already held by process"}} 2018-01-10 20:23:50 =CRASH REPORT==== crasher: initial call: leo_backend_db_server:init/1 pid: <0.17237.1505> registered_name: [] exception exit: {{db_open,"IO error: lock /mnt/avs2/bodies/metadata/10063/LOCK: already held by process"},[{gen_server,init_it,6,[{file,"gen_server.erl"},{line,344}]},{proc_lib,init_p_do_apply,3,[{file,"proc_lib.erl"},{line,247}]}]} ancestors: [leo_backend_db_sup,leo_object_storage_sup,leo_storage_sup,<0.46.0>] messages: [] links: [<0.17234.1505>] dictionary: [] trap_exit: false status: running heap_size: 987 stack_size: 27 reductions: 347 neighbours: ``` Just in case, we have to test/check if all supervisor restarts work as expected.
non_defect
handle supervisor restart properly as reported on the restarts of leo mq sup or leo object storage sup might cause the supervisor process tree to be broken part of worker processes failed to restart the below is the crash log when that happened erlang supervisor report supervisor local leo storage sup context child terminated reason shutdown offender leo storage msg collector restart type permanent shutdown child type supervisor error report module leo backend db eleveldb function open line body db open io error lock mnt bodies metadata lock already held by process crash report crasher initial call leo backend db server init pid registered name exception exit db open io error lock mnt bodies metadata lock already held by process proc lib init p do apply ancestors messages links dictionary trap exit false status running heap size stack size reductions neighbours error report module leo backend db eleveldb function open line body db open io error lock mnt bodies metadata lock already held by process crash report crasher initial call leo backend db server init pid registered name exception exit db open io error lock mnt bodies metadata lock already held by process proc lib init p do apply ancestors messages links dictionary trap exit false status running heap size stack size reductions neighbours just in case we have to test check if all supervisor restarts work as expected
0
96,869
12,168,550,593
IssuesEvent
2020-04-27 12:52:12
nextcloud/talk-ios
https://api.github.com/repos/nextcloud/talk-ios
opened
in-chat-view: allow toggling conversations list if space allows
design enhancement
On a spacious tablet computer, toggling the conversations list should be possible while inside a chat, just like on desktop, so while being active in multiple conversations at the same time gets better managable.
1.0
in-chat-view: allow toggling conversations list if space allows - On a spacious tablet computer, toggling the conversations list should be possible while inside a chat, just like on desktop, so while being active in multiple conversations at the same time gets better managable.
non_defect
in chat view allow toggling conversations list if space allows on a spacious tablet computer toggling the conversations list should be possible while inside a chat just like on desktop so while being active in multiple conversations at the same time gets better managable
0
31,251
6,474,303,448
IssuesEvent
2017-08-17 17:48:11
buildo/scriptoni
https://api.github.com/repos/buildo/scriptoni
closed
dev server configuration should allow browser history location
defect in review
## description currently `/route` is a not found -> not possible to use the history api with scriptoni ## how to reproduce - {optional: describe steps to reproduce defect} ## specs use https://webpack.js.org/configuration/dev-server/#devserver-historyapifallback in https://github.com/buildo/scriptoni/blob/master/src/scripts/webpack/dev.js ## misc {optional: other useful info}
1.0
dev server configuration should allow browser history location - ## description currently `/route` is a not found -> not possible to use the history api with scriptoni ## how to reproduce - {optional: describe steps to reproduce defect} ## specs use https://webpack.js.org/configuration/dev-server/#devserver-historyapifallback in https://github.com/buildo/scriptoni/blob/master/src/scripts/webpack/dev.js ## misc {optional: other useful info}
defect
dev server configuration should allow browser history location description currently route is a not found not possible to use the history api with scriptoni how to reproduce optional describe steps to reproduce defect specs use in misc optional other useful info
1
51,615
13,207,536,897
IssuesEvent
2020-08-14 23:29:51
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
cmake's FindBLAS.cmake reports BLAS_FOUND=TRUE even if dev pkgs aren't installed (Trac #707)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/707">https://code.icecube.wisc.edu/projects/icecube/ticket/707</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:22", "_ts": "1550067082284240", "description": "cmake happily finds libatlas and libblas and reports BLAS_FOUND=TRUE even though no headers are found. \n\nHere's the relevant bits from CMakeCache.txt on the effected system (Ubuntu 12.04).\n\n{{{\n//BLAS binary directory\nBLAS_BIN_DIR:PATH=\n//Tool 'BLAS' found successfully\nBLAS_FOUND:BOOL=FALSE\n//BLAS include dir\nBLAS_INCLUDE_DIR:PATH=/doesnt/exist\n//Libraries for tool BLAS\nBLAS_LIBRARIES:PATH=\nBLAS_atlas_LIBRARY:FILEPATH=/usr/lib/libatlas.so.3gf\nBLAS_f77blas_LIBRARY:FILEPATH=/usr/lib/libf77blas.so.3gf\nBLAS_goto2_LIBRARY:FILEPATH=BLAS_goto2_LIBRARY-NOTFOUND\nLAPACK_Accelerate_LIBRARY:FILEPATH=LAPACK_Accelerate_LIBRARY-NOTFOUND\n//LAPACK binary directory\nLAPACK_BIN_DIR:PATH=\n//Tool 'LAPACK' found successfully\nLAPACK_FOUND:BOOL=FALSE\n//LAPACK include dir\nLAPACK_INCLUDE_DIR:PATH=/doesnt/exist\n//Libraries for tool LAPACK\nLAPACK_LIBRARIES:PATH=\nLAPACK_acml_LIBRARY:FILEPATH=LAPACK_acml_LIBRARY-NOTFOUND\nLAPACK_acml_mp_LIBRARY:FILEPATH=LAPACK_acml_mp_LIBRARY-NOTFOUND\nLAPACK_goto2_LIBRARY:FILEPATH=LAPACK_goto2_LIBRARY-NOTFOUND\nLAPACK_lapack_LIBRARY:FILEPATH=/usr/lib/liblapack.so.3gf\nLAPACK_vecLib_LIBRARY:FILEPATH=LAPACK_vecLib_LIBRARY-NOTFOUND\n//ADVANCED property for variable: BLAS_atlas_LIBRARY\nBLAS_atlas_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: BLAS_f77blas_LIBRARY\nBLAS_f77blas_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: BLAS_f77blas_atlas_WORKS\nBLAS_f77blas_atlas_WORKS-ADVANCED:INTERNAL=1\nBLAS_f77blas_atlas_WORKS:INTERNAL=1\n//ADVANCED property for variable: BLAS_goto2_LIBRARY\nBLAS_goto2_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_Accelerate_LIBRARY\nLAPACK_Accelerate_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_acml_LIBRARY\nLAPACK_acml_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_acml_mp_LIBRARY\nLAPACK_acml_mp_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_goto2_LIBRARY\nLAPACK_goto2_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_lapack_LIBRARY\nLAPACK_lapack_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_lapack_WORKS\nLAPACK_lapack_WORKS-ADVANCED:INTERNAL=1\nLAPACK_lapack_WORKS:INTERNAL=1\n//ADVANCED property for variable: LAPACK_vecLib_LIBRARY\nLAPACK_vecLib_LIBRARY-ADVANCED:INTERNAL=1\n}}}\n\nThis causes linking w/ GSL to fubar, because we were tricked by cmake into thinking we had a BLAS so we didnt include GSL's cblas.\n\nEasy fix is to install GotoBLAS.\n", "reporter": "nega", "cc": "", "resolution": "wontfix", "time": "2013-02-01T20:09:26", "component": "cmake", "summary": "cmake's FindBLAS.cmake reports BLAS_FOUND=TRUE even if dev pkgs aren't installed", "priority": "normal", "keywords": "cmake upstream", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
cmake's FindBLAS.cmake reports BLAS_FOUND=TRUE even if dev pkgs aren't installed (Trac #707) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/707">https://code.icecube.wisc.edu/projects/icecube/ticket/707</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:22", "_ts": "1550067082284240", "description": "cmake happily finds libatlas and libblas and reports BLAS_FOUND=TRUE even though no headers are found. \n\nHere's the relevant bits from CMakeCache.txt on the effected system (Ubuntu 12.04).\n\n{{{\n//BLAS binary directory\nBLAS_BIN_DIR:PATH=\n//Tool 'BLAS' found successfully\nBLAS_FOUND:BOOL=FALSE\n//BLAS include dir\nBLAS_INCLUDE_DIR:PATH=/doesnt/exist\n//Libraries for tool BLAS\nBLAS_LIBRARIES:PATH=\nBLAS_atlas_LIBRARY:FILEPATH=/usr/lib/libatlas.so.3gf\nBLAS_f77blas_LIBRARY:FILEPATH=/usr/lib/libf77blas.so.3gf\nBLAS_goto2_LIBRARY:FILEPATH=BLAS_goto2_LIBRARY-NOTFOUND\nLAPACK_Accelerate_LIBRARY:FILEPATH=LAPACK_Accelerate_LIBRARY-NOTFOUND\n//LAPACK binary directory\nLAPACK_BIN_DIR:PATH=\n//Tool 'LAPACK' found successfully\nLAPACK_FOUND:BOOL=FALSE\n//LAPACK include dir\nLAPACK_INCLUDE_DIR:PATH=/doesnt/exist\n//Libraries for tool LAPACK\nLAPACK_LIBRARIES:PATH=\nLAPACK_acml_LIBRARY:FILEPATH=LAPACK_acml_LIBRARY-NOTFOUND\nLAPACK_acml_mp_LIBRARY:FILEPATH=LAPACK_acml_mp_LIBRARY-NOTFOUND\nLAPACK_goto2_LIBRARY:FILEPATH=LAPACK_goto2_LIBRARY-NOTFOUND\nLAPACK_lapack_LIBRARY:FILEPATH=/usr/lib/liblapack.so.3gf\nLAPACK_vecLib_LIBRARY:FILEPATH=LAPACK_vecLib_LIBRARY-NOTFOUND\n//ADVANCED property for variable: BLAS_atlas_LIBRARY\nBLAS_atlas_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: BLAS_f77blas_LIBRARY\nBLAS_f77blas_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: BLAS_f77blas_atlas_WORKS\nBLAS_f77blas_atlas_WORKS-ADVANCED:INTERNAL=1\nBLAS_f77blas_atlas_WORKS:INTERNAL=1\n//ADVANCED property for variable: BLAS_goto2_LIBRARY\nBLAS_goto2_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_Accelerate_LIBRARY\nLAPACK_Accelerate_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_acml_LIBRARY\nLAPACK_acml_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_acml_mp_LIBRARY\nLAPACK_acml_mp_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_goto2_LIBRARY\nLAPACK_goto2_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_lapack_LIBRARY\nLAPACK_lapack_LIBRARY-ADVANCED:INTERNAL=1\n//ADVANCED property for variable: LAPACK_lapack_WORKS\nLAPACK_lapack_WORKS-ADVANCED:INTERNAL=1\nLAPACK_lapack_WORKS:INTERNAL=1\n//ADVANCED property for variable: LAPACK_vecLib_LIBRARY\nLAPACK_vecLib_LIBRARY-ADVANCED:INTERNAL=1\n}}}\n\nThis causes linking w/ GSL to fubar, because we were tricked by cmake into thinking we had a BLAS so we didnt include GSL's cblas.\n\nEasy fix is to install GotoBLAS.\n", "reporter": "nega", "cc": "", "resolution": "wontfix", "time": "2013-02-01T20:09:26", "component": "cmake", "summary": "cmake's FindBLAS.cmake reports BLAS_FOUND=TRUE even if dev pkgs aren't installed", "priority": "normal", "keywords": "cmake upstream", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
cmake s findblas cmake reports blas found true even if dev pkgs aren t installed trac migrated from json status closed changetime ts description cmake happily finds libatlas and libblas and reports blas found true even though no headers are found n nhere s the relevant bits from cmakecache txt on the effected system ubuntu n n n blas binary directory nblas bin dir path n tool blas found successfully nblas found bool false n blas include dir nblas include dir path doesnt exist n libraries for tool blas nblas libraries path nblas atlas library filepath usr lib libatlas so nblas library filepath usr lib so nblas library filepath blas library notfound nlapack accelerate library filepath lapack accelerate library notfound n lapack binary directory nlapack bin dir path n tool lapack found successfully nlapack found bool false n lapack include dir nlapack include dir path doesnt exist n libraries for tool lapack nlapack libraries path nlapack acml library filepath lapack acml library notfound nlapack acml mp library filepath lapack acml mp library notfound nlapack library filepath lapack library notfound nlapack lapack library filepath usr lib liblapack so nlapack veclib library filepath lapack veclib library notfound n advanced property for variable blas atlas library nblas atlas library advanced internal n advanced property for variable blas library nblas library advanced internal n advanced property for variable blas atlas works nblas atlas works advanced internal nblas atlas works internal n advanced property for variable blas library nblas library advanced internal n advanced property for variable lapack accelerate library nlapack accelerate library advanced internal n advanced property for variable lapack acml library nlapack acml library advanced internal n advanced property for variable lapack acml mp library nlapack acml mp library advanced internal n advanced property for variable lapack library nlapack library advanced internal n advanced property for variable lapack lapack library nlapack lapack library advanced internal n advanced property for variable lapack lapack works nlapack lapack works advanced internal nlapack lapack works internal n advanced property for variable lapack veclib library nlapack veclib library advanced internal n n nthis causes linking w gsl to fubar because we were tricked by cmake into thinking we had a blas so we didnt include gsl s cblas n neasy fix is to install gotoblas n reporter nega cc resolution wontfix time component cmake summary cmake s findblas cmake reports blas found true even if dev pkgs aren t installed priority normal keywords cmake upstream milestone owner nega type defect
1
43,017
11,432,693,404
IssuesEvent
2020-02-04 14:30:30
contao/contao
https://api.github.com/repos/contao/contao
closed
Darstellungsfehler des divs id=#paste_hint bei Neuanlage div. Elemente im Backend
defect
**Affected versions:** 4.8.7 | 4.9.0-RC1 **Description:** Wie von [aasage](https://github.com/asaage) in [diesem Issue](https://github.com/contao/contao/pull/467#issuecomment-489598953) angeführt, befindet sich bei der Anlage eines neuen Elements (im Backend), welches eine notwendige Positionierung des neuen Elements erfordert (z.B. neuer Artikel, neue Seite), der Hinweistext "Legen Sie als Nächstes die (neue) Position des Elements fest" (=div #paste_hint) positionsmäßig HINTER der ersten Zeile des "listing containers". ![image](https://user-images.githubusercontent.com/60215157/72993426-a7d11200-3df5-11ea-95ee-a0f72bc723c2.png) Ist zwar kein Weltungergang, könnte jedoch ggf. in der 4.9er-Version korrigiert werden.
1.0
Darstellungsfehler des divs id=#paste_hint bei Neuanlage div. Elemente im Backend - **Affected versions:** 4.8.7 | 4.9.0-RC1 **Description:** Wie von [aasage](https://github.com/asaage) in [diesem Issue](https://github.com/contao/contao/pull/467#issuecomment-489598953) angeführt, befindet sich bei der Anlage eines neuen Elements (im Backend), welches eine notwendige Positionierung des neuen Elements erfordert (z.B. neuer Artikel, neue Seite), der Hinweistext "Legen Sie als Nächstes die (neue) Position des Elements fest" (=div #paste_hint) positionsmäßig HINTER der ersten Zeile des "listing containers". ![image](https://user-images.githubusercontent.com/60215157/72993426-a7d11200-3df5-11ea-95ee-a0f72bc723c2.png) Ist zwar kein Weltungergang, könnte jedoch ggf. in der 4.9er-Version korrigiert werden.
defect
darstellungsfehler des divs id paste hint bei neuanlage div elemente im backend affected versions description wie von in angeführt befindet sich bei der anlage eines neuen elements im backend welches eine notwendige positionierung des neuen elements erfordert z b neuer artikel neue seite der hinweistext legen sie als nächstes die neue position des elements fest div paste hint positionsmäßig hinter der ersten zeile des listing containers ist zwar kein weltungergang könnte jedoch ggf in der version korrigiert werden
1
262
2,507,028,911
IssuesEvent
2015-01-12 15:41:05
yuji-shimoda/rubyist-connect
https://api.github.com/repos/yuji-shimoda/rubyist-connect
closed
トップページの fontawesome アイコン部分をヘルパーメソッドを使った書き方に変更
design
``` i.fa.fa-github-square ``` のようにHTMLでベタ書きされているところを、 ``` = fa_icon 'github-square' ```` のようにヘルパーメソッドを使った書き方に変更する。 #### 参考 [RailsにFont Awesomeを導入するGem - Qiita](http://qiita.com/succi0303/items/7f76d53ebe54dab41124)
1.0
トップページの fontawesome アイコン部分をヘルパーメソッドを使った書き方に変更 - ``` i.fa.fa-github-square ``` のようにHTMLでベタ書きされているところを、 ``` = fa_icon 'github-square' ```` のようにヘルパーメソッドを使った書き方に変更する。 #### 参考 [RailsにFont Awesomeを導入するGem - Qiita](http://qiita.com/succi0303/items/7f76d53ebe54dab41124)
non_defect
トップページの fontawesome アイコン部分をヘルパーメソッドを使った書き方に変更 i fa fa github square のようにhtmlでベタ書きされているところを、 fa icon github square のようにヘルパーメソッドを使った書き方に変更する。 参考
0
641,944
20,862,786,510
IssuesEvent
2022-03-22 01:46:47
khalidsaadat/soen341
https://api.github.com/repos/khalidsaadat/soen341
closed
[IMPLEMENTATION] Epic Story 1: Browsing Items
High Priority epic story
### SPRINT 2 The following user stories belong to this epic story. - [x] #16 - [x] #23 - [x] #34 - [x] #48 - [x] #44
1.0
[IMPLEMENTATION] Epic Story 1: Browsing Items - ### SPRINT 2 The following user stories belong to this epic story. - [x] #16 - [x] #23 - [x] #34 - [x] #48 - [x] #44
non_defect
epic story browsing items sprint the following user stories belong to this epic story
0
78,229
27,381,840,931
IssuesEvent
2023-02-28 10:33:28
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
data corrupt while reading memory mapped page and writing to file at the same time
Type: Defect
### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | 11 Kernel Version | 5.18 Architecture | amd64 OpenZFS Version | both 2.1.9 and 73c383f ### Describe the problem you're observing When one thread ```write()``` to file and another thread read data through ```mmap()``` of the same file, there is chance of data corruption observed by the reader thread. Corrupted data is all zero and it's seems the race is between reader thread which try to fault in page and writer thread which write file with position corresponding to the same page reader is faulting in. Attached below is the program to reproduce the problem. ### Describe how to reproduce the problem ``` #define _GNU_SOURCE #include <stdio.h> #include <stdlib.h> #include <stdatomic.h> #include <string.h> #include <unistd.h> #include <fcntl.h> #include <sys/mman.h> #include <threads.h> #define MODE_WRITE_FILE 0 #define MODE_WRITE_MMAP 1 #define FILE_SIZE (64*1024*1024) struct thread_info { int mode; atomic_int pos; int fd; char* base; }; char data[] = "cache coherency test\n"; int reader(void* arg) { struct thread_info* ti = (struct thread_info*)arg; char buf[sizeof(data)]; for (int i = 0, offset = 0; offset >= 0; offset = ti->pos) { if (offset == i) continue; char *ptr = buf; if (ti->mode == MODE_WRITE_FILE) ptr = ti->base+sizeof(data)*i; else if (read(ti->fd, buf, sizeof(data)) != sizeof(data)) { perror("read file error"); exit(2); } if (memcmp(ptr, data, sizeof(data)) != 0) { printf("data corrupt, offset: %d\n", i*sizeof(data)); exit(2); } i++; } return 0; } int test(int mode) { struct thread_info ti = { .pos = 0, .mode = mode, }; if ((ti.fd = open("test-cc.bin", O_CREAT|O_RDWR, 0777)) < 0) { perror("open error"); exit(1); } else if (fallocate(ti.fd, 0, 0, FILE_SIZE) < 0) { perror("fallocate error"); exit(1); } else if ((ti.base = mmap(NULL, FILE_SIZE, ti.mode == MODE_WRITE_FILE ? PROT_READ : PROT_WRITE,MAP_SHARED, ti.fd, 0)) == MAP_FAILED) { perror("mmap error"); exit(1); } thrd_t thr; thrd_create(&thr, reader, &ti); for (; ti.pos < FILE_SIZE/sizeof(data); ti.pos++) { if (ti.mode == MODE_WRITE_MMAP) memcpy((char*)ti.base+sizeof(data)*ti.pos, data, sizeof(data)); else if (write(ti.fd, data, sizeof(data)) != sizeof(data)) { perror("write error"); exit(1); } } ti.pos = -1; thrd_join(thr, NULL); return 0; } int main(int argc, char **argv) { test(MODE_WRITE_FILE); test(MODE_WRITE_MMAP); return 0; } ```
1.0
data corrupt while reading memory mapped page and writing to file at the same time - ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Debian Distribution Version | 11 Kernel Version | 5.18 Architecture | amd64 OpenZFS Version | both 2.1.9 and 73c383f ### Describe the problem you're observing When one thread ```write()``` to file and another thread read data through ```mmap()``` of the same file, there is chance of data corruption observed by the reader thread. Corrupted data is all zero and it's seems the race is between reader thread which try to fault in page and writer thread which write file with position corresponding to the same page reader is faulting in. Attached below is the program to reproduce the problem. ### Describe how to reproduce the problem ``` #define _GNU_SOURCE #include <stdio.h> #include <stdlib.h> #include <stdatomic.h> #include <string.h> #include <unistd.h> #include <fcntl.h> #include <sys/mman.h> #include <threads.h> #define MODE_WRITE_FILE 0 #define MODE_WRITE_MMAP 1 #define FILE_SIZE (64*1024*1024) struct thread_info { int mode; atomic_int pos; int fd; char* base; }; char data[] = "cache coherency test\n"; int reader(void* arg) { struct thread_info* ti = (struct thread_info*)arg; char buf[sizeof(data)]; for (int i = 0, offset = 0; offset >= 0; offset = ti->pos) { if (offset == i) continue; char *ptr = buf; if (ti->mode == MODE_WRITE_FILE) ptr = ti->base+sizeof(data)*i; else if (read(ti->fd, buf, sizeof(data)) != sizeof(data)) { perror("read file error"); exit(2); } if (memcmp(ptr, data, sizeof(data)) != 0) { printf("data corrupt, offset: %d\n", i*sizeof(data)); exit(2); } i++; } return 0; } int test(int mode) { struct thread_info ti = { .pos = 0, .mode = mode, }; if ((ti.fd = open("test-cc.bin", O_CREAT|O_RDWR, 0777)) < 0) { perror("open error"); exit(1); } else if (fallocate(ti.fd, 0, 0, FILE_SIZE) < 0) { perror("fallocate error"); exit(1); } else if ((ti.base = mmap(NULL, FILE_SIZE, ti.mode == MODE_WRITE_FILE ? PROT_READ : PROT_WRITE,MAP_SHARED, ti.fd, 0)) == MAP_FAILED) { perror("mmap error"); exit(1); } thrd_t thr; thrd_create(&thr, reader, &ti); for (; ti.pos < FILE_SIZE/sizeof(data); ti.pos++) { if (ti.mode == MODE_WRITE_MMAP) memcpy((char*)ti.base+sizeof(data)*ti.pos, data, sizeof(data)); else if (write(ti.fd, data, sizeof(data)) != sizeof(data)) { perror("write error"); exit(1); } } ti.pos = -1; thrd_join(thr, NULL); return 0; } int main(int argc, char **argv) { test(MODE_WRITE_FILE); test(MODE_WRITE_MMAP); return 0; } ```
defect
data corrupt while reading memory mapped page and writing to file at the same time system information type version name distribution name debian distribution version kernel version architecture openzfs version both and describe the problem you re observing when one thread write to file and another thread read data through mmap of the same file there is chance of data corruption observed by the reader thread corrupted data is all zero and it s seems the race is between reader thread which try to fault in page and writer thread which write file with position corresponding to the same page reader is faulting in attached below is the program to reproduce the problem describe how to reproduce the problem define gnu source include include include include include include include include define mode write file define mode write mmap define file size struct thread info int mode atomic int pos int fd char base char data cache coherency test n int reader void arg struct thread info ti struct thread info arg char buf for int i offset offset offset ti pos if offset i continue char ptr buf if ti mode mode write file ptr ti base sizeof data i else if read ti fd buf sizeof data sizeof data perror read file error exit if memcmp ptr data sizeof data printf data corrupt offset d n i sizeof data exit i return int test int mode struct thread info ti pos mode mode if ti fd open test cc bin o creat o rdwr perror open error exit else if fallocate ti fd file size perror fallocate error exit else if ti base mmap null file size ti mode mode write file prot read prot write map shared ti fd map failed perror mmap error exit thrd t thr thrd create thr reader ti for ti pos file size sizeof data ti pos if ti mode mode write mmap memcpy char ti base sizeof data ti pos data sizeof data else if write ti fd data sizeof data sizeof data perror write error exit ti pos thrd join thr null return int main int argc char argv test mode write file test mode write mmap return
1
28,526
5,285,446,040
IssuesEvent
2017-02-08 04:59:41
extnet/Ext.NET
https://api.github.com/repos/extnet/Ext.NET
opened
4.2.0 mobile: IOS theme has references to font-awesome but no files
4.x defect feature mobile sencha
The font-awesome is missing from the new IOS theme. Probably the best option would be to just point to the existing files to avoid redundant and unnecessary size on `Ext.NET.Mobile.dll`. The theme has been introduced by ExtJS 6.2.1 and the original CSS file and folder structure has not font-awesome theme. It is either provided in a custom, central directory or the files went missing in the ExtJS build process although the font references are left on the built CSS file.
1.0
4.2.0 mobile: IOS theme has references to font-awesome but no files - The font-awesome is missing from the new IOS theme. Probably the best option would be to just point to the existing files to avoid redundant and unnecessary size on `Ext.NET.Mobile.dll`. The theme has been introduced by ExtJS 6.2.1 and the original CSS file and folder structure has not font-awesome theme. It is either provided in a custom, central directory or the files went missing in the ExtJS build process although the font references are left on the built CSS file.
defect
mobile ios theme has references to font awesome but no files the font awesome is missing from the new ios theme probably the best option would be to just point to the existing files to avoid redundant and unnecessary size on ext net mobile dll the theme has been introduced by extjs and the original css file and folder structure has not font awesome theme it is either provided in a custom central directory or the files went missing in the extjs build process although the font references are left on the built css file
1
205,919
23,360,218,518
IssuesEvent
2022-08-10 11:00:44
ShaikUsaf/linux-4.19.72
https://api.github.com/repos/ShaikUsaf/linux-4.19.72
closed
CVE-2022-0480 (Medium) detected in linuxlinux-4.19.238 - autoclosed
security vulnerability
## CVE-2022-0480 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.238</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/ShaikUsaf/linux-4.19.72/commit/63fcdaa32317d63a37988731632cf4e8d4c6072a">63fcdaa32317d63a37988731632cf4e8d4c6072a</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the filelock_init in fs/locks.c function in the Linux kernel. This issue can lead to host memory exhaustion due to memcg not limiting the number of Portable Operating System Interface (POSIX) file locks. <p>Publish Date: 2022-02-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0480>CVE-2022-0480</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-0480">https://www.linuxkernelcves.com/cves/CVE-2022-0480</a></p> <p>Release Date: 2022-02-03</p> <p>Fix Resolution: v5.15-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0480 (Medium) detected in linuxlinux-4.19.238 - autoclosed - ## CVE-2022-0480 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.238</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/ShaikUsaf/linux-4.19.72/commit/63fcdaa32317d63a37988731632cf4e8d4c6072a">63fcdaa32317d63a37988731632cf4e8d4c6072a</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in the filelock_init in fs/locks.c function in the Linux kernel. This issue can lead to host memory exhaustion due to memcg not limiting the number of Portable Operating System Interface (POSIX) file locks. <p>Publish Date: 2022-02-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0480>CVE-2022-0480</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-0480">https://www.linuxkernelcves.com/cves/CVE-2022-0480</a></p> <p>Release Date: 2022-02-03</p> <p>Fix Resolution: v5.15-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details a flaw was found in the filelock init in fs locks c function in the linux kernel this issue can lead to host memory exhaustion due to memcg not limiting the number of portable operating system interface posix file locks publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
26,470
4,725,101,929
IssuesEvent
2016-10-18 04:24:03
catmaid/CATMAID
https://api.github.com/repos/catmaid/CATMAID
closed
When changing a setting in the settings widget, the popup "confirmation" takes too long to disappear
status: done type: defect
There's no need for it to be launched in the first place.
1.0
When changing a setting in the settings widget, the popup "confirmation" takes too long to disappear - There's no need for it to be launched in the first place.
defect
when changing a setting in the settings widget the popup confirmation takes too long to disappear there s no need for it to be launched in the first place
1
647,861
21,157,579,320
IssuesEvent
2022-04-07 06:04:39
harvester/harvester
https://api.github.com/repos/harvester/harvester
closed
[BUG] upgrade_manifest enhancements
bug priority/1 area/upgrade-related
**Describe the bug** <!-- A clear and concise description of what the bug is. --> In [upgrade_manifest](https://github.com/harvester/harvester/blob/master/package%2Fupgrade%2Fupgrade_manifests.sh) stage, we bump Rancher and ManagedChart versions. There are several things that needed be done here to smoothen the upgrade. - Wait for Rancher ready - Wait for rancher, fleet, and system-upgrade-controller pods to be upgraded. - After this, we need to ensure cluster provision is done. This somehow can minimize the impact of [RKE2 restart].(https://github.com/harvester/harvester/issues/1707#issuecomment-998484833) - Pause managedchart: I saw situation managechart bundles go into `ErrApplied` when changing cluster-repo. - Need a way to refresh the cluster-repo. Rancher seems to update chart index only for a certain period. - [By default the index is updated every 5 minutes, we can use `forceUpdate` field to trigger an update ](https://github.com/rancher/rancher/blob/47c22388c5451c74f55e162d1e60b4e6dcfd0800/pkg/controllers/dashboard/helm/repo.go#L290-L294 ) - Wait for Harvester ready - We can't simply bump the Harvester managed chart version, because the upgrade might involve many components upgrade. We need to wait them to settle down. - Wait for Monitoring ready
1.0
[BUG] upgrade_manifest enhancements - **Describe the bug** <!-- A clear and concise description of what the bug is. --> In [upgrade_manifest](https://github.com/harvester/harvester/blob/master/package%2Fupgrade%2Fupgrade_manifests.sh) stage, we bump Rancher and ManagedChart versions. There are several things that needed be done here to smoothen the upgrade. - Wait for Rancher ready - Wait for rancher, fleet, and system-upgrade-controller pods to be upgraded. - After this, we need to ensure cluster provision is done. This somehow can minimize the impact of [RKE2 restart].(https://github.com/harvester/harvester/issues/1707#issuecomment-998484833) - Pause managedchart: I saw situation managechart bundles go into `ErrApplied` when changing cluster-repo. - Need a way to refresh the cluster-repo. Rancher seems to update chart index only for a certain period. - [By default the index is updated every 5 minutes, we can use `forceUpdate` field to trigger an update ](https://github.com/rancher/rancher/blob/47c22388c5451c74f55e162d1e60b4e6dcfd0800/pkg/controllers/dashboard/helm/repo.go#L290-L294 ) - Wait for Harvester ready - We can't simply bump the Harvester managed chart version, because the upgrade might involve many components upgrade. We need to wait them to settle down. - Wait for Monitoring ready
non_defect
upgrade manifest enhancements describe the bug in stage we bump rancher and managedchart versions there are several things that needed be done here to smoothen the upgrade wait for rancher ready wait for rancher fleet and system upgrade controller pods to be upgraded after this we need to ensure cluster provision is done this somehow can minimize the impact of pause managedchart i saw situation managechart bundles go into errapplied when changing cluster repo need a way to refresh the cluster repo rancher seems to update chart index only for a certain period wait for harvester ready we can t simply bump the harvester managed chart version because the upgrade might involve many components upgrade we need to wait them to settle down wait for monitoring ready
0
193,299
22,216,125,609
IssuesEvent
2022-06-08 01:58:35
AlexRogalskiy/github-action-charts
https://api.github.com/repos/AlexRogalskiy/github-action-charts
closed
CVE-2021-37701 (High) detected in tar-6.1.0.tgz - autoclosed
security vulnerability
## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-6.1.0.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.1.0.tgz">https://registry.npmjs.org/tar/-/tar-6.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/tar/package.json</p> <p> Dependency Hierarchy: - editorconfig-checker-3.3.0.tgz (Root Library) - :x: **tar-6.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-charts/commit/943f01a40c71c0d46d7574a64053c0ce00fe11f5">943f01a40c71c0d46d7574a64053c0ce00fe11f5</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.16,5.0.8,6.1.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37701 (High) detected in tar-6.1.0.tgz - autoclosed - ## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-6.1.0.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-6.1.0.tgz">https://registry.npmjs.org/tar/-/tar-6.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/tar/package.json</p> <p> Dependency Hierarchy: - editorconfig-checker-3.3.0.tgz (Root Library) - :x: **tar-6.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-charts/commit/943f01a40c71c0d46d7574a64053c0ce00fe11f5">943f01a40c71c0d46d7574a64053c0ce00fe11f5</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.16,5.0.8,6.1.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in tar tgz autoclosed cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file package json path to vulnerable library node modules tar package json dependency hierarchy editorconfig checker tgz root library x tar tgz vulnerable library found in head commit a href vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems the cache checking logic used both and characters as path separators however is a valid filename character on posix systems by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite additionally a similar confusion could arise on case insensitive filesystems if a tar archive contained a directory at foo followed by a symbolic link named foo then on case insensitive file systems the creation of the symbolic link would remove the directory from the filesystem but not from the internal directory cache as it would not be treated as a cache hit a subsequent file entry within the foo directory would then be placed in the target of the symbolic link thinking that the directory had already been created these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource
0
46,142
13,055,858,485
IssuesEvent
2020-07-30 02:56:47
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
USE_GOLD auto enabled on mac os x (Trac #641)
Incomplete Migration Migrated from Trac cmake defect
Migrated from https://code.icecube.wisc.edu/ticket/641 ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "see: \nhttp://builds.icecube.wisc.edu/builders/quick_osx_10.5.8/builds/410/steps/compile/logs/stdio", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1416713877111216", "component": "cmake", "summary": "USE_GOLD auto enabled on mac os x", "priority": "normal", "keywords": "", "time": "2011-06-02T16:01:41", "milestone": "", "owner": "nega", "type": "defect" } ```
1.0
USE_GOLD auto enabled on mac os x (Trac #641) - Migrated from https://code.icecube.wisc.edu/ticket/641 ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "see: \nhttp://builds.icecube.wisc.edu/builders/quick_osx_10.5.8/builds/410/steps/compile/logs/stdio", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1416713877111216", "component": "cmake", "summary": "USE_GOLD auto enabled on mac os x", "priority": "normal", "keywords": "", "time": "2011-06-02T16:01:41", "milestone": "", "owner": "nega", "type": "defect" } ```
defect
use gold auto enabled on mac os x trac migrated from json status closed changetime description see n reporter nega cc resolution fixed ts component cmake summary use gold auto enabled on mac os x priority normal keywords time milestone owner nega type defect
1
72,527
24,165,153,152
IssuesEvent
2022-09-22 14:31:41
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
FE | Profile | [Markup and meta data] HTML markup isn't valid. (09.01.1)
frontend 508/Accessibility authenticated-experience profile 508-defect-2 collab-cycle-feedback Staging CCIssue09.01 CC-Dashboard bad-address-indicator
### General Information #### VFS team name Profile Team #### VFS product name Profile #### VFS feature name Bad Address Indicator #### Point of Contact/Reviewers Brian DeConinck (@briandeconinck) - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue HTML markup isn't valid. ### Issue Details For each of the "Review your mailing address" alerts, the H2 heading has an `aria-label` attribute. `aria-labels` should only be added to interactive elements. ### Link, screenshot or steps to recreate The "Review your mailing address" alerts are currently coded as: ``` <h2 slot="headline" tabindex="0" role="alert" aria-live="polite" aria-label="The address we have on file for you may not be correct.">Review your mailing address</h2> <p>The mailing address we have on file for you may not be correct.</p> <p><a href="...">...</a></p> ``` ### VA.gov Experience Standard [Category Number 09, Issue Number 01](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 1.3.1_A --- ### Platform Recommendation Typically (although not exclusively!), `aria-label` attributes should only be on interactive elements. It provides something for screen readers to announce as an alternative to the visible text when the visible text is ambiguous or confusing for screen reader users. In this case, I'm not sure what the `aria-label` buys you. It may not be consistently announced by screen readers, and when it is announced it will likely replace the heading text. Since the `aria-label` text is very similar to the body of the alert, you could end up with some repetition. If the goal is to get the body of the alert announced at the same time as the heading on page load (thanks to `role="alert"` on the heading -- but see the other issue about that), consider using `aria-describedby` instead to associate the body text with the heading. That might look something like: ``` <h2 slot="headline" tabindex="0" role="alert" aria-live="polite" aria-describedby="alert-body">Review your mailing address</h2> <p id="alert-body">The mailing address we have on file for you may not be correct.</p> <p><a href="...">...</a></p> ``` (Note that I haven't tested the example code above, so don't trust me too much!) ### VFS Team Tasks to Complete - [x] Comment on the ticket if there are questions or concerns - [x] Close the ticket when the issue has been resolved or validated by your Product Owner. If a team has additional questions or needs Platform help validating the issue, please comment in the ticket.
1.0
FE | Profile | [Markup and meta data] HTML markup isn't valid. (09.01.1) - ### General Information #### VFS team name Profile Team #### VFS product name Profile #### VFS feature name Bad Address Indicator #### Point of Contact/Reviewers Brian DeConinck (@briandeconinck) - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue HTML markup isn't valid. ### Issue Details For each of the "Review your mailing address" alerts, the H2 heading has an `aria-label` attribute. `aria-labels` should only be added to interactive elements. ### Link, screenshot or steps to recreate The "Review your mailing address" alerts are currently coded as: ``` <h2 slot="headline" tabindex="0" role="alert" aria-live="polite" aria-label="The address we have on file for you may not be correct.">Review your mailing address</h2> <p>The mailing address we have on file for you may not be correct.</p> <p><a href="...">...</a></p> ``` ### VA.gov Experience Standard [Category Number 09, Issue Number 01](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 1.3.1_A --- ### Platform Recommendation Typically (although not exclusively!), `aria-label` attributes should only be on interactive elements. It provides something for screen readers to announce as an alternative to the visible text when the visible text is ambiguous or confusing for screen reader users. In this case, I'm not sure what the `aria-label` buys you. It may not be consistently announced by screen readers, and when it is announced it will likely replace the heading text. Since the `aria-label` text is very similar to the body of the alert, you could end up with some repetition. If the goal is to get the body of the alert announced at the same time as the heading on page load (thanks to `role="alert"` on the heading -- but see the other issue about that), consider using `aria-describedby` instead to associate the body text with the heading. That might look something like: ``` <h2 slot="headline" tabindex="0" role="alert" aria-live="polite" aria-describedby="alert-body">Review your mailing address</h2> <p id="alert-body">The mailing address we have on file for you may not be correct.</p> <p><a href="...">...</a></p> ``` (Note that I haven't tested the example code above, so don't trust me too much!) ### VFS Team Tasks to Complete - [x] Comment on the ticket if there are questions or concerns - [x] Close the ticket when the issue has been resolved or validated by your Product Owner. If a team has additional questions or needs Platform help validating the issue, please comment in the ticket.
defect
fe profile html markup isn t valid general information vfs team name profile team vfs product name profile vfs feature name bad address indicator point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue html markup isn t valid issue details for each of the review your mailing address alerts the heading has an aria label attribute aria labels should only be added to interactive elements link screenshot or steps to recreate the review your mailing address alerts are currently coded as review your mailing address the mailing address we have on file for you may not be correct va gov experience standard other references wcag sc a platform recommendation typically although not exclusively aria label attributes should only be on interactive elements it provides something for screen readers to announce as an alternative to the visible text when the visible text is ambiguous or confusing for screen reader users in this case i m not sure what the aria label buys you it may not be consistently announced by screen readers and when it is announced it will likely replace the heading text since the aria label text is very similar to the body of the alert you could end up with some repetition if the goal is to get the body of the alert announced at the same time as the heading on page load thanks to role alert on the heading but see the other issue about that consider using aria describedby instead to associate the body text with the heading that might look something like review your mailing address the mailing address we have on file for you may not be correct note that i haven t tested the example code above so don t trust me too much vfs team tasks to complete comment on the ticket if there are questions or concerns close the ticket when the issue has been resolved or validated by your product owner if a team has additional questions or needs platform help validating the issue please comment in the ticket
1
94,534
11,883,002,522
IssuesEvent
2020-03-27 15:16:17
Opentrons/opentrons
https://api.github.com/repos/Opentrons/opentrons
closed
PD: When switching magnetic module model, preserve engage heights IF they are at default
:spider: SPDDRS protocol designer
## Overview Once we have #5225 and #5226 in place, we can take these features a step farther and avoid having users re-enter `engageHeight` values when they don't need to. ## Acceptance Criteria When a user switches magnetic module model... - [ ] if the value of the engageHeight exactly matches the module labware's `magneticModuleEngageHeight` number, convert its value to the units of the newly-selected module. Note: since this is after #5226, all labware `magneticModuleEngageHeight` numbers will be effectively in real mm, so when switching from GEN1 to GEN2 we'll want to `*2` the `formData.engageHeight` value to convert it to real mm before comparing it to the labware definition's `magneticModuleEngageHeight` - [ ] otherwise, clear out all engageHeight fields as in #5225 (number doesn't match, or there's no labware on module)
1.0
PD: When switching magnetic module model, preserve engage heights IF they are at default - ## Overview Once we have #5225 and #5226 in place, we can take these features a step farther and avoid having users re-enter `engageHeight` values when they don't need to. ## Acceptance Criteria When a user switches magnetic module model... - [ ] if the value of the engageHeight exactly matches the module labware's `magneticModuleEngageHeight` number, convert its value to the units of the newly-selected module. Note: since this is after #5226, all labware `magneticModuleEngageHeight` numbers will be effectively in real mm, so when switching from GEN1 to GEN2 we'll want to `*2` the `formData.engageHeight` value to convert it to real mm before comparing it to the labware definition's `magneticModuleEngageHeight` - [ ] otherwise, clear out all engageHeight fields as in #5225 (number doesn't match, or there's no labware on module)
non_defect
pd when switching magnetic module model preserve engage heights if they are at default overview once we have and in place we can take these features a step farther and avoid having users re enter engageheight values when they don t need to acceptance criteria when a user switches magnetic module model if the value of the engageheight exactly matches the module labware s magneticmoduleengageheight number convert its value to the units of the newly selected module note since this is after all labware magneticmoduleengageheight numbers will be effectively in real mm so when switching from to we ll want to the formdata engageheight value to convert it to real mm before comparing it to the labware definition s magneticmoduleengageheight otherwise clear out all engageheight fields as in number doesn t match or there s no labware on module
0
131,316
12,481,053,456
IssuesEvent
2020-05-29 21:32:46
alexluque/MyMusicApp
https://api.github.com/repos/alexluque/MyMusicApp
closed
Write readme file
documentation
Write down that a proper Google Maps API key should be placed in the code to make the recommendations funcionality work
1.0
Write readme file - Write down that a proper Google Maps API key should be placed in the code to make the recommendations funcionality work
non_defect
write readme file write down that a proper google maps api key should be placed in the code to make the recommendations funcionality work
0
44,034
11,914,130,916
IssuesEvent
2020-03-31 13:09:25
SasView/sasview
https://api.github.com/repos/SasView/sasview
closed
Check package versions in yaml files and setup.py (Trac #1087)
CCX: Infrastructure Migrated from Trac SasView Bug Fixing defect major
In a recent sasview PR (https://github.com/SasView/sasview/pull/150) the version of pyparsing in setup.py is different than the yaml files. We should check the versions match up between the two so local builds and master builds have the same behavior. Migrated from http://trac.sasview.org/ticket/1087 ```json { "status": "new", "changetime": "2018-04-08T16:57:43", "_ts": "2018-04-08 16:57:43.344422+00:00", "description": "In a recent sasview PR (https://github.com/SasView/sasview/pull/150) the version of pyparsing in setup.py is different than the yaml files.\n\nWe should check the versions match up between the two so local builds and master builds have the same behavior.", "reporter": "krzywon", "cc": "", "resolution": "", "workpackage": "SasView Bug Fixing", "time": "2018-04-08T16:57:43", "component": "SasView", "summary": "Check package versions in yaml files and setup.py", "priority": "major", "keywords": "", "milestone": "SasView 4.3.0", "owner": "", "type": "defect" } ```
1.0
Check package versions in yaml files and setup.py (Trac #1087) - In a recent sasview PR (https://github.com/SasView/sasview/pull/150) the version of pyparsing in setup.py is different than the yaml files. We should check the versions match up between the two so local builds and master builds have the same behavior. Migrated from http://trac.sasview.org/ticket/1087 ```json { "status": "new", "changetime": "2018-04-08T16:57:43", "_ts": "2018-04-08 16:57:43.344422+00:00", "description": "In a recent sasview PR (https://github.com/SasView/sasview/pull/150) the version of pyparsing in setup.py is different than the yaml files.\n\nWe should check the versions match up between the two so local builds and master builds have the same behavior.", "reporter": "krzywon", "cc": "", "resolution": "", "workpackage": "SasView Bug Fixing", "time": "2018-04-08T16:57:43", "component": "SasView", "summary": "Check package versions in yaml files and setup.py", "priority": "major", "keywords": "", "milestone": "SasView 4.3.0", "owner": "", "type": "defect" } ```
defect
check package versions in yaml files and setup py trac in a recent sasview pr the version of pyparsing in setup py is different than the yaml files we should check the versions match up between the two so local builds and master builds have the same behavior migrated from json status new changetime ts description in a recent sasview pr the version of pyparsing in setup py is different than the yaml files n nwe should check the versions match up between the two so local builds and master builds have the same behavior reporter krzywon cc resolution workpackage sasview bug fixing time component sasview summary check package versions in yaml files and setup py priority major keywords milestone sasview owner type defect
1
28,158
5,201,777,658
IssuesEvent
2017-01-24 06:46:00
TNGSB/eWallet
https://api.github.com/repos/TNGSB/eWallet
closed
e-Wallet_Mobile App (FPX Reload-iPay88 Email Details) 18012017 #25
Defect - Low (Sev-4) Live Environment TNG
Defect Description : The "Product Description" fields value displayed wrong. Should not be "Pre-Authorization" as the action performed was "Reload". Refer attached screenshot. ![image](https://cloud.githubusercontent.com/assets/23113211/22055293/758f5e0a-dd95-11e6-9206-77a65e62b7f3.png)
1.0
e-Wallet_Mobile App (FPX Reload-iPay88 Email Details) 18012017 #25 - Defect Description : The "Product Description" fields value displayed wrong. Should not be "Pre-Authorization" as the action performed was "Reload". Refer attached screenshot. ![image](https://cloud.githubusercontent.com/assets/23113211/22055293/758f5e0a-dd95-11e6-9206-77a65e62b7f3.png)
defect
e wallet mobile app fpx reload email details defect description the product description fields value displayed wrong should not be pre authorization as the action performed was reload refer attached screenshot
1
193,140
15,368,347,864
IssuesEvent
2021-03-02 05:22:11
microsoft/pxt-arcade
https://api.github.com/repos/microsoft/pxt-arcade
closed
Multiple Levels: The tutorial content is not accurate for step 14
documentation skillmap
**Describe the bug** **Multiple Levels**: The tutorial content is not accurate for **step 14** **Steps to reproduce the behavior** 1.Navigate to https://arcade.makecode.com/beta--skillmap 2.Learn to Make a Platformer----->**Multiple Levels**----->START 3.Click **step 14** **Actual behavior** The tutorial content is not accurate for **step 14**. The tutorial: From the **Arrays** category, grab an **array of sprites of kind [Player]** But in the toolbox, **array of sprites of kind [Player]** is in the **Sprites** category ![image](https://user-images.githubusercontent.com/59192069/109266294-d2846a00-7842-11eb-80bd-6f5bba7723fa.png) **Additional context** 1.OS: Windows(rs6) 2.arcade version: 1.4.7 3.Microsoft MakeCode version: 6.9.12
1.0
Multiple Levels: The tutorial content is not accurate for step 14 - **Describe the bug** **Multiple Levels**: The tutorial content is not accurate for **step 14** **Steps to reproduce the behavior** 1.Navigate to https://arcade.makecode.com/beta--skillmap 2.Learn to Make a Platformer----->**Multiple Levels**----->START 3.Click **step 14** **Actual behavior** The tutorial content is not accurate for **step 14**. The tutorial: From the **Arrays** category, grab an **array of sprites of kind [Player]** But in the toolbox, **array of sprites of kind [Player]** is in the **Sprites** category ![image](https://user-images.githubusercontent.com/59192069/109266294-d2846a00-7842-11eb-80bd-6f5bba7723fa.png) **Additional context** 1.OS: Windows(rs6) 2.arcade version: 1.4.7 3.Microsoft MakeCode version: 6.9.12
non_defect
multiple levels the tutorial content is not accurate for step describe the bug multiple levels the tutorial content is not accurate for step steps to reproduce the behavior navigate to learn to make a platformer multiple levels start click step actual behavior the tutorial content is not accurate for step the tutorial from the arrays category grab an array of sprites of kind but in the toolbox array of sprites of kind is in the sprites category additional context os windows arcade version microsoft makecode version
0
1,226
5,218,843,895
IssuesEvent
2017-01-26 17:27:01
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
apache2_module fails for PHP 5.6 even though it is already enabled
affects_2.2 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME apache2_module ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0.0 config file = /Users/nick/Workspace/-redacted-/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION hostfile & roles_path ##### OS / ENVIRONMENT Running Ansible on macOS Sierra, target server is Ubuntu Xenial ##### SUMMARY Enabling the Apache2 module "[php5.6](https://launchpad.net/~ondrej/+archive/ubuntu/php)" with apache2_module fails even though the module is already enabled. This is the same problem as #5559 and #4744 but with a different package. This module is called `php5.6` but identifies itself in `apache2ctl -M` as `php5_module`. ##### STEPS TO REPRODUCE ``` - name: Enable PHP 5.6 apache2_module: state=present name=php5.6 ``` ##### ACTUAL RESULTS ``` failed: [nicksherlock.com] (item=php5.6) => { "failed": true, "invocation": { "module_args": { "force": false, "name": "php5.6", "state": "present" }, "module_name": "apache2_module" }, "item": "php5.6", "msg": "Failed to set module php5.6 to enabled: Considering dependency mpm_prefork for php5.6:\nConsidering conflict mpm_event for mpm_prefork:\nConsidering conflict mpm_worker for mpm_prefork:\nModule mpm_prefork already enabled\nConsidering conflict php5 for php5.6:\nModule php5.6 already enabled\n", "rc": 0, "stderr": "", "stdout": "Considering dependency mpm_prefork for php5.6:\nConsidering conflict mpm_event for mpm_prefork:\nConsidering conflict mpm_worker for mpm_prefork:\nModule mpm_prefork already enabled\nConsidering conflict php5 for php5.6:\nModule php5.6 already enabled\n", "stdout_lines": [ "Considering dependency mpm_prefork for php5.6:", "Considering conflict mpm_event for mpm_prefork:", "Considering conflict mpm_worker for mpm_prefork:", "Module mpm_prefork already enabled", "Considering conflict php5 for php5.6:", "Module php5.6 already enabled" ] } ``` Running it manually on the server gives: ``` # a2enmod php5.6 Considering dependency mpm_prefork for php5.6: Considering conflict mpm_event for mpm_prefork: Considering conflict mpm_worker for mpm_prefork: Module mpm_prefork already enabled Considering conflict php5 for php5.6: Module php5.6 already enabled # echo $? 0 ``` This is php5.6.load: ``` # Conflicts: php5 # Depends: mpm_prefork LoadModule php5_module /usr/lib/apache2/modules/libphp5.6.so ``` Note that manually running "a2enmod php5.6" on the server directly gives a 0 exit status to signal success, can't apache2_module just check that instead of doing parsing with a regex? What if I wanted several sets of conf files in `mods-available` for the same module? (e.g. php-prod.load, php-dev.load both loading the same module, but with different config) Wouldn't that make it impossible for Ansible to manage those with apache2_module? It just seems odd that Ansible requires that the module's binary name be the same as the name of its .load file.
True
apache2_module fails for PHP 5.6 even though it is already enabled - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME apache2_module ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0.0 config file = /Users/nick/Workspace/-redacted-/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION hostfile & roles_path ##### OS / ENVIRONMENT Running Ansible on macOS Sierra, target server is Ubuntu Xenial ##### SUMMARY Enabling the Apache2 module "[php5.6](https://launchpad.net/~ondrej/+archive/ubuntu/php)" with apache2_module fails even though the module is already enabled. This is the same problem as #5559 and #4744 but with a different package. This module is called `php5.6` but identifies itself in `apache2ctl -M` as `php5_module`. ##### STEPS TO REPRODUCE ``` - name: Enable PHP 5.6 apache2_module: state=present name=php5.6 ``` ##### ACTUAL RESULTS ``` failed: [nicksherlock.com] (item=php5.6) => { "failed": true, "invocation": { "module_args": { "force": false, "name": "php5.6", "state": "present" }, "module_name": "apache2_module" }, "item": "php5.6", "msg": "Failed to set module php5.6 to enabled: Considering dependency mpm_prefork for php5.6:\nConsidering conflict mpm_event for mpm_prefork:\nConsidering conflict mpm_worker for mpm_prefork:\nModule mpm_prefork already enabled\nConsidering conflict php5 for php5.6:\nModule php5.6 already enabled\n", "rc": 0, "stderr": "", "stdout": "Considering dependency mpm_prefork for php5.6:\nConsidering conflict mpm_event for mpm_prefork:\nConsidering conflict mpm_worker for mpm_prefork:\nModule mpm_prefork already enabled\nConsidering conflict php5 for php5.6:\nModule php5.6 already enabled\n", "stdout_lines": [ "Considering dependency mpm_prefork for php5.6:", "Considering conflict mpm_event for mpm_prefork:", "Considering conflict mpm_worker for mpm_prefork:", "Module mpm_prefork already enabled", "Considering conflict php5 for php5.6:", "Module php5.6 already enabled" ] } ``` Running it manually on the server gives: ``` # a2enmod php5.6 Considering dependency mpm_prefork for php5.6: Considering conflict mpm_event for mpm_prefork: Considering conflict mpm_worker for mpm_prefork: Module mpm_prefork already enabled Considering conflict php5 for php5.6: Module php5.6 already enabled # echo $? 0 ``` This is php5.6.load: ``` # Conflicts: php5 # Depends: mpm_prefork LoadModule php5_module /usr/lib/apache2/modules/libphp5.6.so ``` Note that manually running "a2enmod php5.6" on the server directly gives a 0 exit status to signal success, can't apache2_module just check that instead of doing parsing with a regex? What if I wanted several sets of conf files in `mods-available` for the same module? (e.g. php-prod.load, php-dev.load both loading the same module, but with different config) Wouldn't that make it impossible for Ansible to manage those with apache2_module? It just seems odd that Ansible requires that the module's binary name be the same as the name of its .load file.
non_defect
module fails for php even though it is already enabled issue type bug report component name module ansible version ansible config file users nick workspace redacted ansible cfg configured module search path default w o overrides configuration hostfile roles path os environment running ansible on macos sierra target server is ubuntu xenial summary enabling the module with module fails even though the module is already enabled this is the same problem as and but with a different package this module is called but identifies itself in m as module steps to reproduce name enable php module state present name actual results failed item failed true invocation module args force false name state present module name module item msg failed to set module to enabled considering dependency mpm prefork for nconsidering conflict mpm event for mpm prefork nconsidering conflict mpm worker for mpm prefork nmodule mpm prefork already enabled nconsidering conflict for nmodule already enabled n rc stderr stdout considering dependency mpm prefork for nconsidering conflict mpm event for mpm prefork nconsidering conflict mpm worker for mpm prefork nmodule mpm prefork already enabled nconsidering conflict for nmodule already enabled n stdout lines considering dependency mpm prefork for considering conflict mpm event for mpm prefork considering conflict mpm worker for mpm prefork module mpm prefork already enabled considering conflict for module already enabled running it manually on the server gives considering dependency mpm prefork for considering conflict mpm event for mpm prefork considering conflict mpm worker for mpm prefork module mpm prefork already enabled considering conflict for module already enabled echo this is load conflicts depends mpm prefork loadmodule module usr lib modules so note that manually running on the server directly gives a exit status to signal success can t module just check that instead of doing parsing with a regex what if i wanted several sets of conf files in mods available for the same module e g php prod load php dev load both loading the same module but with different config wouldn t that make it impossible for ansible to manage those with module it just seems odd that ansible requires that the module s binary name be the same as the name of its load file
0
14,874
8,695,306,108
IssuesEvent
2018-12-04 14:53:16
owncloud/client
https://api.github.com/repos/owncloud/client
closed
Do parallel local and remote discovery at the same time
Performance p4-low
One could consider to ask for remote and local changes at the same time -- these two actions do not need to happen in sequence. Or maybe it already is done like that? <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/28163435-do-parallel-local-and-remote-discovery-at-the-same-time?utm_campaign=plugin&utm_content=tracker%2F216457&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F216457&utm_medium=issues&utm_source=github). </bountysource-plugin>
True
Do parallel local and remote discovery at the same time - One could consider to ask for remote and local changes at the same time -- these two actions do not need to happen in sequence. Or maybe it already is done like that? <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/28163435-do-parallel-local-and-remote-discovery-at-the-same-time?utm_campaign=plugin&utm_content=tracker%2F216457&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F216457&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_defect
do parallel local and remote discovery at the same time one could consider to ask for remote and local changes at the same time these two actions do not need to happen in sequence or maybe it already is done like that want to back this issue we accept bounties via
0
18,960
13,178,160,058
IssuesEvent
2020-08-12 08:40:35
pyinstaller/pyinstaller
https://api.github.com/repos/pyinstaller/pyinstaller
closed
Add check for correct names of new-fragments
area:project infrastructure feature
While releasing the last version, I was hindered by some news-fragments having the wrong filename. While this is not much of an issue, it costs yet another few minutes to solve. But for a possible (half-)automated release pipeline, this would make some entries vanish from the change log. The check could be part of the linting step. Valid names are in <https://github.com/pyinstaller/pyinstaller/blob/develop/doc/development/changelog-entries.rst> draft pattern: `([^.]+)\.(feature|bugfix|…)\.rst` -> validate with towncrier
1.0
Add check for correct names of new-fragments - While releasing the last version, I was hindered by some news-fragments having the wrong filename. While this is not much of an issue, it costs yet another few minutes to solve. But for a possible (half-)automated release pipeline, this would make some entries vanish from the change log. The check could be part of the linting step. Valid names are in <https://github.com/pyinstaller/pyinstaller/blob/develop/doc/development/changelog-entries.rst> draft pattern: `([^.]+)\.(feature|bugfix|…)\.rst` -> validate with towncrier
non_defect
add check for correct names of new fragments while releasing the last version i was hindered by some news fragments having the wrong filename while this is not much of an issue it costs yet another few minutes to solve but for a possible half automated release pipeline this would make some entries vanish from the change log the check could be part of the linting step valid names are in draft pattern feature bugfix … rst validate with towncrier
0
16,241
2,878,875,594
IssuesEvent
2015-06-10 06:00:51
msg4real/pygooglevoice
https://api.github.com/repos/msg4real/pygooglevoice
closed
Login stopped working as of 9/6 pm
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Try to log in, e.g. user = 'username' pswd = 'passwd' v = Voice() v.login(user, pswd) What is the expected output? What do you see instead? ... File "/usr/local/bin/gvtweak", line 53, in gv_set_athome voice.login(user, pwd) File "/usr/local/lib/python2.6/dist-packages/pygooglevoice-0.5-py2.6.egg/googlevoice/voice.py", line 78, in login raise LoginError googlevoice.util.LoginError What version of the product are you using? On what operating system? 0.5 ubuntu Please provide any additional information below. - Run via a cron job successfully and regularly for months - No recent config/account changes - Can still log in via browser interface - Additional troubleshooting suggestions requested ``` Original issue reported on code.google.com by `ser...@gleizer.net` on 7 Sep 2011 at 3:27
1.0
Login stopped working as of 9/6 pm - ``` What steps will reproduce the problem? 1. Try to log in, e.g. user = 'username' pswd = 'passwd' v = Voice() v.login(user, pswd) What is the expected output? What do you see instead? ... File "/usr/local/bin/gvtweak", line 53, in gv_set_athome voice.login(user, pwd) File "/usr/local/lib/python2.6/dist-packages/pygooglevoice-0.5-py2.6.egg/googlevoice/voice.py", line 78, in login raise LoginError googlevoice.util.LoginError What version of the product are you using? On what operating system? 0.5 ubuntu Please provide any additional information below. - Run via a cron job successfully and regularly for months - No recent config/account changes - Can still log in via browser interface - Additional troubleshooting suggestions requested ``` Original issue reported on code.google.com by `ser...@gleizer.net` on 7 Sep 2011 at 3:27
defect
login stopped working as of pm what steps will reproduce the problem try to log in e g user username pswd passwd v voice v login user pswd what is the expected output what do you see instead file usr local bin gvtweak line in gv set athome voice login user pwd file usr local lib dist packages pygooglevoice egg googlevoice voice py line in login raise loginerror googlevoice util loginerror what version of the product are you using on what operating system ubuntu please provide any additional information below run via a cron job successfully and regularly for months no recent config account changes can still log in via browser interface additional troubleshooting suggestions requested original issue reported on code google com by ser gleizer net on sep at
1
378,523
26,324,546,008
IssuesEvent
2023-01-10 04:46:15
fern-api/fern
https://api.github.com/repos/fern-api/fern
closed
[decomp] Change headings in "Defining your API"
documentation
> In the Defining your API do you think it’s beneficial to use the same categorizations as in the second sentence of the intro (i.e. endpoints, types, errors, and examples)? > -> We use services now which is a new concept.
1.0
[decomp] Change headings in "Defining your API" - > In the Defining your API do you think it’s beneficial to use the same categorizations as in the second sentence of the intro (i.e. endpoints, types, errors, and examples)? > -> We use services now which is a new concept.
non_defect
change headings in defining your api in the defining your api do you think it’s beneficial to use the same categorizations as in the second sentence of the intro i e endpoints types errors and examples we use services now which is a new concept
0
35,858
7,810,734,976
IssuesEvent
2018-06-12 07:55:41
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Calendar: selecting date requires doubleclick if minDate or maxDate are assigned a new Date object
defect
**I'm submitting a ...** (check one with "x") ``` [x] bug report => Search github for a similar issue or PR before submitting [ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap [ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35 ``` **Current behavior** If the minDate or maxDate field of a p-calendar is assigned a newly created Date object, a double-click is then required to select a date from the popup **Expected behavior** User should be able to select a date with a single click, as per normal **Minimal reproduction of the problem with instructions** Plunker: http://plnkr.co/edit/yvk40ILn5TNJl4vK9GpH?p=preview The minimal repro example in the calendar just assigns a new Date() to the minDate field, to show this bug in operation. After that I included an example more along the lines of what I am trying to do - have calculated min/max dates based on values at run time. In this case, have a calendar representing a "start date" that must be at least one day before the value of the "end date" calendar, and vice-versa for end date being at least one day after start. The start/end example lets the user select one date as per normal just fine, but after that every subsequent selection requires two clicks. **What is the motivation / use case for changing the behavior?** This is an interface bug **Please tell us about your environment:** * **Angular version:** 4.4.4 * **PrimeNG version:** 4.1.2 * **Browser:** checked in [Chrome 66.0.3359.26 | Safari 11.0.3 ] but probably all * **Language:** [all]
1.0
Calendar: selecting date requires doubleclick if minDate or maxDate are assigned a new Date object - **I'm submitting a ...** (check one with "x") ``` [x] bug report => Search github for a similar issue or PR before submitting [ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap [ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35 ``` **Current behavior** If the minDate or maxDate field of a p-calendar is assigned a newly created Date object, a double-click is then required to select a date from the popup **Expected behavior** User should be able to select a date with a single click, as per normal **Minimal reproduction of the problem with instructions** Plunker: http://plnkr.co/edit/yvk40ILn5TNJl4vK9GpH?p=preview The minimal repro example in the calendar just assigns a new Date() to the minDate field, to show this bug in operation. After that I included an example more along the lines of what I am trying to do - have calculated min/max dates based on values at run time. In this case, have a calendar representing a "start date" that must be at least one day before the value of the "end date" calendar, and vice-versa for end date being at least one day after start. The start/end example lets the user select one date as per normal just fine, but after that every subsequent selection requires two clicks. **What is the motivation / use case for changing the behavior?** This is an interface bug **Please tell us about your environment:** * **Angular version:** 4.4.4 * **PrimeNG version:** 4.1.2 * **Browser:** checked in [Chrome 66.0.3359.26 | Safari 11.0.3 ] but probably all * **Language:** [all]
defect
calendar selecting date requires doubleclick if mindate or maxdate are assigned a new date object i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see current behavior if the mindate or maxdate field of a p calendar is assigned a newly created date object a double click is then required to select a date from the popup expected behavior user should be able to select a date with a single click as per normal minimal reproduction of the problem with instructions plunker the minimal repro example in the calendar just assigns a new date to the mindate field to show this bug in operation after that i included an example more along the lines of what i am trying to do have calculated min max dates based on values at run time in this case have a calendar representing a start date that must be at least one day before the value of the end date calendar and vice versa for end date being at least one day after start the start end example lets the user select one date as per normal just fine but after that every subsequent selection requires two clicks what is the motivation use case for changing the behavior this is an interface bug please tell us about your environment angular version primeng version browser checked in but probably all language
1
39,189
9,303,969,267
IssuesEvent
2019-03-24 21:27:48
Jonah-Roesler-Edu/ABJ-SoftSol_A02
https://api.github.com/repos/Jonah-Roesler-Edu/ABJ-SoftSol_A02
closed
Clean up the Code repository
defect
Please consider cleaning up your code repository, you have like 5 different branches for iteration 2, you should probably merge or remove some...
1.0
Clean up the Code repository - Please consider cleaning up your code repository, you have like 5 different branches for iteration 2, you should probably merge or remove some...
defect
clean up the code repository please consider cleaning up your code repository you have like different branches for iteration you should probably merge or remove some
1
12,352
2,693,354,629
IssuesEvent
2015-04-01 14:21:18
akvo/akvo-flow-mobile
https://api.github.com/repos/akvo/akvo-flow-mobile
closed
Use explicit error messages for resource download failures
Defect
# Overview When a cascade resource download fails, the notification error is rather generic: *Form resources have been fully downloaded*. This seems to cause confusion. Instead, we'll show a more explicit message, directly referencing the cascade.
1.0
Use explicit error messages for resource download failures - # Overview When a cascade resource download fails, the notification error is rather generic: *Form resources have been fully downloaded*. This seems to cause confusion. Instead, we'll show a more explicit message, directly referencing the cascade.
defect
use explicit error messages for resource download failures overview when a cascade resource download fails the notification error is rather generic form resources have been fully downloaded this seems to cause confusion instead we ll show a more explicit message directly referencing the cascade
1
309,799
26,679,007,722
IssuesEvent
2023-01-26 16:15:14
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix miscellaneous_ops.test_torch_renorm
PyTorch Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix miscellaneous_ops.test_torch_renorm - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4014333560/jobs/6894720191" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
non_defect
fix miscellaneous ops test torch renorm tensorflow img src torch img src numpy img src jax img src
0
245,624
7,888,760,854
IssuesEvent
2018-06-27 23:44:44
Baystation12/Baystation12
https://api.github.com/repos/Baystation12/Baystation12
closed
Runtime in drill.dm,238: Cannot read .len
bug has known steps to reproduce priority: low runtime error ☢
Runtime at drill.dm,238 on 07cc2830df40bbb609d4f836701c7ae8f53d5b7a: ``` Runtime in drill.dm,238: Cannot read .len proc name: check supports (/obj/machinery/mining/drill/proc/check_supports) usr: Unknown (as Scrappie) (/mob/living/carbon/human) ([0x30000ff]) (the plating) (144,43,9) (/turf/simulated/floor/plating) (noblecaos) usr.loc: The plating (/turf/simulated/floor/plating) ([0x105035f]) (144,43,9) (/area/guppy_hangar/start) ```
1.0
Runtime in drill.dm,238: Cannot read .len - Runtime at drill.dm,238 on 07cc2830df40bbb609d4f836701c7ae8f53d5b7a: ``` Runtime in drill.dm,238: Cannot read .len proc name: check supports (/obj/machinery/mining/drill/proc/check_supports) usr: Unknown (as Scrappie) (/mob/living/carbon/human) ([0x30000ff]) (the plating) (144,43,9) (/turf/simulated/floor/plating) (noblecaos) usr.loc: The plating (/turf/simulated/floor/plating) ([0x105035f]) (144,43,9) (/area/guppy_hangar/start) ```
non_defect
runtime in drill dm cannot read len runtime at drill dm on runtime in drill dm cannot read len proc name check supports obj machinery mining drill proc check supports usr unknown as scrappie mob living carbon human the plating turf simulated floor plating noblecaos usr loc the plating turf simulated floor plating area guppy hangar start
0
176,587
21,411,783,471
IssuesEvent
2022-04-22 06:58:22
AlexRogalskiy/java-patterns
https://api.github.com/repos/AlexRogalskiy/java-patterns
opened
CVE-2018-1107 (Medium) detected in is-my-json-valid-2.13.1.tgz
security vulnerability
## CVE-2018-1107 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.13.1.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.13.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.13.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npmi/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - gitbook-cli-2.3.2.tgz (Root Library) - npmi-1.0.1.tgz - npm-2.15.12.tgz - request-2.74.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.13.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java-patterns/commit/0e3f838823fb09cc237bb3fc8f2e2651a2d0f0e6">0e3f838823fb09cc237bb3fc8f2e2651a2d0f0e6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> It was discovered that the is-my-json-valid JavaScript library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated. <p>Publish Date: 2021-03-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-4hpf-3wq7-5rpr">https://github.com/advisories/GHSA-4hpf-3wq7-5rpr</a></p> <p>Release Date: 2021-03-30</p> <p>Fix Resolution: is-my-json-valid - 1.4.1, 2.17.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-1107 (Medium) detected in is-my-json-valid-2.13.1.tgz - ## CVE-2018-1107 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.13.1.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.13.1.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.13.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/npmi/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - gitbook-cli-2.3.2.tgz (Root Library) - npmi-1.0.1.tgz - npm-2.15.12.tgz - request-2.74.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.13.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java-patterns/commit/0e3f838823fb09cc237bb3fc8f2e2651a2d0f0e6">0e3f838823fb09cc237bb3fc8f2e2651a2d0f0e6</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> It was discovered that the is-my-json-valid JavaScript library used an inefficient regular expression to validate JSON fields defined to have email format. A specially crafted JSON file could cause it to consume an excessive amount of CPU time when validated. <p>Publish Date: 2021-03-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107>CVE-2018-1107</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-4hpf-3wq7-5rpr">https://github.com/advisories/GHSA-4hpf-3wq7-5rpr</a></p> <p>Release Date: 2021-03-30</p> <p>Fix Resolution: is-my-json-valid - 1.4.1, 2.17.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in is my json valid tgz cve medium severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file package json path to vulnerable library node modules npmi node modules npm node modules request node modules har validator node modules is my json valid package json dependency hierarchy gitbook cli tgz root library npmi tgz npm tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch master vulnerability details it was discovered that the is my json valid javascript library used an inefficient regular expression to validate json fields defined to have email format a specially crafted json file could cause it to consume an excessive amount of cpu time when validated publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid step up your open source security game with whitesource
0
545,450
15,950,900,975
IssuesEvent
2021-04-15 09:12:16
web-platform-tests/wpt
https://api.github.com/repos/web-platform-tests/wpt
closed
Mechanism to pin browser versions
Azure Pipelines Taskcluster infra priority:backlog
For both Chrome Dev and Safari Technology Preview, new versions have occasionally had so serious bugs that we've had to revert to using an older version in Taskcluster and Azure Pipelines respectively. See, https://github.com/web-platform-tests/wpt/issues/17186 + https://github.com/web-platform-tests/wpt/issues/19362 for those cases, and https://github.com/web-platform-tests/wpt/pull/17187 + https://github.com/web-platform-tests/wpt/pull/19360 for how the actual pinning was done. Pinning by directly changing the code that installs the browser works (if you know about it) but I want to propose something similar to requirements.txt to encode the version information, while retaining the ability to just say "latest" when pinning isn't wanted. This could layer on top of ideas that @jgraham has to pass the browser download URL to wptrunner, and picking the version in a decision task to ensure the same version is tested in every shard. Advantages: - easier to pin in response to some breakage if the mechanism exists and is regularly exercised - this or the supporting code for it would make it easier to get runs of browser version N and N+1 from same state of WPT for comparison Whether pinning the WebDriver versions is valuable I don't know, so not suggesting that now. cc @jugglinmike who spent some effort on reproducible builds for https://github.com/web-platform-tests/results-collection.
1.0
Mechanism to pin browser versions - For both Chrome Dev and Safari Technology Preview, new versions have occasionally had so serious bugs that we've had to revert to using an older version in Taskcluster and Azure Pipelines respectively. See, https://github.com/web-platform-tests/wpt/issues/17186 + https://github.com/web-platform-tests/wpt/issues/19362 for those cases, and https://github.com/web-platform-tests/wpt/pull/17187 + https://github.com/web-platform-tests/wpt/pull/19360 for how the actual pinning was done. Pinning by directly changing the code that installs the browser works (if you know about it) but I want to propose something similar to requirements.txt to encode the version information, while retaining the ability to just say "latest" when pinning isn't wanted. This could layer on top of ideas that @jgraham has to pass the browser download URL to wptrunner, and picking the version in a decision task to ensure the same version is tested in every shard. Advantages: - easier to pin in response to some breakage if the mechanism exists and is regularly exercised - this or the supporting code for it would make it easier to get runs of browser version N and N+1 from same state of WPT for comparison Whether pinning the WebDriver versions is valuable I don't know, so not suggesting that now. cc @jugglinmike who spent some effort on reproducible builds for https://github.com/web-platform-tests/results-collection.
non_defect
mechanism to pin browser versions for both chrome dev and safari technology preview new versions have occasionally had so serious bugs that we ve had to revert to using an older version in taskcluster and azure pipelines respectively see for those cases and for how the actual pinning was done pinning by directly changing the code that installs the browser works if you know about it but i want to propose something similar to requirements txt to encode the version information while retaining the ability to just say latest when pinning isn t wanted this could layer on top of ideas that jgraham has to pass the browser download url to wptrunner and picking the version in a decision task to ensure the same version is tested in every shard advantages easier to pin in response to some breakage if the mechanism exists and is regularly exercised this or the supporting code for it would make it easier to get runs of browser version n and n from same state of wpt for comparison whether pinning the webdriver versions is valuable i don t know so not suggesting that now cc jugglinmike who spent some effort on reproducible builds for
0