Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
110,201
11,694,119,216
IssuesEvent
2020-03-06 02:53:37
AGHSEagleRobotics/frc1388-2020
https://api.github.com/repos/AGHSEagleRobotics/frc1388-2020
closed
Wiring List (spreadsheet)
documentation
The wiring and placement of electronic components and sensors needs to be tracked. This will be done with a spreadsheet linked here: https://docs.google.com/spreadsheets/d/1-_TZ13xsfVZVFzq8rJBhIH7Ht_RBKVlgj4VI8T67ISo/edit#gid=0 This needs to be updated regularly to ensure good communication and planning.
1.0
Wiring List (spreadsheet) - The wiring and placement of electronic components and sensors needs to be tracked. This will be done with a spreadsheet linked here: https://docs.google.com/spreadsheets/d/1-_TZ13xsfVZVFzq8rJBhIH7Ht_RBKVlgj4VI8T67ISo/edit#gid=0 This needs to be updated regularly to ensure good communication and planning.
non_defect
wiring list spreadsheet the wiring and placement of electronic components and sensors needs to be tracked this will be done with a spreadsheet linked here this needs to be updated regularly to ensure good communication and planning
0
22,842
3,711,518,664
IssuesEvent
2016-03-02 10:40:19
OpenMS/OpenMS
https://api.github.com/repos/OpenMS/OpenMS
closed
MzTabExporter crash on idXML export
defect major
The `MzTabExporter` tool crashes on this `idXML` [1] with ``` Error: Unexpected internal error (Could not convert non-StringList DataValue to StringList) ``` [1] https://gist.github.com/lars20070/adfebc5fc6c44c641601
1.0
MzTabExporter crash on idXML export - The `MzTabExporter` tool crashes on this `idXML` [1] with ``` Error: Unexpected internal error (Could not convert non-StringList DataValue to StringList) ``` [1] https://gist.github.com/lars20070/adfebc5fc6c44c641601
defect
mztabexporter crash on idxml export the mztabexporter tool crashes on this idxml with error unexpected internal error could not convert non stringlist datavalue to stringlist
1
279,400
24,222,497,929
IssuesEvent
2022-09-26 12:04:10
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Add tests for Console's context menu and its actions
Feature:Console Feature:Dev Tools Team:Deployment Management test-coverage
## Summary Console offers a context menu per request, which enables the user to perform a number of actions with the relevant request. Let's add functional tests to document and verify this behavior: - [ ] Request-specific context menu - [ ] Copy the request as cURL - [ ] Open a documentation link for the relevant API - [ ] Auto-indent ## Content menu and its actions ### Request-specific context menu You can use this context menu to copy the request as cURL, open a documentation link for the relevant API, and auto-indent the request body. ![image](https://user-images.githubusercontent.com/1238659/61909736-c9554d80-aee7-11e9-879f-018af82b91e7.png) ### Auto-indent Auto-indent will format the request body so that it has indentation at each nesting level, and empty objects will be collapsed so that the opening and closing brackets are on the same line. ![image](https://user-images.githubusercontent.com/1238659/61909983-4b457680-aee8-11e9-95ed-7abe3d8b73e3.png) When auto-indented, the above becomes nicely formatted like this: ![image](https://user-images.githubusercontent.com/1238659/61909984-4b457680-aee8-11e9-902f-fe082b55af0d.png) Auto-indenting again will condense the JSON so that it only occupies a single line and has no whitespace: ![image](https://user-images.githubusercontent.com/1238659/61910078-7def6f00-aee8-11e9-8977-346a5f30d845.png) Subsequent auto-indents alternate between the condensed form and the nicely indented form. Note that auto-indenting a triple-quoted string like this: ``` POST test/1 { "test": """a b""" } ``` Will replace newlines with a newline character: ``` POST test/1 {"test":"a\n b"} ``` Auto-indenting again will return to the original triple-quoted string form.
1.0
Add tests for Console's context menu and its actions - ## Summary Console offers a context menu per request, which enables the user to perform a number of actions with the relevant request. Let's add functional tests to document and verify this behavior: - [ ] Request-specific context menu - [ ] Copy the request as cURL - [ ] Open a documentation link for the relevant API - [ ] Auto-indent ## Content menu and its actions ### Request-specific context menu You can use this context menu to copy the request as cURL, open a documentation link for the relevant API, and auto-indent the request body. ![image](https://user-images.githubusercontent.com/1238659/61909736-c9554d80-aee7-11e9-879f-018af82b91e7.png) ### Auto-indent Auto-indent will format the request body so that it has indentation at each nesting level, and empty objects will be collapsed so that the opening and closing brackets are on the same line. ![image](https://user-images.githubusercontent.com/1238659/61909983-4b457680-aee8-11e9-95ed-7abe3d8b73e3.png) When auto-indented, the above becomes nicely formatted like this: ![image](https://user-images.githubusercontent.com/1238659/61909984-4b457680-aee8-11e9-902f-fe082b55af0d.png) Auto-indenting again will condense the JSON so that it only occupies a single line and has no whitespace: ![image](https://user-images.githubusercontent.com/1238659/61910078-7def6f00-aee8-11e9-8977-346a5f30d845.png) Subsequent auto-indents alternate between the condensed form and the nicely indented form. Note that auto-indenting a triple-quoted string like this: ``` POST test/1 { "test": """a b""" } ``` Will replace newlines with a newline character: ``` POST test/1 {"test":"a\n b"} ``` Auto-indenting again will return to the original triple-quoted string form.
non_defect
add tests for console s context menu and its actions summary console offers a context menu per request which enables the user to perform a number of actions with the relevant request let s add functional tests to document and verify this behavior request specific context menu copy the request as curl open a documentation link for the relevant api auto indent content menu and its actions request specific context menu you can use this context menu to copy the request as curl open a documentation link for the relevant api and auto indent the request body auto indent auto indent will format the request body so that it has indentation at each nesting level and empty objects will be collapsed so that the opening and closing brackets are on the same line when auto indented the above becomes nicely formatted like this auto indenting again will condense the json so that it only occupies a single line and has no whitespace subsequent auto indents alternate between the condensed form and the nicely indented form note that auto indenting a triple quoted string like this post test test a b will replace newlines with a newline character post test test a n b auto indenting again will return to the original triple quoted string form
0
15,352
19,522,523,109
IssuesEvent
2021-12-29 21:41:53
metabase/metabase
https://api.github.com/repos/metabase/metabase
opened
Error shown for certain column types when getting results from cache
Type:Bug Priority:P2 Database/Redshift Querying/Processor .Backend
**Describe the bug** When returning an `interval` column type via cached results, then an error is shown in the frontend instead of the results. I'm unsure if this is specific to Redshift or generally for unknown column types. I've not been able to reproduce on Postgres. **To Reproduce** 1. Admin > Settings > Caching: Duration=`0.0001` and TTL=`10000` 2. Native query > **Redshift** > `select interval '1 day'` - run query - see results - save question 3. Do browser refresh a couple of time, so it starts returning cached results (or is supposed to) The returned "result" is: `{"nippy/unthawable":{"type":"serializable","cause":"quarantined","class-name":"com.amazon.redshift.util.RedshiftInterval","content":"0xACED0005"}}` And the log only contains the following related to this, no stacktraces or anything real errors: `WARN metabase.driver.common Don't know how to map class 'class com.amazon.redshift.util.RedshiftInterval' to a Field base_type, falling back to :type/*.` ![image](https://user-images.githubusercontent.com/1447303/147705140-d518edcf-1bd4-4ea2-bce1-84f0ba7c577f.png) **Expected behavior** 1. Correct results from cache, of course 2. An error or stracktrace in the log. **Information about your Metabase Installation:** Tested 1.41.5 and master `b9bee5d`
1.0
Error shown for certain column types when getting results from cache - **Describe the bug** When returning an `interval` column type via cached results, then an error is shown in the frontend instead of the results. I'm unsure if this is specific to Redshift or generally for unknown column types. I've not been able to reproduce on Postgres. **To Reproduce** 1. Admin > Settings > Caching: Duration=`0.0001` and TTL=`10000` 2. Native query > **Redshift** > `select interval '1 day'` - run query - see results - save question 3. Do browser refresh a couple of time, so it starts returning cached results (or is supposed to) The returned "result" is: `{"nippy/unthawable":{"type":"serializable","cause":"quarantined","class-name":"com.amazon.redshift.util.RedshiftInterval","content":"0xACED0005"}}` And the log only contains the following related to this, no stacktraces or anything real errors: `WARN metabase.driver.common Don't know how to map class 'class com.amazon.redshift.util.RedshiftInterval' to a Field base_type, falling back to :type/*.` ![image](https://user-images.githubusercontent.com/1447303/147705140-d518edcf-1bd4-4ea2-bce1-84f0ba7c577f.png) **Expected behavior** 1. Correct results from cache, of course 2. An error or stracktrace in the log. **Information about your Metabase Installation:** Tested 1.41.5 and master `b9bee5d`
non_defect
error shown for certain column types when getting results from cache describe the bug when returning an interval column type via cached results then an error is shown in the frontend instead of the results i m unsure if this is specific to redshift or generally for unknown column types i ve not been able to reproduce on postgres to reproduce admin settings caching duration and ttl native query redshift select interval day run query see results save question do browser refresh a couple of time so it starts returning cached results or is supposed to the returned result is nippy unthawable type serializable cause quarantined class name com amazon redshift util redshiftinterval content and the log only contains the following related to this no stacktraces or anything real errors warn metabase driver common don t know how to map class class com amazon redshift util redshiftinterval to a field base type falling back to type expected behavior correct results from cache of course an error or stracktrace in the log information about your metabase installation tested and master
0
23,731
3,851,867,304
IssuesEvent
2016-04-06 05:28:43
GPF/imame4all
https://api.github.com/repos/GPF/imame4all
closed
Not issue, just Thanks for your new 0.134
auto-migrated Priority-Medium Type-Defect
``` Thanks! Thanks! Thanks! Great Man! hope can see you keep update! ``` Original issue reported on code.google.com by `fatcatma...@gmail.com` on 5 Mar 2012 at 6:14
1.0
Not issue, just Thanks for your new 0.134 - ``` Thanks! Thanks! Thanks! Great Man! hope can see you keep update! ``` Original issue reported on code.google.com by `fatcatma...@gmail.com` on 5 Mar 2012 at 6:14
defect
not issue just thanks for your new thanks thanks thanks great man hope can see you keep update original issue reported on code google com by fatcatma gmail com on mar at
1
79,776
10,141,769,542
IssuesEvent
2019-08-03 17:12:26
microsoft/Requirements
https://api.github.com/repos/microsoft/Requirements
opened
Better Docs
documentation
Topics that need to be covered * What is a Requirement? * What does a single Requirement look like? * What are all the properties on a Requirement? * Which properties are mandatory/optional? * What are the major patterns for building Requirements? * Dynamically generating Requirements using control flow (`if`, `foreach`, etc) * "Getter/setter" pattern for getting properties of your system * "Generator function" pattern for returning Requirements from a function based on parameters
1.0
Better Docs - Topics that need to be covered * What is a Requirement? * What does a single Requirement look like? * What are all the properties on a Requirement? * Which properties are mandatory/optional? * What are the major patterns for building Requirements? * Dynamically generating Requirements using control flow (`if`, `foreach`, etc) * "Getter/setter" pattern for getting properties of your system * "Generator function" pattern for returning Requirements from a function based on parameters
non_defect
better docs topics that need to be covered what is a requirement what does a single requirement look like what are all the properties on a requirement which properties are mandatory optional what are the major patterns for building requirements dynamically generating requirements using control flow if foreach etc getter setter pattern for getting properties of your system generator function pattern for returning requirements from a function based on parameters
0
532,867
15,572,387,535
IssuesEvent
2021-03-17 06:57:47
datastax/cassandra-quarkus
https://api.github.com/repos/datastax/cassandra-quarkus
closed
Cannot disable metrics in native mode
priority:critical type:bug
Steps to reproduce: * Create a quickstart project on quarkus.io with the C* extension; * Disable metrics in application.properties; * Run the native tests with `mvn clean verify -Dnative`. Expected outcome: the tests pass. Observed outcome: ``` ERROR [io.qua.run.Application] (main) Failed to start application (with profile prod): java.lang.IllegalArgumentException: Can't find class DropwizardMetricsFactory (specified by advanced.metrics.factory.class) at com.datastax.oss.driver.internal.core.util.Reflection.buildFromConfig(Reflection.java:219) at com.datastax.oss.driver.internal.core.util.Reflection.buildFromConfig(Reflection.java:108) at com.datastax.oss.driver.internal.core.context.DefaultDriverContext.buildMetricsFactory(DefaultDriverContext.java:605) ``` This is due to `DropwizardMetricsFactory` not being registered for reflection, but it should be since that's the driver's default implementation of `MetricsFactory`, and it also serves as the de facto no-op implementation when metrics are disabled.
1.0
Cannot disable metrics in native mode - Steps to reproduce: * Create a quickstart project on quarkus.io with the C* extension; * Disable metrics in application.properties; * Run the native tests with `mvn clean verify -Dnative`. Expected outcome: the tests pass. Observed outcome: ``` ERROR [io.qua.run.Application] (main) Failed to start application (with profile prod): java.lang.IllegalArgumentException: Can't find class DropwizardMetricsFactory (specified by advanced.metrics.factory.class) at com.datastax.oss.driver.internal.core.util.Reflection.buildFromConfig(Reflection.java:219) at com.datastax.oss.driver.internal.core.util.Reflection.buildFromConfig(Reflection.java:108) at com.datastax.oss.driver.internal.core.context.DefaultDriverContext.buildMetricsFactory(DefaultDriverContext.java:605) ``` This is due to `DropwizardMetricsFactory` not being registered for reflection, but it should be since that's the driver's default implementation of `MetricsFactory`, and it also serves as the de facto no-op implementation when metrics are disabled.
non_defect
cannot disable metrics in native mode steps to reproduce create a quickstart project on quarkus io with the c extension disable metrics in application properties run the native tests with mvn clean verify dnative expected outcome the tests pass observed outcome error main failed to start application with profile prod java lang illegalargumentexception can t find class dropwizardmetricsfactory specified by advanced metrics factory class at com datastax oss driver internal core util reflection buildfromconfig reflection java at com datastax oss driver internal core util reflection buildfromconfig reflection java at com datastax oss driver internal core context defaultdrivercontext buildmetricsfactory defaultdrivercontext java this is due to dropwizardmetricsfactory not being registered for reflection but it should be since that s the driver s default implementation of metricsfactory and it also serves as the de facto no op implementation when metrics are disabled
0
65,146
12,533,998,156
IssuesEvent
2020-06-04 18:36:51
pi-hole/AdminLTE
https://api.github.com/repos/pi-hole/AdminLTE
opened
Remove `initCheckboxRadioStyle()`
Code maintenance
footer.js is loaded synchronously, which means that `initCheckboxRadioStyle()` fires up after the DOM has loaded. This should be removed and the logic moved to PHP.
1.0
Remove `initCheckboxRadioStyle()` - footer.js is loaded synchronously, which means that `initCheckboxRadioStyle()` fires up after the DOM has loaded. This should be removed and the logic moved to PHP.
non_defect
remove initcheckboxradiostyle footer js is loaded synchronously which means that initcheckboxradiostyle fires up after the dom has loaded this should be removed and the logic moved to php
0
46,693
13,055,960,432
IssuesEvent
2020-07-30 03:14:32
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
building multiple copies of HTML docs (Trac #1723)
Incomplete Migration Migrated from Trac defect other
Migrated from https://code.icecube.wisc.edu/ticket/1723 ```json { "status": "closed", "changetime": "2016-06-09T14:50:36", "description": "with the addition of sphinx-apidoc in r2580/IceTray we're now building two copies of the docs that step on each other and barf warnings everywhere", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1465483836277867", "component": "other", "summary": "building multiple copies of HTML docs", "priority": "normal", "keywords": "documentation sphinx html", "time": "2016-06-01T21:10:22", "milestone": "", "owner": "kjmeagher", "type": "defect" } ```
1.0
building multiple copies of HTML docs (Trac #1723) - Migrated from https://code.icecube.wisc.edu/ticket/1723 ```json { "status": "closed", "changetime": "2016-06-09T14:50:36", "description": "with the addition of sphinx-apidoc in r2580/IceTray we're now building two copies of the docs that step on each other and barf warnings everywhere", "reporter": "nega", "cc": "", "resolution": "fixed", "_ts": "1465483836277867", "component": "other", "summary": "building multiple copies of HTML docs", "priority": "normal", "keywords": "documentation sphinx html", "time": "2016-06-01T21:10:22", "milestone": "", "owner": "kjmeagher", "type": "defect" } ```
defect
building multiple copies of html docs trac migrated from json status closed changetime description with the addition of sphinx apidoc in icetray we re now building two copies of the docs that step on each other and barf warnings everywhere reporter nega cc resolution fixed ts component other summary building multiple copies of html docs priority normal keywords documentation sphinx html time milestone owner kjmeagher type defect
1
29,277
5,632,301,722
IssuesEvent
2017-04-05 16:16:31
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
Call to a member function schema() on null ... AclComponent.php on line 298
Defect On hold
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 2.0.5 * Platform and Target: Windows 10 xampp ### What you did this is de 298 line in AclComponent: `$permKeys = $this->_getAcoKeys($this->Aro->Permission->schema()); ` ### What happened after login in my cakephp 2 with acl i get this error `Fatal error: Call to a member function schema() on null in D:\ProgramFiles\xampp\htdocs\cgp\lib\Cake\Controller\Component\AclComponent.php on line 298 ` any help for this?
1.0
Call to a member function schema() on null ... AclComponent.php on line 298 - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 2.0.5 * Platform and Target: Windows 10 xampp ### What you did this is de 298 line in AclComponent: `$permKeys = $this->_getAcoKeys($this->Aro->Permission->schema()); ` ### What happened after login in my cakephp 2 with acl i get this error `Fatal error: Call to a member function schema() on null in D:\ProgramFiles\xampp\htdocs\cgp\lib\Cake\Controller\Component\AclComponent.php on line 298 ` any help for this?
defect
call to a member function schema on null aclcomponent php on line this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target windows xampp what you did this is de line in aclcomponent permkeys this getacokeys this aro permission schema what happened after login in my cakephp with acl i get this error fatal error call to a member function schema on null in d programfiles xampp htdocs cgp lib cake controller component aclcomponent php on line any help for this
1
34,244
12,258,886,846
IssuesEvent
2020-05-06 15:45:36
DbugTech/dbugtech.github.io
https://api.github.com/repos/DbugTech/dbugtech.github.io
opened
CVE-2020-11022 (Medium) detected in jquery-2.1.4.min.js, jquery-1.11.3.min.js
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.11.3.min.js</b></p></summary> <p> <details><summary><b>jquery-2.1.4.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/dbugtech.github.io/portfolio/work-9.html</p> <p>Path to vulnerable library: /dbugtech.github.io/portfolio/../js/jquery-2.1.4.min.js,/dbugtech.github.io/js/jquery-2.1.4.min.js,/dbugtech.github.io/js/jquery-2.1.4.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-2.1.4.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.3.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js</a></p> <p>Path to vulnerable library: /dbugtech.github.io/showcase/assets/js/jquery.min.js,/dbugtech.github.io/showcase/assets/js/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.3.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/DbugTech/dbugtech.github.io/commit/d662c7405f30a0b3bd84a5ce6d443879cac3063d">d662c7405f30a0b3bd84a5ce6d443879cac3063d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11022 (Medium) detected in jquery-2.1.4.min.js, jquery-1.11.3.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.11.3.min.js</b></p></summary> <p> <details><summary><b>jquery-2.1.4.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/dbugtech.github.io/portfolio/work-9.html</p> <p>Path to vulnerable library: /dbugtech.github.io/portfolio/../js/jquery-2.1.4.min.js,/dbugtech.github.io/js/jquery-2.1.4.min.js,/dbugtech.github.io/js/jquery-2.1.4.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-2.1.4.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.3.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js</a></p> <p>Path to vulnerable library: /dbugtech.github.io/showcase/assets/js/jquery.min.js,/dbugtech.github.io/showcase/assets/js/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.3.min.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/DbugTech/dbugtech.github.io/commit/d662c7405f30a0b3bd84a5ce6d443879cac3063d">d662c7405f30a0b3bd84a5ce6d443879cac3063d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in jquery min js jquery min js cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm dbugtech github io portfolio work html path to vulnerable library dbugtech github io portfolio js jquery min js dbugtech github io js jquery min js dbugtech github io js jquery min js dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to vulnerable library dbugtech github io showcase assets js jquery min js dbugtech github io showcase assets js jquery min js dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details in jquery before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
777,638
27,289,102,780
IssuesEvent
2023-02-23 15:25:31
PIP-Technical-Team/pipapi
https://api.github.com/repos/PIP-Technical-Team/pipapi
closed
review potential issue with ui_cp_download
Priority: 2-High Type: 1-Bug
`ui_cp_download_single` sometimes returns data.frame with duplciated rows l594 of `ui_functions.R` Need to figure out why this is happening.
1.0
review potential issue with ui_cp_download - `ui_cp_download_single` sometimes returns data.frame with duplciated rows l594 of `ui_functions.R` Need to figure out why this is happening.
non_defect
review potential issue with ui cp download ui cp download single sometimes returns data frame with duplciated rows of ui functions r need to figure out why this is happening
0
688,786
23,596,307,596
IssuesEvent
2022-08-23 19:36:01
GoogleChrome/lighthouse
https://api.github.com/repos/GoogleChrome/lighthouse
closed
PageSpeed Insights showing as same resource(file) loaded multiple times
needs-priority PSI/LR
Site URL: https://getgist.com/ I have tried analysing the performance of our site in the [PageSpeed Insights developers site](https://developers.google.com/speed/pagespeed/insights/) and found that the results showing as same resource loaded multiple times and the **total requests count as 422** with around **10MB** of data transferred which doesn't looks to be correct. Also, I have analysed the same URL in [GTMetrix](https://gtmetrix.com/) and it shows the **total requests count as 49** with **889KB** of data transferred. **Screenshots from PageSpeed insights,** <img width="1077" alt="Screenshot 2019-05-18 at 5 48 09 PM" src="https://user-images.githubusercontent.com/11786333/57969499-24713e00-7995-11e9-8ada-6afbd2cd9dca.png"> <img width="840" alt="Screenshot 2019-05-18 at 5 33 18 PM" src="https://user-images.githubusercontent.com/11786333/57969399-1f5fbf00-7994-11e9-8372-35ece5d04243.png"> <img width="836" alt="Screenshot 2019-05-18 at 5 33 27 PM" src="https://user-images.githubusercontent.com/11786333/57969400-1ff85580-7994-11e9-90e4-93748f3cc769.png"> <img width="829" alt="Screenshot 2019-05-18 at 5 33 38 PM" src="https://user-images.githubusercontent.com/11786333/57969401-1ff85580-7994-11e9-93f4-87236d9aa86d.png"> <img width="822" alt="Screenshot 2019-05-18 at 5 33 57 PM" src="https://user-images.githubusercontent.com/11786333/57969402-1ff85580-7994-11e9-9b47-258f983e1abe.png"> <img width="837" alt="Screenshot 2019-05-18 at 5 34 09 PM" src="https://user-images.githubusercontent.com/11786333/57969403-2090ec00-7994-11e9-80f3-2548f9bf9e02.png"> From the second and third screenshots we can see that it showing as **cf-assets-###.min.js** and **.woff** files loaded as multiple times, not sure why it is showing up like this. I even looked up the browser network tab and it doesn't seems to be loading these much resources. **And, here is the GTmetrix report** - https://gtmetrix.com/reports/getgist.com/jYUi6H9W Can anyone please help?
1.0
PageSpeed Insights showing as same resource(file) loaded multiple times - Site URL: https://getgist.com/ I have tried analysing the performance of our site in the [PageSpeed Insights developers site](https://developers.google.com/speed/pagespeed/insights/) and found that the results showing as same resource loaded multiple times and the **total requests count as 422** with around **10MB** of data transferred which doesn't looks to be correct. Also, I have analysed the same URL in [GTMetrix](https://gtmetrix.com/) and it shows the **total requests count as 49** with **889KB** of data transferred. **Screenshots from PageSpeed insights,** <img width="1077" alt="Screenshot 2019-05-18 at 5 48 09 PM" src="https://user-images.githubusercontent.com/11786333/57969499-24713e00-7995-11e9-8ada-6afbd2cd9dca.png"> <img width="840" alt="Screenshot 2019-05-18 at 5 33 18 PM" src="https://user-images.githubusercontent.com/11786333/57969399-1f5fbf00-7994-11e9-8372-35ece5d04243.png"> <img width="836" alt="Screenshot 2019-05-18 at 5 33 27 PM" src="https://user-images.githubusercontent.com/11786333/57969400-1ff85580-7994-11e9-90e4-93748f3cc769.png"> <img width="829" alt="Screenshot 2019-05-18 at 5 33 38 PM" src="https://user-images.githubusercontent.com/11786333/57969401-1ff85580-7994-11e9-93f4-87236d9aa86d.png"> <img width="822" alt="Screenshot 2019-05-18 at 5 33 57 PM" src="https://user-images.githubusercontent.com/11786333/57969402-1ff85580-7994-11e9-9b47-258f983e1abe.png"> <img width="837" alt="Screenshot 2019-05-18 at 5 34 09 PM" src="https://user-images.githubusercontent.com/11786333/57969403-2090ec00-7994-11e9-80f3-2548f9bf9e02.png"> From the second and third screenshots we can see that it showing as **cf-assets-###.min.js** and **.woff** files loaded as multiple times, not sure why it is showing up like this. I even looked up the browser network tab and it doesn't seems to be loading these much resources. **And, here is the GTmetrix report** - https://gtmetrix.com/reports/getgist.com/jYUi6H9W Can anyone please help?
non_defect
pagespeed insights showing as same resource file loaded multiple times site url i have tried analysing the performance of our site in the and found that the results showing as same resource loaded multiple times and the total requests count as with around of data transferred which doesn t looks to be correct also i have analysed the same url in and it shows the total requests count as with of data transferred screenshots from pagespeed insights img width alt screenshot at pm src img width alt screenshot at pm src img width alt screenshot at pm src img width alt screenshot at pm src img width alt screenshot at pm src img width alt screenshot at pm src from the second and third screenshots we can see that it showing as cf assets min js and woff files loaded as multiple times not sure why it is showing up like this i even looked up the browser network tab and it doesn t seems to be loading these much resources and here is the gtmetrix report can anyone please help
0
3,053
2,607,970,564
IssuesEvent
2015-02-26 00:44:23
chrsmithdemos/leveldb
https://api.github.com/repos/chrsmithdemos/leveldb
opened
corruption_test segfaults on ARM
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Compile leveldb's corruption test for armv7a 2. Execute on some such lower-powered device (e.g. beaglebone) 3. Segmentation fault! What is the expected output? What do you see instead? >>EXPECTED>> ==== Test CorruptionTest.Recovery expected=100..100; got=100; bad_keys=0; bad_values=0; missed=0 expected=36..36; got=36; bad_keys=0; bad_values=0; missed=64 ==== Test CorruptionTest.RecoverWriteError ==== Test CorruptionTest.NewFileErrorDuringWrite ==== Test CorruptionTest.TableFile expected=90..99; got=99; bad_keys=0; bad_values=1; missed=0 ==== Test CorruptionTest.TableFileRepair expected=95..99; got=98; bad_keys=0; bad_values=0; missed=2 ==== Test CorruptionTest.TableFileIndexData expected=5000..9999; got=8056; bad_keys=0; bad_values=0; missed=0 ==== Test CorruptionTest.MissingDescriptor expected=1000..1000; got=1000; bad_keys=0; bad_values=0; missed=0 ==== Test CorruptionTest.SequenceNumberRecovery ==== Test CorruptionTest.CorruptedDescriptor ==== Test CorruptionTest.CompactionInputError expected=5..9; got=9; bad_keys=0; bad_values=1; missed=0 expected=10000..10000; got=10000; bad_keys=0; bad_values=0; missed=0 ==== Test CorruptionTest.CompactionInputErrorParanoid ==== Test CorruptionTest.UnrelatedKeys ==== PASSED 12 tests <<EXPECTED<< >>>ACTUAL>>> ==== Test CorruptionTest.Recovery expected=100..100; got=100; bad_keys=0; bad_values=0; missed=0 expected=36..36; got=36; bad_keys=0; bad_values=0; missed=64 ==== Test CorruptionTest.RecoverWriteError ==== Test CorruptionTest.NewFileErrorDuringWrite ==== Test CorruptionTest.TableFile expected=90..99; got=99; bad_keys=0; bad_values=1; missed=0 ==== Test CorruptionTest.TableFileRepair expected=95..99; got=98; bad_keys=0; bad_values=0; missed=2 ==== Test CorruptionTest.TableFileIndexData Segmentation fault <<<ACTUAL<<< What version of the product are you using? On what operating system? LevelDB versions: both 1.15 and git-e353fbc7ea81f12a5694991b708f8f45343594b1 (claims to be 1.17) OS: both Debian 7.5 (wheezy) for beaglebone and chromiumos for beaglebone (not sure what version -- based on kernel 3.8.13 + libstdc++ 6.0.18) Please provide any additional information below. This testcase is flaky (segfaulting approximately 30% of the time) with the following compiler/linker flags: FLAGs: -g2 -ggdb -no-canonical-prefixes CXXFLAGS: -ffunction-sections -fdata-sections -fno-math-errno -fno-buildtin-memcmp -pthread -std=gnu++0x LDFLAGS: -Wl,--gc-sections,--no-wchar-size-warning -std=gnu++0x The testcase is segfaults 100% (unless run under gdb) of the time with the following compiler/linker flags: FLAGs: -O2 -D_FORTIFY_SOURCE=2 -no-canonical-prefixes CXXFLAGS: -ffunction-sections -fdata-sections -fno-math-errno -fno-strict-aliasing -fstack-protector-all -fPIC -fno-buildtin-memcmp -z noexecstack -pthread -std=gnu++0x LDFLAGS: -Wl,--gc-sections,--no-wchar-size-warning -pie -Wl,-z,relrow,-z,now -std=gnu++0x The segfault's back trace is is as follows: Program terminated with signal 11, Segmentation fault. #0 0x0002f79a in leveldb::DecodeEntry(char const*, char const*, unsigned int*, unsigned int*, unsigned int*) () at ./leveldb/table/block.cc:58 58 ./leveldb/table/block.cc: No such file or directory. (gdb) bt #0 0x0002f79a in leveldb::DecodeEntry(char const*, char const*, unsigned int*, unsigned int*, unsigned int*) () at ./leveldb/table/block.cc:58 #1 0x0002fe3e in leveldb::Block::Iter::ParseNextKey() () at ./leveldb/table/block.cc:239 #2 0x0002fd02 in leveldb::Block::Iter::SeekToFirst() () at ./leveldb/table/block.cc:207 #3 0x00032734 in leveldb::IteratorWrapper::SeekToFirst() () at ./leveldb/table/iterator_wrapper.h:45 #4 0x00034e72 in leveldb::(anonymous namespace)::TwoLevelIterator::SeekToFirst() () at ./leveldb/table/two_level_iterator.cc:99 #5 0x00032734 in leveldb::IteratorWrapper::SeekToFirst() () at ./leveldb/table/iterator_wrapper.h:45 #6 0x00034fee in leveldb::(anonymous namespace)::TwoLevelIterator::SkipEmptyDataBlocksForward() () at ./leveldb/table/two_level_iterator.cc:134 #7 0x00034f38 in leveldb::(anonymous namespace)::TwoLevelIterator::Next() () at ./leveldb/table/two_level_iterator.cc:115 #8 0x0003264c in leveldb::IteratorWrapper::Next() () at ./leveldb/table/iterator_wrapper.h:42 #9 0x00032b24 in leveldb::(anonymous namespace)::MergingIterator::Next() () at ./leveldb/table/merger.cc:81 #10 0x0001fff0 in leveldb::(anonymous namespace)::DBIter::FindNextUserEntry(bool, std::string*) () at ./leveldb/db/db_iter.cc:199 #11 0x0001feb4 in leveldb::(anonymous namespace)::DBIter::Next() () at ./leveldb/db/db_iter.cc:170 #12 0x0001076c in leveldb::CorruptionTest::Check(int, int) () at ./leveldb/db/corruption_test.cc:96 #13 0x000115da in leveldb::_Test_TableFileIndexData::_Run() () (gdb) disassemble Dump of assembler code for function _ZN7leveldbL11DecodeEntryEPKcS1_PjS2_S2_: 0x0002f77c <+0>: push {r7, lr} 0x0002f77e <+2>: subne sp, #16 0x0002f780 <+4>: addeq r7, sp, #0 0x0002f782 <+6>: str r0, [r7, #12] 0x0002f784 <+8>: str r1, [r7, #8] 0x0002f786 <+10>: str r2, [r7, #4] 0x0002f788 <+12>: str r3, [r7, #0] 0x0002f78a <+14>: ldr r2, [r7, #8] 0x0002f78c <+16>: ldr r3, [r7, #12] 0x0002f78e <+18>: subs r3, r2, r3 0x0002f790 <+20>: cmp r3, #2 0x0002f792 <+22>: bgt.n 0x2f798 <_ZN7leveldbL11DecodeEntryEPKcS1_PjS2_S2_+28> 0x0002f794 <+24>: movs r3, #0 0x0002f796 <+26>: b.n 0x2f858 <_ZN7leveldbL11DecodeEntryEPKcS1_PjS2_S2_+220> 0x0002f798 <+28>: ldr r3, [r7, #12] => 0x0002f79a <+30>: ldrb r3, [r3, #0] 0x0002f79c <+32>: mov r2, r3 ``` ----- Original issue reported on code.google.com by `vlankh...@google.com` on 22 Aug 2014 at 12:40
1.0
corruption_test segfaults on ARM - ``` What steps will reproduce the problem? 1. Compile leveldb's corruption test for armv7a 2. Execute on some such lower-powered device (e.g. beaglebone) 3. Segmentation fault! What is the expected output? What do you see instead? >>EXPECTED>> ==== Test CorruptionTest.Recovery expected=100..100; got=100; bad_keys=0; bad_values=0; missed=0 expected=36..36; got=36; bad_keys=0; bad_values=0; missed=64 ==== Test CorruptionTest.RecoverWriteError ==== Test CorruptionTest.NewFileErrorDuringWrite ==== Test CorruptionTest.TableFile expected=90..99; got=99; bad_keys=0; bad_values=1; missed=0 ==== Test CorruptionTest.TableFileRepair expected=95..99; got=98; bad_keys=0; bad_values=0; missed=2 ==== Test CorruptionTest.TableFileIndexData expected=5000..9999; got=8056; bad_keys=0; bad_values=0; missed=0 ==== Test CorruptionTest.MissingDescriptor expected=1000..1000; got=1000; bad_keys=0; bad_values=0; missed=0 ==== Test CorruptionTest.SequenceNumberRecovery ==== Test CorruptionTest.CorruptedDescriptor ==== Test CorruptionTest.CompactionInputError expected=5..9; got=9; bad_keys=0; bad_values=1; missed=0 expected=10000..10000; got=10000; bad_keys=0; bad_values=0; missed=0 ==== Test CorruptionTest.CompactionInputErrorParanoid ==== Test CorruptionTest.UnrelatedKeys ==== PASSED 12 tests <<EXPECTED<< >>>ACTUAL>>> ==== Test CorruptionTest.Recovery expected=100..100; got=100; bad_keys=0; bad_values=0; missed=0 expected=36..36; got=36; bad_keys=0; bad_values=0; missed=64 ==== Test CorruptionTest.RecoverWriteError ==== Test CorruptionTest.NewFileErrorDuringWrite ==== Test CorruptionTest.TableFile expected=90..99; got=99; bad_keys=0; bad_values=1; missed=0 ==== Test CorruptionTest.TableFileRepair expected=95..99; got=98; bad_keys=0; bad_values=0; missed=2 ==== Test CorruptionTest.TableFileIndexData Segmentation fault <<<ACTUAL<<< What version of the product are you using? On what operating system? LevelDB versions: both 1.15 and git-e353fbc7ea81f12a5694991b708f8f45343594b1 (claims to be 1.17) OS: both Debian 7.5 (wheezy) for beaglebone and chromiumos for beaglebone (not sure what version -- based on kernel 3.8.13 + libstdc++ 6.0.18) Please provide any additional information below. This testcase is flaky (segfaulting approximately 30% of the time) with the following compiler/linker flags: FLAGs: -g2 -ggdb -no-canonical-prefixes CXXFLAGS: -ffunction-sections -fdata-sections -fno-math-errno -fno-buildtin-memcmp -pthread -std=gnu++0x LDFLAGS: -Wl,--gc-sections,--no-wchar-size-warning -std=gnu++0x The testcase is segfaults 100% (unless run under gdb) of the time with the following compiler/linker flags: FLAGs: -O2 -D_FORTIFY_SOURCE=2 -no-canonical-prefixes CXXFLAGS: -ffunction-sections -fdata-sections -fno-math-errno -fno-strict-aliasing -fstack-protector-all -fPIC -fno-buildtin-memcmp -z noexecstack -pthread -std=gnu++0x LDFLAGS: -Wl,--gc-sections,--no-wchar-size-warning -pie -Wl,-z,relrow,-z,now -std=gnu++0x The segfault's back trace is is as follows: Program terminated with signal 11, Segmentation fault. #0 0x0002f79a in leveldb::DecodeEntry(char const*, char const*, unsigned int*, unsigned int*, unsigned int*) () at ./leveldb/table/block.cc:58 58 ./leveldb/table/block.cc: No such file or directory. (gdb) bt #0 0x0002f79a in leveldb::DecodeEntry(char const*, char const*, unsigned int*, unsigned int*, unsigned int*) () at ./leveldb/table/block.cc:58 #1 0x0002fe3e in leveldb::Block::Iter::ParseNextKey() () at ./leveldb/table/block.cc:239 #2 0x0002fd02 in leveldb::Block::Iter::SeekToFirst() () at ./leveldb/table/block.cc:207 #3 0x00032734 in leveldb::IteratorWrapper::SeekToFirst() () at ./leveldb/table/iterator_wrapper.h:45 #4 0x00034e72 in leveldb::(anonymous namespace)::TwoLevelIterator::SeekToFirst() () at ./leveldb/table/two_level_iterator.cc:99 #5 0x00032734 in leveldb::IteratorWrapper::SeekToFirst() () at ./leveldb/table/iterator_wrapper.h:45 #6 0x00034fee in leveldb::(anonymous namespace)::TwoLevelIterator::SkipEmptyDataBlocksForward() () at ./leveldb/table/two_level_iterator.cc:134 #7 0x00034f38 in leveldb::(anonymous namespace)::TwoLevelIterator::Next() () at ./leveldb/table/two_level_iterator.cc:115 #8 0x0003264c in leveldb::IteratorWrapper::Next() () at ./leveldb/table/iterator_wrapper.h:42 #9 0x00032b24 in leveldb::(anonymous namespace)::MergingIterator::Next() () at ./leveldb/table/merger.cc:81 #10 0x0001fff0 in leveldb::(anonymous namespace)::DBIter::FindNextUserEntry(bool, std::string*) () at ./leveldb/db/db_iter.cc:199 #11 0x0001feb4 in leveldb::(anonymous namespace)::DBIter::Next() () at ./leveldb/db/db_iter.cc:170 #12 0x0001076c in leveldb::CorruptionTest::Check(int, int) () at ./leveldb/db/corruption_test.cc:96 #13 0x000115da in leveldb::_Test_TableFileIndexData::_Run() () (gdb) disassemble Dump of assembler code for function _ZN7leveldbL11DecodeEntryEPKcS1_PjS2_S2_: 0x0002f77c <+0>: push {r7, lr} 0x0002f77e <+2>: subne sp, #16 0x0002f780 <+4>: addeq r7, sp, #0 0x0002f782 <+6>: str r0, [r7, #12] 0x0002f784 <+8>: str r1, [r7, #8] 0x0002f786 <+10>: str r2, [r7, #4] 0x0002f788 <+12>: str r3, [r7, #0] 0x0002f78a <+14>: ldr r2, [r7, #8] 0x0002f78c <+16>: ldr r3, [r7, #12] 0x0002f78e <+18>: subs r3, r2, r3 0x0002f790 <+20>: cmp r3, #2 0x0002f792 <+22>: bgt.n 0x2f798 <_ZN7leveldbL11DecodeEntryEPKcS1_PjS2_S2_+28> 0x0002f794 <+24>: movs r3, #0 0x0002f796 <+26>: b.n 0x2f858 <_ZN7leveldbL11DecodeEntryEPKcS1_PjS2_S2_+220> 0x0002f798 <+28>: ldr r3, [r7, #12] => 0x0002f79a <+30>: ldrb r3, [r3, #0] 0x0002f79c <+32>: mov r2, r3 ``` ----- Original issue reported on code.google.com by `vlankh...@google.com` on 22 Aug 2014 at 12:40
defect
corruption test segfaults on arm what steps will reproduce the problem compile leveldb s corruption test for execute on some such lower powered device e g beaglebone segmentation fault what is the expected output what do you see instead expected test corruptiontest recovery expected got bad keys bad values missed expected got bad keys bad values missed test corruptiontest recoverwriteerror test corruptiontest newfileerrorduringwrite test corruptiontest tablefile expected got bad keys bad values missed test corruptiontest tablefilerepair expected got bad keys bad values missed test corruptiontest tablefileindexdata expected got bad keys bad values missed test corruptiontest missingdescriptor expected got bad keys bad values missed test corruptiontest sequencenumberrecovery test corruptiontest corrupteddescriptor test corruptiontest compactioninputerror expected got bad keys bad values missed expected got bad keys bad values missed test corruptiontest compactioninputerrorparanoid test corruptiontest unrelatedkeys passed tests expected actual test corruptiontest recovery expected got bad keys bad values missed expected got bad keys bad values missed test corruptiontest recoverwriteerror test corruptiontest newfileerrorduringwrite test corruptiontest tablefile expected got bad keys bad values missed test corruptiontest tablefilerepair expected got bad keys bad values missed test corruptiontest tablefileindexdata segmentation fault actual what version of the product are you using on what operating system leveldb versions both and git claims to be os both debian wheezy for beaglebone and chromiumos for beaglebone not sure what version based on kernel libstdc please provide any additional information below this testcase is flaky segfaulting approximately of the time with the following compiler linker flags flags ggdb no canonical prefixes cxxflags ffunction sections fdata sections fno math errno fno buildtin memcmp pthread std gnu ldflags wl gc sections no wchar size warning std gnu the testcase is segfaults unless run under gdb of the time with the following compiler linker flags flags d fortify source no canonical prefixes cxxflags ffunction sections fdata sections fno math errno fno strict aliasing fstack protector all fpic fno buildtin memcmp z noexecstack pthread std gnu ldflags wl gc sections no wchar size warning pie wl z relrow z now std gnu the segfault s back trace is is as follows program terminated with signal segmentation fault in leveldb decodeentry char const char const unsigned int unsigned int unsigned int at leveldb table block cc leveldb table block cc no such file or directory gdb bt in leveldb decodeentry char const char const unsigned int unsigned int unsigned int at leveldb table block cc in leveldb block iter parsenextkey at leveldb table block cc in leveldb block iter seektofirst at leveldb table block cc in leveldb iteratorwrapper seektofirst at leveldb table iterator wrapper h in leveldb anonymous namespace twoleveliterator seektofirst at leveldb table two level iterator cc in leveldb iteratorwrapper seektofirst at leveldb table iterator wrapper h in leveldb anonymous namespace twoleveliterator skipemptydatablocksforward at leveldb table two level iterator cc in leveldb anonymous namespace twoleveliterator next at leveldb table two level iterator cc in leveldb iteratorwrapper next at leveldb table iterator wrapper h in leveldb anonymous namespace mergingiterator next at leveldb table merger cc in leveldb anonymous namespace dbiter findnextuserentry bool std string at leveldb db db iter cc in leveldb anonymous namespace dbiter next at leveldb db db iter cc in leveldb corruptiontest check int int at leveldb db corruption test cc in leveldb test tablefileindexdata run gdb disassemble dump of assembler code for function push lr subne sp addeq sp str str str str ldr ldr subs cmp bgt n movs b n ldr ldrb mov original issue reported on code google com by vlankh google com on aug at
1
70,276
23,091,115,953
IssuesEvent
2022-07-26 15:17:24
SAP/fundamental-ngx
https://api.github.com/repos/SAP/fundamental-ngx
closed
Defect Hunting: Wizard
bug Defect Hunting wizard
#### Is this a bug, enhancement, or feature request? bug/doc #### Briefly describe your proposal. - [ ] Wizard mobile examples - I am not familiar with any cell phone screens this tall, let's make these examples shorter ![Screen Shot 2021-09-01 at 9 49 37 AM](https://user-images.githubusercontent.com/2471874/131703063-b05a4314-be98-4622-9395-114d5a0c8a64.png)
1.0
Defect Hunting: Wizard - #### Is this a bug, enhancement, or feature request? bug/doc #### Briefly describe your proposal. - [ ] Wizard mobile examples - I am not familiar with any cell phone screens this tall, let's make these examples shorter ![Screen Shot 2021-09-01 at 9 49 37 AM](https://user-images.githubusercontent.com/2471874/131703063-b05a4314-be98-4622-9395-114d5a0c8a64.png)
defect
defect hunting wizard is this a bug enhancement or feature request bug doc briefly describe your proposal wizard mobile examples i am not familiar with any cell phone screens this tall let s make these examples shorter
1
180,548
30,520,640,801
IssuesEvent
2023-07-19 07:50:41
goalapa/spring-cacamuca
https://api.github.com/repos/goalapa/spring-cacamuca
closed
[fix] [3주차]-공통페이지 헤더 수정
🎨design ✅fix
✏️Description - 헤더 대륙>드롭다운>국가 형식으로 수정 ✅TODO - - [x] 헤더 대륙 수정 🐾ETC -
1.0
[fix] [3주차]-공통페이지 헤더 수정 - ✏️Description - 헤더 대륙>드롭다운>국가 형식으로 수정 ✅TODO - - [x] 헤더 대륙 수정 🐾ETC -
non_defect
공통페이지 헤더 수정 ✏️description 헤더 대륙 드롭다운 국가 형식으로 수정 ✅todo 헤더 대륙 수정 🐾etc
0
80,863
30,568,504,531
IssuesEvent
2023-07-20 19:51:06
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Tree: Nodes with long text are displaced
:lady_beetle: defect workaround
### Describe the bug Nodes with long text are displaced ![image](https://github.com/primefaces/primefaces/assets/2945329/2bfb23da-3fbb-4bd4-9fd0-f9da53fb2bf2) ### Reproducer ``` <h:form id="form"> <div style="width: 500px"> <p:tree id="testTree" value="#{indexController.root}" var="node"> <p:treeNode> <h:outputText value="#{node}"/> </p:treeNode> </p:tree> </div> </h:form> ``` ``` @Named @ViewScoped public class IndexController implements Serializable { private TreeNode<String> root; public void init() { root = new DefaultTreeNode<>(); new DefaultTreeNode<>("Test 1", root); new DefaultTreeNode<>("Test 2 extra long text (Lorem ipsum dolor sit amet, consectetur adipisici elit)", root); new DefaultTreeNode<>("Test 3", root); TreeNode<String> test3 = new DefaultTreeNode<>("Test 4", root); new DefaultTreeNode<>("Test 5", root); TreeNode<String> test3_1 = new DefaultTreeNode<>("Child 3.1 extra long text (Lorem ipsum dolor sit amet, consectetur adipisici elit)", test3); new DefaultTreeNode<>("Child 3.1.1", test3_1); new DefaultTreeNode<>("Child 3.1.2 extra long text (Lorem ipsum dolor sit amet, consectetur adipisici elit)", test3_1); new DefaultTreeNode<>("Child 3.1.3", test3_1); } public TreeNode<String> getRoot() { return root; } public void setRoot(TreeNode<String> root) { this.root = root; } } ``` ### Expected behavior It looks clean and not displaced ![image](https://github.com/primefaces/primefaces/assets/2945329/87984334-4418-4546-be92-a7fe3e8276ef) Here is a possible bugfix code ``` .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-treenode-leaf-icon, .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-chkbox, .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-treenode-icon, .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-tree-toggler { flex: none; } .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-treenode-label { flex-grow: 1; } ``` ### PrimeFaces edition None ### PrimeFaces version 13.0.0, 14.0.0-SNAPSHOT ### Theme _No response_ ### JSF implementation MyFaces ### JSF version 2.3 ### Java version 17 ### Browser(s) _No response_
1.0
Tree: Nodes with long text are displaced - ### Describe the bug Nodes with long text are displaced ![image](https://github.com/primefaces/primefaces/assets/2945329/2bfb23da-3fbb-4bd4-9fd0-f9da53fb2bf2) ### Reproducer ``` <h:form id="form"> <div style="width: 500px"> <p:tree id="testTree" value="#{indexController.root}" var="node"> <p:treeNode> <h:outputText value="#{node}"/> </p:treeNode> </p:tree> </div> </h:form> ``` ``` @Named @ViewScoped public class IndexController implements Serializable { private TreeNode<String> root; public void init() { root = new DefaultTreeNode<>(); new DefaultTreeNode<>("Test 1", root); new DefaultTreeNode<>("Test 2 extra long text (Lorem ipsum dolor sit amet, consectetur adipisici elit)", root); new DefaultTreeNode<>("Test 3", root); TreeNode<String> test3 = new DefaultTreeNode<>("Test 4", root); new DefaultTreeNode<>("Test 5", root); TreeNode<String> test3_1 = new DefaultTreeNode<>("Child 3.1 extra long text (Lorem ipsum dolor sit amet, consectetur adipisici elit)", test3); new DefaultTreeNode<>("Child 3.1.1", test3_1); new DefaultTreeNode<>("Child 3.1.2 extra long text (Lorem ipsum dolor sit amet, consectetur adipisici elit)", test3_1); new DefaultTreeNode<>("Child 3.1.3", test3_1); } public TreeNode<String> getRoot() { return root; } public void setRoot(TreeNode<String> root) { this.root = root; } } ``` ### Expected behavior It looks clean and not displaced ![image](https://github.com/primefaces/primefaces/assets/2945329/87984334-4418-4546-be92-a7fe3e8276ef) Here is a possible bugfix code ``` .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-treenode-leaf-icon, .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-chkbox, .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-treenode-icon, .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-tree-toggler { flex: none; } .ui-tree .ui-tree-container .ui-treenode .ui-treenode-content .ui-treenode-label { flex-grow: 1; } ``` ### PrimeFaces edition None ### PrimeFaces version 13.0.0, 14.0.0-SNAPSHOT ### Theme _No response_ ### JSF implementation MyFaces ### JSF version 2.3 ### Java version 17 ### Browser(s) _No response_
defect
tree nodes with long text are displaced describe the bug nodes with long text are displaced reproducer named viewscoped public class indexcontroller implements serializable private treenode root public void init root new defaulttreenode new defaulttreenode test root new defaulttreenode test extra long text lorem ipsum dolor sit amet consectetur adipisici elit root new defaulttreenode test root treenode new defaulttreenode test root new defaulttreenode test root treenode new defaulttreenode child extra long text lorem ipsum dolor sit amet consectetur adipisici elit new defaulttreenode child new defaulttreenode child extra long text lorem ipsum dolor sit amet consectetur adipisici elit new defaulttreenode child public treenode getroot return root public void setroot treenode root this root root expected behavior it looks clean and not displaced here is a possible bugfix code ui tree ui tree container ui treenode ui treenode content ui treenode leaf icon ui tree ui tree container ui treenode ui treenode content ui chkbox ui tree ui tree container ui treenode ui treenode content ui treenode icon ui tree ui tree container ui treenode ui treenode content ui tree toggler flex none ui tree ui tree container ui treenode ui treenode content ui treenode label flex grow primefaces edition none primefaces version snapshot theme no response jsf implementation myfaces jsf version java version browser s no response
1
42,999
11,423,166,462
IssuesEvent
2020-02-03 15:27:05
extnet/Ext.NET
https://api.github.com/repos/extnet/Ext.NET
closed
Desktop.addModule()'s added shortcut position is reset on window resize
5.x defect
Found: 5.1.0 Ext.NET Forum Thread: [Desktop module icon overlap after window resize](https://forums.ext.net/showthread.php?62833) No Sencha thread - feature is exclusive to Ext.NET. Desktop window resize calls desktop's `arrangeShortcuts(ignorePosition: false, ignoreTemp: true)`, thus icons just added via `addModule()` position would get reset to the top-left corner of the desktop, even if there are other icons placed.
1.0
Desktop.addModule()'s added shortcut position is reset on window resize - Found: 5.1.0 Ext.NET Forum Thread: [Desktop module icon overlap after window resize](https://forums.ext.net/showthread.php?62833) No Sencha thread - feature is exclusive to Ext.NET. Desktop window resize calls desktop's `arrangeShortcuts(ignorePosition: false, ignoreTemp: true)`, thus icons just added via `addModule()` position would get reset to the top-left corner of the desktop, even if there are other icons placed.
defect
desktop addmodule s added shortcut position is reset on window resize found ext net forum thread no sencha thread feature is exclusive to ext net desktop window resize calls desktop s arrangeshortcuts ignoreposition false ignoretemp true thus icons just added via addmodule position would get reset to the top left corner of the desktop even if there are other icons placed
1
64,825
18,939,725,761
IssuesEvent
2021-11-18 00:28:53
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Custom Jitsi server support
T-Defect
### Steps to reproduce I have setup a custom Jitsi server according to [this](https://github.com/vector-im/element-web/blob/develop/docs/jitsi.md#configuring-element-to-use-your-self-hosted-jitsi-server) documentation. I setup a videocall using the 'Camera' call button but the client used jitsi.riot.im rather than my custom server. Here is my nginx config which sets the jitsi domain name, ``` server { listen 443 ssl; listen 8448 ssl; server_name matrix.mydomain.com; client_max_body_size 5000M; location /_matrix { root html; index index.html index.htm; proxy_set_header Host $host; proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header X-Forwarded-Proto $scheme; proxy_pass http://192.168.1.2:8008; } location /.well-known/matrix/client { add_header Access-Control-Allow-Origin "*"; add_header Content-Type "application/json"; return 200 '{"im.vector.riot.jitsi": {"preferredDomain": "meet.mydomain.com"}}'; } ssl_certificate /etc/letsencrypt/live/mydomain.com/fullchain.pem; # managed by Certbot ssl_certificate_key /etc/letsencrypt/live/mydomain.com/privkey.pem; # managed by Certbot } ``` ### Outcome #### What did you expect? Call will be setup using meet.mydomain.com #### What happened instead? Call was setup using jitsi.riot.im ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store Element 1.3.7 ### Homeserver _No response_ ### Will you send logs? No
1.0
Custom Jitsi server support - ### Steps to reproduce I have setup a custom Jitsi server according to [this](https://github.com/vector-im/element-web/blob/develop/docs/jitsi.md#configuring-element-to-use-your-self-hosted-jitsi-server) documentation. I setup a videocall using the 'Camera' call button but the client used jitsi.riot.im rather than my custom server. Here is my nginx config which sets the jitsi domain name, ``` server { listen 443 ssl; listen 8448 ssl; server_name matrix.mydomain.com; client_max_body_size 5000M; location /_matrix { root html; index index.html index.htm; proxy_set_header Host $host; proxy_set_header X-Real-IP $remote_addr; proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for; proxy_set_header X-Forwarded-Proto $scheme; proxy_pass http://192.168.1.2:8008; } location /.well-known/matrix/client { add_header Access-Control-Allow-Origin "*"; add_header Content-Type "application/json"; return 200 '{"im.vector.riot.jitsi": {"preferredDomain": "meet.mydomain.com"}}'; } ssl_certificate /etc/letsencrypt/live/mydomain.com/fullchain.pem; # managed by Certbot ssl_certificate_key /etc/letsencrypt/live/mydomain.com/privkey.pem; # managed by Certbot } ``` ### Outcome #### What did you expect? Call will be setup using meet.mydomain.com #### What happened instead? Call was setup using jitsi.riot.im ### Your phone model _No response_ ### Operating system version _No response_ ### Application version and app store Element 1.3.7 ### Homeserver _No response_ ### Will you send logs? No
defect
custom jitsi server support steps to reproduce i have setup a custom jitsi server according to documentation i setup a videocall using the camera call button but the client used jitsi riot im rather than my custom server here is my nginx config which sets the jitsi domain name server listen ssl listen ssl server name matrix mydomain com client max body size location matrix root html index index html index htm proxy set header host host proxy set header x real ip remote addr proxy set header x forwarded for proxy add x forwarded for proxy set header x forwarded proto scheme proxy pass location well known matrix client add header access control allow origin add header content type application json return im vector riot jitsi preferreddomain meet mydomain com ssl certificate etc letsencrypt live mydomain com fullchain pem managed by certbot ssl certificate key etc letsencrypt live mydomain com privkey pem managed by certbot outcome what did you expect call will be setup using meet mydomain com what happened instead call was setup using jitsi riot im your phone model no response operating system version no response application version and app store element homeserver no response will you send logs no
1
51,559
13,207,526,832
IssuesEvent
2020-08-14 23:27:20
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
can't find python libs > python 2.6.x (Trac #627)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/627">https://code.icecube.wisc.edu/projects/icecube/ticket/627</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2011-05-11T19:50:31", "_ts": "1305143431000000", "description": "cmake 2.6.x only has support for python < 2.6.x in its PythonFindX.cmake modules. (hard coded values). this seems to have been fixed in cmake 2.8.x, but is currently untested by !IceCube. (i personally use cmake 2.8.x but not python 2.7.x)\n\ncurrently i solve this on the build bots by hard coding python paths in a I3_CITE_CMAKE_DIR file on akuma (rhel4 + external python 2.7) and beastie (freebsd 9RC + freebsd port install of python 2.7.1 in /usr/local).\n\nthis '''can not''' be solved by including the cmake 2.8 FindPythonX modules, as we'd get into dependency hell w/ the cmake modules.\n\nprobably be able to solve this w/ i3-tools-v4 (#278)", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2011-04-30T16:22:01", "component": "cmake", "summary": "can't find python libs > python 2.6.x", "priority": "major", "keywords": "cmake python freebsd rhel", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
can't find python libs > python 2.6.x (Trac #627) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/627">https://code.icecube.wisc.edu/projects/icecube/ticket/627</a>, reported by negaand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2011-05-11T19:50:31", "_ts": "1305143431000000", "description": "cmake 2.6.x only has support for python < 2.6.x in its PythonFindX.cmake modules. (hard coded values). this seems to have been fixed in cmake 2.8.x, but is currently untested by !IceCube. (i personally use cmake 2.8.x but not python 2.7.x)\n\ncurrently i solve this on the build bots by hard coding python paths in a I3_CITE_CMAKE_DIR file on akuma (rhel4 + external python 2.7) and beastie (freebsd 9RC + freebsd port install of python 2.7.1 in /usr/local).\n\nthis '''can not''' be solved by including the cmake 2.8 FindPythonX modules, as we'd get into dependency hell w/ the cmake modules.\n\nprobably be able to solve this w/ i3-tools-v4 (#278)", "reporter": "nega", "cc": "", "resolution": "fixed", "time": "2011-04-30T16:22:01", "component": "cmake", "summary": "can't find python libs > python 2.6.x", "priority": "major", "keywords": "cmake python freebsd rhel", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
can t find python libs python x trac migrated from json status closed changetime ts description cmake x only has support for python x in its pythonfindx cmake modules hard coded values this seems to have been fixed in cmake x but is currently untested by icecube i personally use cmake x but not python x n ncurrently i solve this on the build bots by hard coding python paths in a cite cmake dir file on akuma external python and beastie freebsd freebsd port install of python in usr local n nthis can not be solved by including the cmake findpythonx modules as we d get into dependency hell w the cmake modules n nprobably be able to solve this w tools reporter nega cc resolution fixed time component cmake summary can t find python libs python x priority major keywords cmake python freebsd rhel milestone owner nega type defect
1
31,623
7,430,523,144
IssuesEvent
2018-03-25 02:56:41
dsherret/ts-simple-ast
https://api.github.com/repos/dsherret/ts-simple-ast
closed
Remove "Base" variables from declaration file
code improvement
The declaration file for this project shouldn't have the "Base" variable declarations in it.
1.0
Remove "Base" variables from declaration file - The declaration file for this project shouldn't have the "Base" variable declarations in it.
non_defect
remove base variables from declaration file the declaration file for this project shouldn t have the base variable declarations in it
0
150,269
11,954,067,218
IssuesEvent
2020-04-03 22:26:03
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
System.Net.Sockets.Tests.DualModeAcceptAsync.AcceptAsyncV4BoundToSpecificV6_CantConnect test failed in CI
area-System.Net.Sockets test bug test-run-core
https://dnceng.visualstudio.com/public/_build/results?buildId=393221&view=ms.vss-test-web.build-test-results-tab&runId=12222448&resultId=177679&paneView=debug Configuration: `netcoreapp-Windows_NT-Release-x86-Windows.10.Amd64.Client19H1.Open` ``` Assert.Throws() Failure\r\nExpected: typeof(System.Net.Sockets.SocketException)\r\nActual: typeof(System.TimeoutException): Timed out while waiting for either of client and server connections...\r\n---- System.TimeoutException : Timed out while waiting for either of client and server connections... at System.Net.Sockets.Tests.DualModeAcceptAsync.DualModeConnect_AcceptAsync_Helper(IPAddress listenOn, IPAddress connectTo) in /_/src/System.Net.Sockets/tests/FunctionalTests/DualModeSocketTest.cs:line 973 at System.Net.Sockets.Tests.DualModeAcceptAsync.<AcceptAsyncV4BoundToSpecificV6_CantConnect>b__5_0() in /_/src/System.Net.Sockets/tests/FunctionalTests/DualModeSocketTest.cs:line 924 ----- Inner Stack Trace ----- at System.Net.Sockets.Tests.DualModeAcceptAsync.DualModeConnect_AcceptAsync_Helper(IPAddress listenOn, IPAddress connectTo) in /_/src/System.Net.Sockets/tests/FunctionalTests/DualModeSocketTest.cs:line 973 at System.Net.Sockets.Tests.DualModeAcceptAsync.<AcceptAsyncV4BoundToSpecificV6_CantConnect>b__5_0() in /_/src/System.Net.Sockets/tests/FunctionalTests/DualModeSocketTest.cs:line 924 ``` cc @davidsh, @wfurt, @scalablecory, @eiriktsarpalis, @karelz
2.0
System.Net.Sockets.Tests.DualModeAcceptAsync.AcceptAsyncV4BoundToSpecificV6_CantConnect test failed in CI - https://dnceng.visualstudio.com/public/_build/results?buildId=393221&view=ms.vss-test-web.build-test-results-tab&runId=12222448&resultId=177679&paneView=debug Configuration: `netcoreapp-Windows_NT-Release-x86-Windows.10.Amd64.Client19H1.Open` ``` Assert.Throws() Failure\r\nExpected: typeof(System.Net.Sockets.SocketException)\r\nActual: typeof(System.TimeoutException): Timed out while waiting for either of client and server connections...\r\n---- System.TimeoutException : Timed out while waiting for either of client and server connections... at System.Net.Sockets.Tests.DualModeAcceptAsync.DualModeConnect_AcceptAsync_Helper(IPAddress listenOn, IPAddress connectTo) in /_/src/System.Net.Sockets/tests/FunctionalTests/DualModeSocketTest.cs:line 973 at System.Net.Sockets.Tests.DualModeAcceptAsync.<AcceptAsyncV4BoundToSpecificV6_CantConnect>b__5_0() in /_/src/System.Net.Sockets/tests/FunctionalTests/DualModeSocketTest.cs:line 924 ----- Inner Stack Trace ----- at System.Net.Sockets.Tests.DualModeAcceptAsync.DualModeConnect_AcceptAsync_Helper(IPAddress listenOn, IPAddress connectTo) in /_/src/System.Net.Sockets/tests/FunctionalTests/DualModeSocketTest.cs:line 973 at System.Net.Sockets.Tests.DualModeAcceptAsync.<AcceptAsyncV4BoundToSpecificV6_CantConnect>b__5_0() in /_/src/System.Net.Sockets/tests/FunctionalTests/DualModeSocketTest.cs:line 924 ``` cc @davidsh, @wfurt, @scalablecory, @eiriktsarpalis, @karelz
non_defect
system net sockets tests dualmodeacceptasync cantconnect test failed in ci configuration netcoreapp windows nt release windows open assert throws failure r nexpected typeof system net sockets socketexception r nactual typeof system timeoutexception timed out while waiting for either of client and server connections r n system timeoutexception timed out while waiting for either of client and server connections at system net sockets tests dualmodeacceptasync dualmodeconnect acceptasync helper ipaddress listenon ipaddress connectto in src system net sockets tests functionaltests dualmodesockettest cs line at system net sockets tests dualmodeacceptasync b in src system net sockets tests functionaltests dualmodesockettest cs line inner stack trace at system net sockets tests dualmodeacceptasync dualmodeconnect acceptasync helper ipaddress listenon ipaddress connectto in src system net sockets tests functionaltests dualmodesockettest cs line at system net sockets tests dualmodeacceptasync b in src system net sockets tests functionaltests dualmodesockettest cs line cc davidsh   wfurt   scalablecory   eiriktsarpalis   karelz
0
23,251
3,783,890,464
IssuesEvent
2016-03-19 12:46:04
VladSerdobintsev/zfcore
https://api.github.com/repos/VladSerdobintsev/zfcore
closed
Login issue with "Remember me" flag.
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1.Login with "Remember me" flag. 2.Logout 3.Login without "Remember me" flag. 4.Close and open browser. Issue : User still logged in. Caused by: When user logging out session id and expires date doesn't change after logout. Resolve: Add to class Users_Model_User_Manager function logout: Zend_Session::forgetMe(); ``` Original issue reported on code.google.com by `ad...@sunset.dp.ua` on 26 Jul 2012 at 10:08
1.0
Login issue with "Remember me" flag. - ``` What steps will reproduce the problem? 1.Login with "Remember me" flag. 2.Logout 3.Login without "Remember me" flag. 4.Close and open browser. Issue : User still logged in. Caused by: When user logging out session id and expires date doesn't change after logout. Resolve: Add to class Users_Model_User_Manager function logout: Zend_Session::forgetMe(); ``` Original issue reported on code.google.com by `ad...@sunset.dp.ua` on 26 Jul 2012 at 10:08
defect
login issue with remember me flag what steps will reproduce the problem login with remember me flag logout login without remember me flag close and open browser issue user still logged in caused by when user logging out session id and expires date doesn t change after logout resolve add to class users model user manager function logout zend session forgetme original issue reported on code google com by ad sunset dp ua on jul at
1
342,643
30,633,022,011
IssuesEvent
2023-07-24 15:42:17
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
schemachanger,sctest: backup tests are slightly broken
C-bug C-test-failure A-testing T-sql-foundations A-schema-changer-impl
In an effort to understand how the `TestBackup*` cumulative tests for the declarative schema changer work, I noticed a subtle bug: we back up _after_ executing each post-commit stage, which is fine but means we never test backups after the stmt txn commits but before the job runs, which is arguably the most important case. ~As it turns out, fixing this reveals other longstanding bugs.~ This issue tracks the need to refactor the `sctest` framework, especially the backups tests. Jira issue: CRDB-29767
2.0
schemachanger,sctest: backup tests are slightly broken - In an effort to understand how the `TestBackup*` cumulative tests for the declarative schema changer work, I noticed a subtle bug: we back up _after_ executing each post-commit stage, which is fine but means we never test backups after the stmt txn commits but before the job runs, which is arguably the most important case. ~As it turns out, fixing this reveals other longstanding bugs.~ This issue tracks the need to refactor the `sctest` framework, especially the backups tests. Jira issue: CRDB-29767
non_defect
schemachanger sctest backup tests are slightly broken in an effort to understand how the testbackup cumulative tests for the declarative schema changer work i noticed a subtle bug we back up after executing each post commit stage which is fine but means we never test backups after the stmt txn commits but before the job runs which is arguably the most important case as it turns out fixing this reveals other longstanding bugs this issue tracks the need to refactor the sctest framework especially the backups tests jira issue crdb
0
46,822
13,055,982,575
IssuesEvent
2020-07-30 03:18:07
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[codereview] grbllh (Trac #1935)
Incomplete Migration Migrated from Trac cmake defect
Migrated from https://code.icecube.wisc.edu/ticket/1935 ```json { "status": "closed", "changetime": "2019-02-13T14:12:55", "description": "", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1550067175380821", "component": "cmake", "summary": "[codereview] grbllh", "priority": "normal", "keywords": "", "time": "2017-01-20T09:19:12", "milestone": "", "owner": "kjmeagher", "type": "defect" } ```
1.0
[codereview] grbllh (Trac #1935) - Migrated from https://code.icecube.wisc.edu/ticket/1935 ```json { "status": "closed", "changetime": "2019-02-13T14:12:55", "description": "", "reporter": "kjmeagher", "cc": "", "resolution": "fixed", "_ts": "1550067175380821", "component": "cmake", "summary": "[codereview] grbllh", "priority": "normal", "keywords": "", "time": "2017-01-20T09:19:12", "milestone": "", "owner": "kjmeagher", "type": "defect" } ```
defect
grbllh trac migrated from json status closed changetime description reporter kjmeagher cc resolution fixed ts component cmake summary grbllh priority normal keywords time milestone owner kjmeagher type defect
1
59,511
17,023,148,139
IssuesEvent
2021-07-03 00:35:26
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Website using ISO-8859-1 encoding conflicts with UTF-8 data.
Component: website Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 5.27am, Monday, 19th March 2007]** The website uses iso-8859-1 (default www) encoding, which results in all kinds of issues when using the browser to read data from or input data into the OSM project. This is no problem when handling ASCII-data but using the applet to edit things the input data is not correct UTF8 for example.
1.0
Website using ISO-8859-1 encoding conflicts with UTF-8 data. - **[Submitted to the original trac issue database at 5.27am, Monday, 19th March 2007]** The website uses iso-8859-1 (default www) encoding, which results in all kinds of issues when using the browser to read data from or input data into the OSM project. This is no problem when handling ASCII-data but using the applet to edit things the input data is not correct UTF8 for example.
defect
website using iso encoding conflicts with utf data the website uses iso default www encoding which results in all kinds of issues when using the browser to read data from or input data into the osm project this is no problem when handling ascii data but using the applet to edit things the input data is not correct for example
1
61,400
17,023,684,731
IssuesEvent
2021-07-03 03:17:31
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Keyboard commands not responsive
Component: potlatch2 Priority: minor Resolution: invalid Type: defect
**[Submitted to the original trac issue database at 2.28pm, Thursday, 3rd March 2011]** (Andy, potlatch-dev, 4 mar 2011) "I still find the shift key stopping working every now and then, but like Richard I don't have steps to reproduce this one." Fwiw, I haven't seen this. Does it also happen with other modifiers? Does it stay out of action for the whole session? Any workaround?
1.0
Keyboard commands not responsive - **[Submitted to the original trac issue database at 2.28pm, Thursday, 3rd March 2011]** (Andy, potlatch-dev, 4 mar 2011) "I still find the shift key stopping working every now and then, but like Richard I don't have steps to reproduce this one." Fwiw, I haven't seen this. Does it also happen with other modifiers? Does it stay out of action for the whole session? Any workaround?
defect
keyboard commands not responsive andy potlatch dev mar i still find the shift key stopping working every now and then but like richard i don t have steps to reproduce this one fwiw i haven t seen this does it also happen with other modifiers does it stay out of action for the whole session any workaround
1
473,467
13,642,868,207
IssuesEvent
2020-09-25 16:10:29
dwyl/smart-home-auth-server
https://api.github.com/repos/dwyl/smart-home-auth-server
opened
Locks aren't implemented with a mode through GUI
bug priority-3
While testing the new lock GUI for RBAC, I noticed that new locks aren't set a default mode, resulting in the error: `16:08:04.856 [info] Unimplemented mode: nil/Lock not configured` This should be an easy fix. Through the API this works fine.
1.0
Locks aren't implemented with a mode through GUI - While testing the new lock GUI for RBAC, I noticed that new locks aren't set a default mode, resulting in the error: `16:08:04.856 [info] Unimplemented mode: nil/Lock not configured` This should be an easy fix. Through the API this works fine.
non_defect
locks aren t implemented with a mode through gui while testing the new lock gui for rbac i noticed that new locks aren t set a default mode resulting in the error unimplemented mode nil lock not configured this should be an easy fix through the api this works fine
0
98,869
30,206,655,843
IssuesEvent
2023-07-05 09:53:02
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
[mono][ios] Unknown format in import in System.Security.Cryptography.Tests
area-System.Security disabled-test os-ios in-pr Known Build Error
## Build Information Build: https://dev.azure.com/dnceng-public/public/_build/results?buildId=320749&view=results Build error leg or test failing: iossimulator-x64 Release AllSubsets_Mono Pull request: https://github.com/dotnet/runtime/pull/88042 <!-- Error message template --> ## Error Message Fill the error message using [step by step known issues guidance](https://github.com/dotnet/arcade/blob/main/Documentation/Projects/Build%20Analysis/KnownIssues.md#how-to-fill-out-a-known-issue-error-section). <!-- Use ErrorMessage for String.Contains matches. Use ErrorPattern for regex matches (single line/no backtracking). Set BuildRetry to `true` to retry builds with this error. Set ExcludeConsoleLog to `true` to skip helix logs analysis. --> ```json { "ErrorMessage": "Interop+AppleCrypto+AppleCommonCryptoCryptographicException : Unknown format in import", "ErrorPattern": "", "BuildRetry": false, "ExcludeConsoleLog": false } ``` <!-- Known issue validation start --> ### Known issue validation **Build: :mag_right:** https://dev.azure.com/dnceng-public/public/_build/results?buildId=320749 **Result validation: :white_check_mark:** Known issue matched with the provided build. <!-- Known issue validation end --> <!--Known issue error report start --> ### Report |Build|Definition|Test|Pull Request| |---|---|---|---| |[325432](https://dev.azure.com/dnceng-public/public/_build/results?buildId=325432)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=325432&view=ms.vss-test-web.build-test-results-tab&runId=6693096&resultId=193828)|| |[324452](https://dev.azure.com/dnceng-public/public/_build/results?buildId=324452)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=324452&view=ms.vss-test-web.build-test-results-tab&runId=6670824&resultId=193410)|dotnet/runtime#88167| |[324371](https://dev.azure.com/dnceng-public/public/_build/results?buildId=324371)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=324371&view=ms.vss-test-web.build-test-results-tab&runId=6668604&resultId=193410)|| |[323893](https://dev.azure.com/dnceng-public/public/_build/results?buildId=323893)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=323893&view=ms.vss-test-web.build-test-results-tab&runId=6657996&resultId=193373)|dotnet/runtime#88113| |[323521](https://dev.azure.com/dnceng-public/public/_build/results?buildId=323521)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=323521&view=ms.vss-test-web.build-test-results-tab&runId=6648200&resultId=193371)|| |[322934](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322934)|dotnet/runtime|[System.Security.Cryptography.X509Certificates.Tests.PfxIterationCountTests_X509Certificate2.Import_BlobHasMoreThanOnePfx_LoadsOnlyOne](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322934&view=ms.vss-test-web.build-test-results-tab&runId=6634534&resultId=168163)|| |[322681](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322681)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322681&view=ms.vss-test-web.build-test-results-tab&runId=6628760&resultId=193373)|dotnet/runtime#87260| |[322518](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322518)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322518&view=ms.vss-test-web.build-test-results-tab&runId=6623814&resultId=193373)|dotnet/runtime#88094| |[321994](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321994)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321994&view=ms.vss-test-web.build-test-results-tab&runId=6610714&resultId=193373)|| |[321687](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321687)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321687&view=ms.vss-test-web.build-test-results-tab&runId=6602456&resultId=193828)|| |[321510](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321510)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321510&view=ms.vss-test-web.build-test-results-tab&runId=6599782&resultId=193828)|| |[321392](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321392)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321392&view=ms.vss-test-web.build-test-results-tab&runId=6594820&resultId=193373)|| |[320749](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320749)|dotnet/runtime|[System.Security.Cryptography.X509Certificates.Tests.PfxIterationCountTests_X509Certificate2Collection.Import_BlobHasMoreThanOnePfx_LoadsOnlyOne](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320749&view=ms.vss-test-web.build-test-results-tab&runId=6583430&resultId=187768)|dotnet/runtime#88042| |[320642](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320642)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320642&view=ms.vss-test-web.build-test-results-tab&runId=6581770&resultId=193803)|dotnet/runtime#88036| |[320422](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320422)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320422&view=ms.vss-test-web.build-test-results-tab&runId=6574344&resultId=193803)|| |[320144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320144)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320144&view=ms.vss-test-web.build-test-results-tab&runId=6564330&resultId=193803)|| #### Summary |24-Hour Hit Count|7-Day Hit Count|1-Month Count| |---|---|---| |0|5|16| <!--Known issue error report end -->
1.0
[mono][ios] Unknown format in import in System.Security.Cryptography.Tests - ## Build Information Build: https://dev.azure.com/dnceng-public/public/_build/results?buildId=320749&view=results Build error leg or test failing: iossimulator-x64 Release AllSubsets_Mono Pull request: https://github.com/dotnet/runtime/pull/88042 <!-- Error message template --> ## Error Message Fill the error message using [step by step known issues guidance](https://github.com/dotnet/arcade/blob/main/Documentation/Projects/Build%20Analysis/KnownIssues.md#how-to-fill-out-a-known-issue-error-section). <!-- Use ErrorMessage for String.Contains matches. Use ErrorPattern for regex matches (single line/no backtracking). Set BuildRetry to `true` to retry builds with this error. Set ExcludeConsoleLog to `true` to skip helix logs analysis. --> ```json { "ErrorMessage": "Interop+AppleCrypto+AppleCommonCryptoCryptographicException : Unknown format in import", "ErrorPattern": "", "BuildRetry": false, "ExcludeConsoleLog": false } ``` <!-- Known issue validation start --> ### Known issue validation **Build: :mag_right:** https://dev.azure.com/dnceng-public/public/_build/results?buildId=320749 **Result validation: :white_check_mark:** Known issue matched with the provided build. <!-- Known issue validation end --> <!--Known issue error report start --> ### Report |Build|Definition|Test|Pull Request| |---|---|---|---| |[325432](https://dev.azure.com/dnceng-public/public/_build/results?buildId=325432)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=325432&view=ms.vss-test-web.build-test-results-tab&runId=6693096&resultId=193828)|| |[324452](https://dev.azure.com/dnceng-public/public/_build/results?buildId=324452)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=324452&view=ms.vss-test-web.build-test-results-tab&runId=6670824&resultId=193410)|dotnet/runtime#88167| |[324371](https://dev.azure.com/dnceng-public/public/_build/results?buildId=324371)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=324371&view=ms.vss-test-web.build-test-results-tab&runId=6668604&resultId=193410)|| |[323893](https://dev.azure.com/dnceng-public/public/_build/results?buildId=323893)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=323893&view=ms.vss-test-web.build-test-results-tab&runId=6657996&resultId=193373)|dotnet/runtime#88113| |[323521](https://dev.azure.com/dnceng-public/public/_build/results?buildId=323521)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=323521&view=ms.vss-test-web.build-test-results-tab&runId=6648200&resultId=193371)|| |[322934](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322934)|dotnet/runtime|[System.Security.Cryptography.X509Certificates.Tests.PfxIterationCountTests_X509Certificate2.Import_BlobHasMoreThanOnePfx_LoadsOnlyOne](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322934&view=ms.vss-test-web.build-test-results-tab&runId=6634534&resultId=168163)|| |[322681](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322681)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322681&view=ms.vss-test-web.build-test-results-tab&runId=6628760&resultId=193373)|dotnet/runtime#87260| |[322518](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322518)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=322518&view=ms.vss-test-web.build-test-results-tab&runId=6623814&resultId=193373)|dotnet/runtime#88094| |[321994](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321994)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321994&view=ms.vss-test-web.build-test-results-tab&runId=6610714&resultId=193373)|| |[321687](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321687)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321687&view=ms.vss-test-web.build-test-results-tab&runId=6602456&resultId=193828)|| |[321510](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321510)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321510&view=ms.vss-test-web.build-test-results-tab&runId=6599782&resultId=193828)|| |[321392](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321392)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=321392&view=ms.vss-test-web.build-test-results-tab&runId=6594820&resultId=193373)|| |[320749](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320749)|dotnet/runtime|[System.Security.Cryptography.X509Certificates.Tests.PfxIterationCountTests_X509Certificate2Collection.Import_BlobHasMoreThanOnePfx_LoadsOnlyOne](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320749&view=ms.vss-test-web.build-test-results-tab&runId=6583430&resultId=187768)|dotnet/runtime#88042| |[320642](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320642)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320642&view=ms.vss-test-web.build-test-results-tab&runId=6581770&resultId=193803)|dotnet/runtime#88036| |[320422](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320422)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320422&view=ms.vss-test-web.build-test-results-tab&runId=6574344&resultId=193803)|| |[320144](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320144)|dotnet/runtime|[System.Security.Cryptography.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=320144&view=ms.vss-test-web.build-test-results-tab&runId=6564330&resultId=193803)|| #### Summary |24-Hour Hit Count|7-Day Hit Count|1-Month Count| |---|---|---| |0|5|16| <!--Known issue error report end -->
non_defect
unknown format in import in system security cryptography tests build information build build error leg or test failing iossimulator release allsubsets mono pull request error message fill the error message using json errormessage interop applecrypto applecommoncryptocryptographicexception unknown format in import errorpattern buildretry false excludeconsolelog false known issue validation build mag right result validation white check mark known issue matched with the provided build report build definition test pull request summary hour hit count day hit count month count
0
465,448
13,385,994,963
IssuesEvent
2020-09-02 14:10:45
strapi/strapi
https://api.github.com/repos/strapi/strapi
closed
Strapi admin roles&permissions no longer able to click anywhere on row to get into edit screen
priority: high source: admin status: confirmed type: bug
Version 3.1.0, worked in 3.0.5. Now being forced to click on the small edit icon which is next to delete icon - each time feels intimidating. In Collection Types clicking anywhere on row still works to edit. Maybe this bug should be filled for plugin? Don't know which part is responsible for UI behavior.
1.0
Strapi admin roles&permissions no longer able to click anywhere on row to get into edit screen - Version 3.1.0, worked in 3.0.5. Now being forced to click on the small edit icon which is next to delete icon - each time feels intimidating. In Collection Types clicking anywhere on row still works to edit. Maybe this bug should be filled for plugin? Don't know which part is responsible for UI behavior.
non_defect
strapi admin roles permissions no longer able to click anywhere on row to get into edit screen version worked in now being forced to click on the small edit icon which is next to delete icon each time feels intimidating in collection types clicking anywhere on row still works to edit maybe this bug should be filled for plugin don t know which part is responsible for ui behavior
0
71,954
23,867,521,731
IssuesEvent
2022-09-07 12:22:09
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Some Hazelcast log statements missing when used via Spring Boot [HZ-822]
Type: Defect Team: Core Source: Internal Module: Spring to-jira
**Describe the bug** When Hazelcast is used via Spring Boot not all log statements are printed. All debug/trace logs are missing, some info statements are missing. **Expected behavior** All log statements are visible according to Spring Boot Logging configuration **To Reproduce** Use any starter, e.g. https://github.com/leszko/springboot-caching Change logging level for com.hazelcast to `debug`. Start the application **Additional context** Some INFO and WARN logs are visible. I haven't investigated this in detail, but I suppose the logging chain is (for the default spring boot starter) Hazelcast -> JUL (default logger) -> jul-over-slf4j -> logback Maybe the check for enabled log level is lost in this chain. The following can be used as workaround ``` -Dhazelcast.logging.type=slf4j ```
1.0
Some Hazelcast log statements missing when used via Spring Boot [HZ-822] - **Describe the bug** When Hazelcast is used via Spring Boot not all log statements are printed. All debug/trace logs are missing, some info statements are missing. **Expected behavior** All log statements are visible according to Spring Boot Logging configuration **To Reproduce** Use any starter, e.g. https://github.com/leszko/springboot-caching Change logging level for com.hazelcast to `debug`. Start the application **Additional context** Some INFO and WARN logs are visible. I haven't investigated this in detail, but I suppose the logging chain is (for the default spring boot starter) Hazelcast -> JUL (default logger) -> jul-over-slf4j -> logback Maybe the check for enabled log level is lost in this chain. The following can be used as workaround ``` -Dhazelcast.logging.type=slf4j ```
defect
some hazelcast log statements missing when used via spring boot describe the bug when hazelcast is used via spring boot not all log statements are printed all debug trace logs are missing some info statements are missing expected behavior all log statements are visible according to spring boot logging configuration to reproduce use any starter e g change logging level for com hazelcast to debug start the application additional context some info and warn logs are visible i haven t investigated this in detail but i suppose the logging chain is for the default spring boot starter hazelcast jul default logger jul over logback maybe the check for enabled log level is lost in this chain the following can be used as workaround dhazelcast logging type
1
19,037
3,129,397,136
IssuesEvent
2015-09-09 00:53:33
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
opened
VM Service: null instances don't have a class field
Area-Observatory Type-Defect
According to `Instance`'s documentation, "`Instance` references always include their class". However, for a null instance I get the following JSON: ```json { "type": "@Instance", "_vmType": "null", "kind": "Null", "fixedId": true, "id": "objects/null", "valueAsString": "null" } ```
1.0
VM Service: null instances don't have a class field - According to `Instance`'s documentation, "`Instance` references always include their class". However, for a null instance I get the following JSON: ```json { "type": "@Instance", "_vmType": "null", "kind": "Null", "fixedId": true, "id": "objects/null", "valueAsString": "null" } ```
defect
vm service null instances don t have a class field according to instance s documentation instance references always include their class however for a null instance i get the following json json type instance vmtype null kind null fixedid true id objects null valueasstring null
1
8,252
2,611,473,470
IssuesEvent
2015-02-27 05:17:35
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
team name disappears when switching to fullscreen in OS X Lion
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. start a game (windowed by default) 2. switch to fullscreen (a team name disappears) 3. go back to windowed mode (the problem persists) What is the expected output? What do you see instead? What version of the product are you using? On what operating system? 0.9.16 OS X Lion Please provide any additional information below. ``` Original issue reported on code.google.com by `iiigerar...@gmail.com` on 25 Sep 2011 at 2:32 Attachments: * [1,PlayingNormally.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-289/comment-0/1,PlayingNormally.png) * [2,SwitchingToFullscreen.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-289/comment-0/2,SwitchingToFullscreen.png) * [3,GoingBackToWindowed.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-289/comment-0/3,GoingBackToWindowed.png)
1.0
team name disappears when switching to fullscreen in OS X Lion - ``` What steps will reproduce the problem? 1. start a game (windowed by default) 2. switch to fullscreen (a team name disappears) 3. go back to windowed mode (the problem persists) What is the expected output? What do you see instead? What version of the product are you using? On what operating system? 0.9.16 OS X Lion Please provide any additional information below. ``` Original issue reported on code.google.com by `iiigerar...@gmail.com` on 25 Sep 2011 at 2:32 Attachments: * [1,PlayingNormally.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-289/comment-0/1,PlayingNormally.png) * [2,SwitchingToFullscreen.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-289/comment-0/2,SwitchingToFullscreen.png) * [3,GoingBackToWindowed.png](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-289/comment-0/3,GoingBackToWindowed.png)
defect
team name disappears when switching to fullscreen in os x lion what steps will reproduce the problem start a game windowed by default switch to fullscreen a team name disappears go back to windowed mode the problem persists what is the expected output what do you see instead what version of the product are you using on what operating system os x lion please provide any additional information below original issue reported on code google com by iiigerar gmail com on sep at attachments
1
176,162
14,565,334,478
IssuesEvent
2020-12-17 07:04:09
MarkEdmondson1234/googleAnalyticsR
https://api.github.com/repos/MarkEdmondson1234/googleAnalyticsR
closed
[DOCS]
documentation
Dear Mark, Thanks for your package. When reading the [tutorials section](https://code.markedmondson.me/googleAnalyticsR/#tutorials) of the package's documentation, I found some broken links: * Bert uses googleAnalyticsR to detect bot traffic * Fredrik forecasts sales from Google Analytics and sends it to your own Slack bot [Swedish] * Omar shows some heatmaps and ggplot2 using googleAnalyticsR in Spanish * Mararoff.com shows how to compare your CRM data with Google Analytics in Russian
1.0
[DOCS] - Dear Mark, Thanks for your package. When reading the [tutorials section](https://code.markedmondson.me/googleAnalyticsR/#tutorials) of the package's documentation, I found some broken links: * Bert uses googleAnalyticsR to detect bot traffic * Fredrik forecasts sales from Google Analytics and sends it to your own Slack bot [Swedish] * Omar shows some heatmaps and ggplot2 using googleAnalyticsR in Spanish * Mararoff.com shows how to compare your CRM data with Google Analytics in Russian
non_defect
dear mark thanks for your package when reading the of the package s documentation i found some broken links bert uses googleanalyticsr to detect bot traffic fredrik forecasts sales from google analytics and sends it to your own slack bot omar shows some heatmaps and using googleanalyticsr in spanish mararoff com shows how to compare your crm data with google analytics in russian
0
42,310
10,960,782,758
IssuesEvent
2019-11-27 14:14:09
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
MacOS Catalina requires all PKGs to be notarized
C: Documentation T: defect
## Bug Description MacOS Catalina requires all .pkg bundles to be notarized. Ref: https://developer.apple.com/documentation/xcode/notarizing_macos_software_before_distribution ## Steps to Reproduce Download the app using a web browser. This will apply a quarantine attribute to the downloaded file. Which, triggers the notarizing halt. ## Impact Without documentation, a user may not know how to circumvent GateKeeper.
1.0
MacOS Catalina requires all PKGs to be notarized - ## Bug Description MacOS Catalina requires all .pkg bundles to be notarized. Ref: https://developer.apple.com/documentation/xcode/notarizing_macos_software_before_distribution ## Steps to Reproduce Download the app using a web browser. This will apply a quarantine attribute to the downloaded file. Which, triggers the notarizing halt. ## Impact Without documentation, a user may not know how to circumvent GateKeeper.
defect
macos catalina requires all pkgs to be notarized bug description macos catalina requires all pkg bundles to be notarized ref steps to reproduce download the app using a web browser this will apply a quarantine attribute to the downloaded file which triggers the notarizing halt impact without documentation a user may not know how to circumvent gatekeeper
1
57,883
16,115,783,561
IssuesEvent
2021-04-28 07:12:45
line/armeria
https://api.github.com/repos/line/armeria
closed
`ClosedStreamException` triggered by `ArmeriaServerCall.doSendMessage()` not handled by Armeria
defect
We got the following exception report from a user at our Slack workspace: ``` 2021-04-12 14:14:02.858 WARN 8324 --- [-worker-nio-2-8] i.n.u.concurrent.AbstractEventExecutor : A task raised an exception. Task: com.linecorp.armeria.common.RequestContext$$Lambda$1068/66855464@6d21f471 com.linecorp.armeria.common.stream.ClosedStreamException: null at com.linecorp.armeria.common.stream.ClosedStreamException.get(ClosedStreamException.java:37) ~[armeria-1.5.0.jar:na] at com.linecorp.armeria.common.stream.StreamWriter.write(StreamWriter.java:71) ~[armeria-1.5.0.jar:na] at com.linecorp.armeria.server.grpc.ArmeriaServerCall.doSendMessage(ArmeriaServerCall.java:287) ~[armeria-grpc-1.5.0.jar:na] at com.linecorp.armeria.server.grpc.ArmeriaServerCall.lambda$sendMessage$4(ArmeriaServerCall.java:270) ~[armeria-grpc-1.5.0.jar:na] at com.linecorp.armeria.common.RequestContext.lambda$makeContextAware$3(RequestContext.java:522) ~[armeria-1.5.0.jar:na] at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164) ~[netty-common-4.1.55.Final.jar:4.1.55.Final] at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472) [netty-common-4.1.55.Final.jar:4.1.55.Final] at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:497) [netty-transport-4.1.55.Final.jar:4.1.55.Final] at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) [netty-common-4.1.55.Final.jar:4.1.55.Final] at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) [netty-common-4.1.55.Final.jar:4.1.55.Final] at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) [netty-common-4.1.55.Final.jar:4.1.55.Final] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_281] ``` According to the user, the exception is triggered when a server attempts to write something to the stream cancelled by a client. The stack trace shows that we did not catch the exception but just left it propagated to Netty. We should instead let it go to a `RequestLog` or discard it if that makes sense.
1.0
`ClosedStreamException` triggered by `ArmeriaServerCall.doSendMessage()` not handled by Armeria - We got the following exception report from a user at our Slack workspace: ``` 2021-04-12 14:14:02.858 WARN 8324 --- [-worker-nio-2-8] i.n.u.concurrent.AbstractEventExecutor : A task raised an exception. Task: com.linecorp.armeria.common.RequestContext$$Lambda$1068/66855464@6d21f471 com.linecorp.armeria.common.stream.ClosedStreamException: null at com.linecorp.armeria.common.stream.ClosedStreamException.get(ClosedStreamException.java:37) ~[armeria-1.5.0.jar:na] at com.linecorp.armeria.common.stream.StreamWriter.write(StreamWriter.java:71) ~[armeria-1.5.0.jar:na] at com.linecorp.armeria.server.grpc.ArmeriaServerCall.doSendMessage(ArmeriaServerCall.java:287) ~[armeria-grpc-1.5.0.jar:na] at com.linecorp.armeria.server.grpc.ArmeriaServerCall.lambda$sendMessage$4(ArmeriaServerCall.java:270) ~[armeria-grpc-1.5.0.jar:na] at com.linecorp.armeria.common.RequestContext.lambda$makeContextAware$3(RequestContext.java:522) ~[armeria-1.5.0.jar:na] at io.netty.util.concurrent.AbstractEventExecutor.safeExecute(AbstractEventExecutor.java:164) ~[netty-common-4.1.55.Final.jar:4.1.55.Final] at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:472) [netty-common-4.1.55.Final.jar:4.1.55.Final] at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:497) [netty-transport-4.1.55.Final.jar:4.1.55.Final] at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989) [netty-common-4.1.55.Final.jar:4.1.55.Final] at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) [netty-common-4.1.55.Final.jar:4.1.55.Final] at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) [netty-common-4.1.55.Final.jar:4.1.55.Final] at java.lang.Thread.run(Thread.java:748) [na:1.8.0_281] ``` According to the user, the exception is triggered when a server attempts to write something to the stream cancelled by a client. The stack trace shows that we did not catch the exception but just left it propagated to Netty. We should instead let it go to a `RequestLog` or discard it if that makes sense.
defect
closedstreamexception triggered by armeriaservercall dosendmessage not handled by armeria we got the following exception report from a user at our slack workspace warn i n u concurrent abstracteventexecutor a task raised an exception task com linecorp armeria common requestcontext lambda com linecorp armeria common stream closedstreamexception null at com linecorp armeria common stream closedstreamexception get closedstreamexception java at com linecorp armeria common stream streamwriter write streamwriter java at com linecorp armeria server grpc armeriaservercall dosendmessage armeriaservercall java at com linecorp armeria server grpc armeriaservercall lambda sendmessage armeriaservercall java at com linecorp armeria common requestcontext lambda makecontextaware requestcontext java at io netty util concurrent abstracteventexecutor safeexecute abstracteventexecutor java at io netty util concurrent singlethreadeventexecutor runalltasks singlethreadeventexecutor java at io netty channel nio nioeventloop run nioeventloop java at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java at io netty util internal threadexecutormap run threadexecutormap java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java lang thread run thread java according to the user the exception is triggered when a server attempts to write something to the stream cancelled by a client the stack trace shows that we did not catch the exception but just left it propagated to netty we should instead let it go to a requestlog or discard it if that makes sense
1
153,658
12,156,448,111
IssuesEvent
2020-04-25 17:22:38
SAP/spartacus
https://api.github.com/repos/SAP/spartacus
closed
E2E: additional test case
applied-promotions e2e-tests promotions
Test scenario: -> add two products from the price range 100-150$ to the cart -> test if you see the order promotion for order with total sum of > 200$ -> remove one of the product -> test that you should not see this order promotion anymore
1.0
E2E: additional test case - Test scenario: -> add two products from the price range 100-150$ to the cart -> test if you see the order promotion for order with total sum of > 200$ -> remove one of the product -> test that you should not see this order promotion anymore
non_defect
additional test case test scenario add two products from the price range to the cart test if you see the order promotion for order with total sum of remove one of the product test that you should not see this order promotion anymore
0
15,296
2,850,599,479
IssuesEvent
2015-05-31 18:21:31
damonkohler/sl4a
https://api.github.com/repos/damonkohler/sl4a
opened
SL4A Force Close on droid.startActivityIntent(chooserIntent)
auto-migrated Priority-Medium Type-Defect
_From @GoogleCodeExporter on May 31, 2015 11:30_ ``` What device(s) are you experiencing the problem on? Samsung Vibrant (SGH-T959) What firmware version are you running on the device? 2.2 What steps will reproduce the problem? 1. Run the attached Python script on an Android device with SL4Ar4 -or- 1. Make an intent with action="android.intent.action.SEND", uriType = "text/plain", extras = {"EXTRA_TEXT":"my message", "EXTRA_SUBJECT":"and my subject"}, and everything else = None. 2. Make another intent with action = "android.intent.action.CHOOSER", extras = {"EXTRA_INTENT":theFirstIntent}, and everything else = None. 3. droid.startActivityIntent(yourSecondIntent) 4. SL4A force closes What is the expected output? What do you see instead? I'm trying to get a chooser for sharing text via other apps (Gmail, Messaging, etc.). It should be just like the one from the "Share" option in the menu of SL4A's script editor activity. Instead, SL4A force closes, and there's no indication whatsoever to the Python interpreter. What version of the product are you using? On what operating system? I'm using SL4Ar4 on Samung's OEM build FROYO.UVKB5 My development machine is Windows 7 64-bit, but running the script directly on the phone still crashes, so that's probably not the issue. Please provide any additional information below. I tried to adapt the code from the "Share" option of the SL4A script editor by manually making the same intent as Intent.createChooser() as described in the Android documentation. My script is adapted from lines 198-202 here: http://code.google.com/p/android-scripting/source/diff?spec=svndeba523e8a9b44ea6 ef6e2497e21914084b109a8&r=deba523e8a9b44ea6ef6e2497e21914084b109a8&format=side&p ath=/android/ScriptingLayerForAndroid/src/com/googlecode/android_scripting/activ ity/ScriptEditor.java And this describes how to make the chooser intent: http://developer.android.com/reference/android/content/Intent.html#ACTION_CHOOSE R ``` Original issue reported on code.google.com by `armadaje...@gmail.com` on 5 Dec 2011 at 12:55 Attachments: * [logcat.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/logcat.txt) * [chooser.py](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/chooser.py) _Copied from original issue: damonkohler/android-scripting#588_
1.0
SL4A Force Close on droid.startActivityIntent(chooserIntent) - _From @GoogleCodeExporter on May 31, 2015 11:30_ ``` What device(s) are you experiencing the problem on? Samsung Vibrant (SGH-T959) What firmware version are you running on the device? 2.2 What steps will reproduce the problem? 1. Run the attached Python script on an Android device with SL4Ar4 -or- 1. Make an intent with action="android.intent.action.SEND", uriType = "text/plain", extras = {"EXTRA_TEXT":"my message", "EXTRA_SUBJECT":"and my subject"}, and everything else = None. 2. Make another intent with action = "android.intent.action.CHOOSER", extras = {"EXTRA_INTENT":theFirstIntent}, and everything else = None. 3. droid.startActivityIntent(yourSecondIntent) 4. SL4A force closes What is the expected output? What do you see instead? I'm trying to get a chooser for sharing text via other apps (Gmail, Messaging, etc.). It should be just like the one from the "Share" option in the menu of SL4A's script editor activity. Instead, SL4A force closes, and there's no indication whatsoever to the Python interpreter. What version of the product are you using? On what operating system? I'm using SL4Ar4 on Samung's OEM build FROYO.UVKB5 My development machine is Windows 7 64-bit, but running the script directly on the phone still crashes, so that's probably not the issue. Please provide any additional information below. I tried to adapt the code from the "Share" option of the SL4A script editor by manually making the same intent as Intent.createChooser() as described in the Android documentation. My script is adapted from lines 198-202 here: http://code.google.com/p/android-scripting/source/diff?spec=svndeba523e8a9b44ea6 ef6e2497e21914084b109a8&r=deba523e8a9b44ea6ef6e2497e21914084b109a8&format=side&p ath=/android/ScriptingLayerForAndroid/src/com/googlecode/android_scripting/activ ity/ScriptEditor.java And this describes how to make the chooser intent: http://developer.android.com/reference/android/content/Intent.html#ACTION_CHOOSE R ``` Original issue reported on code.google.com by `armadaje...@gmail.com` on 5 Dec 2011 at 12:55 Attachments: * [logcat.txt](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/logcat.txt) * [chooser.py](https://storage.googleapis.com/google-code-attachments/android-scripting/issue-588/comment-0/chooser.py) _Copied from original issue: damonkohler/android-scripting#588_
defect
force close on droid startactivityintent chooserintent from googlecodeexporter on may what device s are you experiencing the problem on samsung vibrant sgh what firmware version are you running on the device what steps will reproduce the problem run the attached python script on an android device with or make an intent with action android intent action send uritype text plain extras extra text my message extra subject and my subject and everything else none make another intent with action android intent action chooser extras extra intent thefirstintent and everything else none droid startactivityintent yoursecondintent force closes what is the expected output what do you see instead i m trying to get a chooser for sharing text via other apps gmail messaging etc it should be just like the one from the share option in the menu of s script editor activity instead force closes and there s no indication whatsoever to the python interpreter what version of the product are you using on what operating system i m using on samung s oem build froyo my development machine is windows bit but running the script directly on the phone still crashes so that s probably not the issue please provide any additional information below i tried to adapt the code from the share option of the script editor by manually making the same intent as intent createchooser as described in the android documentation my script is adapted from lines here r format side p ath android scriptinglayerforandroid src com googlecode android scripting activ ity scripteditor java and this describes how to make the chooser intent r original issue reported on code google com by armadaje gmail com on dec at attachments copied from original issue damonkohler android scripting
1
69,661
22,597,048,163
IssuesEvent
2022-06-29 04:58:34
unascribed/Fabrication
https://api.github.com/repos/unascribed/Fabrication
opened
[REQUEST] Would You like to add KEYBIND to instantly open Fabrication Settings menu?
k: Defect n: Fabric s: New
Instead of opening ModMenu first. Nice idea cool isn't? THANK YOU! :)
1.0
[REQUEST] Would You like to add KEYBIND to instantly open Fabrication Settings menu? - Instead of opening ModMenu first. Nice idea cool isn't? THANK YOU! :)
defect
would you like to add keybind to instantly open fabrication settings menu instead of opening modmenu first nice idea cool isn t thank you
1
294,752
22,162,120,910
IssuesEvent
2022-06-04 16:56:03
Irishbecky91/TheWaggingTailor
https://api.github.com/repos/Irishbecky91/TheWaggingTailor
opened
TASK: README Requirements
documentation
- [ ] - Evidence of either a real Facebook site or a mockup of one for digital marketing purposes. - [ ] - A description of the e-commerce business model including marketing strategies in the README file. - [ ] - Detailed testing write-ups, beyond results of validation tools. - [ ] - Detailed ERD image displayed on README, showcasing relation between models.
1.0
TASK: README Requirements - - [ ] - Evidence of either a real Facebook site or a mockup of one for digital marketing purposes. - [ ] - A description of the e-commerce business model including marketing strategies in the README file. - [ ] - Detailed testing write-ups, beyond results of validation tools. - [ ] - Detailed ERD image displayed on README, showcasing relation between models.
non_defect
task readme requirements evidence of either a real facebook site or a mockup of one for digital marketing purposes a description of the e commerce business model including marketing strategies in the readme file detailed testing write ups beyond results of validation tools detailed erd image displayed on readme showcasing relation between models
0
78,754
27,746,229,984
IssuesEvent
2023-03-15 17:13:24
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Warning about inviting unknown users is not shown when creating a room via "Start chat"
T-Defect
### Steps to reproduce 1. Click the + "Start chat" button next to the header of the People section of the room list. 2. Attempt to start a chat with a user that you know does not exist (and alternatively, on a homeserver you know does exist). 3. Click "Go" 4. You are let through to the next stage without any warnings. 5. Send a message to create the room. If you chose a non-existent localpart on a real matrix homeserver domain, then that non-existent user will have an invite (which the homeserver received). If a user is ever created under that localpart, then they will find that invite waiting for them when they log in for the first time. If you chose a non-existent homeserver domain, then you will end up in a broken room. ### Outcome ### What did you expect? The following errors appear when attempting to invite a non-existent user through the "Invite" dialog in an existing Matrix room. When attempting to invite a non-existent localpart on an existing homeserver: ![image](https://user-images.githubusercontent.com/1342360/225382856-5bcf3aed-5eee-4739-aac6-28b57ad78122.png) When attempting to invite any user on a non-existent homeserver: ![image](https://user-images.githubusercontent.com/1342360/225382769-7de48e29-043f-4075-aacd-a672895c931d.png) ### What happened instead? For a non-existent user on an existing homeserver, no warning is presented on clicking Go here: ![Screenshot_20230315_165033](https://user-images.githubusercontent.com/1342360/225382284-af0f4564-052b-431c-bdbf-85a5612e5720.png) When attempting to contact a user on a non-existent homeserver, you end up with a broken room. Clicking retry does nothing: ![image](https://user-images.githubusercontent.com/1342360/225383940-03336908-1c37-4e15-94eb-c6a49a415773.png) ### Operating system NixOS ### Application version Element version: 1.11.24 Olm version: 3.2.12 ### How did you install the app? nixpkgs package repository ### Homeserver _No response_ ### Will you send logs? No
1.0
Warning about inviting unknown users is not shown when creating a room via "Start chat" - ### Steps to reproduce 1. Click the + "Start chat" button next to the header of the People section of the room list. 2. Attempt to start a chat with a user that you know does not exist (and alternatively, on a homeserver you know does exist). 3. Click "Go" 4. You are let through to the next stage without any warnings. 5. Send a message to create the room. If you chose a non-existent localpart on a real matrix homeserver domain, then that non-existent user will have an invite (which the homeserver received). If a user is ever created under that localpart, then they will find that invite waiting for them when they log in for the first time. If you chose a non-existent homeserver domain, then you will end up in a broken room. ### Outcome ### What did you expect? The following errors appear when attempting to invite a non-existent user through the "Invite" dialog in an existing Matrix room. When attempting to invite a non-existent localpart on an existing homeserver: ![image](https://user-images.githubusercontent.com/1342360/225382856-5bcf3aed-5eee-4739-aac6-28b57ad78122.png) When attempting to invite any user on a non-existent homeserver: ![image](https://user-images.githubusercontent.com/1342360/225382769-7de48e29-043f-4075-aacd-a672895c931d.png) ### What happened instead? For a non-existent user on an existing homeserver, no warning is presented on clicking Go here: ![Screenshot_20230315_165033](https://user-images.githubusercontent.com/1342360/225382284-af0f4564-052b-431c-bdbf-85a5612e5720.png) When attempting to contact a user on a non-existent homeserver, you end up with a broken room. Clicking retry does nothing: ![image](https://user-images.githubusercontent.com/1342360/225383940-03336908-1c37-4e15-94eb-c6a49a415773.png) ### Operating system NixOS ### Application version Element version: 1.11.24 Olm version: 3.2.12 ### How did you install the app? nixpkgs package repository ### Homeserver _No response_ ### Will you send logs? No
defect
warning about inviting unknown users is not shown when creating a room via start chat steps to reproduce click the start chat button next to the header of the people section of the room list attempt to start a chat with a user that you know does not exist and alternatively on a homeserver you know does exist click go you are let through to the next stage without any warnings send a message to create the room if you chose a non existent localpart on a real matrix homeserver domain then that non existent user will have an invite which the homeserver received if a user is ever created under that localpart then they will find that invite waiting for them when they log in for the first time if you chose a non existent homeserver domain then you will end up in a broken room outcome what did you expect the following errors appear when attempting to invite a non existent user through the invite dialog in an existing matrix room when attempting to invite a non existent localpart on an existing homeserver when attempting to invite any user on a non existent homeserver what happened instead for a non existent user on an existing homeserver no warning is presented on clicking go here when attempting to contact a user on a non existent homeserver you end up with a broken room clicking retry does nothing operating system nixos application version element version olm version how did you install the app nixpkgs package repository homeserver no response will you send logs no
1
26,687
4,777,576,680
IssuesEvent
2016-10-27 16:40:31
wheeler-microfluidics/microdrop
https://api.github.com/repos/wheeler-microfluidics/microdrop
opened
gstreamer help message overriding argparse message (Trac #166)
defect Incomplete Migration microdrop Migrated from Trac
Migrated from http://microfluidics.utoronto.ca/microdrop/ticket/166 ```json { "status": "new", "changetime": "2015-01-06T14:23:40", "description": "{{{\n#!md\nSee [here][1].\n\n[1]: http://stackoverflow.com/questions/12059806/gstreamer-help-message-overriding-my-argparse-message#answer-12417626\n}}}", "reporter": "cfobel", "cc": "", "resolution": "", "_ts": "1420554220756600", "component": "microdrop", "summary": "gstreamer help message overriding argparse message", "priority": "minor", "keywords": "", "version": "0.1", "time": "2015-01-06T14:23:40", "milestone": "", "owner": "", "type": "defect" } ```
1.0
gstreamer help message overriding argparse message (Trac #166) - Migrated from http://microfluidics.utoronto.ca/microdrop/ticket/166 ```json { "status": "new", "changetime": "2015-01-06T14:23:40", "description": "{{{\n#!md\nSee [here][1].\n\n[1]: http://stackoverflow.com/questions/12059806/gstreamer-help-message-overriding-my-argparse-message#answer-12417626\n}}}", "reporter": "cfobel", "cc": "", "resolution": "", "_ts": "1420554220756600", "component": "microdrop", "summary": "gstreamer help message overriding argparse message", "priority": "minor", "keywords": "", "version": "0.1", "time": "2015-01-06T14:23:40", "milestone": "", "owner": "", "type": "defect" } ```
defect
gstreamer help message overriding argparse message trac migrated from json status new changetime description n md nsee n n reporter cfobel cc resolution ts component microdrop summary gstreamer help message overriding argparse message priority minor keywords version time milestone owner type defect
1
2,885
2,607,964,509
IssuesEvent
2015-02-26 00:41:43
chrsmithdemos/leveldb
https://api.github.com/repos/chrsmithdemos/leveldb
closed
Support for DragonFlyBSD/NetBSD/OpenBSD
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Try to compile on one of the above operating systems 2. The compile will fail because of missing includes and missing fdatasync, etc What is the expected output? What do you see instead? leveldb should compile correctly, instead it fails. Please provide any additional information below. Patch is attached. This supersedes issue 31. ``` ----- Original issue reported on code.google.com by `and...@hijacked.us` on 3 Dec 2011 at 3:42 Attachments: * [leveldb_bsd_support.diff](https://storage.googleapis.com/google-code-attachments/leveldb/issue-59/comment-0/leveldb_bsd_support.diff)
1.0
Support for DragonFlyBSD/NetBSD/OpenBSD - ``` What steps will reproduce the problem? 1. Try to compile on one of the above operating systems 2. The compile will fail because of missing includes and missing fdatasync, etc What is the expected output? What do you see instead? leveldb should compile correctly, instead it fails. Please provide any additional information below. Patch is attached. This supersedes issue 31. ``` ----- Original issue reported on code.google.com by `and...@hijacked.us` on 3 Dec 2011 at 3:42 Attachments: * [leveldb_bsd_support.diff](https://storage.googleapis.com/google-code-attachments/leveldb/issue-59/comment-0/leveldb_bsd_support.diff)
defect
support for dragonflybsd netbsd openbsd what steps will reproduce the problem try to compile on one of the above operating systems the compile will fail because of missing includes and missing fdatasync etc what is the expected output what do you see instead leveldb should compile correctly instead it fails please provide any additional information below patch is attached this supersedes issue original issue reported on code google com by and hijacked us on dec at attachments
1
84,716
3,670,937,163
IssuesEvent
2016-02-22 02:46:33
coollog/sublite
https://api.github.com/repos/coollog/sublite
opened
Show most popular job listings
2 Difficulty 3 Length 5 Priority Category: Jobs Type: Feature
Not much detail on this yet, so need a design for where it should go first.
1.0
Show most popular job listings - Not much detail on this yet, so need a design for where it should go first.
non_defect
show most popular job listings not much detail on this yet so need a design for where it should go first
0
414,309
27,983,938,134
IssuesEvent
2023-03-26 13:32:49
OpenTabletDriver/OpenTabletDriver.Web
https://api.github.com/repos/OpenTabletDriver/OpenTabletDriver.Web
opened
Document DE's that handle monitor mapping (KDE, more?)
documentation
Problem: DE sees tablet and maps tablet for you. However, since OTD also maps the tablet for you, the common problem is along the lines of "tablet only reachable in the upper left" or similar. The fix is to either map KDE to the full monitor layout (OTD devs: can we work around this?), or mapping to the entire monitor layout in OpenTabletDriver. This can likely happen on Windows too in some setups, so this likely needs to be included in a generic FAQ.
1.0
Document DE's that handle monitor mapping (KDE, more?) - Problem: DE sees tablet and maps tablet for you. However, since OTD also maps the tablet for you, the common problem is along the lines of "tablet only reachable in the upper left" or similar. The fix is to either map KDE to the full monitor layout (OTD devs: can we work around this?), or mapping to the entire monitor layout in OpenTabletDriver. This can likely happen on Windows too in some setups, so this likely needs to be included in a generic FAQ.
non_defect
document de s that handle monitor mapping kde more problem de sees tablet and maps tablet for you however since otd also maps the tablet for you the common problem is along the lines of tablet only reachable in the upper left or similar the fix is to either map kde to the full monitor layout otd devs can we work around this or mapping to the entire monitor layout in opentabletdriver this can likely happen on windows too in some setups so this likely needs to be included in a generic faq
0
70,904
23,366,122,153
IssuesEvent
2022-08-10 15:29:55
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Subspaces aren't indented
T-Defect X-Regression S-Minor X-Release-Blocker A-Spaces O-Frequent
### Steps to reproduce 1. Expand the space panel 2. Click the 'expand' arrow next to a top-level space to reveal its subspaces ### Outcome #### What did you expect? The subspaces should be indented further than the top-level space #### What happened instead? They have matching indentation ![Screenshot 2022-08-10 at 08-42-38 Element 1 tooling](https://user-images.githubusercontent.com/48614497/183904420-b6802050-5823-42ad-87ec-b740f6bbf32b.png) This was probably regressed by https://github.com/matrix-org/matrix-react-sdk/pull/9157 ### Operating system NixOS unstable ### Browser information Firefox 102.0.1 ### URL for webapp develop.element.io ### Application version Element version: 9c36adc4dfe8-react-3d0982e9a603-js-8f13df2dd986 Olm version: 3.2.12 ### Homeserver Synapse 1.63.1 ### Will you send logs? No
1.0
Subspaces aren't indented - ### Steps to reproduce 1. Expand the space panel 2. Click the 'expand' arrow next to a top-level space to reveal its subspaces ### Outcome #### What did you expect? The subspaces should be indented further than the top-level space #### What happened instead? They have matching indentation ![Screenshot 2022-08-10 at 08-42-38 Element 1 tooling](https://user-images.githubusercontent.com/48614497/183904420-b6802050-5823-42ad-87ec-b740f6bbf32b.png) This was probably regressed by https://github.com/matrix-org/matrix-react-sdk/pull/9157 ### Operating system NixOS unstable ### Browser information Firefox 102.0.1 ### URL for webapp develop.element.io ### Application version Element version: 9c36adc4dfe8-react-3d0982e9a603-js-8f13df2dd986 Olm version: 3.2.12 ### Homeserver Synapse 1.63.1 ### Will you send logs? No
defect
subspaces aren t indented steps to reproduce expand the space panel click the expand arrow next to a top level space to reveal its subspaces outcome what did you expect the subspaces should be indented further than the top level space what happened instead they have matching indentation this was probably regressed by operating system nixos unstable browser information firefox url for webapp develop element io application version element version react js olm version homeserver synapse will you send logs no
1
23,253
3,783,890,467
IssuesEvent
2016-03-19 12:46:04
VladSerdobintsev/zfcore
https://api.github.com/repos/VladSerdobintsev/zfcore
closed
Windows Install Process
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? Try to install ZFCore on WAMP What is the expected output? What do you see instead? Инсталяция на windows прошла с ошибками: 1. install.bat в отличии от install.sh не преименовывает .htaccess.sample - .htaccess, пришлось вручную после чего запустилась веб инсталятор 2. application.yaml не содается - тоже создал вручную и скопировал конф текст в конце устаноки 3. и rm -r .....modules/install - такой команди нету в windows. И еще неплохо было вставлять в install module DIR_SEPARATOR, потомучто путя в установцике виглядят примерно вот так c:\a\b/app/conf Миграции поднялись, админка работает ``` Original issue reported on code.google.com by `AntonShe...@gmail.com` on 31 Jul 2012 at 11:44
1.0
Windows Install Process - ``` What steps will reproduce the problem? Try to install ZFCore on WAMP What is the expected output? What do you see instead? Инсталяция на windows прошла с ошибками: 1. install.bat в отличии от install.sh не преименовывает .htaccess.sample - .htaccess, пришлось вручную после чего запустилась веб инсталятор 2. application.yaml не содается - тоже создал вручную и скопировал конф текст в конце устаноки 3. и rm -r .....modules/install - такой команди нету в windows. И еще неплохо было вставлять в install module DIR_SEPARATOR, потомучто путя в установцике виглядят примерно вот так c:\a\b/app/conf Миграции поднялись, админка работает ``` Original issue reported on code.google.com by `AntonShe...@gmail.com` on 31 Jul 2012 at 11:44
defect
windows install process what steps will reproduce the problem try to install zfcore on wamp what is the expected output what do you see instead инсталяция на windows прошла с ошибками install bat в отличии от install sh не преименовывает htaccess sample htaccess пришлось вручную после чего запустилась веб инсталятор application yaml не содается тоже создал вручную и скопировал конф текст в конце устаноки и rm r modules install такой команди нету в windows и еще неплохо было вставлять в install module dir separator потомучто путя в установцике виглядят примерно вот так c a b app conf миграции поднялись админка работает original issue reported on code google com by antonshe gmail com on jul at
1
50,773
13,187,734,787
IssuesEvent
2020-08-13 04:24:14
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
[topsimulator] I3TimeShiffter (Trac #1345)
Migrated from Trac combo simulation defect
The segment still uses the C++ version of I3TimeShifter, which was removed. Why is the segment using this at all? Why not just use the I3TriggerSim segment? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1345">https://code.icecube.wisc.edu/ticket/1345</a>, reported by olivas and owned by jgonzalez</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:03", "description": "The segment still uses the C++ version of I3TimeShifter, which was removed. Why is the segment using this at all? Why not just use the I3TriggerSim segment?", "reporter": "olivas", "cc": "", "resolution": "fixed", "_ts": "1458335643235016", "component": "combo simulation", "summary": "[topsimulator] I3TimeShiffter", "priority": "blocker", "keywords": "", "time": "2015-09-15T06:02:55", "milestone": "", "owner": "jgonzalez", "type": "defect" } ``` </p> </details>
1.0
[topsimulator] I3TimeShiffter (Trac #1345) - The segment still uses the C++ version of I3TimeShifter, which was removed. Why is the segment using this at all? Why not just use the I3TriggerSim segment? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1345">https://code.icecube.wisc.edu/ticket/1345</a>, reported by olivas and owned by jgonzalez</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:03", "description": "The segment still uses the C++ version of I3TimeShifter, which was removed. Why is the segment using this at all? Why not just use the I3TriggerSim segment?", "reporter": "olivas", "cc": "", "resolution": "fixed", "_ts": "1458335643235016", "component": "combo simulation", "summary": "[topsimulator] I3TimeShiffter", "priority": "blocker", "keywords": "", "time": "2015-09-15T06:02:55", "milestone": "", "owner": "jgonzalez", "type": "defect" } ``` </p> </details>
defect
trac the segment still uses the c version of which was removed why is the segment using this at all why not just use the segment migrated from json status closed changetime description the segment still uses the c version of which was removed why is the segment using this at all why not just use the segment reporter olivas cc resolution fixed ts component combo simulation summary priority blocker keywords time milestone owner jgonzalez type defect
1
595,500
18,067,678,174
IssuesEvent
2021-09-20 21:13:18
OpenMandrivaAssociation/test2
https://api.github.com/repos/OpenMandrivaAssociation/test2
closed
ruby 2.0 and rpm macro (Bugzilla Bug 241)
bug high priority major
This issue was created automatically with bugzilla2github # Bugzilla Bug 241 Date: 2013-10-28 13:35:59 +0000 From: Alexander &lt;<nobodydead@gmail.com>&gt; To: OpenMandriva QA &lt;<bugs@openmandriva.org>&gt; CC: @berolinux, @cris-b, @itchka, nix.or.die@gmail.com, raul.liota@gmail.com, @robxu9, @tpgxyz Last updated: 2018-03-03 03:32:47 +0000 ## Comment 1700 Date: 2013-10-28 13:35:59 +0000 From: Alexander &lt;<nobodydead@gmail.com>&gt; In Mageia gem_helper.rb is deprecated since ruby version 2.0 And mga have a few good macros for it /etc/rpm/macros.d/20build.macros like ============================================================ %gem_setup \ %setup -q \ tar xmf data.tar.gz \ if [ ! -f %{__ruby_module_name}.gemspec ]; then \ %{_bindir}/gem specification -l --ruby %{SOURCE0} > %{__ruby_module_name}.gemspec \ fi \ sed -i -e 's/`git ls-files`/`find -type f`/' \\\ -e 's/`git ls-files -- \\\(.*\\\)`/`find \\\1 -type f`/' \\\ -e 's/`git ls-files \\\(.*\\\)`/`find \\\1 -type f`/' \\\ -e 's/ s.version = .*/ s.version = "%{version}"/' \\\ %{__ruby_module_name}.gemspec =============================================================== %gem_build \ if [ ! -f %{__ruby_module_name}.gemspec ]; then \ %{_bindir}/gem specification -l --ruby %{SOURCE0} > %{__ruby_module_name}.gemspec \ fi \ %{_bindir}/gem build %{__ruby_module_name}.gemspec =============================================================== %gem_install(d) \ %{_bindir}/gem install %{__ruby_module_name}-%{version}.gem --local \\\ --env-shebang --rdoc --ri --force --ignore-dependencies \\\ --install-dir %{buildroot}%{ruby_gemdir} \\\ --bindir %{buildroot}%{_bindir} \ %{-d: \ mkdir -p %{buildroot}%{ruby_vendorarchdir} \ for f in $(find %{buildroot}%{ruby_gemdir}/gems/%{__ruby_module_name}-%{version}/ext -name *.h); do \ mv $f %{buildroot}%{ruby_vendorarchdir}/ \ done \ } \ (cd %{buildroot}%{ruby_gemdir}/gems/%{__ruby_module_name}-%{version}; \\\ for f in $(find -name '*.so'); do \\\ dest=${f##./ext/}; \\\ dest=${dest##./lib/}; \\\ mkdir -p %{buildroot}%{ruby_sitearchdir}/$(dirname $dest); \\\ mv $f %{buildroot}%{ruby_sitearchdir}/$dest; done ) \ rm -fr %{buildroot}%{ruby_gemdir}/{cache,gems/%{__ruby_module_name}-%{version}/ext,gems/%{__ruby_module_name}-%{version}/%{__ruby_module_name}.gemspec} \ find %{buildroot}%{ruby_gemdir}/gems/%{__ruby_module_name}-%{version} -maxdepth 1 -type f -delete \ =============================================================== We need it too. In a short 1.we need same macros 2.drop gem_helper.rb as depreacted feature ## Comment 3728 Date: 2014-03-11 14:57:00 +0000 From: @robxu9 What is the status of this? ## Comment 4732 Date: 2014-04-01 11:35:12 +0000 From: @itchka Will this be actioned? ## Comment 4733 Date: 2014-04-01 12:04:16 +0000 From: Alexander &lt;<nobodydead@gmail.com>&gt; Need to fix ;) ## Comment 4757 Date: 2014-04-02 12:05:05 +0000 From: @tpgxyz Look this only affects development and particulary build packages on ABF. For an end user look and feel this is almost invisible. Fedya you are enough skilled to add these macros to rpm packages by yourself :p TPG Cooker Team ## Comment 6536 Date: 2014-07-12 23:23:39 +0000 From: @itchka What is the status of this bug? ## Comment 9095 Date: 2015-04-14 07:20:18 +0000 From: @itchka Has this bug been resolved?
1.0
ruby 2.0 and rpm macro (Bugzilla Bug 241) - This issue was created automatically with bugzilla2github # Bugzilla Bug 241 Date: 2013-10-28 13:35:59 +0000 From: Alexander &lt;<nobodydead@gmail.com>&gt; To: OpenMandriva QA &lt;<bugs@openmandriva.org>&gt; CC: @berolinux, @cris-b, @itchka, nix.or.die@gmail.com, raul.liota@gmail.com, @robxu9, @tpgxyz Last updated: 2018-03-03 03:32:47 +0000 ## Comment 1700 Date: 2013-10-28 13:35:59 +0000 From: Alexander &lt;<nobodydead@gmail.com>&gt; In Mageia gem_helper.rb is deprecated since ruby version 2.0 And mga have a few good macros for it /etc/rpm/macros.d/20build.macros like ============================================================ %gem_setup \ %setup -q \ tar xmf data.tar.gz \ if [ ! -f %{__ruby_module_name}.gemspec ]; then \ %{_bindir}/gem specification -l --ruby %{SOURCE0} > %{__ruby_module_name}.gemspec \ fi \ sed -i -e 's/`git ls-files`/`find -type f`/' \\\ -e 's/`git ls-files -- \\\(.*\\\)`/`find \\\1 -type f`/' \\\ -e 's/`git ls-files \\\(.*\\\)`/`find \\\1 -type f`/' \\\ -e 's/ s.version = .*/ s.version = "%{version}"/' \\\ %{__ruby_module_name}.gemspec =============================================================== %gem_build \ if [ ! -f %{__ruby_module_name}.gemspec ]; then \ %{_bindir}/gem specification -l --ruby %{SOURCE0} > %{__ruby_module_name}.gemspec \ fi \ %{_bindir}/gem build %{__ruby_module_name}.gemspec =============================================================== %gem_install(d) \ %{_bindir}/gem install %{__ruby_module_name}-%{version}.gem --local \\\ --env-shebang --rdoc --ri --force --ignore-dependencies \\\ --install-dir %{buildroot}%{ruby_gemdir} \\\ --bindir %{buildroot}%{_bindir} \ %{-d: \ mkdir -p %{buildroot}%{ruby_vendorarchdir} \ for f in $(find %{buildroot}%{ruby_gemdir}/gems/%{__ruby_module_name}-%{version}/ext -name *.h); do \ mv $f %{buildroot}%{ruby_vendorarchdir}/ \ done \ } \ (cd %{buildroot}%{ruby_gemdir}/gems/%{__ruby_module_name}-%{version}; \\\ for f in $(find -name '*.so'); do \\\ dest=${f##./ext/}; \\\ dest=${dest##./lib/}; \\\ mkdir -p %{buildroot}%{ruby_sitearchdir}/$(dirname $dest); \\\ mv $f %{buildroot}%{ruby_sitearchdir}/$dest; done ) \ rm -fr %{buildroot}%{ruby_gemdir}/{cache,gems/%{__ruby_module_name}-%{version}/ext,gems/%{__ruby_module_name}-%{version}/%{__ruby_module_name}.gemspec} \ find %{buildroot}%{ruby_gemdir}/gems/%{__ruby_module_name}-%{version} -maxdepth 1 -type f -delete \ =============================================================== We need it too. In a short 1.we need same macros 2.drop gem_helper.rb as depreacted feature ## Comment 3728 Date: 2014-03-11 14:57:00 +0000 From: @robxu9 What is the status of this? ## Comment 4732 Date: 2014-04-01 11:35:12 +0000 From: @itchka Will this be actioned? ## Comment 4733 Date: 2014-04-01 12:04:16 +0000 From: Alexander &lt;<nobodydead@gmail.com>&gt; Need to fix ;) ## Comment 4757 Date: 2014-04-02 12:05:05 +0000 From: @tpgxyz Look this only affects development and particulary build packages on ABF. For an end user look and feel this is almost invisible. Fedya you are enough skilled to add these macros to rpm packages by yourself :p TPG Cooker Team ## Comment 6536 Date: 2014-07-12 23:23:39 +0000 From: @itchka What is the status of this bug? ## Comment 9095 Date: 2015-04-14 07:20:18 +0000 From: @itchka Has this bug been resolved?
non_defect
ruby and rpm macro bugzilla bug this issue was created automatically with bugzilla bug date from alexander lt gt to openmandriva qa lt gt cc berolinux cris b itchka nix or die gmail com raul liota gmail com tpgxyz last updated comment date from alexander lt gt in mageia gem helper rb is deprecated since ruby version and mga have a few good macros for it etc rpm macros d macros like gem setup setup q tar xmf data tar gz if then bindir gem specification l ruby ruby module name gemspec fi sed i e s git ls files find type f e s git ls files find type f e s git ls files find type f e s s version s version version ruby module name gemspec gem build if then bindir gem specification l ruby ruby module name gemspec fi bindir gem build ruby module name gemspec gem install d bindir gem install ruby module name version gem local env shebang rdoc ri force ignore dependencies install dir buildroot ruby gemdir bindir buildroot bindir d mkdir p buildroot ruby vendorarchdir for f in find buildroot ruby gemdir gems ruby module name version ext name h do mv f buildroot ruby vendorarchdir done cd buildroot ruby gemdir gems ruby module name version for f in find name so do dest f ext dest dest lib mkdir p buildroot ruby sitearchdir dirname dest mv f buildroot ruby sitearchdir dest done rm fr buildroot ruby gemdir cache gems ruby module name version ext gems ruby module name version ruby module name gemspec find buildroot ruby gemdir gems ruby module name version maxdepth type f delete we need it too in a short we need same macros drop gem helper rb as depreacted feature comment date from what is the status of this comment date from itchka will this be actioned comment date from alexander lt gt need to fix comment date from tpgxyz look this only affects development and particulary build packages on abf for an end user look and feel this is almost invisible fedya you are enough skilled to add these macros to rpm packages by yourself p tpg cooker team comment date from itchka what is the status of this bug comment date from itchka has this bug been resolved
0
699,927
24,037,515,966
IssuesEvent
2022-09-15 20:40:58
plexiondev/plexiondev.github.io
https://api.github.com/repos/plexiondev/plexiondev.github.io
closed
make project types clearer
area:project-library priority:1 type:styling type:accessibility
it is currently not the most obvious what project type these are: ![image](https://user-images.githubusercontent.com/46572320/190486209-0bc70d44-8249-4d3c-9b85-baa74cbe876c.png) (or even that the icons represent anything) related to #78
1.0
make project types clearer - it is currently not the most obvious what project type these are: ![image](https://user-images.githubusercontent.com/46572320/190486209-0bc70d44-8249-4d3c-9b85-baa74cbe876c.png) (or even that the icons represent anything) related to #78
non_defect
make project types clearer it is currently not the most obvious what project type these are or even that the icons represent anything related to
0
68,645
21,775,565,465
IssuesEvent
2022-05-13 13:30:16
matrix-org/synapse
https://api.github.com/repos/matrix-org/synapse
closed
Server notice rooms are created rapidly if triggered by maybe_send_server_notice_to_user
S-Minor T-Defect
In the case where a server has reached it's MAU limit, the ResourceLimitsServerNotices class is meant to send a notice to users informing them of this. However, it ends up creating a ton of duplicate rooms *and* never actually invites the target for the notice to the room. - `ServerNoticesManager.get_or_create_notice_room_for_user` is used to create the room for the user but does not invite them, that happens in `maybe_invite_user_to_room` which is called together in [`send_notice`](https://github.com/matrix-org/synapse/blob/2e2d8cc2f9b9af5f8b48d75e22c474e08feca236/synapse/server_notices/server_notices_manager.py#L49) - `ResourceLimitsServerNotices` is cheeky and calls `ServerNoticesManager.get_or_create_notice_room_for_user` in [`maybe_send_server_notice_to_user`](https://github.com/matrix-org/synapse/blob/e24ff8ebe3d4119d377355402245947f7de61c00/synapse/server_notices/resource_limits_server_notices.py#L55), presumably making the fatal mistake that `get_or_create_notice_room_for_user` does NOT invite users. - Hence, if your server is over the MAU limit, it's probably calling `ResourceLimitsServerNotices.maybe_send_server_notice_to_user` often, and therefore creating new rooms often. This is causing some servers to bloat their database with empty rooms, and of course any interfaces polling the admin API for rooms will see a load of extra redundant rooms.
1.0
Server notice rooms are created rapidly if triggered by maybe_send_server_notice_to_user - In the case where a server has reached it's MAU limit, the ResourceLimitsServerNotices class is meant to send a notice to users informing them of this. However, it ends up creating a ton of duplicate rooms *and* never actually invites the target for the notice to the room. - `ServerNoticesManager.get_or_create_notice_room_for_user` is used to create the room for the user but does not invite them, that happens in `maybe_invite_user_to_room` which is called together in [`send_notice`](https://github.com/matrix-org/synapse/blob/2e2d8cc2f9b9af5f8b48d75e22c474e08feca236/synapse/server_notices/server_notices_manager.py#L49) - `ResourceLimitsServerNotices` is cheeky and calls `ServerNoticesManager.get_or_create_notice_room_for_user` in [`maybe_send_server_notice_to_user`](https://github.com/matrix-org/synapse/blob/e24ff8ebe3d4119d377355402245947f7de61c00/synapse/server_notices/resource_limits_server_notices.py#L55), presumably making the fatal mistake that `get_or_create_notice_room_for_user` does NOT invite users. - Hence, if your server is over the MAU limit, it's probably calling `ResourceLimitsServerNotices.maybe_send_server_notice_to_user` often, and therefore creating new rooms often. This is causing some servers to bloat their database with empty rooms, and of course any interfaces polling the admin API for rooms will see a load of extra redundant rooms.
defect
server notice rooms are created rapidly if triggered by maybe send server notice to user in the case where a server has reached it s mau limit the resourcelimitsservernotices class is meant to send a notice to users informing them of this however it ends up creating a ton of duplicate rooms and never actually invites the target for the notice to the room servernoticesmanager get or create notice room for user is used to create the room for the user but does not invite them that happens in maybe invite user to room which is called together in resourcelimitsservernotices is cheeky and calls servernoticesmanager get or create notice room for user in presumably making the fatal mistake that get or create notice room for user does not invite users hence if your server is over the mau limit it s probably calling resourcelimitsservernotices maybe send server notice to user often and therefore creating new rooms often this is causing some servers to bloat their database with empty rooms and of course any interfaces polling the admin api for rooms will see a load of extra redundant rooms
1
108,912
4,363,688,708
IssuesEvent
2016-08-03 01:55:52
idevelopment/Hcrm
https://api.github.com/repos/idevelopment/Hcrm
closed
Add Password routes
bug enhancement High Priority
For now the routes for the password reset system. Are not in the routes file. We need to add them.
1.0
Add Password routes - For now the routes for the password reset system. Are not in the routes file. We need to add them.
non_defect
add password routes for now the routes for the password reset system are not in the routes file we need to add them
0
55,223
14,285,480,370
IssuesEvent
2020-11-23 13:58:42
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
JSON_TABLE emulation in PostgreSQL broken
T: Defect
jOOQ 3.14.3 openjdk version "11.0.8" 2020-07-14 Postgres 13 Using **_exactly_** the example from https://www.jooq.org/doc/3.14/manual/sql-building/table-expressions/json-table-function/: ```java // For more information on imports and data types, click on the "Help" icon // These imports, and possibly others, are implied: import static org.jooq.impl.DSL.*; import static org.jooq.impl.SQLDataType.*; import org.jooq.*; import org.jooq.impl.*; // The "create" reference is an instance of DSLContext create.select() .from(jsonTable( JSON.valueOf("[{\"a\":5,\"b\":{\"x\":10}}," + "{\"a\":7,\"b\":{\"y\":20}}]"), "$.[*]" ) .column("id").forOrdinality() .column("a", INTEGER) .column("x", INTEGER).path("$.b.x") .column("y", INTEGER).path("$.b.y")) .fetch(); ``` fails with following error: ``` ERROR: subquery in FROM must have an alias at character 33 HINT: For example, FROM (SELECT ...) [AS] foo. STATEMENT: select "id", "a", "x", "y" from (select o as "id", (jsonb_path_query_first(j, '$.a'::jsonpath)->>0)::int as "a", (jsonb_path_query_first(j, $1::jsonpath)->>0)::int as "x", (jsonb_path_query_first(j, $2::jsonpath)->>0)::int as "y" from jsonb_path_query(cast($3 as jsonb), $4::jsonpath) with ordinality as t(j, o)) ``` BTW: The imports in the example are also not correct. The following two imports have a conflict: ```java import static org.jooq.impl.SQLDataType.*; import org.jooq.*; ``` both provide `JSON` as a datatype and as class. The fix is to either just import ```java import static org.jooq.impl.SQLDataType.INTEGER; ``` because only `INTEGER` is used in the concret example, or to use ```java ... org.jooq.JSON.valueOf("[{\"a\":5,\"b\":{\"x\":10}}," ... ``` in the example.
1.0
JSON_TABLE emulation in PostgreSQL broken - jOOQ 3.14.3 openjdk version "11.0.8" 2020-07-14 Postgres 13 Using **_exactly_** the example from https://www.jooq.org/doc/3.14/manual/sql-building/table-expressions/json-table-function/: ```java // For more information on imports and data types, click on the "Help" icon // These imports, and possibly others, are implied: import static org.jooq.impl.DSL.*; import static org.jooq.impl.SQLDataType.*; import org.jooq.*; import org.jooq.impl.*; // The "create" reference is an instance of DSLContext create.select() .from(jsonTable( JSON.valueOf("[{\"a\":5,\"b\":{\"x\":10}}," + "{\"a\":7,\"b\":{\"y\":20}}]"), "$.[*]" ) .column("id").forOrdinality() .column("a", INTEGER) .column("x", INTEGER).path("$.b.x") .column("y", INTEGER).path("$.b.y")) .fetch(); ``` fails with following error: ``` ERROR: subquery in FROM must have an alias at character 33 HINT: For example, FROM (SELECT ...) [AS] foo. STATEMENT: select "id", "a", "x", "y" from (select o as "id", (jsonb_path_query_first(j, '$.a'::jsonpath)->>0)::int as "a", (jsonb_path_query_first(j, $1::jsonpath)->>0)::int as "x", (jsonb_path_query_first(j, $2::jsonpath)->>0)::int as "y" from jsonb_path_query(cast($3 as jsonb), $4::jsonpath) with ordinality as t(j, o)) ``` BTW: The imports in the example are also not correct. The following two imports have a conflict: ```java import static org.jooq.impl.SQLDataType.*; import org.jooq.*; ``` both provide `JSON` as a datatype and as class. The fix is to either just import ```java import static org.jooq.impl.SQLDataType.INTEGER; ``` because only `INTEGER` is used in the concret example, or to use ```java ... org.jooq.JSON.valueOf("[{\"a\":5,\"b\":{\"x\":10}}," ... ``` in the example.
defect
json table emulation in postgresql broken jooq openjdk version postgres using exactly the example from java for more information on imports and data types click on the help icon these imports and possibly others are implied import static org jooq impl dsl import static org jooq impl sqldatatype import org jooq import org jooq impl the create reference is an instance of dslcontext create select from jsontable json valueof a b x a b y column id forordinality column a integer column x integer path b x column y integer path b y fetch fails with following error error subquery in from must have an alias at character hint for example from select foo statement select id a x y from select o as id jsonb path query first j a jsonpath int as a jsonb path query first j jsonpath int as x jsonb path query first j jsonpath int as y from jsonb path query cast as jsonb jsonpath with ordinality as t j o btw the imports in the example are also not correct the following two imports have a conflict java import static org jooq impl sqldatatype import org jooq both provide json as a datatype and as class the fix is to either just import java import static org jooq impl sqldatatype integer because only integer is used in the concret example or to use java org jooq json valueof a b x in the example
1
615,973
19,287,131,068
IssuesEvent
2021-12-11 05:48:53
Vyxal/Vyxal
https://api.github.com/repos/Vyxal/Vyxal
closed
Integers can't be sorted
bug difficulty: easy priority:medium
When trying to sort numbers with the `s` command, it errors. Example: [Try it Online!](https://vyxal.pythonanywhere.com/#WyIiLCIiLCIxMjMgcyAiLCIiLCIiXQ==)
1.0
Integers can't be sorted - When trying to sort numbers with the `s` command, it errors. Example: [Try it Online!](https://vyxal.pythonanywhere.com/#WyIiLCIiLCIxMjMgcyAiLCIiLCIiXQ==)
non_defect
integers can t be sorted when trying to sort numbers with the s command it errors example
0
81,404
30,829,673,525
IssuesEvent
2023-08-01 23:56:01
dotCMS/core
https://api.github.com/repos/dotCMS/core
closed
LogViewer connection times out.
Type : Defect QA : Approved Merged QA : Passed Internal LTS: Next Team : Scout Triage Release : 23.06
### Parent Issue _No response_ ### Problem Statement The log viewer disconnects when the load balancer times out. The old implementation used a servlet to feed the front end and used a keep-alive strategy sending a blank space every 20 seconds. That implementation got replaced by a much more modern Server-Sent Events approach. However, we still need to keep the connection alive ### Steps to Reproduce since this is happening behind an LB reproducing the original problem has implications, But apparently, Firefox offers an option to simulate a time-out setting the value of `network.http.connection-timeout` in `about:config` ### Acceptance Criteria The log viewer should remain connected and functional even after long periods of inactivity ### dotCMS Version 23.01 ### Proposed Objective Customer Support ### Proposed Priority Priority 3 - Average ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. _No response_ ### Assumptions & Initiation Needs _No response_ ### Quality Assurance Notes & Workarounds _No response_ ### Sub-Tasks & Estimates _No response_
1.0
LogViewer connection times out. - ### Parent Issue _No response_ ### Problem Statement The log viewer disconnects when the load balancer times out. The old implementation used a servlet to feed the front end and used a keep-alive strategy sending a blank space every 20 seconds. That implementation got replaced by a much more modern Server-Sent Events approach. However, we still need to keep the connection alive ### Steps to Reproduce since this is happening behind an LB reproducing the original problem has implications, But apparently, Firefox offers an option to simulate a time-out setting the value of `network.http.connection-timeout` in `about:config` ### Acceptance Criteria The log viewer should remain connected and functional even after long periods of inactivity ### dotCMS Version 23.01 ### Proposed Objective Customer Support ### Proposed Priority Priority 3 - Average ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. _No response_ ### Assumptions & Initiation Needs _No response_ ### Quality Assurance Notes & Workarounds _No response_ ### Sub-Tasks & Estimates _No response_
defect
logviewer connection times out parent issue no response problem statement the log viewer disconnects when the load balancer times out the old implementation used a servlet to feed the front end and used a keep alive strategy sending a blank space every seconds that implementation got replaced by a much more modern server sent events approach however we still need to keep the connection alive steps to reproduce since this is happening behind an lb reproducing the original problem has implications but apparently firefox offers an option to simulate a time out setting the value of network http connection timeout in about config acceptance criteria the log viewer should remain connected and functional even after long periods of inactivity dotcms version proposed objective customer support proposed priority priority average external links slack conversations support tickets figma designs etc no response assumptions initiation needs no response quality assurance notes workarounds no response sub tasks estimates no response
1
70,898
23,363,709,762
IssuesEvent
2022-08-10 13:47:58
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]: WebdriverJS FluentWait
C-nodejs I-defect
### What happened? WebdriverJS FluentWait works in Javascript as per the example in the [documentation](https://www.selenium.dev/documentation/webdriver/waits/#fluentwait). However in TypeScript we are missing the types for the 4th parameter (the frequency at which it polls). Working JavaScript example (which is polling and including the frequency at which it polls: ```shell const { Builder } = require("selenium-webdriver"); function myFunc() { console.log("hello world"); } (async function example() { let driver = await new Builder().forBrowser("chrome").build(); await driver.get("http://google.com"); // should log 5 times, polling every 1 second then times out await driver.wait(myFunc, 5000, "Timed out after 5 seconds", 1000); await driver.close(); })(); ``` ### How can we reproduce the issue? ```shell import { Builder } from "selenium-webdriver"; function myFunc() { console.log("hello world"); } (async function example() { let driver = await new Builder().forBrowser("chrome").build(); await driver.get("http://google.com"); // should log 5 times, polling every 1 second then times out await driver.wait(myFunc, 5000, "Timed out after 5 seconds", 1000); await driver.close(); })(); ``` Note: if we cast the `driver` to `any` we are able to get it polling correctly: ``` await (driver as any).wait(myFunc, 5000, "Timed out after 5 seconds", 1000); ``` ### Relevant log output ```shell index.ts:11:65 - error TS2554: Expected 1-3 arguments, but got 4. 11 await driver.wait(myFunc, 5000, "Timed out after 5 seconds", 1000); ``` Issue is within `types.d.ts:` ``` wait(condition: WebElementCondition, opt_timeout?: number, opt_message?: string): WebElementPromise; ``` ### Operating System macOS ### Selenium version 4.1.1 ### What are the browser(s) and version(s) where you see this issue? Chrome ### What are the browser driver(s) and version(s) where you see this issue? 97 ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: WebdriverJS FluentWait - ### What happened? WebdriverJS FluentWait works in Javascript as per the example in the [documentation](https://www.selenium.dev/documentation/webdriver/waits/#fluentwait). However in TypeScript we are missing the types for the 4th parameter (the frequency at which it polls). Working JavaScript example (which is polling and including the frequency at which it polls: ```shell const { Builder } = require("selenium-webdriver"); function myFunc() { console.log("hello world"); } (async function example() { let driver = await new Builder().forBrowser("chrome").build(); await driver.get("http://google.com"); // should log 5 times, polling every 1 second then times out await driver.wait(myFunc, 5000, "Timed out after 5 seconds", 1000); await driver.close(); })(); ``` ### How can we reproduce the issue? ```shell import { Builder } from "selenium-webdriver"; function myFunc() { console.log("hello world"); } (async function example() { let driver = await new Builder().forBrowser("chrome").build(); await driver.get("http://google.com"); // should log 5 times, polling every 1 second then times out await driver.wait(myFunc, 5000, "Timed out after 5 seconds", 1000); await driver.close(); })(); ``` Note: if we cast the `driver` to `any` we are able to get it polling correctly: ``` await (driver as any).wait(myFunc, 5000, "Timed out after 5 seconds", 1000); ``` ### Relevant log output ```shell index.ts:11:65 - error TS2554: Expected 1-3 arguments, but got 4. 11 await driver.wait(myFunc, 5000, "Timed out after 5 seconds", 1000); ``` Issue is within `types.d.ts:` ``` wait(condition: WebElementCondition, opt_timeout?: number, opt_message?: string): WebElementPromise; ``` ### Operating System macOS ### Selenium version 4.1.1 ### What are the browser(s) and version(s) where you see this issue? Chrome ### What are the browser driver(s) and version(s) where you see this issue? 97 ### Are you using Selenium Grid? _No response_
defect
webdriverjs fluentwait what happened webdriverjs fluentwait works in javascript as per the example in the however in typescript we are missing the types for the parameter the frequency at which it polls working javascript example which is polling and including the frequency at which it polls shell const builder require selenium webdriver function myfunc console log hello world async function example let driver await new builder forbrowser chrome build await driver get should log times polling every second then times out await driver wait myfunc timed out after seconds await driver close how can we reproduce the issue shell import builder from selenium webdriver function myfunc console log hello world async function example let driver await new builder forbrowser chrome build await driver get should log times polling every second then times out await driver wait myfunc timed out after seconds await driver close note if we cast the driver to any we are able to get it polling correctly await driver as any wait myfunc timed out after seconds relevant log output shell index ts error expected arguments but got await driver wait myfunc timed out after seconds issue is within types d ts wait condition webelementcondition opt timeout number opt message string webelementpromise operating system macos selenium version what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue are you using selenium grid no response
1
356,093
25,176,107,653
IssuesEvent
2022-11-11 09:24:12
jeromepui/pe
https://api.github.com/repos/jeromepui/pe
opened
[User Guide] Did not say all kinds of roles in description
severity.Low type.DocumentationBug
There are multiple roles for the professor such as Coordinator, Lecturer, etc. but these were not mentioned in the description for the command. ![Screenshot 2022-11-11 at 5.23.08 PM.png](https://raw.githubusercontent.com/jeromepui/pe/main/files/62e49ad7-e2e2-431e-932c-da80ff66ade7.png) <!--session: 1668153057787-9e9283b8-9304-4dc3-8744-f7c54d9ba55b--> <!--Version: Web v3.4.4-->
1.0
[User Guide] Did not say all kinds of roles in description - There are multiple roles for the professor such as Coordinator, Lecturer, etc. but these were not mentioned in the description for the command. ![Screenshot 2022-11-11 at 5.23.08 PM.png](https://raw.githubusercontent.com/jeromepui/pe/main/files/62e49ad7-e2e2-431e-932c-da80ff66ade7.png) <!--session: 1668153057787-9e9283b8-9304-4dc3-8744-f7c54d9ba55b--> <!--Version: Web v3.4.4-->
non_defect
did not say all kinds of roles in description there are multiple roles for the professor such as coordinator lecturer etc but these were not mentioned in the description for the command
0
133,664
18,927,300,022
IssuesEvent
2021-11-17 10:53:28
GwaChaePah/front
https://api.github.com/repos/GwaChaePah/front
closed
mobileWeb post menu fix-up- [ji-kim]
design
## 제목 ### fix mobileWeb post menu ## 작업 상세 내용 - [x] change sidebar navigation to topbar fixed navigation - [x] move price info to top as well ## 참고 사항
1.0
mobileWeb post menu fix-up- [ji-kim] - ## 제목 ### fix mobileWeb post menu ## 작업 상세 내용 - [x] change sidebar navigation to topbar fixed navigation - [x] move price info to top as well ## 참고 사항
non_defect
mobileweb post menu fix up 제목 fix mobileweb post menu 작업 상세 내용 change sidebar navigation to topbar fixed navigation move price info to top as well 참고 사항
0
5,205
2,610,183,173
IssuesEvent
2015-02-26 18:58:20
chrsmith/quchuseban
https://api.github.com/repos/chrsmith/quchuseban
opened
剖析脸长色斑怎么办
auto-migrated Priority-Medium Type-Defect
``` 《摘要》 当事业遭受挫折、生活艰难不堪、人士处于低谷时,悲观脆�� �者,要么自暴自弃,一蹶不振,要么低头认输,境况越来越� ��;而乐观豁达、直面人生者,能把平凡的日子过得精彩,能 把沉重的生活变得轻松,能把苦难的体验变得生动,能够不�� �去开辟人生的新境界,享受生活赋予的一切酸甜苦辣,从而� ��正去感受人生的真谛和生命的意义。脸长色斑怎么办, 《客户案例》   我是做软件开发的,很多男生都是做这个的,做我们这�� �的都是离不开电脑的,每天的工作就是对着电脑 ,一忙起来更是没什么时间概念了,有时候一天就能在电脑�� �呆十多个小时,就这样长时间面对电脑屏幕 ,每天都要承受很多的电脑辐射,时间长了脸上就出现了和�� �多斑斑点点。本来以为一个大男人长斑没什 么大不了的,也没在乎那么多,可我发现女朋友不要求我陪�� �了,时间长了,我才重视起来,对着镜子看 看自己,衣服不修边幅,头发乱蓬蓬的,还满脸的斑斑点点�� �要我是个女孩也不愿意,为了我的爱情,我 开始加入祛斑大军了,开始买各种各样的祛斑的化妆品,可�� �了一段时间,却看不到效果。一次在网上跟 一位网友聊天,从她那儿得知一款很好的祛斑产品—「黛芙�� �尔精华液」,她说她就是用的这个把斑去掉的,我听了后特� ��兴奋,马上给她要了联系方式,她给了我一个客服的Q号,�� �先在网上咨询了下客服, 说这个是纯精华的去斑精华液,是外用的,感觉没化妆品那�� �麻烦,感觉不错。便订购了2个周期,使用完 后脸上的斑真的不见了,皮肤也白多了了,当我拿着玫瑰站�� �女朋友面前,我们又和好如初了。 阅读了脸长色斑怎么办,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 脸长色斑怎么办,同时为您分享祛斑小方法 西瓜面膜 西瓜可去油脂,改善皮肤出油状况。 材料:吃剩的西瓜(几片)。 做法:把西瓜的果肉剔除,露出青色的果皮,敷在脸上5-10�� �钟再洗干净。 注意: 敷完脸后记得洗干净脸皮上西瓜留下的甜味,否则可能会吸�� �小蚂蚁来野餐。 南瓜番茄胡萝卜汤 用料:南瓜220克,番茄110克,胡萝卜110克,瘦猪肉100克,生�� �15克,调料适量。 做法:上料洗净后加水煮汤,瓜熟烂即可。 食法:吃菜喝汤,每日1剂。30日为一疗程。 功效:除黑斑,美肌肤。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:01
1.0
剖析脸长色斑怎么办 - ``` 《摘要》 当事业遭受挫折、生活艰难不堪、人士处于低谷时,悲观脆�� �者,要么自暴自弃,一蹶不振,要么低头认输,境况越来越� ��;而乐观豁达、直面人生者,能把平凡的日子过得精彩,能 把沉重的生活变得轻松,能把苦难的体验变得生动,能够不�� �去开辟人生的新境界,享受生活赋予的一切酸甜苦辣,从而� ��正去感受人生的真谛和生命的意义。脸长色斑怎么办, 《客户案例》   我是做软件开发的,很多男生都是做这个的,做我们这�� �的都是离不开电脑的,每天的工作就是对着电脑 ,一忙起来更是没什么时间概念了,有时候一天就能在电脑�� �呆十多个小时,就这样长时间面对电脑屏幕 ,每天都要承受很多的电脑辐射,时间长了脸上就出现了和�� �多斑斑点点。本来以为一个大男人长斑没什 么大不了的,也没在乎那么多,可我发现女朋友不要求我陪�� �了,时间长了,我才重视起来,对着镜子看 看自己,衣服不修边幅,头发乱蓬蓬的,还满脸的斑斑点点�� �要我是个女孩也不愿意,为了我的爱情,我 开始加入祛斑大军了,开始买各种各样的祛斑的化妆品,可�� �了一段时间,却看不到效果。一次在网上跟 一位网友聊天,从她那儿得知一款很好的祛斑产品—「黛芙�� �尔精华液」,她说她就是用的这个把斑去掉的,我听了后特� ��兴奋,马上给她要了联系方式,她给了我一个客服的Q号,�� �先在网上咨询了下客服, 说这个是纯精华的去斑精华液,是外用的,感觉没化妆品那�� �麻烦,感觉不错。便订购了2个周期,使用完 后脸上的斑真的不见了,皮肤也白多了了,当我拿着玫瑰站�� �女朋友面前,我们又和好如初了。 阅读了脸长色斑怎么办,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 脸长色斑怎么办,同时为您分享祛斑小方法 西瓜面膜 西瓜可去油脂,改善皮肤出油状况。 材料:吃剩的西瓜(几片)。 做法:把西瓜的果肉剔除,露出青色的果皮,敷在脸上5-10�� �钟再洗干净。 注意: 敷完脸后记得洗干净脸皮上西瓜留下的甜味,否则可能会吸�� �小蚂蚁来野餐。 南瓜番茄胡萝卜汤 用料:南瓜220克,番茄110克,胡萝卜110克,瘦猪肉100克,生�� �15克,调料适量。 做法:上料洗净后加水煮汤,瓜熟烂即可。 食法:吃菜喝汤,每日1剂。30日为一疗程。 功效:除黑斑,美肌肤。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 3:01
defect
剖析脸长色斑怎么办 《摘要》 当事业遭受挫折、生活艰难不堪、人士处于低谷时,悲观脆�� �者,要么自暴自弃,一蹶不振,要么低头认输,境况越来越� ��;而乐观豁达、直面人生者,能把平凡的日子过得精彩,能 把沉重的生活变得轻松,能把苦难的体验变得生动,能够不�� �去开辟人生的新境界,享受生活赋予的一切酸甜苦辣,从而� ��正去感受人生的真谛和生命的意义。脸长色斑怎么办, 《客户案例》   我是做软件开发的,很多男生都是做这个的,做我们这�� �的都是离不开电脑的,每天的工作就是对着电脑 ,一忙起来更是没什么时间概念了,有时候一天就能在电脑�� �呆十多个小时,就这样长时间面对电脑屏幕 ,每天都要承受很多的电脑辐射,时间长了脸上就出现了和�� �多斑斑点点。本来以为一个大男人长斑没什 么大不了的,也没在乎那么多,可我发现女朋友不要求我陪�� �了,时间长了,我才重视起来,对着镜子看 看自己,衣服不修边幅,头发乱蓬蓬的,还满脸的斑斑点点�� �要我是个女孩也不愿意,为了我的爱情,我 开始加入祛斑大军了,开始买各种各样的祛斑的化妆品,可�� �了一段时间,却看不到效果。一次在网上跟 一位网友聊天,从她那儿得知一款很好的祛斑产品—「黛芙�� �尔精华液」,她说她就是用的这个把斑去掉的,我听了后特� ��兴奋,马上给她要了联系方式,她给了我一个客服的q号,�� �先在网上咨询了下客服, 说这个是纯精华的去斑精华液,是外用的,感觉没化妆品那�� �麻烦,感觉不错。 ,使用完 后脸上的斑真的不见了,皮肤也白多了了,当我拿着玫瑰站�� �女朋友面前,我们又和好如初了。 阅读了脸长色斑怎么办,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》    黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗   答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来    ,服用黛芙薇尔美白,会伤身体吗 有副作用吗   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖    ,去除黄褐斑之后,会反弹吗   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗    ,你们的价格有点贵,能不能便宜一点   答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗    ,我适合用黛芙薇尔精华液吗   答:黛芙薇尔适用人群:    、生理紊乱引起的黄褐斑人群    、生育引起的妊娠斑人群    、年纪增长引起的老年斑人群    、化妆品色素沉积、辐射斑人群    、长期日照引起的日晒斑人群    、肌肤暗淡急需美白的人群 《祛斑小方法》 脸长色斑怎么办,同时为您分享祛斑小方法 西瓜面膜 西瓜可去油脂,改善皮肤出油状况。 材料:吃剩的西瓜(几片)。 做法:把西瓜的果肉剔除,露出青色的果皮, - �� �钟再洗干净。 注意: 敷完脸后记得洗干净脸皮上西瓜留下的甜味,否则可能会吸�� �小蚂蚁来野餐。 南瓜番茄胡萝卜汤 用料: , , , ,生�� � ,调料适量。 做法:上料洗净后加水煮汤,瓜熟烂即可。 食法:吃菜喝汤, 。 。 功效:除黑斑,美肌肤。 original issue reported on code google com by additive gmail com on jul at
1
103,385
12,892,254,752
IssuesEvent
2020-07-13 19:16:04
chavarera/python-mini-projects
https://api.github.com/repos/chavarera/python-mini-projects
closed
create a simple stopwatch
GUI design
**Problem Statement** Create a simple stopwatch using tkinter or PYQT5 **Operations** start/pause/stop timer
1.0
create a simple stopwatch - **Problem Statement** Create a simple stopwatch using tkinter or PYQT5 **Operations** start/pause/stop timer
non_defect
create a simple stopwatch problem statement create a simple stopwatch using tkinter or operations start pause stop timer
0
63,757
17,930,053,001
IssuesEvent
2021-09-10 08:02:26
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
p-checkbox causes layout issues when used inside an overflow: auto div
defect
If you have a PrimeNG PRO Support subscription please post your issue at; https://pro.primefaces.org where our team will respond within 4 business hours. If you do not have a PrimeNG PRO Support subscription, fill-in the report below. Please note that your issue will be added to the waiting list of community issues and will be reviewed on a first-come first-serve basis, as a result, the support team is unable to guarantee a specific schedule on when it will be reviewed. Thank you for your understanding. Current Queue Time for Review Without PRO Support: ~8-12 weeks. With PRO Support: 1 hour **I'm submitting a ...** (check one with "x") ``` [X] bug report => Search github for a similar issue or PR before submitting [ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap [ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35 ``` **Plunkr Case (Bug Reports)** Please demonstrate your case at stackblitz by using the issue template below. Issues without a test case have much less possibility to be reviewd in detail and assisted. https://stackblitz.com/edit/primeng-checkbox-flex-issue **Current behavior** <!-- Describe how the bug manifests. --> When the checkbox is used in a div with overflow:auto and requires scrolling in the div to make the checkbox visible, extra space is added at the bottom of the page. When the checkbox is clicked, the page jumps. **Expected behavior** No extra space, and page should not scroll on interaction with the checkbox. **Minimal reproduction of the problem with instructions** <!-- See link above. --> **What is the motivation / use case for changing the behavior?** Checkbox does not work if using overflow: auto to create scrollable panels. * **Angular version:** 11.x (latest) * **PrimeNG version:** 11.x * **Browser:** Chrome 87, Edge 87 <!-- All browsers where this could be reproduced --> * **Language:** TS 4.03
1.0
p-checkbox causes layout issues when used inside an overflow: auto div - If you have a PrimeNG PRO Support subscription please post your issue at; https://pro.primefaces.org where our team will respond within 4 business hours. If you do not have a PrimeNG PRO Support subscription, fill-in the report below. Please note that your issue will be added to the waiting list of community issues and will be reviewed on a first-come first-serve basis, as a result, the support team is unable to guarantee a specific schedule on when it will be reviewed. Thank you for your understanding. Current Queue Time for Review Without PRO Support: ~8-12 weeks. With PRO Support: 1 hour **I'm submitting a ...** (check one with "x") ``` [X] bug report => Search github for a similar issue or PR before submitting [ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap [ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35 ``` **Plunkr Case (Bug Reports)** Please demonstrate your case at stackblitz by using the issue template below. Issues without a test case have much less possibility to be reviewd in detail and assisted. https://stackblitz.com/edit/primeng-checkbox-flex-issue **Current behavior** <!-- Describe how the bug manifests. --> When the checkbox is used in a div with overflow:auto and requires scrolling in the div to make the checkbox visible, extra space is added at the bottom of the page. When the checkbox is clicked, the page jumps. **Expected behavior** No extra space, and page should not scroll on interaction with the checkbox. **Minimal reproduction of the problem with instructions** <!-- See link above. --> **What is the motivation / use case for changing the behavior?** Checkbox does not work if using overflow: auto to create scrollable panels. * **Angular version:** 11.x (latest) * **PrimeNG version:** 11.x * **Browser:** Chrome 87, Edge 87 <!-- All browsers where this could be reproduced --> * **Language:** TS 4.03
defect
p checkbox causes layout issues when used inside an overflow auto div if you have a primeng pro support subscription please post your issue at where our team will respond within business hours if you do not have a primeng pro support subscription fill in the report below please note that your issue will be added to the waiting list of community issues and will be reviewed on a first come first serve basis as a result the support team is unable to guarantee a specific schedule on when it will be reviewed thank you for your understanding current queue time for review without pro support weeks with pro support hour i m submitting a check one with x bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports please demonstrate your case at stackblitz by using the issue template below issues without a test case have much less possibility to be reviewd in detail and assisted current behavior when the checkbox is used in a div with overflow auto and requires scrolling in the div to make the checkbox visible extra space is added at the bottom of the page when the checkbox is clicked the page jumps expected behavior no extra space and page should not scroll on interaction with the checkbox minimal reproduction of the problem with instructions see link above what is the motivation use case for changing the behavior checkbox does not work if using overflow auto to create scrollable panels angular version x latest primeng version x browser chrome edge language ts
1
94,663
8,507,873,892
IssuesEvent
2018-10-30 20:18:56
zowe/api-layer
https://api.github.com/repos/zowe/api-layer
closed
[CLOSED] Setup idea local params
passed CA tests under CA testing
<a href="https://github.com/vojoup"><img src="https://avatars2.githubusercontent.com/u/24936972?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [vojoup](https://github.com/vojoup)** _Thursday Aug 02, 2018 at 09:05 GMT_ _Originally opened as https://github.com/gizafoundation/api-layer/pull/22_ ---- This PR should make it easy to use local config for services while using 'Run Dashboard' in IntelliJ ---- _**[vojoup](https://github.com/vojoup)** included the following code: https://github.com/gizafoundation/api-layer/pull/22/commits_
2.0
[CLOSED] Setup idea local params - <a href="https://github.com/vojoup"><img src="https://avatars2.githubusercontent.com/u/24936972?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [vojoup](https://github.com/vojoup)** _Thursday Aug 02, 2018 at 09:05 GMT_ _Originally opened as https://github.com/gizafoundation/api-layer/pull/22_ ---- This PR should make it easy to use local config for services while using 'Run Dashboard' in IntelliJ ---- _**[vojoup](https://github.com/vojoup)** included the following code: https://github.com/gizafoundation/api-layer/pull/22/commits_
non_defect
setup idea local params issue by thursday aug at gmt originally opened as this pr should make it easy to use local config for services while using run dashboard in intellij included the following code
0
34,659
7,458,528,222
IssuesEvent
2018-03-30 10:49:10
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
dev-mini jaoks disaini- ja funktsionaalsuse muudatused
C: AVAR P: highest R: fixed T: defect
**Reported by mehis muldma on 12 Jan 2015 10:15 UTC** '''Objekt''' Rakendusse teha dev-mini keskkonna jaoks disaini- ja funktsionaalsuse muudatusi. '''TODO''' 1) Väljendada suurelt ja punaselt päises, et tegu on väikse baasiga 2) Impordi ajaks panna sait lukku ja näidata kirjakest, et "Import käib" 3) indekseerimise ajaks teha sait lahti, päises näidata kirjakest, et "Indekseerimine käib ja otsing võib kelbast panna"
1.0
dev-mini jaoks disaini- ja funktsionaalsuse muudatused - **Reported by mehis muldma on 12 Jan 2015 10:15 UTC** '''Objekt''' Rakendusse teha dev-mini keskkonna jaoks disaini- ja funktsionaalsuse muudatusi. '''TODO''' 1) Väljendada suurelt ja punaselt päises, et tegu on väikse baasiga 2) Impordi ajaks panna sait lukku ja näidata kirjakest, et "Import käib" 3) indekseerimise ajaks teha sait lahti, päises näidata kirjakest, et "Indekseerimine käib ja otsing võib kelbast panna"
defect
dev mini jaoks disaini ja funktsionaalsuse muudatused reported by mehis muldma on jan utc objekt rakendusse teha dev mini keskkonna jaoks disaini ja funktsionaalsuse muudatusi todo väljendada suurelt ja punaselt päises et tegu on väikse baasiga impordi ajaks panna sait lukku ja näidata kirjakest et import käib indekseerimise ajaks teha sait lahti päises näidata kirjakest et indekseerimine käib ja otsing võib kelbast panna
1
320,797
23,825,929,720
IssuesEvent
2022-09-05 14:51:47
osism/issues
https://api.github.com/repos/osism/issues
closed
Neutron Upgrade failing due to missing Neutron SSH key
bug documentation
Neutron Upgrade Failed, because the Neutron SSH key was undefined. We had to create the key and the ansible var/dict ourselves. Traceback (most recent call last): File "/usr/local/lib/python3.8/dist-packages/ansible/template/__init__.py", line 1160, in do_template res = j2_concat(rf) File "<template>", line 12, in root File "/usr/local/lib/python3.8/dist-packages/jinja2/runtime.py", line 852, in _fail_with_undefined_error raise self._undefined_exception(self._undefined_message) jinja2.exceptions.UndefinedError: 'neutron_ssh_key' is undefined sudo -E kolla_set_configs INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json INFO:__main__:Validating config file INFO:__main__:Kolla config strategy set to: COPY_ALWAYS INFO:__main__:Copying service configuration files INFO:__main__:Copying /var/lib/kolla/config_files/neutron.conf to /etc/neutron/neutron.conf INFO:__main__:Setting permission for /etc/neutron/neutron.conf INFO:__main__:Copying /var/lib/kolla/config_files/neutron_vpnaas.conf to /etc/neutron/neutron_vpnaas.conf INFO:__main__:Setting permission for /etc/neutron/neutron_vpnaas.conf INFO:__main__:Copying /var/lib/kolla/config_files/ml2_conf.ini to /etc/neutron/plugins/ml2/ml2_conf.ini INFO:__main__:Setting permission for /etc/neutron/plugins/ml2/ml2_conf.ini ERROR:__main__:MissingRequiredSource: /var/lib/kolla/config_files/id_rsa file is not found
1.0
Neutron Upgrade failing due to missing Neutron SSH key - Neutron Upgrade Failed, because the Neutron SSH key was undefined. We had to create the key and the ansible var/dict ourselves. Traceback (most recent call last): File "/usr/local/lib/python3.8/dist-packages/ansible/template/__init__.py", line 1160, in do_template res = j2_concat(rf) File "<template>", line 12, in root File "/usr/local/lib/python3.8/dist-packages/jinja2/runtime.py", line 852, in _fail_with_undefined_error raise self._undefined_exception(self._undefined_message) jinja2.exceptions.UndefinedError: 'neutron_ssh_key' is undefined sudo -E kolla_set_configs INFO:__main__:Loading config file at /var/lib/kolla/config_files/config.json INFO:__main__:Validating config file INFO:__main__:Kolla config strategy set to: COPY_ALWAYS INFO:__main__:Copying service configuration files INFO:__main__:Copying /var/lib/kolla/config_files/neutron.conf to /etc/neutron/neutron.conf INFO:__main__:Setting permission for /etc/neutron/neutron.conf INFO:__main__:Copying /var/lib/kolla/config_files/neutron_vpnaas.conf to /etc/neutron/neutron_vpnaas.conf INFO:__main__:Setting permission for /etc/neutron/neutron_vpnaas.conf INFO:__main__:Copying /var/lib/kolla/config_files/ml2_conf.ini to /etc/neutron/plugins/ml2/ml2_conf.ini INFO:__main__:Setting permission for /etc/neutron/plugins/ml2/ml2_conf.ini ERROR:__main__:MissingRequiredSource: /var/lib/kolla/config_files/id_rsa file is not found
non_defect
neutron upgrade failing due to missing neutron ssh key neutron upgrade failed because the neutron ssh key was undefined we had to create the key and the ansible var dict ourselves traceback most recent call last file usr local lib dist packages ansible template init py line in do template res concat rf file line in root file usr local lib dist packages runtime py line in fail with undefined error raise self undefined exception self undefined message exceptions undefinederror neutron ssh key is undefined sudo e kolla set configs info main loading config file at var lib kolla config files config json info main validating config file info main kolla config strategy set to copy always info main copying service configuration files info main copying var lib kolla config files neutron conf to etc neutron neutron conf info main setting permission for etc neutron neutron conf info main copying var lib kolla config files neutron vpnaas conf to etc neutron neutron vpnaas conf info main setting permission for etc neutron neutron vpnaas conf info main copying var lib kolla config files conf ini to etc neutron plugins conf ini info main setting permission for etc neutron plugins conf ini error main missingrequiredsource var lib kolla config files id rsa file is not found
0
505,002
14,625,780,446
IssuesEvent
2020-12-23 09:10:56
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.netflix.com - video or audio doesn't play
browser-firefox engine-gecko ml-needsdiagnosis-false priority-critical
<!-- @browser: Firefox 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile VR; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64176 --> **URL**: https://www.netflix.com/browse **Browser / Version**: Firefox 81.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: There is no video **Steps to Reproduce**: inop <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.netflix.com - video or audio doesn't play - <!-- @browser: Firefox 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile VR; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64176 --> **URL**: https://www.netflix.com/browse **Browser / Version**: Firefox 81.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: There is no video **Steps to Reproduce**: inop <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_defect
video or audio doesn t play url browser version firefox operating system android tested another browser no problem type video or audio doesn t play description there is no video steps to reproduce inop browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❤️
0
213,209
16,515,775,573
IssuesEvent
2021-05-26 09:35:31
AlexanderNorup/SemesterProjekt-2
https://api.github.com/repos/AlexanderNorup/SemesterProjekt-2
closed
PLPGSQL functionTypeByName & categoryByName
documentation
Jeg smider lige funktionerne til databasen her, så vi ikke glemmer hvor de er. ## PLPGSQL Functions functionTypeByName: ``` CREATE OR REPLACE FUNCTION functionTypeByName(nameToFind varchar) RETURNS integer AS $$ DECLARE functionTypeId integer := 42; -- Uknown BEGIN SELECT FT.id INTO functionTypeId FROM function_types FT WHERE FT.name = nameToFind; RETURN functionTypeId; END; $$ LANGUAGE plpgsql; ``` categoryByName: ``` CREATE OR REPLACE FUNCTION categoryByName(nameToFind varchar) RETURNS integer AS $$ DECLARE categoryId integer := 1; -- Uknown BEGIN SELECT CA.id INTO categoryId FROM categories CA WHERE CA.name = nameToFind; RETURN categoryId; END; $$ LANGUAGE plpgsql;
1.0
PLPGSQL functionTypeByName & categoryByName - Jeg smider lige funktionerne til databasen her, så vi ikke glemmer hvor de er. ## PLPGSQL Functions functionTypeByName: ``` CREATE OR REPLACE FUNCTION functionTypeByName(nameToFind varchar) RETURNS integer AS $$ DECLARE functionTypeId integer := 42; -- Uknown BEGIN SELECT FT.id INTO functionTypeId FROM function_types FT WHERE FT.name = nameToFind; RETURN functionTypeId; END; $$ LANGUAGE plpgsql; ``` categoryByName: ``` CREATE OR REPLACE FUNCTION categoryByName(nameToFind varchar) RETURNS integer AS $$ DECLARE categoryId integer := 1; -- Uknown BEGIN SELECT CA.id INTO categoryId FROM categories CA WHERE CA.name = nameToFind; RETURN categoryId; END; $$ LANGUAGE plpgsql;
non_defect
plpgsql functiontypebyname categorybyname jeg smider lige funktionerne til databasen her så vi ikke glemmer hvor de er plpgsql functions functiontypebyname create or replace function functiontypebyname nametofind varchar returns integer as declare functiontypeid integer uknown begin select ft id into functiontypeid from function types ft where ft name nametofind return functiontypeid end language plpgsql categorybyname create or replace function categorybyname nametofind varchar returns integer as declare categoryid integer uknown begin select ca id into categoryid from categories ca where ca name nametofind return categoryid end language plpgsql
0
50,361
12,503,084,741
IssuesEvent
2020-06-02 06:27:20
microsoft/fluentui
https://api.github.com/repos/microsoft/fluentui
closed
unable to build
Area: Build System Status: In PR Type: Bug :bug:
Repro steps: $ git clone <repo> $ yarn $ yarn build Expected: built repo Actual: ``` @fluentui/ability-attributes: [11:32:40] Requiring external module @uifabric/build/babel/register @fluentui/ability-attributes: internal/modules/cjs/loader.js:492 @fluentui/ability-attributes: throw new ERR_PACKAGE_PATH_NOT_EXPORTED(basePath, mappingKey); @fluentui/ability-attributes: ^ @fluentui/ability-attributes: Error [ERR_PACKAGE_PATH_NOT_EXPORTED]: No "exports" main resolved in /Users/jdh/projects/office-ui-fabric-react/node_modules/@babel/helper-compilation-targets/package.json @fluentui/ability-attributes: at applyExports (internal/modules/cjs/loader.js:492:9) @fluentui/ability-attributes: at resolveExports (internal/modules/cjs/loader.js:508:23) @fluentui/ability-attributes: at Function.Module._findPath (internal/modules/cjs/loader.js:632:31) @fluentui/ability-attributes: at Function.Module._resolveFilename (internal/modules/cjs/loader.js:1001:27) @fluentui/ability-attributes: at Function.Module._load (internal/modules/cjs/loader.js:884:27) @fluentui/ability-attributes: at Module.require (internal/modules/cjs/loader.js:1074:19) @fluentui/ability-attributes: at require (internal/modules/cjs/helpers.js:72:18) @fluentui/ability-attributes: at Object.<anonymous> (/Users/jdh/projects/office-ui-fabric-react/node_modules/@babel/preset-env/lib/debug.js:8:33) @fluentui/ability-attributes: at Module._compile (internal/modules/cjs/loader.js:1185:30) @fluentui/ability-attributes: at Module._compile (/Users/jdh/projects/office-ui-fabric-react/node_modules/pirates/lib/index.js:99:24) { @fluentui/ability-attributes: code: 'ERR_PACKAGE_PATH_NOT_EXPORTED' @fluentui/ability-attributes: } @fluentui/ability-attributes: error Command failed with exit code 1. @fluentui/ability-attributes: info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command. ```
1.0
unable to build - Repro steps: $ git clone <repo> $ yarn $ yarn build Expected: built repo Actual: ``` @fluentui/ability-attributes: [11:32:40] Requiring external module @uifabric/build/babel/register @fluentui/ability-attributes: internal/modules/cjs/loader.js:492 @fluentui/ability-attributes: throw new ERR_PACKAGE_PATH_NOT_EXPORTED(basePath, mappingKey); @fluentui/ability-attributes: ^ @fluentui/ability-attributes: Error [ERR_PACKAGE_PATH_NOT_EXPORTED]: No "exports" main resolved in /Users/jdh/projects/office-ui-fabric-react/node_modules/@babel/helper-compilation-targets/package.json @fluentui/ability-attributes: at applyExports (internal/modules/cjs/loader.js:492:9) @fluentui/ability-attributes: at resolveExports (internal/modules/cjs/loader.js:508:23) @fluentui/ability-attributes: at Function.Module._findPath (internal/modules/cjs/loader.js:632:31) @fluentui/ability-attributes: at Function.Module._resolveFilename (internal/modules/cjs/loader.js:1001:27) @fluentui/ability-attributes: at Function.Module._load (internal/modules/cjs/loader.js:884:27) @fluentui/ability-attributes: at Module.require (internal/modules/cjs/loader.js:1074:19) @fluentui/ability-attributes: at require (internal/modules/cjs/helpers.js:72:18) @fluentui/ability-attributes: at Object.<anonymous> (/Users/jdh/projects/office-ui-fabric-react/node_modules/@babel/preset-env/lib/debug.js:8:33) @fluentui/ability-attributes: at Module._compile (internal/modules/cjs/loader.js:1185:30) @fluentui/ability-attributes: at Module._compile (/Users/jdh/projects/office-ui-fabric-react/node_modules/pirates/lib/index.js:99:24) { @fluentui/ability-attributes: code: 'ERR_PACKAGE_PATH_NOT_EXPORTED' @fluentui/ability-attributes: } @fluentui/ability-attributes: error Command failed with exit code 1. @fluentui/ability-attributes: info Visit https://yarnpkg.com/en/docs/cli/run for documentation about this command. ```
non_defect
unable to build repro steps git clone yarn yarn build expected built repo actual fluentui ability attributes requiring external module uifabric build babel register fluentui ability attributes internal modules cjs loader js fluentui ability attributes throw new err package path not exported basepath mappingkey fluentui ability attributes fluentui ability attributes error no exports main resolved in users jdh projects office ui fabric react node modules babel helper compilation targets package json fluentui ability attributes at applyexports internal modules cjs loader js fluentui ability attributes at resolveexports internal modules cjs loader js fluentui ability attributes at function module findpath internal modules cjs loader js fluentui ability attributes at function module resolvefilename internal modules cjs loader js fluentui ability attributes at function module load internal modules cjs loader js fluentui ability attributes at module require internal modules cjs loader js fluentui ability attributes at require internal modules cjs helpers js fluentui ability attributes at object users jdh projects office ui fabric react node modules babel preset env lib debug js fluentui ability attributes at module compile internal modules cjs loader js fluentui ability attributes at module compile users jdh projects office ui fabric react node modules pirates lib index js fluentui ability attributes code err package path not exported fluentui ability attributes fluentui ability attributes error command failed with exit code fluentui ability attributes info visit for documentation about this command
0
153,522
13,508,363,042
IssuesEvent
2020-09-14 07:36:12
kyma-project/control-plane
https://api.github.com/repos/kyma-project/control-plane
closed
Create and align existing documentation about upgrade
area/control-plane documentation
<!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** Create and align existing documentation about the upgrade: - how to make an upgrade via KEB tutorial - any specific info about upgrade process - info about new auth scopes - update existing documentation, for example doc about steps epic: #118
1.0
Create and align existing documentation about upgrade - <!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** Create and align existing documentation about the upgrade: - how to make an upgrade via KEB tutorial - any specific info about upgrade process - info about new auth scopes - update existing documentation, for example doc about steps epic: #118
non_defect
create and align existing documentation about upgrade thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description create and align existing documentation about the upgrade how to make an upgrade via keb tutorial any specific info about upgrade process info about new auth scopes update existing documentation for example doc about steps epic
0
175,908
14,544,857,340
IssuesEvent
2020-12-15 18:47:25
bounswe/bounswe2020group4
https://api.github.com/repos/bounswe/bounswe2020group4
opened
(BKND) Api documentation
Backend Effort: Medium Priority: High Status: In-Progress Type: Documentation
An api documentation should be created for the endpoints. Deadline: 22/12/20
1.0
(BKND) Api documentation - An api documentation should be created for the endpoints. Deadline: 22/12/20
non_defect
bknd api documentation an api documentation should be created for the endpoints deadline
0
86,068
10,473,149,832
IssuesEvent
2019-09-23 11:58:35
Denmads/3-semesterprojekt-fitness
https://api.github.com/repos/Denmads/3-semesterprojekt-fitness
closed
Create Project Foundation Structure dokument
documentation
start op the Project Foundation Structure dokument
1.0
Create Project Foundation Structure dokument - start op the Project Foundation Structure dokument
non_defect
create project foundation structure dokument start op the project foundation structure dokument
0
76,648
26,535,450,434
IssuesEvent
2023-01-19 15:20:53
SasView/sasview
https://api.github.com/repos/SasView/sasview
closed
5.x: Parameter limits in the Fit Page are not respected.
Defect
This defect was originally reported by User AdrianR using 5.0.3 on 28 Dec 2020 but I cannot find a corresponding issue (@krzywon might...). However, User GregS has now reported the same issue using 5.0.4 which suggests no fix was implemented. So am logging the issue here. Adrian said: >I have been using SasView 5.03 under Windows 10 and noticed an anomaly when fitting and applying constraints (simply minimum or maximum values of parameters) that these are not always applied after some iterations of fitting. The attached screen shot shows 'thickness1' exceeding the maximum but my impression is that this can happen with various models/parameters. I am not certain from the various web pages whether this corresponds to other comments for which tickets have been generated or whether this may arise from unusual values. <img width="886" alt="ScreenShot_Parameters" src="https://user-images.githubusercontent.com/10938679/149184040-33ab66eb-95dc-494a-8f50-88a87c871cf2.png"> Greg says: >I was again trying to fit SESANS data, but the fit value of some SLDs went beyond my pre-defined limits. For models with fewer parameters, it seemed fine, but for this highly parameterised model, it didn’t restrict the fit to these limits. <img width="1792" alt="Screenshot 2021-12-09 at 12 16 55" src="https://user-images.githubusercontent.com/10938679/149184171-9a4d5a2d-3553-4a94-9c57-687bdffb8c2f.png">
1.0
5.x: Parameter limits in the Fit Page are not respected. - This defect was originally reported by User AdrianR using 5.0.3 on 28 Dec 2020 but I cannot find a corresponding issue (@krzywon might...). However, User GregS has now reported the same issue using 5.0.4 which suggests no fix was implemented. So am logging the issue here. Adrian said: >I have been using SasView 5.03 under Windows 10 and noticed an anomaly when fitting and applying constraints (simply minimum or maximum values of parameters) that these are not always applied after some iterations of fitting. The attached screen shot shows 'thickness1' exceeding the maximum but my impression is that this can happen with various models/parameters. I am not certain from the various web pages whether this corresponds to other comments for which tickets have been generated or whether this may arise from unusual values. <img width="886" alt="ScreenShot_Parameters" src="https://user-images.githubusercontent.com/10938679/149184040-33ab66eb-95dc-494a-8f50-88a87c871cf2.png"> Greg says: >I was again trying to fit SESANS data, but the fit value of some SLDs went beyond my pre-defined limits. For models with fewer parameters, it seemed fine, but for this highly parameterised model, it didn’t restrict the fit to these limits. <img width="1792" alt="Screenshot 2021-12-09 at 12 16 55" src="https://user-images.githubusercontent.com/10938679/149184171-9a4d5a2d-3553-4a94-9c57-687bdffb8c2f.png">
defect
x parameter limits in the fit page are not respected this defect was originally reported by user adrianr using on dec but i cannot find a corresponding issue krzywon might however user gregs has now reported the same issue using which suggests no fix was implemented so am logging the issue here adrian said i have been using sasview under windows and noticed an anomaly when fitting and applying constraints simply minimum or maximum values of parameters that these are not always applied after some iterations of fitting the attached screen shot shows exceeding the maximum but my impression is that this can happen with various models parameters i am not certain from the various web pages whether this corresponds to other comments for which tickets have been generated or whether this may arise from unusual values img width alt screenshot parameters src greg says i was again trying to fit sesans data but the fit value of some slds went beyond my pre defined limits for models with fewer parameters it seemed fine but for this highly parameterised model it didn’t restrict the fit to these limits img width alt screenshot at src
1
132,453
18,722,817,426
IssuesEvent
2021-11-03 13:37:57
CDCgov/prime-reportstream
https://api.github.com/repos/CDCgov/prime-reportstream
opened
Research & Design - Organizations Can Request Support Online
design research experience
As the Experience Team, we need some research and design for the feature to allow Organizations to request support online as outlined in this Epic [https://app.zenhub.com/workspaces/experience-607d9d5e68b95200150fec37/issues/cdcgov/prime-reportstream/2906](url).
1.0
Research & Design - Organizations Can Request Support Online - As the Experience Team, we need some research and design for the feature to allow Organizations to request support online as outlined in this Epic [https://app.zenhub.com/workspaces/experience-607d9d5e68b95200150fec37/issues/cdcgov/prime-reportstream/2906](url).
non_defect
research design organizations can request support online as the experience team we need some research and design for the feature to allow organizations to request support online as outlined in this epic url
0
411,731
12,031,152,786
IssuesEvent
2020-04-13 09:01:27
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Ecopedia Item linking is overly sensitive and has a weird broken version
Priority: Low Status: Fixed Week Task
![image](https://user-images.githubusercontent.com/14983165/74691895-93d0c400-5199-11ea-8e3e-19addfeca9f9.png) Ideally we can make it match the sensitivity level in admin commands, so you can just do [coppergolditem] and have it correctly link the item with the friendly name.
1.0
Ecopedia Item linking is overly sensitive and has a weird broken version - ![image](https://user-images.githubusercontent.com/14983165/74691895-93d0c400-5199-11ea-8e3e-19addfeca9f9.png) Ideally we can make it match the sensitivity level in admin commands, so you can just do [coppergolditem] and have it correctly link the item with the friendly name.
non_defect
ecopedia item linking is overly sensitive and has a weird broken version ideally we can make it match the sensitivity level in admin commands so you can just do and have it correctly link the item with the friendly name
0
24,295
4,074,882,028
IssuesEvent
2016-05-28 19:36:57
vertigo17/Cerberus
https://api.github.com/repos/vertigo17/Cerberus
closed
[TestCase] Countries checked by default when property is created
enhancement Perim : GUITest
In the TestCase,jsp page. when a property is created, all countries should be checked by default, unless there is a property with the same name already defined, in which case no countries will be checked.
1.0
[TestCase] Countries checked by default when property is created - In the TestCase,jsp page. when a property is created, all countries should be checked by default, unless there is a property with the same name already defined, in which case no countries will be checked.
non_defect
countries checked by default when property is created in the testcase jsp page when a property is created all countries should be checked by default unless there is a property with the same name already defined in which case no countries will be checked
0
529,611
15,392,367,515
IssuesEvent
2021-03-03 15:33:18
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
CustomStats with numeric name do not work
Category: Laws Category: Refactoring Priority: Low Staging Type: Bug
While debugging https://github.com/StrangeLoopGames/Eco/pull/7948, I found out that attempting to register a custom stat with all numeric name will fail to setup database. It can happen with any name starting with a number and can even be possible if the name contains number anywhere. It's related with StatInfo.ShortName generation that uses `ShortNameMap`, that uses `StringUtils.ShortName`. ``` Error inserting custom stat 555 into databaseLiteDB.LiteException: Invalid collection name '55': Use only [a-Z$_] at LiteDB.Engine.CollectionService.CheckName(String name, HeaderPage header) at LiteDB.Engine.CollectionService.Add(String name, CollectionPage& collectionPage) at LiteDB.Engine.CollectionService.Get(String name, Boolean addIfNotExists, CollectionPage& collectionPage) at LiteDB.Engine.Snapshot..ctor(LockMode mode, String collectionName, HeaderPage header, UInt32 transactionID, TransactionPages transPages, LockService locker, WalIndexService walIndex, DiskReader reader, Boolean addIfNotExists) at LiteDB.Engine.TransactionService.<CreateSnapshot>g__create|42_0(<>c__DisplayClass42_0& ) at LiteDB.Engine.TransactionService.CreateSnapshot(LockMode mode, String collection, Boolean addIfNotExists) at LiteDB.Engine.LiteEngine.<>c__DisplayClass7_0.<Insert>b__0(TransactionService transaction) at LiteDB.Engine.LiteEngine.AutoTransaction[T](Func`2 fn) at LiteDB.Engine.LiteEngine.Insert(String collection, IEnumerable`1 docs, BsonAutoId autoId) at LiteDB.LiteCollection`1.Insert(T entity) at Eco.Stats.Stats.RecordCustom(String statName, Int32 time, Single count, Int32 userId, Nullable`1 value) in C:\Eco\Server\Eco.Stats\Stats.cs:line 99 ```
1.0
CustomStats with numeric name do not work - While debugging https://github.com/StrangeLoopGames/Eco/pull/7948, I found out that attempting to register a custom stat with all numeric name will fail to setup database. It can happen with any name starting with a number and can even be possible if the name contains number anywhere. It's related with StatInfo.ShortName generation that uses `ShortNameMap`, that uses `StringUtils.ShortName`. ``` Error inserting custom stat 555 into databaseLiteDB.LiteException: Invalid collection name '55': Use only [a-Z$_] at LiteDB.Engine.CollectionService.CheckName(String name, HeaderPage header) at LiteDB.Engine.CollectionService.Add(String name, CollectionPage& collectionPage) at LiteDB.Engine.CollectionService.Get(String name, Boolean addIfNotExists, CollectionPage& collectionPage) at LiteDB.Engine.Snapshot..ctor(LockMode mode, String collectionName, HeaderPage header, UInt32 transactionID, TransactionPages transPages, LockService locker, WalIndexService walIndex, DiskReader reader, Boolean addIfNotExists) at LiteDB.Engine.TransactionService.<CreateSnapshot>g__create|42_0(<>c__DisplayClass42_0& ) at LiteDB.Engine.TransactionService.CreateSnapshot(LockMode mode, String collection, Boolean addIfNotExists) at LiteDB.Engine.LiteEngine.<>c__DisplayClass7_0.<Insert>b__0(TransactionService transaction) at LiteDB.Engine.LiteEngine.AutoTransaction[T](Func`2 fn) at LiteDB.Engine.LiteEngine.Insert(String collection, IEnumerable`1 docs, BsonAutoId autoId) at LiteDB.LiteCollection`1.Insert(T entity) at Eco.Stats.Stats.RecordCustom(String statName, Int32 time, Single count, Int32 userId, Nullable`1 value) in C:\Eco\Server\Eco.Stats\Stats.cs:line 99 ```
non_defect
customstats with numeric name do not work while debugging i found out that attempting to register a custom stat with all numeric name will fail to setup database it can happen with any name starting with a number and can even be possible if the name contains number anywhere it s related with statinfo shortname generation that uses shortnamemap that uses stringutils shortname error inserting custom stat into databaselitedb liteexception invalid collection name use only at litedb engine collectionservice checkname string name headerpage header at litedb engine collectionservice add string name collectionpage collectionpage at litedb engine collectionservice get string name boolean addifnotexists collectionpage collectionpage at litedb engine snapshot ctor lockmode mode string collectionname headerpage header transactionid transactionpages transpages lockservice locker walindexservice walindex diskreader reader boolean addifnotexists at litedb engine transactionservice g create c at litedb engine transactionservice createsnapshot lockmode mode string collection boolean addifnotexists at litedb engine liteengine c b transactionservice transaction at litedb engine liteengine autotransaction func fn at litedb engine liteengine insert string collection ienumerable docs bsonautoid autoid at litedb litecollection insert t entity at eco stats stats recordcustom string statname time single count userid nullable value in c eco server eco stats stats cs line
0
160,371
12,509,192,610
IssuesEvent
2020-06-02 16:44:01
rancher/k3s
https://api.github.com/repos/rancher/k3s
closed
Addressing K8S CVE - CVE 2020-10749 and CVE-2020-8555
[zube]: To Test
We are working to patch k3s to address the following CVEs: - [CVE-2020-10749](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10749) - IPv4 only clusters susceptible to MitM attacks via IPv6 rogue router advertisements - [CVE-2020-8555](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8555): kube-controller-manager SSRF CVE-2020-10749 is being addressed by upgrading the version of the https://github.com/containernetworking/plugins library used by k3s to v0.8.6. See the following PRs for reference: - 1.18 - https://github.com/rancher/k3s/pull/1839 - 1.17 - https://github.com/rancher/k3s/pull/1838 - 1.16 - https://github.com/rancher/k3s/pull/1837 CVE-2020-8555 was patched in the following versions of Kubernetes: v1.18.1+, v1.17.5+, v1.16.9+. Thus, this vulnerability is already patched in the K3s releases: - v1.18.2+k3s1 - v1.17.5+k3s1 - v1.16.9+k3s1 To ensure you have the latest stable version of Kubernetes for each release when upgrading to address these CVEs, we'll be releasing the following: - v1.18.3+k3s1 - v1.17.6+k3s1 - v1.16.10+k3s1 These releases will contain patches for both CVEs.
1.0
Addressing K8S CVE - CVE 2020-10749 and CVE-2020-8555 - We are working to patch k3s to address the following CVEs: - [CVE-2020-10749](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-10749) - IPv4 only clusters susceptible to MitM attacks via IPv6 rogue router advertisements - [CVE-2020-8555](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8555): kube-controller-manager SSRF CVE-2020-10749 is being addressed by upgrading the version of the https://github.com/containernetworking/plugins library used by k3s to v0.8.6. See the following PRs for reference: - 1.18 - https://github.com/rancher/k3s/pull/1839 - 1.17 - https://github.com/rancher/k3s/pull/1838 - 1.16 - https://github.com/rancher/k3s/pull/1837 CVE-2020-8555 was patched in the following versions of Kubernetes: v1.18.1+, v1.17.5+, v1.16.9+. Thus, this vulnerability is already patched in the K3s releases: - v1.18.2+k3s1 - v1.17.5+k3s1 - v1.16.9+k3s1 To ensure you have the latest stable version of Kubernetes for each release when upgrading to address these CVEs, we'll be releasing the following: - v1.18.3+k3s1 - v1.17.6+k3s1 - v1.16.10+k3s1 These releases will contain patches for both CVEs.
non_defect
addressing cve cve and cve we are working to patch to address the following cves only clusters susceptible to mitm attacks via rogue router advertisements kube controller manager ssrf cve is being addressed by upgrading the version of the library used by to see the following prs for reference cve was patched in the following versions of kubernetes thus this vulnerability is already patched in the releases to ensure you have the latest stable version of kubernetes for each release when upgrading to address these cves we ll be releasing the following these releases will contain patches for both cves
0
237,067
26,078,775,048
IssuesEvent
2022-12-25 01:10:25
arabaske/Ceres
https://api.github.com/repos/arabaske/Ceres
opened
CVE-2022-23540 (Medium) detected in jsonwebtoken-8.1.0.tgz
security vulnerability
## CVE-2022-23540 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonwebtoken-8.1.0.tgz</b></p></summary> <p>JSON Web Token implementation (symmetric and asymmetric)</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-8.1.0.tgz">https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-8.1.0.tgz</a></p> <p>Path to dependency file: /Ceres/package.json</p> <p>Path to vulnerable library: /node_modules/jsonwebtoken/package.json</p> <p> Dependency Hierarchy: - firebase-admin-5.12.1.tgz (Root Library) - :x: **jsonwebtoken-8.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/arabaske/Ceres/commits/b31d728670f7b1cea140b9a346bf71d1a9771fb2">b31d728670f7b1cea140b9a346bf71d1a9771fb2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In versions `<=8.5.1` of `jsonwebtoken` library, lack of algorithm definition in the `jwt.verify()` function can lead to signature validation bypass due to defaulting to the `none` algorithm for signature verification. Users are affected if you do not specify algorithms in the `jwt.verify()` function. This issue has been fixed, please update to version 9.0.0 which removes the default support for the none algorithm in the `jwt.verify()` method. There will be no impact, if you update to version 9.0.0 and you don’t need to allow for the `none` algorithm. If you need 'none' algorithm, you have to explicitly specify that in `jwt.verify()` options. <p>Publish Date: 2022-12-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23540>CVE-2022-23540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23540">https://www.cve.org/CVERecord?id=CVE-2022-23540</a></p> <p>Release Date: 2022-12-22</p> <p>Fix Resolution: jsonwebtoken - 9.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-23540 (Medium) detected in jsonwebtoken-8.1.0.tgz - ## CVE-2022-23540 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonwebtoken-8.1.0.tgz</b></p></summary> <p>JSON Web Token implementation (symmetric and asymmetric)</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-8.1.0.tgz">https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-8.1.0.tgz</a></p> <p>Path to dependency file: /Ceres/package.json</p> <p>Path to vulnerable library: /node_modules/jsonwebtoken/package.json</p> <p> Dependency Hierarchy: - firebase-admin-5.12.1.tgz (Root Library) - :x: **jsonwebtoken-8.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/arabaske/Ceres/commits/b31d728670f7b1cea140b9a346bf71d1a9771fb2">b31d728670f7b1cea140b9a346bf71d1a9771fb2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In versions `<=8.5.1` of `jsonwebtoken` library, lack of algorithm definition in the `jwt.verify()` function can lead to signature validation bypass due to defaulting to the `none` algorithm for signature verification. Users are affected if you do not specify algorithms in the `jwt.verify()` function. This issue has been fixed, please update to version 9.0.0 which removes the default support for the none algorithm in the `jwt.verify()` method. There will be no impact, if you update to version 9.0.0 and you don’t need to allow for the `none` algorithm. If you need 'none' algorithm, you have to explicitly specify that in `jwt.verify()` options. <p>Publish Date: 2022-12-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23540>CVE-2022-23540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-23540">https://www.cve.org/CVERecord?id=CVE-2022-23540</a></p> <p>Release Date: 2022-12-22</p> <p>Fix Resolution: jsonwebtoken - 9.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in jsonwebtoken tgz cve medium severity vulnerability vulnerable library jsonwebtoken tgz json web token implementation symmetric and asymmetric library home page a href path to dependency file ceres package json path to vulnerable library node modules jsonwebtoken package json dependency hierarchy firebase admin tgz root library x jsonwebtoken tgz vulnerable library found in head commit a href vulnerability details in versions of jsonwebtoken library lack of algorithm definition in the jwt verify function can lead to signature validation bypass due to defaulting to the none algorithm for signature verification users are affected if you do not specify algorithms in the jwt verify function this issue has been fixed please update to version which removes the default support for the none algorithm in the jwt verify method there will be no impact if you update to version and you don’t need to allow for the none algorithm if you need none algorithm you have to explicitly specify that in jwt verify options publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact high availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jsonwebtoken step up your open source security game with mend
0
76,305
26,354,683,100
IssuesEvent
2023-01-11 08:48:11
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
Incorrect leadership rebalancing after remove CP member, add a new member and add new CP groups
Type: Defect Team: Core Source: Internal Module: CP Subsystem
**Description:** I saw incorrect leadership rebalancing for following scenario: - `cp-size=5` and `group-size=3`, 3 nodes with priority 1, 2 nodes with priority 2. - 3 CP groups for every possible combination of groups are added (i.e. there are 30 groups; only 2 new CP groups for combination which is used for METADATA group are added). Leadership is correctly rebalanced - each member with priority 1 is leader of 10 groups. - 1 member with priority 1 is removed. Related CP groups are updated to have expected group size 3 again. Leadership is correctly rebalanced - each member with priority 1 is leader of 15 groups. - 1 other member with priority 1 is promoted to be CP member (it doesn't participate in any of already created CP groups) - 2 CP groups for every possible combination of groups are added (i.e. there are 20 new groups). The expectation is that the new member will be leader of all new CP groups which include this member (it means 12 groups). The rest of new groups should be divided between the two original nodes with priority 1 (it means 4 new groups for original member). It means each of original members with priority 1 should be leader of 19 groups and the new member should be leader of 12 groups. However I saw leadership divided like 20groups, 18groups and 12groups - it's not rebalanced properly. Final leadership rebalancing: ``` 2023-01-11 08:04:13,702 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-6^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Current leadership claims: CPMember{uuid=17561857-da0e-4620-9c64-41dcea2e3db9, address=[hz5]:5701} priority 2 has 12, CPMember{uuid=0a3eeaee-f20a-4cb7-90ce-90bc18c9299c, address=[hz0]:5701} priority 1 has 0, CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} priority 2 has 18, CPMember{uuid=b515ac5b-30c1-4bec-9fa9-05edf8f1765b, address=[hz3]:5701} priority 2 has 20, CPMember{uuid=e3546068-b1f7-421e-b011-f9a712ffd41a, address=[hz1]:5701} priority 1 has 0, leaderships. 2023-01-11 08:04:13,702 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-6^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] CPGroup leadership balance is fine, cannot rebalance further... ``` It seems that reason for it is that it started to search new leaderships for a member which doesn't contain the highest numbers of leaderships, see: ``` 2023-01-11 08:03:13,910 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-1^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Current leadership claims: CPMember{uuid=17561857-da0e-4620-9c64-41dcea2e3db9, address=[hz5]:5701} priority 2 has 11, CPMember{uuid=0a3eeaee-f20a-4cb7-90ce-90bc18c9299c, address=[hz0]:5701} priority 1 has 0, CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} priority 2 has 19, CPMember{uuid=b515ac5b-30c1-4bec-9fa9-05edf8f1765b, address=[hz3]:5701} priority 2 has 20, CPMember{uuid=e3546068-b1f7-421e-b011-f9a712ffd41a, address=[hz1]:5701} priority 1 has 0, leaderships. 2023-01-11 08:03:13,910 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-1^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Searching a candidate transfer leadership from CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} with 19 leaderships. ... 2023-01-11 08:03:13,944 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-1^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Current leadership claims: CPMember{uuid=17561857-da0e-4620-9c64-41dcea2e3db9, address=[hz5]:5701} priority 2 has 11, CPMember{uuid=0a3eeaee-f20a-4cb7-90ce-90bc18c9299c, address=[hz0]:5701} priority 1 has 0, CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} priority 2 has 18, CPMember{uuid=b515ac5b-30c1-4bec-9fa9-05edf8f1765b, address=[hz3]:5701} priority 2 has 20, CPMember{uuid=e3546068-b1f7-421e-b011-f9a712ffd41a, address=[hz1]:5701} priority 1 has 0, leaderships. 2023-01-11 08:03:13,945 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-1^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Searching a candidate transfer leadership from CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} with 18 leaderships. ``` **Logs:** See the logs for the test when I met this issue (by test described in How to reproduce section): [LeadershipPrioriotyTest_testRebalancingAfterRemoveAndAddHighPriorityCpMember.zip](https://github.com/hazelcast/hazelcast/files/10390350/LeadershipPrioriotyTest_testRebalancingAfterRemoveAndAddHighPriorityCpMember.zip) **How to reproduce:** - this is intermittent failure and it doesn't occur every time - following branches can be used (those tests are still WIP so please ignore various `System.out.println`, commented code, strange code etc.): - HZ OS: https://github.com/olukas/hazelcast/tree/HZ-651-it - HZ EE: https://github.com/olukas/hazelcast-enterprise/tree/HZ-651-it-testRebalancingAfterRemoveAndAddHighPriorityCpMember-reproducer - build HZ OS and then build HZ EE by `mvn clean install -DskipTests -Pdocker`. Then go to `hazelcast-enterprise-it/enterprise-tests-using-docker/cp-subsystem-it` and run the test by `mvn clean test -Pnightly-build -Dtest=LeadershipPrioriotyTest -Dhazelcast.test.defaultTestTimeoutInSeconds=7200`
1.0
Incorrect leadership rebalancing after remove CP member, add a new member and add new CP groups - **Description:** I saw incorrect leadership rebalancing for following scenario: - `cp-size=5` and `group-size=3`, 3 nodes with priority 1, 2 nodes with priority 2. - 3 CP groups for every possible combination of groups are added (i.e. there are 30 groups; only 2 new CP groups for combination which is used for METADATA group are added). Leadership is correctly rebalanced - each member with priority 1 is leader of 10 groups. - 1 member with priority 1 is removed. Related CP groups are updated to have expected group size 3 again. Leadership is correctly rebalanced - each member with priority 1 is leader of 15 groups. - 1 other member with priority 1 is promoted to be CP member (it doesn't participate in any of already created CP groups) - 2 CP groups for every possible combination of groups are added (i.e. there are 20 new groups). The expectation is that the new member will be leader of all new CP groups which include this member (it means 12 groups). The rest of new groups should be divided between the two original nodes with priority 1 (it means 4 new groups for original member). It means each of original members with priority 1 should be leader of 19 groups and the new member should be leader of 12 groups. However I saw leadership divided like 20groups, 18groups and 12groups - it's not rebalanced properly. Final leadership rebalancing: ``` 2023-01-11 08:04:13,702 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-6^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Current leadership claims: CPMember{uuid=17561857-da0e-4620-9c64-41dcea2e3db9, address=[hz5]:5701} priority 2 has 12, CPMember{uuid=0a3eeaee-f20a-4cb7-90ce-90bc18c9299c, address=[hz0]:5701} priority 1 has 0, CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} priority 2 has 18, CPMember{uuid=b515ac5b-30c1-4bec-9fa9-05edf8f1765b, address=[hz3]:5701} priority 2 has 20, CPMember{uuid=e3546068-b1f7-421e-b011-f9a712ffd41a, address=[hz1]:5701} priority 1 has 0, leaderships. 2023-01-11 08:04:13,702 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-6^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] CPGroup leadership balance is fine, cannot rebalance further... ``` It seems that reason for it is that it started to search new leaderships for a member which doesn't contain the highest numbers of leaderships, see: ``` 2023-01-11 08:03:13,910 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-1^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Current leadership claims: CPMember{uuid=17561857-da0e-4620-9c64-41dcea2e3db9, address=[hz5]:5701} priority 2 has 11, CPMember{uuid=0a3eeaee-f20a-4cb7-90ce-90bc18c9299c, address=[hz0]:5701} priority 1 has 0, CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} priority 2 has 19, CPMember{uuid=b515ac5b-30c1-4bec-9fa9-05edf8f1765b, address=[hz3]:5701} priority 2 has 20, CPMember{uuid=e3546068-b1f7-421e-b011-f9a712ffd41a, address=[hz1]:5701} priority 1 has 0, leaderships. 2023-01-11 08:03:13,910 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-1^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Searching a candidate transfer leadership from CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} with 19 leaderships. ... 2023-01-11 08:03:13,944 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-1^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Current leadership claims: CPMember{uuid=17561857-da0e-4620-9c64-41dcea2e3db9, address=[hz5]:5701} priority 2 has 11, CPMember{uuid=0a3eeaee-f20a-4cb7-90ce-90bc18c9299c, address=[hz0]:5701} priority 1 has 0, CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} priority 2 has 18, CPMember{uuid=b515ac5b-30c1-4bec-9fa9-05edf8f1765b, address=[hz3]:5701} priority 2 has 20, CPMember{uuid=e3546068-b1f7-421e-b011-f9a712ffd41a, address=[hz1]:5701} priority 1 has 0, leaderships. 2023-01-11 08:03:13,945 [ INFO^[[m] [hz.eloquent_ramanujan.cached.thread-1^[[m] [c.h.c.i.RaftGroupMembershipManager^[[m]: [hz3]:5701 [cp-subsystem-leadership-priority] [5.3.0-SNAPSHOT] Searching a candidate transfer leadership from CPMember{uuid=11935946-5666-4e2c-aa2f-92525c1c6d8c, address=[hz4]:5701} with 18 leaderships. ``` **Logs:** See the logs for the test when I met this issue (by test described in How to reproduce section): [LeadershipPrioriotyTest_testRebalancingAfterRemoveAndAddHighPriorityCpMember.zip](https://github.com/hazelcast/hazelcast/files/10390350/LeadershipPrioriotyTest_testRebalancingAfterRemoveAndAddHighPriorityCpMember.zip) **How to reproduce:** - this is intermittent failure and it doesn't occur every time - following branches can be used (those tests are still WIP so please ignore various `System.out.println`, commented code, strange code etc.): - HZ OS: https://github.com/olukas/hazelcast/tree/HZ-651-it - HZ EE: https://github.com/olukas/hazelcast-enterprise/tree/HZ-651-it-testRebalancingAfterRemoveAndAddHighPriorityCpMember-reproducer - build HZ OS and then build HZ EE by `mvn clean install -DskipTests -Pdocker`. Then go to `hazelcast-enterprise-it/enterprise-tests-using-docker/cp-subsystem-it` and run the test by `mvn clean test -Pnightly-build -Dtest=LeadershipPrioriotyTest -Dhazelcast.test.defaultTestTimeoutInSeconds=7200`
defect
incorrect leadership rebalancing after remove cp member add a new member and add new cp groups description i saw incorrect leadership rebalancing for following scenario cp size and group size nodes with priority nodes with priority cp groups for every possible combination of groups are added i e there are groups only new cp groups for combination which is used for metadata group are added leadership is correctly rebalanced each member with priority is leader of groups member with priority is removed related cp groups are updated to have expected group size again leadership is correctly rebalanced each member with priority is leader of groups other member with priority is promoted to be cp member it doesn t participate in any of already created cp groups cp groups for every possible combination of groups are added i e there are new groups the expectation is that the new member will be leader of all new cp groups which include this member it means groups the rest of new groups should be divided between the two original nodes with priority it means new groups for original member it means each of original members with priority should be leader of groups and the new member should be leader of groups however i saw leadership divided like and it s not rebalanced properly final leadership rebalancing current leadership claims cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has leaderships cpgroup leadership balance is fine cannot rebalance further it seems that reason for it is that it started to search new leaderships for a member which doesn t contain the highest numbers of leaderships see current leadership claims cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has leaderships searching a candidate transfer leadership from cpmember uuid address with leaderships current leadership claims cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has cpmember uuid address priority has leaderships searching a candidate transfer leadership from cpmember uuid address with leaderships logs see the logs for the test when i met this issue by test described in how to reproduce section how to reproduce this is intermittent failure and it doesn t occur every time following branches can be used those tests are still wip so please ignore various system out println commented code strange code etc hz os hz ee build hz os and then build hz ee by mvn clean install dskiptests pdocker then go to hazelcast enterprise it enterprise tests using docker cp subsystem it and run the test by mvn clean test pnightly build dtest leadershipprioriotytest dhazelcast test defaulttesttimeoutinseconds
1
626,078
19,784,628,786
IssuesEvent
2022-01-18 04:15:37
lokka30/LevelledMobs
https://api.github.com/repos/lokka30/LevelledMobs
closed
Player Levelling Change ; Scale Multipliers
type: improvement status: unconfirmed status: unassigned target version status: on hold priority: unknown status: waiting for approval
> ParseBadData — Today at 1:47 PM > Well I want for the levels and multipliers to change, like having them be calculated separately. Like a formulla that I have right now that is basically (archery+fighting+defense)/0.33 = monster's level, but for their HP multiplier to change with (archery+fighting)/0.5, so its like if the formulla gives 2, it would change the HP scaling to 2.0, but if it's something like 97 the HP scaling would be 15.0 https://discord.com/channels/752310043214479462/760051505067458590/905906023611646032 TLDR: The ability for Player Levelling to make changes to the multipliers independently to the levels. I'll have to return to this topic with potential config arrangements for such a system, as well as how it might be implemented.
1.0
Player Levelling Change ; Scale Multipliers - > ParseBadData — Today at 1:47 PM > Well I want for the levels and multipliers to change, like having them be calculated separately. Like a formulla that I have right now that is basically (archery+fighting+defense)/0.33 = monster's level, but for their HP multiplier to change with (archery+fighting)/0.5, so its like if the formulla gives 2, it would change the HP scaling to 2.0, but if it's something like 97 the HP scaling would be 15.0 https://discord.com/channels/752310043214479462/760051505067458590/905906023611646032 TLDR: The ability for Player Levelling to make changes to the multipliers independently to the levels. I'll have to return to this topic with potential config arrangements for such a system, as well as how it might be implemented.
non_defect
player levelling change scale multipliers parsebaddata — today at pm well i want for the levels and multipliers to change like having them be calculated separately like a formulla that i have right now that is basically archery fighting defense monster s level but for their hp multiplier to change with archery fighting so its like if the formulla gives it would change the hp scaling to but if it s something like the hp scaling would be tldr the ability for player levelling to make changes to the multipliers independently to the levels i ll have to return to this topic with potential config arrangements for such a system as well as how it might be implemented
0
425,162
12,336,468,155
IssuesEvent
2020-05-14 13:38:17
googleapis/nodejs-bigquery-storage
https://api.github.com/repos/googleapis/nodejs-bigquery-storage
opened
Synthesis failed for nodejs-bigquery-storage
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate nodejs-bigquery-storage. :broken_heart: Here's the output from running `synth.py`: ``` 2020-05-14 06:37:10,870 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/nodejs-bigquery-storage 2020-05-14 06:37:10,873 synthtool [ERROR] > Failed executing git clone --single-branch https://github.com/googleapis/nodejs-bigquery-storage.git /home/kbuilder/.cache/synthtool/nodejs-bigquery-storage: fatal: could not create work tree dir '/home/kbuilder/.cache/synthtool/nodejs-bigquery-storage': No space left on device Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 600, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 472, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 519, in _inner_main working_repo_path = synthtool_git.clone(f"https://github.com/{args.repository}.git") File "/tmpfs/src/github/synthtool/synthtool/sources/git.py", line 83, in clone shell.run(cmd, check=True) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['git', 'clone', '--single-branch', 'https://github.com/googleapis/nodejs-bigquery-storage.git', PosixPath('/home/kbuilder/.cache/synthtool/nodejs-bigquery-storage')]' returned non-zero exit status 128. ``` Google internal developers can see the full log [here](https://sponge/ea51e8ba-def4-460f-9acd-a09a559fce5d).
1.0
Synthesis failed for nodejs-bigquery-storage - Hello! Autosynth couldn't regenerate nodejs-bigquery-storage. :broken_heart: Here's the output from running `synth.py`: ``` 2020-05-14 06:37:10,870 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/nodejs-bigquery-storage 2020-05-14 06:37:10,873 synthtool [ERROR] > Failed executing git clone --single-branch https://github.com/googleapis/nodejs-bigquery-storage.git /home/kbuilder/.cache/synthtool/nodejs-bigquery-storage: fatal: could not create work tree dir '/home/kbuilder/.cache/synthtool/nodejs-bigquery-storage': No space left on device Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 600, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 472, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 519, in _inner_main working_repo_path = synthtool_git.clone(f"https://github.com/{args.repository}.git") File "/tmpfs/src/github/synthtool/synthtool/sources/git.py", line 83, in clone shell.run(cmd, check=True) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['git', 'clone', '--single-branch', 'https://github.com/googleapis/nodejs-bigquery-storage.git', PosixPath('/home/kbuilder/.cache/synthtool/nodejs-bigquery-storage')]' returned non-zero exit status 128. ``` Google internal developers can see the full log [here](https://sponge/ea51e8ba-def4-460f-9acd-a09a559fce5d).
non_defect
synthesis failed for nodejs bigquery storage hello autosynth couldn t regenerate nodejs bigquery storage broken heart here s the output from running synth py autosynth logs will be written to tmpfs src github synthtool logs googleapis nodejs bigquery storage synthtool failed executing git clone single branch home kbuilder cache synthtool nodejs bigquery storage fatal could not create work tree dir home kbuilder cache synthtool nodejs bigquery storage no space left on device traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main working repo path synthtool git clone f file tmpfs src github synthtool synthtool sources git py line in clone shell run cmd check true file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
388,966
11,495,827,365
IssuesEvent
2020-02-12 06:13:07
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
Cash register balance should only include cash transaction totals
Effort: small Feature Module: dispensary Priority: high
## Is your feature request related to a problem? Please describe. The cash register balance currently does not filter transactions by payment type. This differs from desktop behaviour which only sums transactions with cash payment types. ## Describe the solution you'd like Filter transactions by payment type before summing totals. ## Implementation See above. ## Describe alternatives you've considered N/A. ## Additional context N/A.
1.0
Cash register balance should only include cash transaction totals - ## Is your feature request related to a problem? Please describe. The cash register balance currently does not filter transactions by payment type. This differs from desktop behaviour which only sums transactions with cash payment types. ## Describe the solution you'd like Filter transactions by payment type before summing totals. ## Implementation See above. ## Describe alternatives you've considered N/A. ## Additional context N/A.
non_defect
cash register balance should only include cash transaction totals is your feature request related to a problem please describe the cash register balance currently does not filter transactions by payment type this differs from desktop behaviour which only sums transactions with cash payment types describe the solution you d like filter transactions by payment type before summing totals implementation see above describe alternatives you ve considered n a additional context n a
0
56,040
14,910,688,203
IssuesEvent
2021-01-22 09:56:53
STEllAR-GROUP/phylanx
https://api.github.com/repos/STEllAR-GROUP/phylanx
closed
physl issues
category: PhySL type: defect
So I've been debugging phsyl, trying to figure out what's going wrong with it. There are several things. First, if you run with multiple localities, each of them tries to write the same file. Should each locality write a file, or just one locality? IDK. It's easiest just have them all write it out. Second, writing out the data_csv fails. The critical place is line 206 of src/util/performance_data.cpp ``` 187 // Reuse get_counter_values_array calls 188 // key: primitive type 189 // value: vector of performance counter values 190 std::map<std::string, std::vector<std::vector<std::int64_t>>> 191 counter_values_pile; 192 193 // Collect the values from the futures 194 for (auto& values_futures : values_futures_pile) 195 { 196 // Performance counter values 197 std::vector<std::vector<std::int64_t>>& counter_values = 198 counter_values_pile[values_futures.first]; 199 200 // Preallocate memory 201 counter_values.reserve(counter_name_last_parts.size()); 202 203 // Collect the performance counter values 204 for (auto& f : values_futures.second) 205 { 206 counter_values.push_back(f.get().values_); 207 } 208 } ``` The following horrible hack fixes these issues and allows JetLag to run. Obviously, it is not the correct solution. ``` diff --git a/examples/interpreter/physl.cpp b/examples/interpreter/physl.cpp index 66a52d84..90be39d0 100644 --- a/examples/interpreter/physl.cpp +++ b/examples/interpreter/physl.cpp @@ -513,9 +513,30 @@ void print_newick_tree(std::string const& code_source_name, void print_performance_profile( phylanx::execution_tree::compiler::function_list& snippets, - std::string const& code_source_name, std::string const& dot_file, - std::string const& newick_tree_file, std::string const& counter_file) + std::string const& code_source_name, std::string const& _dot_file, + std::string const& _newick_tree_file, std::string const& _counter_file) { + std::int64_t _locality = + hpx::naming::get_locality_id_from_id(hpx::find_here()); + + std::ostringstream cf; + cf << _counter_file; + if(_locality > 0) + cf << "." << _locality; + std::string counter_file = cf.str(); + + std::ostringstream df; + df << _dot_file; + if(_locality > 0) + df << "." << _locality; + std::string dot_file = df.str(); + + std::ostringstream nf; + nf << _newick_tree_file; + if(_locality > 0) + nf << "." << _locality; + std::string newick_tree_file = nf.str(); + std::set<std::string> resolve_children; for (auto const& ep : snippets.program_.entry_points()) { @@ -681,7 +702,14 @@ void interpreter(po::variables_map const& vm) // if requested if (vm.count("print") != 0) { - std::string const result_file = vm["print"].as<std::string>(); + std::string const _result_file = vm["print"].as<std::string>(); + std::int64_t _locality = + hpx::naming::get_locality_id_from_id(hpx::find_here()); + std::ostringstream rf; + rf << _result_file; + if(_locality > 0) + rf << "." << _locality; + std::string result_file = rf.str(); std::string format = "plain"; if (vm.count("print-format") != 0) { diff --git a/src/util/performance_data.cpp b/src/util/performance_data.cpp index 0703d7a3..999d26b6 100644 --- a/src/util/performance_data.cpp +++ b/src/util/performance_data.cpp @@ -203,7 +203,10 @@ namespace phylanx { namespace util { // Collect the performance counter values for (auto& f : values_futures.second) { - counter_values.push_back(f.get().values_); + try { + counter_values.push_back(f.get().values_); + } catch(...) { + } } } ```
1.0
physl issues - So I've been debugging phsyl, trying to figure out what's going wrong with it. There are several things. First, if you run with multiple localities, each of them tries to write the same file. Should each locality write a file, or just one locality? IDK. It's easiest just have them all write it out. Second, writing out the data_csv fails. The critical place is line 206 of src/util/performance_data.cpp ``` 187 // Reuse get_counter_values_array calls 188 // key: primitive type 189 // value: vector of performance counter values 190 std::map<std::string, std::vector<std::vector<std::int64_t>>> 191 counter_values_pile; 192 193 // Collect the values from the futures 194 for (auto& values_futures : values_futures_pile) 195 { 196 // Performance counter values 197 std::vector<std::vector<std::int64_t>>& counter_values = 198 counter_values_pile[values_futures.first]; 199 200 // Preallocate memory 201 counter_values.reserve(counter_name_last_parts.size()); 202 203 // Collect the performance counter values 204 for (auto& f : values_futures.second) 205 { 206 counter_values.push_back(f.get().values_); 207 } 208 } ``` The following horrible hack fixes these issues and allows JetLag to run. Obviously, it is not the correct solution. ``` diff --git a/examples/interpreter/physl.cpp b/examples/interpreter/physl.cpp index 66a52d84..90be39d0 100644 --- a/examples/interpreter/physl.cpp +++ b/examples/interpreter/physl.cpp @@ -513,9 +513,30 @@ void print_newick_tree(std::string const& code_source_name, void print_performance_profile( phylanx::execution_tree::compiler::function_list& snippets, - std::string const& code_source_name, std::string const& dot_file, - std::string const& newick_tree_file, std::string const& counter_file) + std::string const& code_source_name, std::string const& _dot_file, + std::string const& _newick_tree_file, std::string const& _counter_file) { + std::int64_t _locality = + hpx::naming::get_locality_id_from_id(hpx::find_here()); + + std::ostringstream cf; + cf << _counter_file; + if(_locality > 0) + cf << "." << _locality; + std::string counter_file = cf.str(); + + std::ostringstream df; + df << _dot_file; + if(_locality > 0) + df << "." << _locality; + std::string dot_file = df.str(); + + std::ostringstream nf; + nf << _newick_tree_file; + if(_locality > 0) + nf << "." << _locality; + std::string newick_tree_file = nf.str(); + std::set<std::string> resolve_children; for (auto const& ep : snippets.program_.entry_points()) { @@ -681,7 +702,14 @@ void interpreter(po::variables_map const& vm) // if requested if (vm.count("print") != 0) { - std::string const result_file = vm["print"].as<std::string>(); + std::string const _result_file = vm["print"].as<std::string>(); + std::int64_t _locality = + hpx::naming::get_locality_id_from_id(hpx::find_here()); + std::ostringstream rf; + rf << _result_file; + if(_locality > 0) + rf << "." << _locality; + std::string result_file = rf.str(); std::string format = "plain"; if (vm.count("print-format") != 0) { diff --git a/src/util/performance_data.cpp b/src/util/performance_data.cpp index 0703d7a3..999d26b6 100644 --- a/src/util/performance_data.cpp +++ b/src/util/performance_data.cpp @@ -203,7 +203,10 @@ namespace phylanx { namespace util { // Collect the performance counter values for (auto& f : values_futures.second) { - counter_values.push_back(f.get().values_); + try { + counter_values.push_back(f.get().values_); + } catch(...) { + } } } ```
defect
physl issues so i ve been debugging phsyl trying to figure out what s going wrong with it there are several things first if you run with multiple localities each of them tries to write the same file should each locality write a file or just one locality idk it s easiest just have them all write it out second writing out the data csv fails the critical place is line of src util performance data cpp reuse get counter values array calls key primitive type value vector of performance counter values std map counter values pile collect the values from the futures for auto values futures values futures pile performance counter values std vector counter values counter values pile preallocate memory counter values reserve counter name last parts size collect the performance counter values for auto f values futures second counter values push back f get values the following horrible hack fixes these issues and allows jetlag to run obviously it is not the correct solution diff git a examples interpreter physl cpp b examples interpreter physl cpp index a examples interpreter physl cpp b examples interpreter physl cpp void print newick tree std string const code source name void print performance profile phylanx execution tree compiler function list snippets std string const code source name std string const dot file std string const newick tree file std string const counter file std string const code source name std string const dot file std string const newick tree file std string const counter file std t locality hpx naming get locality id from id hpx find here std ostringstream cf cf counter file if locality cf locality std string counter file cf str std ostringstream df df dot file if locality df locality std string dot file df str std ostringstream nf nf newick tree file if locality nf locality std string newick tree file nf str std set resolve children for auto const ep snippets program entry points void interpreter po variables map const vm if requested if vm count print std string const result file vm as std string const result file vm as std t locality hpx naming get locality id from id hpx find here std ostringstream rf rf result file if locality rf locality std string result file rf str std string format plain if vm count print format diff git a src util performance data cpp b src util performance data cpp index a src util performance data cpp b src util performance data cpp namespace phylanx namespace util collect the performance counter values for auto f values futures second counter values push back f get values try counter values push back f get values catch
1
58,674
16,680,099,546
IssuesEvent
2021-06-07 21:56:40
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
quick theme toggle not working
A-Themes-Official S-Major T-Defect X-Regression X-Release-Blocker
<!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P --> <!-- Please report security issues by email to security@matrix.org --> <!-- This is a bug report template. By following the instructions below and filling out the sections with your information, you will help the us to get all the necessary data to fix your issue. You can also preview your report before submitting it. You may remove sections that aren't relevant to your particular case. Text between <!-- and --​> marks will be invisible in the report. --> ### Description This thing ![image](https://user-images.githubusercontent.com/2803622/121069252-8e9fea00-c7cd-11eb-8976-a5b1f14bf220.png) isn't working for me at all. The light theme will become selected in the menu like this ![image](https://user-images.githubusercontent.com/2803622/121069162-76c86600-c7cd-11eb-9d26-63af780eb676.png) but theme is only applied when switching via this menu. #17581 could be related. ### Steps to reproduce - click light/dark theme quick toggle in the top left user menu thing - potentially use these labs settings, especially custom themes - no idea since when this is, i don't normally use this feature ![image](https://user-images.githubusercontent.com/2803622/121069203-83e55500-c7cd-11eb-857c-261ca77cb8e2.png) <!-- Please send us logs for your bug report. They're very important for bugs which are hard to reproduce. To do this, create this issue then go to your account settings and click 'Submit Debug Logs' from the Help & About tab --> Logs being sent: yes <!-- Include screenshots if possible: you can drag and drop images below. --> ### Version information - **Platform**: desktop For the desktop app: - **OS**: Arch - **Version**: aur/element-desktop-nightly-bin 2021060702-1
1.0
quick theme toggle not working - <!-- A picture's worth a thousand words: PLEASE INCLUDE A SCREENSHOT :P --> <!-- Please report security issues by email to security@matrix.org --> <!-- This is a bug report template. By following the instructions below and filling out the sections with your information, you will help the us to get all the necessary data to fix your issue. You can also preview your report before submitting it. You may remove sections that aren't relevant to your particular case. Text between <!-- and --​> marks will be invisible in the report. --> ### Description This thing ![image](https://user-images.githubusercontent.com/2803622/121069252-8e9fea00-c7cd-11eb-8976-a5b1f14bf220.png) isn't working for me at all. The light theme will become selected in the menu like this ![image](https://user-images.githubusercontent.com/2803622/121069162-76c86600-c7cd-11eb-9d26-63af780eb676.png) but theme is only applied when switching via this menu. #17581 could be related. ### Steps to reproduce - click light/dark theme quick toggle in the top left user menu thing - potentially use these labs settings, especially custom themes - no idea since when this is, i don't normally use this feature ![image](https://user-images.githubusercontent.com/2803622/121069203-83e55500-c7cd-11eb-857c-261ca77cb8e2.png) <!-- Please send us logs for your bug report. They're very important for bugs which are hard to reproduce. To do this, create this issue then go to your account settings and click 'Submit Debug Logs' from the Help & About tab --> Logs being sent: yes <!-- Include screenshots if possible: you can drag and drop images below. --> ### Version information - **Platform**: desktop For the desktop app: - **OS**: Arch - **Version**: aur/element-desktop-nightly-bin 2021060702-1
defect
quick theme toggle not working this is a bug report template by following the instructions below and filling out the sections with your information you will help the us to get all the necessary data to fix your issue you can also preview your report before submitting it you may remove sections that aren t relevant to your particular case text between marks will be invisible in the report description this thing isn t working for me at all the light theme will become selected in the menu like this but theme is only applied when switching via this menu could be related steps to reproduce click light dark theme quick toggle in the top left user menu thing potentially use these labs settings especially custom themes no idea since when this is i don t normally use this feature please send us logs for your bug report they re very important for bugs which are hard to reproduce to do this create this issue then go to your account settings and click submit debug logs from the help about tab logs being sent yes version information platform desktop for the desktop app os arch version aur element desktop nightly bin
1
55,691
14,632,217,590
IssuesEvent
2020-12-23 21:48:12
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
closed
Redirect ignores Alpine -context parameter
defect p2 pending release
### Current Behavior: Having the Alpine parameter set on the Commandline _-context /dtrack_ in the previous version of DT and now migrating to 3.8 leads to a redirect to _/login_ without taking the _context-path_ into account. ### Steps to Reproduce: ``` java $JAVA_OPTIONS -DdependencyTrack.logging.level=$LOGGING_LEVEL -jar dependency-track-embedded.war -context /dtrack ``` ### Expected Behavior: I would expect that _-context_ is taken into account when redirects are made. This seems to be a regression because we had previously also problems with the _-context_ parameter. ### Environment: - Dependency-Track Version: 3.8 - Distribution: Executable WAR - Browser: Chrome ### Additional Information: Running it locally (_java -jar dependency-track-embedded.war -context=/dtrack_) shows that the _-context_ path is totally ignored. Everything is directly served via _/_ like _/dashboards /projects /licenses_ and so on. This might be a regression that came in with Alpine 1.8.0-SNAPSHOT.
1.0
Redirect ignores Alpine -context parameter - ### Current Behavior: Having the Alpine parameter set on the Commandline _-context /dtrack_ in the previous version of DT and now migrating to 3.8 leads to a redirect to _/login_ without taking the _context-path_ into account. ### Steps to Reproduce: ``` java $JAVA_OPTIONS -DdependencyTrack.logging.level=$LOGGING_LEVEL -jar dependency-track-embedded.war -context /dtrack ``` ### Expected Behavior: I would expect that _-context_ is taken into account when redirects are made. This seems to be a regression because we had previously also problems with the _-context_ parameter. ### Environment: - Dependency-Track Version: 3.8 - Distribution: Executable WAR - Browser: Chrome ### Additional Information: Running it locally (_java -jar dependency-track-embedded.war -context=/dtrack_) shows that the _-context_ path is totally ignored. Everything is directly served via _/_ like _/dashboards /projects /licenses_ and so on. This might be a regression that came in with Alpine 1.8.0-SNAPSHOT.
defect
redirect ignores alpine context parameter current behavior having the alpine parameter set on the commandline context dtrack in the previous version of dt and now migrating to leads to a redirect to login without taking the context path into account steps to reproduce java java options ddependencytrack logging level logging level jar dependency track embedded war context dtrack expected behavior i would expect that context is taken into account when redirects are made this seems to be a regression because we had previously also problems with the context parameter environment dependency track version distribution executable war browser chrome additional information running it locally java jar dependency track embedded war context dtrack shows that the context path is totally ignored everything is directly served via like dashboards projects licenses and so on this might be a regression that came in with alpine snapshot
1
22,639
3,670,980,094
IssuesEvent
2016-02-22 03:09:52
gperftools/gperftools
https://api.github.com/repos/gperftools/gperftools
closed
Functions names in a relatively large shared library are not shown in cpu profile
Priority-Medium Status-New Type-Defect
Originally reported on Google Code with ID 622 ``` What steps will reproduce the problem? 1. Build a dynamic linking library with -g and -fno-pie. 2. Build an executable that is also compiled with -g and -fno-pie, and also linked with the library with -g. 3. Link to libprofiler. ProfilerStart/ProfilerStop in the executable, then pprof --pdf <application> <profile> > a.pdf What is the expected output? What do you see instead? Expected: All function names are retrieved from the dynamic library and demangled names are shown in the profile. Actual: Many function names in the shared library are shown as physical addresses like "00007f54c09cbbf7" What version of the product are you using? On what operating system? Fedora 20, x86_64. gcc 4.8.2. libunwind 1.1.3. Tried two versions of gperftools 1. gperftools/gperftools-devel-2.1.4-fc20.x86_64 (from the repo) 2. gperftools 2.2, build from source myself. Please provide any additional information below. Both the shared library and the executable have debug symbols. Attached nm results as well as the profile. When I build the library as a static library and link the executable to it, I can get the correct function names. I tried several things, but no avail so far. I guess it's related to the following issues. https://code.google.com/p/gperftools/issues/detail?id=562 https://code.google.com/p/gperftools/issues/detail?id=586 So, I tried -fno-pie and the newest gperftools (2.2) from source build. Nope, the problem is still there. The symptoms described in Issue 586 seem quite similar to what I get. This problem happens only with shared library. I didn't have this issue until my shared library became relatively big (8MB). I had no problems using gperftools/pprof otherwise. It seems like gperftools fails to retrieve symbols from shared libraries for some reason. I vaguely remember there is some option to get gperftools into verbose mode and describe what symbols/why it couldn't load, is there? The workaround so far is to build the library as a static library when I want to profile, but not ideal:-( ``` Reported by `hideaki.kimura` on 2014-05-15 18:28:26 <hr> * *Attachment: [exec.nm](https://storage.googleapis.com/google-code-attachments/gperftools/issue-622/comment-0/exec.nm)* * *Attachment: [lib.nm](https://storage.googleapis.com/google-code-attachments/gperftools/issue-622/comment-0/lib.nm)* * *Attachment: [tpcb_experiment.prof](https://storage.googleapis.com/google-code-attachments/gperftools/issue-622/comment-0/tpcb_experiment.prof)*
1.0
Functions names in a relatively large shared library are not shown in cpu profile - Originally reported on Google Code with ID 622 ``` What steps will reproduce the problem? 1. Build a dynamic linking library with -g and -fno-pie. 2. Build an executable that is also compiled with -g and -fno-pie, and also linked with the library with -g. 3. Link to libprofiler. ProfilerStart/ProfilerStop in the executable, then pprof --pdf <application> <profile> > a.pdf What is the expected output? What do you see instead? Expected: All function names are retrieved from the dynamic library and demangled names are shown in the profile. Actual: Many function names in the shared library are shown as physical addresses like "00007f54c09cbbf7" What version of the product are you using? On what operating system? Fedora 20, x86_64. gcc 4.8.2. libunwind 1.1.3. Tried two versions of gperftools 1. gperftools/gperftools-devel-2.1.4-fc20.x86_64 (from the repo) 2. gperftools 2.2, build from source myself. Please provide any additional information below. Both the shared library and the executable have debug symbols. Attached nm results as well as the profile. When I build the library as a static library and link the executable to it, I can get the correct function names. I tried several things, but no avail so far. I guess it's related to the following issues. https://code.google.com/p/gperftools/issues/detail?id=562 https://code.google.com/p/gperftools/issues/detail?id=586 So, I tried -fno-pie and the newest gperftools (2.2) from source build. Nope, the problem is still there. The symptoms described in Issue 586 seem quite similar to what I get. This problem happens only with shared library. I didn't have this issue until my shared library became relatively big (8MB). I had no problems using gperftools/pprof otherwise. It seems like gperftools fails to retrieve symbols from shared libraries for some reason. I vaguely remember there is some option to get gperftools into verbose mode and describe what symbols/why it couldn't load, is there? The workaround so far is to build the library as a static library when I want to profile, but not ideal:-( ``` Reported by `hideaki.kimura` on 2014-05-15 18:28:26 <hr> * *Attachment: [exec.nm](https://storage.googleapis.com/google-code-attachments/gperftools/issue-622/comment-0/exec.nm)* * *Attachment: [lib.nm](https://storage.googleapis.com/google-code-attachments/gperftools/issue-622/comment-0/lib.nm)* * *Attachment: [tpcb_experiment.prof](https://storage.googleapis.com/google-code-attachments/gperftools/issue-622/comment-0/tpcb_experiment.prof)*
defect
functions names in a relatively large shared library are not shown in cpu profile originally reported on google code with id what steps will reproduce the problem build a dynamic linking library with g and fno pie build an executable that is also compiled with g and fno pie and also linked with the library with g link to libprofiler profilerstart profilerstop in the executable then pprof pdf a pdf what is the expected output what do you see instead expected all function names are retrieved from the dynamic library and demangled names are shown in the profile actual many function names in the shared library are shown as physical addresses like what version of the product are you using on what operating system fedora gcc libunwind tried two versions of gperftools gperftools gperftools devel from the repo gperftools build from source myself please provide any additional information below both the shared library and the executable have debug symbols attached nm results as well as the profile when i build the library as a static library and link the executable to it i can get the correct function names i tried several things but no avail so far i guess it s related to the following issues so i tried fno pie and the newest gperftools from source build nope the problem is still there the symptoms described in issue seem quite similar to what i get this problem happens only with shared library i didn t have this issue until my shared library became relatively big i had no problems using gperftools pprof otherwise it seems like gperftools fails to retrieve symbols from shared libraries for some reason i vaguely remember there is some option to get gperftools into verbose mode and describe what symbols why it couldn t load is there the workaround so far is to build the library as a static library when i want to profile but not ideal reported by hideaki kimura on attachment attachment attachment
1
11,587
2,659,667,405
IssuesEvent
2015-03-18 22:26:35
notspiff/visualization.nastyfft
https://api.github.com/repos/notspiff/visualization.nastyfft
closed
Eden patch
auto-migrated Priority-Medium Type-Defect
``` Not really an issue just a patch if you want to produce an Eden version of the Visualization. ``` Original issue reported on code.google.com by `giftie61@gmail.com` on 2 Oct 2011 at 11:28 Attachments: * [pre-Eden.patch](https://storage.googleapis.com/google-code-attachments/nastyfft/issue-1/comment-0/pre-Eden.patch)
1.0
Eden patch - ``` Not really an issue just a patch if you want to produce an Eden version of the Visualization. ``` Original issue reported on code.google.com by `giftie61@gmail.com` on 2 Oct 2011 at 11:28 Attachments: * [pre-Eden.patch](https://storage.googleapis.com/google-code-attachments/nastyfft/issue-1/comment-0/pre-Eden.patch)
defect
eden patch not really an issue just a patch if you want to produce an eden version of the visualization original issue reported on code google com by gmail com on oct at attachments
1
22,981
3,985,505,231
IssuesEvent
2016-05-07 22:53:42
servo/servo
https://api.github.com/repos/servo/servo
closed
Intermittent failure in /_mozilla/mozilla/iframe_contentDocument.html
A-testing E-easy I-intermittent L-javascript
``` ▶ Unexpected subtest result in /_mozilla/mozilla/iframe_contentDocument.html: │ FAIL [expected PASS] Iframe contentDocument test. │ → iframe.contentDocument.getElementById(...) is null │ │ test_contentWindow@http://web-platform.test:8000/_mozilla/mozilla/iframe_contentDocument.html:11:19 │ Test.prototype.step@http://web-platform.test:8000/resources/testharness.js:1397:20 └ Test.prototype.step_func/&lt;@http://web-platform.test:8000/resources/testharness.js:1421:20 ``` This test was written before the load event for iframes was implemented. It should be rewritten to use that instead of using setTimeout. Code: `tests/wpt/mozilla/tests/mozilla/iframe_contentDocument.html` Test: `./mach test-wpt tests/wpt/mozilla/tests/mozilla/iframe_contentDocument.html`
1.0
Intermittent failure in /_mozilla/mozilla/iframe_contentDocument.html - ``` ▶ Unexpected subtest result in /_mozilla/mozilla/iframe_contentDocument.html: │ FAIL [expected PASS] Iframe contentDocument test. │ → iframe.contentDocument.getElementById(...) is null │ │ test_contentWindow@http://web-platform.test:8000/_mozilla/mozilla/iframe_contentDocument.html:11:19 │ Test.prototype.step@http://web-platform.test:8000/resources/testharness.js:1397:20 └ Test.prototype.step_func/&lt;@http://web-platform.test:8000/resources/testharness.js:1421:20 ``` This test was written before the load event for iframes was implemented. It should be rewritten to use that instead of using setTimeout. Code: `tests/wpt/mozilla/tests/mozilla/iframe_contentDocument.html` Test: `./mach test-wpt tests/wpt/mozilla/tests/mozilla/iframe_contentDocument.html`
non_defect
intermittent failure in mozilla mozilla iframe contentdocument html ▶ unexpected subtest result in mozilla mozilla iframe contentdocument html │ fail iframe contentdocument test │ → iframe contentdocument getelementbyid is null │ │ test contentwindow │ test prototype step └ test prototype step func lt this test was written before the load event for iframes was implemented it should be rewritten to use that instead of using settimeout code tests wpt mozilla tests mozilla iframe contentdocument html test mach test wpt tests wpt mozilla tests mozilla iframe contentdocument html
0
4,259
2,610,090,357
IssuesEvent
2015-02-26 18:27:21
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳痘痘怎么样消除最好
auto-migrated Priority-Medium Type-Defect
``` 深圳痘痘怎么样消除最好【深圳韩方科颜全国热线400-869-1818�� �24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以�� �国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品� ��韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反 弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国�� �专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸� ��的痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:41
1.0
深圳痘痘怎么样消除最好 - ``` 深圳痘痘怎么样消除最好【深圳韩方科颜全国热线400-869-1818�� �24小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以�� �国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品� ��韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反 弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国�� �专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸� ��的痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:41
defect
深圳痘痘怎么样消除最好 深圳痘痘怎么样消除最好【 �� � 】深圳韩方科颜专业祛痘连锁机构,机构以�� �国秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品� ��韩方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反 弹”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国�� �专业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸� ��的痘痘。 original issue reported on code google com by szft com on may at
1
197,581
22,596,252,405
IssuesEvent
2022-06-29 03:34:01
scriptex/initial-commit
https://api.github.com/repos/scriptex/initial-commit
closed
CVE-2022-2217 (High) detected in parse-url-5.0.2.tgz
security vulnerability
## CVE-2022-2217 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.2.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/parse-url/package.json</p> <p> Dependency Hierarchy: - release-it-15.1.1.tgz (Root Library) - git-url-parse-11.6.0.tgz - git-up-4.0.2.tgz - :x: **parse-url-5.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/scriptex/initial-commit/commit/ab82099b14d398d093bec389dcfaee3fd8bb6f07">ab82099b14d398d093bec389dcfaee3fd8bb6f07</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-site Scripting (XSS) - Generic in GitHub repository ionicabizau/parse-url prior to 7.0.0. <p>Publish Date: 2022-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2217>CVE-2022-2217</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/">https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/</a></p> <p>Release Date: 2022-06-27</p> <p>Fix Resolution: parse-url - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-2217 (High) detected in parse-url-5.0.2.tgz - ## CVE-2022-2217 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.2.tgz</b></p></summary> <p>An advanced url parser supporting git urls too.</p> <p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/parse-url/package.json</p> <p> Dependency Hierarchy: - release-it-15.1.1.tgz (Root Library) - git-url-parse-11.6.0.tgz - git-up-4.0.2.tgz - :x: **parse-url-5.0.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/scriptex/initial-commit/commit/ab82099b14d398d093bec389dcfaee3fd8bb6f07">ab82099b14d398d093bec389dcfaee3fd8bb6f07</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Cross-site Scripting (XSS) - Generic in GitHub repository ionicabizau/parse-url prior to 7.0.0. <p>Publish Date: 2022-06-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2217>CVE-2022-2217</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/">https://huntr.dev/bounties/4e046c63-b1ca-4bcc-b418-29796918a71b/</a></p> <p>Release Date: 2022-06-27</p> <p>Fix Resolution: parse-url - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in parse url tgz cve high severity vulnerability vulnerable library parse url tgz an advanced url parser supporting git urls too library home page a href path to dependency file package json path to vulnerable library node modules parse url package json dependency hierarchy release it tgz root library git url parse tgz git up tgz x parse url tgz vulnerable library found in head commit a href found in base branch master vulnerability details cross site scripting xss generic in github repository ionicabizau parse url prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution parse url step up your open source security game with mend
0
55,619
14,597,328,698
IssuesEvent
2020-12-20 19:38:17
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
`scipy.ncx2.sf` should be monotone decreasing
defect scipy.special scipy.stats
I'm getting unexpected behavior from `scipy.ncx2.sf`. The survival function should eventually go to zero for large enough inputs, but instead I'm seeing a kind of "foot", i.e., it is bottoming out at a positive value. The following interactive I/O demonstrates the issue. ### Reproducing code **example:** (Pdb) v_T_norm[::50] array([ 4.99515382e+00, 1.07617327e+01, 2.31854502e+01, 4.99515382e+01, 1.07617327e+02, 2.31854502e+02, 4.99515382e+02, 1.07617327e+03, 2.31854502e+03, 4.99515382e+03, 1.07617327e+04, 2.31854502e+04, 4.99515382e+04]) (Pdb) nu 20.0 (Pdb) lam 499.51538166556196 (Pdb) ncx2.sf(v_T_norm[::50], df=nu, nc=lam) array([ 1.00000000e+00, 1.00000000e+00, 1.00000000e+00, 1.00000000e+00, 1.00000000e+00, 1.00000000e+00, 6.64666099e-01, 4.16325934e-05, 4.16325934e-05, 4.16325934e-05, 4.16325934e-05, 4.16325934e-05, 4.16325934e-05])
1.0
`scipy.ncx2.sf` should be monotone decreasing - I'm getting unexpected behavior from `scipy.ncx2.sf`. The survival function should eventually go to zero for large enough inputs, but instead I'm seeing a kind of "foot", i.e., it is bottoming out at a positive value. The following interactive I/O demonstrates the issue. ### Reproducing code **example:** (Pdb) v_T_norm[::50] array([ 4.99515382e+00, 1.07617327e+01, 2.31854502e+01, 4.99515382e+01, 1.07617327e+02, 2.31854502e+02, 4.99515382e+02, 1.07617327e+03, 2.31854502e+03, 4.99515382e+03, 1.07617327e+04, 2.31854502e+04, 4.99515382e+04]) (Pdb) nu 20.0 (Pdb) lam 499.51538166556196 (Pdb) ncx2.sf(v_T_norm[::50], df=nu, nc=lam) array([ 1.00000000e+00, 1.00000000e+00, 1.00000000e+00, 1.00000000e+00, 1.00000000e+00, 1.00000000e+00, 6.64666099e-01, 4.16325934e-05, 4.16325934e-05, 4.16325934e-05, 4.16325934e-05, 4.16325934e-05, 4.16325934e-05])
defect
scipy sf should be monotone decreasing i m getting unexpected behavior from scipy sf the survival function should eventually go to zero for large enough inputs but instead i m seeing a kind of foot i e it is bottoming out at a positive value the following interactive i o demonstrates the issue reproducing code example pdb v t norm array pdb nu pdb lam pdb sf v t norm df nu nc lam array
1
49,345
13,186,625,418
IssuesEvent
2020-08-13 00:47:21
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[clast] TODOs (Trac #1207)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1207">https://code.icecube.wisc.edu/ticket/1207</a>, reported by david.schultz and owned by markw04</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:55", "description": "Looks like there is more to be said. Let's fill that in or make some tickets.\n\n{{{\npublic/clast/I3CLastModule.h: * @todo There is probably more to be said here as David B. would put it.\npublic/clast/I3CLastCalculator.h: * @todo There is probably more to be said here as David B. would put it.\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1550067295757382", "component": "combo reconstruction", "summary": "[clast] TODOs", "priority": "critical", "keywords": "", "time": "2015-08-19T19:25:33", "milestone": "", "owner": "markw04", "type": "defect" } ``` </p> </details>
1.0
[clast] TODOs (Trac #1207) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1207">https://code.icecube.wisc.edu/ticket/1207</a>, reported by david.schultz and owned by markw04</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:55", "description": "Looks like there is more to be said. Let's fill that in or make some tickets.\n\n{{{\npublic/clast/I3CLastModule.h: * @todo There is probably more to be said here as David B. would put it.\npublic/clast/I3CLastCalculator.h: * @todo There is probably more to be said here as David B. would put it.\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1550067295757382", "component": "combo reconstruction", "summary": "[clast] TODOs", "priority": "critical", "keywords": "", "time": "2015-08-19T19:25:33", "milestone": "", "owner": "markw04", "type": "defect" } ``` </p> </details>
defect
todos trac migrated from json status closed changetime description looks like there is more to be said let s fill that in or make some tickets n n npublic clast h todo there is probably more to be said here as david b would put it npublic clast h todo there is probably more to be said here as david b would put it n reporter david schultz cc resolution fixed ts component combo reconstruction summary todos priority critical keywords time milestone owner type defect
1
278,493
21,082,111,596
IssuesEvent
2022-04-03 03:25:44
nartc/mapper
https://api.github.com/repos/nartc/mapper
closed
Please document ts/node version requirements
documentation enhancement
### Is your feature request related to a problem? Please describe. Many companies are still stuck using Typescript 3.9, but `AutoMapper` does not compile properly when using this older version of typescript. Errors include things like: - `node_modules/@automapper/core/src/lib/types/core.d.ts:149:24 - error TS1005: ',' expected.` - `node_modules/@automapper/core/src/lib/types/core.d.ts:150:7 - error TS1011: An element access expression should take an argument.` - `node_modules/@automapper/core/src/lib/types/core.d.ts:151:1 - error TS1109: Expression expected.` ### Describe the solution you'd like These compile errors go away with `v4.1.x`, so it seems this is an issue with the typescript version rather than a problem with `AutoMapper` itself. It would be nice to put minimum version requirements somewhere in the docs, if not in the `package.json` file itself. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
1.0
Please document ts/node version requirements - ### Is your feature request related to a problem? Please describe. Many companies are still stuck using Typescript 3.9, but `AutoMapper` does not compile properly when using this older version of typescript. Errors include things like: - `node_modules/@automapper/core/src/lib/types/core.d.ts:149:24 - error TS1005: ',' expected.` - `node_modules/@automapper/core/src/lib/types/core.d.ts:150:7 - error TS1011: An element access expression should take an argument.` - `node_modules/@automapper/core/src/lib/types/core.d.ts:151:1 - error TS1109: Expression expected.` ### Describe the solution you'd like These compile errors go away with `v4.1.x`, so it seems this is an issue with the typescript version rather than a problem with `AutoMapper` itself. It would be nice to put minimum version requirements somewhere in the docs, if not in the `package.json` file itself. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
non_defect
please document ts node version requirements is your feature request related to a problem please describe many companies are still stuck using typescript but automapper does not compile properly when using this older version of typescript errors include things like node modules automapper core src lib types core d ts error expected node modules automapper core src lib types core d ts error an element access expression should take an argument node modules automapper core src lib types core d ts error expression expected describe the solution you d like these compile errors go away with x so it seems this is an issue with the typescript version rather than a problem with automapper itself it would be nice to put minimum version requirements somewhere in the docs if not in the package json file itself describe alternatives you ve considered no response additional context no response
0
69,995
22,781,572,905
IssuesEvent
2022-07-08 20:28:24
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: `sparse.linalg` docstrings
defect scipy.sparse.linalg Documentation
### Describe your issue. Reported by Tyler in https://github.com/scipy/scipy/pull/16082#issuecomment-1114052991: ``` =================== scipy.sparse.linalg =================== scipy.sparse.linalg.bicgstab ---------------------------- File "build/testenv/lib/python3.9/site-packages/scipy/sparse/linalg/_isolve/iterative.py", line 62, in bicgstab Failed example: np.allclose(A.dot(x), b) Expected: True Got: False ``` The failure seems to be sporadic and/or platform dependent (cannot repro locally). A few things to check / fix: - The web docs, https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.linalg.bicgstab.html, do not show a docstring example at all - The source, https://github.com/scipy/scipy/blob/main/scipy/sparse/linalg/_isolve/iterative.py#L215, shows some non-standard manipulations with docstrings --- which seem to include some usage examples. - These examples use random numbers *without seeding*. So that might be the cause of sporadic failures? - The web docs do not have a `source` link.
1.0
BUG: `sparse.linalg` docstrings - ### Describe your issue. Reported by Tyler in https://github.com/scipy/scipy/pull/16082#issuecomment-1114052991: ``` =================== scipy.sparse.linalg =================== scipy.sparse.linalg.bicgstab ---------------------------- File "build/testenv/lib/python3.9/site-packages/scipy/sparse/linalg/_isolve/iterative.py", line 62, in bicgstab Failed example: np.allclose(A.dot(x), b) Expected: True Got: False ``` The failure seems to be sporadic and/or platform dependent (cannot repro locally). A few things to check / fix: - The web docs, https://docs.scipy.org/doc/scipy/reference/generated/scipy.sparse.linalg.bicgstab.html, do not show a docstring example at all - The source, https://github.com/scipy/scipy/blob/main/scipy/sparse/linalg/_isolve/iterative.py#L215, shows some non-standard manipulations with docstrings --- which seem to include some usage examples. - These examples use random numbers *without seeding*. So that might be the cause of sporadic failures? - The web docs do not have a `source` link.
defect
bug sparse linalg docstrings describe your issue reported by tyler in scipy sparse linalg scipy sparse linalg bicgstab file build testenv lib site packages scipy sparse linalg isolve iterative py line in bicgstab failed example np allclose a dot x b expected true got false the failure seems to be sporadic and or platform dependent cannot repro locally a few things to check fix the web docs do not show a docstring example at all the source shows some non standard manipulations with docstrings which seem to include some usage examples these examples use random numbers without seeding so that might be the cause of sporadic failures the web docs do not have a source link
1
114,637
17,227,068,567
IssuesEvent
2021-07-20 04:24:45
atlslscsrv-app/package.json-yarn.lock
https://api.github.com/repos/atlslscsrv-app/package.json-yarn.lock
closed
CVE-2020-7662 (High) detected in websocket-extensions-0.1.3.tgz - autoclosed
security vulnerability
## CVE-2020-7662 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>websocket-extensions-0.1.3.tgz</b></p></summary> <p>Generic extension manager for WebSocket connections</p> <p>Library home page: <a href="https://registry.npmjs.org/websocket-extensions/-/websocket-extensions-0.1.3.tgz">https://registry.npmjs.org/websocket-extensions/-/websocket-extensions-0.1.3.tgz</a></p> <p>Path to dependency file: package.json-yarn.lock/package.json</p> <p>Path to vulnerable library: package.json-yarn.lock/node_modules/websocket-extensions/package.json</p> <p> Dependency Hierarchy: - live-server-1.2.0.tgz (Root Library) - faye-websocket-0.11.1.tgz - websocket-driver-0.7.0.tgz - :x: **websocket-extensions-0.1.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/atlslscsrv-app/package.json-yarn.lock/commits/7cfdab90cc50359bbbd31f931845f2b1770973b5">7cfdab90cc50359bbbd31f931845f2b1770973b5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> websocket-extensions npm module prior to 0.1.4 allows Denial of Service (DoS) via Regex Backtracking. The extension parser may take quadratic time when parsing a header containing an unclosed string parameter value whose content is a repeating two-byte sequence of a backslash and some other character. This could be abused by an attacker to conduct Regex Denial Of Service (ReDoS) on a single-threaded server by providing a malicious payload with the Sec-WebSocket-Extensions header. <p>Publish Date: 2020-06-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7662>CVE-2020-7662</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-g78m-2chm-r7qv">https://github.com/advisories/GHSA-g78m-2chm-r7qv</a></p> <p>Release Date: 2020-06-02</p> <p>Fix Resolution: websocket-extensions - 0.1.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7662 (High) detected in websocket-extensions-0.1.3.tgz - autoclosed - ## CVE-2020-7662 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>websocket-extensions-0.1.3.tgz</b></p></summary> <p>Generic extension manager for WebSocket connections</p> <p>Library home page: <a href="https://registry.npmjs.org/websocket-extensions/-/websocket-extensions-0.1.3.tgz">https://registry.npmjs.org/websocket-extensions/-/websocket-extensions-0.1.3.tgz</a></p> <p>Path to dependency file: package.json-yarn.lock/package.json</p> <p>Path to vulnerable library: package.json-yarn.lock/node_modules/websocket-extensions/package.json</p> <p> Dependency Hierarchy: - live-server-1.2.0.tgz (Root Library) - faye-websocket-0.11.1.tgz - websocket-driver-0.7.0.tgz - :x: **websocket-extensions-0.1.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/atlslscsrv-app/package.json-yarn.lock/commits/7cfdab90cc50359bbbd31f931845f2b1770973b5">7cfdab90cc50359bbbd31f931845f2b1770973b5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> websocket-extensions npm module prior to 0.1.4 allows Denial of Service (DoS) via Regex Backtracking. The extension parser may take quadratic time when parsing a header containing an unclosed string parameter value whose content is a repeating two-byte sequence of a backslash and some other character. This could be abused by an attacker to conduct Regex Denial Of Service (ReDoS) on a single-threaded server by providing a malicious payload with the Sec-WebSocket-Extensions header. <p>Publish Date: 2020-06-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7662>CVE-2020-7662</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-g78m-2chm-r7qv">https://github.com/advisories/GHSA-g78m-2chm-r7qv</a></p> <p>Release Date: 2020-06-02</p> <p>Fix Resolution: websocket-extensions - 0.1.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in websocket extensions tgz autoclosed cve high severity vulnerability vulnerable library websocket extensions tgz generic extension manager for websocket connections library home page a href path to dependency file package json yarn lock package json path to vulnerable library package json yarn lock node modules websocket extensions package json dependency hierarchy live server tgz root library faye websocket tgz websocket driver tgz x websocket extensions tgz vulnerable library found in head commit a href found in base branch master vulnerability details websocket extensions npm module prior to allows denial of service dos via regex backtracking the extension parser may take quadratic time when parsing a header containing an unclosed string parameter value whose content is a repeating two byte sequence of a backslash and some other character this could be abused by an attacker to conduct regex denial of service redos on a single threaded server by providing a malicious payload with the sec websocket extensions header publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution websocket extensions step up your open source security game with whitesource
0
67,868
3,282,378,122
IssuesEvent
2015-10-28 05:54:25
The-Compiler/qutebrowser
https://api.github.com/repos/The-Compiler/qutebrowser
opened
Add a Qt wrapper module
priority: 2 - low qt
This might be useful when [PySide2](https://github.com/PySide/pyside2) is more usable - and we definitely should have it in place before starting with the plugin API. We also could simplify `@pyqtSlot` by using type annotations, like in [this example](https://gist.github.com/empyrical/a103c46f454a433e8798) (we probably don't need all the overloading logic).
1.0
Add a Qt wrapper module - This might be useful when [PySide2](https://github.com/PySide/pyside2) is more usable - and we definitely should have it in place before starting with the plugin API. We also could simplify `@pyqtSlot` by using type annotations, like in [this example](https://gist.github.com/empyrical/a103c46f454a433e8798) (we probably don't need all the overloading logic).
non_defect
add a qt wrapper module this might be useful when is more usable and we definitely should have it in place before starting with the plugin api we also could simplify pyqtslot by using type annotations like in we probably don t need all the overloading logic
0
26,138
7,786,436,413
IssuesEvent
2018-06-06 18:56:42
habitat-sh/builder
https://api.github.com/repos/habitat-sh/builder
opened
Make auto-rebuild of packages opt-in
A-builder C-feature V-bldr V-front-end
Currently, once a plan is connected to builder, it will get auto-rebuilt any time a dependency re-builds. While desirable for things like enterprise packages, it may not always be what is appropriate (eg, if you have a demo app that depends on core/node..). We should make the auto-rebuild capability opt-in, and default it to off. From a UI standpoint, this would be one more setting in the Settings page, that could go before or after the Public / Private package option.
1.0
Make auto-rebuild of packages opt-in - Currently, once a plan is connected to builder, it will get auto-rebuilt any time a dependency re-builds. While desirable for things like enterprise packages, it may not always be what is appropriate (eg, if you have a demo app that depends on core/node..). We should make the auto-rebuild capability opt-in, and default it to off. From a UI standpoint, this would be one more setting in the Settings page, that could go before or after the Public / Private package option.
non_defect
make auto rebuild of packages opt in currently once a plan is connected to builder it will get auto rebuilt any time a dependency re builds while desirable for things like enterprise packages it may not always be what is appropriate eg if you have a demo app that depends on core node we should make the auto rebuild capability opt in and default it to off from a ui standpoint this would be one more setting in the settings page that could go before or after the public private package option
0
50,658
13,187,666,612
IssuesEvent
2020-08-13 04:10:03
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
neutrino-generator index.dox should be converted to index.rst (Trac #1146)
Migrated from Trac combo simulation defect
Convert the Doxygen created index.dox file to a Sphinx documentation file index.rst. Only Sphinx (.rst) files get build as part of the online documentation. This is why currently there is no documentation for neutrino-generator to be found at http://software.icecube.wisc.edu/simulation_trunk/ <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1146">https://code.icecube.wisc.edu/ticket/1146</a>, reported by jtatar and owned by Kotoyo Hoshina</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:13:35", "description": "Convert the Doxygen created index.dox file to a Sphinx documentation file index.rst. Only Sphinx (.rst) files get build as part of the online documentation. This is why currently there is no documentation for neutrino-generator to be found at http://software.icecube.wisc.edu/simulation_trunk/", "reporter": "jtatar", "cc": "", "resolution": "fixed", "_ts": "1550067215093672", "component": "combo simulation", "summary": "neutrino-generator index.dox should be converted to index.rst", "priority": "blocker", "keywords": "", "time": "2015-08-18T00:13:26", "milestone": "", "owner": "Kotoyo Hoshina", "type": "defect" } ``` </p> </details>
1.0
neutrino-generator index.dox should be converted to index.rst (Trac #1146) - Convert the Doxygen created index.dox file to a Sphinx documentation file index.rst. Only Sphinx (.rst) files get build as part of the online documentation. This is why currently there is no documentation for neutrino-generator to be found at http://software.icecube.wisc.edu/simulation_trunk/ <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1146">https://code.icecube.wisc.edu/ticket/1146</a>, reported by jtatar and owned by Kotoyo Hoshina</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:13:35", "description": "Convert the Doxygen created index.dox file to a Sphinx documentation file index.rst. Only Sphinx (.rst) files get build as part of the online documentation. This is why currently there is no documentation for neutrino-generator to be found at http://software.icecube.wisc.edu/simulation_trunk/", "reporter": "jtatar", "cc": "", "resolution": "fixed", "_ts": "1550067215093672", "component": "combo simulation", "summary": "neutrino-generator index.dox should be converted to index.rst", "priority": "blocker", "keywords": "", "time": "2015-08-18T00:13:26", "milestone": "", "owner": "Kotoyo Hoshina", "type": "defect" } ``` </p> </details>
defect
neutrino generator index dox should be converted to index rst trac convert the doxygen created index dox file to a sphinx documentation file index rst only sphinx rst files get build as part of the online documentation this is why currently there is no documentation for neutrino generator to be found at migrated from json status closed changetime description convert the doxygen created index dox file to a sphinx documentation file index rst only sphinx rst files get build as part of the online documentation this is why currently there is no documentation for neutrino generator to be found at reporter jtatar cc resolution fixed ts component combo simulation summary neutrino generator index dox should be converted to index rst priority blocker keywords time milestone owner kotoyo hoshina type defect
1
7,507
2,610,392,056
IssuesEvent
2015-02-26 20:06:49
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Remapped mouse buttons do not work in fullscreen mode
auto-migrated Priority-Medium Type-Defect
``` I use trackball and remap buttons following way: Section "InputDevice" Identifier "Configured Mouse" Driver "evdev" Option "CorePointer" Option "Device" "/dev/input/by-id/usb-Logitech_USB_Trackball-event-mouse" Option "Protocol" "Auto" Option "Buttons" "9" Option "EmulateWheel" "True" Option "EmulateWheelButton" "3" Option "XAxisMapping" "6 7" Option "YAxisMapping" "4 5" Option "ZAxisMapping" "4 5" Option "ButtonMapping" "1 8 9 4 5 6 7 2 3" EndSection The game works well in window mode, but doesn't recognize right button clicks in fullscreen mode. ``` ----- Original issue reported on code.google.com by `andy1...@gmail.com` on 17 Jun 2012 at 6:56
1.0
Remapped mouse buttons do not work in fullscreen mode - ``` I use trackball and remap buttons following way: Section "InputDevice" Identifier "Configured Mouse" Driver "evdev" Option "CorePointer" Option "Device" "/dev/input/by-id/usb-Logitech_USB_Trackball-event-mouse" Option "Protocol" "Auto" Option "Buttons" "9" Option "EmulateWheel" "True" Option "EmulateWheelButton" "3" Option "XAxisMapping" "6 7" Option "YAxisMapping" "4 5" Option "ZAxisMapping" "4 5" Option "ButtonMapping" "1 8 9 4 5 6 7 2 3" EndSection The game works well in window mode, but doesn't recognize right button clicks in fullscreen mode. ``` ----- Original issue reported on code.google.com by `andy1...@gmail.com` on 17 Jun 2012 at 6:56
defect
remapped mouse buttons do not work in fullscreen mode i use trackball and remap buttons following way section inputdevice identifier configured mouse driver evdev option corepointer option device dev input by id usb logitech usb trackball event mouse option protocol auto option buttons option emulatewheel true option emulatewheelbutton option xaxismapping option yaxismapping option zaxismapping option buttonmapping endsection the game works well in window mode but doesn t recognize right button clicks in fullscreen mode original issue reported on code google com by gmail com on jun at
1
1,772
2,603,971,115
IssuesEvent
2015-02-24 19:00:16
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳阴茎有颗粒怎么回事
auto-migrated Priority-Medium Type-Defect
``` 沈阳阴茎有颗粒怎么回事〓沈陽軍區政治部醫院性病〓TEL:02 4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療� ��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:31
1.0
沈阳阴茎有颗粒怎么回事 - ``` 沈阳阴茎有颗粒怎么回事〓沈陽軍區政治部醫院性病〓TEL:02 4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療� ��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 7:31
defect
沈阳阴茎有颗粒怎么回事 沈阳阴茎有颗粒怎么回事〓沈陽軍區政治部醫院性病〓tel: 〓 , � �� 。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 original issue reported on code google com by gmail com on jun at
1
63,208
17,463,155,799
IssuesEvent
2021-08-06 13:24:17
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
SEVERE: HazelcastSerializationException: Problem while reading IdentifiedDataSerializable, namespace: -18, classId: 6, exception: null
Type: Defect Team: Core Source: Internal Module: IMap
http://jenkins.hazelcast.com/view/rollup/job/rollup-x2/8/console /disk1/workspace/rollup-x2/5.0-SNAPSHOT/2021_07_14-10_56_13/projection Failed fail HzClient5HZ _idx-heavy_sql_queryTargetTxnId_mapBak1HDSqlIdx hzcmd.map.sql.predicate.txn.model.queryTargetTxnId threadId=7 global.AssertionException: map mapBak1HDSqlIdx_sql_idx-heavy key=5411 value==null ``` SEVERE: [10.0.0.76]:5701 [HZ] [4.2] Failed to process: Packet{partitionId=23, frameLength=53229, conn=Connection[id=14, /10.0.0.76:5701->/10.0.0.42:38173, qualifier=null, endpoint=[10.0.0.42]:5701, alive=true, connectionType=MEMBER, planeIndex=0], rawFlags=10010001, isUrgent=true, packetType=OPERATION, typeSpecificFlags=[isResponse=false, isOpControl=false]} on: hz.affectionate_rubin.partition-operation.thread-3 com.hazelcast.nio.serialization.HazelcastSerializationException: Problem while reading IdentifiedDataSerializable, namespace: -18, classId: 6, exception: null at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.rethrowIdsReadException(EnterpriseDataSerializableSerializer.java:315) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.readIdentifiedDataSerializable(EnterpriseDataSerializableSerializer.java:168) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.doRead(EnterpriseDataSerializableSerializer.java:130) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:124) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:52) at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44) at com.hazelcast.internal.serialization.impl.AbstractSerializationService.readObject(AbstractSerializationService.java:289) at com.hazelcast.internal.serialization.impl.ByteArrayObjectDataInput.readObject(ByteArrayObjectDataInput.java:592) at com.hazelcast.internal.partition.ReplicaFragmentMigrationState.readData(ReplicaFragmentMigrationState.java:97) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.readIdentifiedDataSerializable(EnterpriseDataSerializableSerializer.java:161) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.doRead(EnterpriseDataSerializableSerializer.java:130) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:124) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:52) at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44) at com.hazelcast.internal.serialization.impl.AbstractSerializationService.readObject(AbstractSerializationService.java:289) at com.hazelcast.internal.serialization.impl.ByteArrayObjectDataInput.readObject(ByteArrayObjectDataInput.java:592) at com.hazelcast.internal.partition.operation.MigrationOperation.readInternal(MigrationOperation.java:249) at com.hazelcast.spi.impl.operationservice.Operation.readData(Operation.java:755) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.readIdentifiedDataSerializable(EnterpriseDataSerializableSerializer.java:161) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.doRead(EnterpriseDataSerializableSerializer.java:130) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:124) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:52) at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44) at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toObject(AbstractSerializationService.java:208) at com.hazelcast.spi.impl.NodeEngineImpl.toObject(NodeEngineImpl.java:355) at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:454) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:197) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:137) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.executeRun(OperationThread.java:123) at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) Caused by: java.lang.NullPointerException at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.readIdentifiedDataSerializable(EnterpriseDataSerializableSerializer.java:161) ... 28 more ```
1.0
SEVERE: HazelcastSerializationException: Problem while reading IdentifiedDataSerializable, namespace: -18, classId: 6, exception: null - http://jenkins.hazelcast.com/view/rollup/job/rollup-x2/8/console /disk1/workspace/rollup-x2/5.0-SNAPSHOT/2021_07_14-10_56_13/projection Failed fail HzClient5HZ _idx-heavy_sql_queryTargetTxnId_mapBak1HDSqlIdx hzcmd.map.sql.predicate.txn.model.queryTargetTxnId threadId=7 global.AssertionException: map mapBak1HDSqlIdx_sql_idx-heavy key=5411 value==null ``` SEVERE: [10.0.0.76]:5701 [HZ] [4.2] Failed to process: Packet{partitionId=23, frameLength=53229, conn=Connection[id=14, /10.0.0.76:5701->/10.0.0.42:38173, qualifier=null, endpoint=[10.0.0.42]:5701, alive=true, connectionType=MEMBER, planeIndex=0], rawFlags=10010001, isUrgent=true, packetType=OPERATION, typeSpecificFlags=[isResponse=false, isOpControl=false]} on: hz.affectionate_rubin.partition-operation.thread-3 com.hazelcast.nio.serialization.HazelcastSerializationException: Problem while reading IdentifiedDataSerializable, namespace: -18, classId: 6, exception: null at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.rethrowIdsReadException(EnterpriseDataSerializableSerializer.java:315) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.readIdentifiedDataSerializable(EnterpriseDataSerializableSerializer.java:168) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.doRead(EnterpriseDataSerializableSerializer.java:130) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:124) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:52) at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44) at com.hazelcast.internal.serialization.impl.AbstractSerializationService.readObject(AbstractSerializationService.java:289) at com.hazelcast.internal.serialization.impl.ByteArrayObjectDataInput.readObject(ByteArrayObjectDataInput.java:592) at com.hazelcast.internal.partition.ReplicaFragmentMigrationState.readData(ReplicaFragmentMigrationState.java:97) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.readIdentifiedDataSerializable(EnterpriseDataSerializableSerializer.java:161) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.doRead(EnterpriseDataSerializableSerializer.java:130) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:124) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:52) at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44) at com.hazelcast.internal.serialization.impl.AbstractSerializationService.readObject(AbstractSerializationService.java:289) at com.hazelcast.internal.serialization.impl.ByteArrayObjectDataInput.readObject(ByteArrayObjectDataInput.java:592) at com.hazelcast.internal.partition.operation.MigrationOperation.readInternal(MigrationOperation.java:249) at com.hazelcast.spi.impl.operationservice.Operation.readData(Operation.java:755) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.readIdentifiedDataSerializable(EnterpriseDataSerializableSerializer.java:161) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.doRead(EnterpriseDataSerializableSerializer.java:130) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:124) at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.read(EnterpriseDataSerializableSerializer.java:52) at com.hazelcast.internal.serialization.impl.StreamSerializerAdapter.read(StreamSerializerAdapter.java:44) at com.hazelcast.internal.serialization.impl.AbstractSerializationService.toObject(AbstractSerializationService.java:208) at com.hazelcast.spi.impl.NodeEngineImpl.toObject(NodeEngineImpl.java:355) at com.hazelcast.spi.impl.operationservice.impl.OperationRunnerImpl.run(OperationRunnerImpl.java:454) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:197) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.process(OperationThread.java:137) at com.hazelcast.spi.impl.operationexecutor.impl.OperationThread.executeRun(OperationThread.java:123) at com.hazelcast.internal.util.executor.HazelcastManagedThread.run(HazelcastManagedThread.java:102) Caused by: java.lang.NullPointerException at com.hazelcast.internal.serialization.impl.EnterpriseDataSerializableSerializer.readIdentifiedDataSerializable(EnterpriseDataSerializableSerializer.java:161) ... 28 more ```
defect
severe hazelcastserializationexception problem while reading identifieddataserializable namespace classid exception null workspace rollup snapshot projection failed fail idx heavy sql querytargettxnid hzcmd map sql predicate txn model querytargettxnid threadid global assertionexception map sql idx heavy key value null severe failed to process packet partitionid framelength conn connection alive true connectiontype member planeindex rawflags isurgent true packettype operation typespecificflags on hz affectionate rubin partition operation thread com hazelcast nio serialization hazelcastserializationexception problem while reading identifieddataserializable namespace classid exception null at com hazelcast internal serialization impl enterprisedataserializableserializer rethrowidsreadexception enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer readidentifieddataserializable enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer doread enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer read enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer read enterprisedataserializableserializer java at com hazelcast internal serialization impl streamserializeradapter read streamserializeradapter java at com hazelcast internal serialization impl abstractserializationservice readobject abstractserializationservice java at com hazelcast internal serialization impl bytearrayobjectdatainput readobject bytearrayobjectdatainput java at com hazelcast internal partition replicafragmentmigrationstate readdata replicafragmentmigrationstate java at com hazelcast internal serialization impl enterprisedataserializableserializer readidentifieddataserializable enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer doread enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer read enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer read enterprisedataserializableserializer java at com hazelcast internal serialization impl streamserializeradapter read streamserializeradapter java at com hazelcast internal serialization impl abstractserializationservice readobject abstractserializationservice java at com hazelcast internal serialization impl bytearrayobjectdatainput readobject bytearrayobjectdatainput java at com hazelcast internal partition operation migrationoperation readinternal migrationoperation java at com hazelcast spi impl operationservice operation readdata operation java at com hazelcast internal serialization impl enterprisedataserializableserializer readidentifieddataserializable enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer doread enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer read enterprisedataserializableserializer java at com hazelcast internal serialization impl enterprisedataserializableserializer read enterprisedataserializableserializer java at com hazelcast internal serialization impl streamserializeradapter read streamserializeradapter java at com hazelcast internal serialization impl abstractserializationservice toobject abstractserializationservice java at com hazelcast spi impl nodeengineimpl toobject nodeengineimpl java at com hazelcast spi impl operationservice impl operationrunnerimpl run operationrunnerimpl java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread process operationthread java at com hazelcast spi impl operationexecutor impl operationthread executerun operationthread java at com hazelcast internal util executor hazelcastmanagedthread run hazelcastmanagedthread java caused by java lang nullpointerexception at com hazelcast internal serialization impl enterprisedataserializableserializer readidentifieddataserializable enterprisedataserializableserializer java more
1