Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
435,854
12,542,339,935
IssuesEvent
2020-06-05 13:53:35
UltimateCodeMonkeys/CodeMonkeysMVVM
https://api.github.com/repos/UltimateCodeMonkeys/CodeMonkeysMVVM
opened
Support async conditions for WPF and Xamarin.Forms navigation
Feature Request Navigation Navigation.Forms Navigation.WPF Priority: Low
Instead of `Func<bool>`, it would be nice to have `Func<Task<bool>>` in order to be able to pass async method calls as `ShowAsync` condition.
1.0
Support async conditions for WPF and Xamarin.Forms navigation - Instead of `Func<bool>`, it would be nice to have `Func<Task<bool>>` in order to be able to pass async method calls as `ShowAsync` condition.
non_defect
support async conditions for wpf and xamarin forms navigation instead of func it would be nice to have func in order to be able to pass async method calls as showasync condition
0
243,139
26,277,933,754
IssuesEvent
2023-01-07 01:32:21
tlkh/serverless-transformers
https://api.github.com/repos/tlkh/serverless-transformers
opened
CVE-2022-40897 (Medium) detected in setuptools-44.1.1-py2.py3-none-any.whl
security vulnerability
## CVE-2022-40897 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>setuptools-44.1.1-py2.py3-none-any.whl</b></p></summary> <p>Easily download, build, install, upgrade, and uninstall Python packages</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e1/b7/182161210a13158cd3ccc41ee19aadef54496b74f2817cc147006ec932b4/setuptools-44.1.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/e1/b7/182161210a13158cd3ccc41ee19aadef54496b74f2817cc147006ec932b4/setuptools-44.1.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt,/backend/requirements.txt</p> <p> Dependency Hierarchy: - streamlit-0.55.2-py2.py3-none-any.whl (Root Library) - altair-3.3.0-py2.py3-none-any.whl - jsonschema-3.2.0-py2.py3-none-any.whl - :x: **setuptools-44.1.1-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Python Packaging Authority (PyPA) setuptools before 65.5.1 allows remote attackers to cause a denial of service via HTML in a crafted package or custom PackageIndex page. There is a Regular Expression Denial of Service (ReDoS) in package_index.py. <p>Publish Date: 2022-12-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40897>CVE-2022-40897</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pyup.io/posts/pyup-discovers-redos-vulnerabilities-in-top-python-packages/">https://pyup.io/posts/pyup-discovers-redos-vulnerabilities-in-top-python-packages/</a></p> <p>Release Date: 2022-12-23</p> <p>Fix Resolution: setuptools - 65.5.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-40897 (Medium) detected in setuptools-44.1.1-py2.py3-none-any.whl - ## CVE-2022-40897 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>setuptools-44.1.1-py2.py3-none-any.whl</b></p></summary> <p>Easily download, build, install, upgrade, and uninstall Python packages</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e1/b7/182161210a13158cd3ccc41ee19aadef54496b74f2817cc147006ec932b4/setuptools-44.1.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/e1/b7/182161210a13158cd3ccc41ee19aadef54496b74f2817cc147006ec932b4/setuptools-44.1.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt,/backend/requirements.txt</p> <p> Dependency Hierarchy: - streamlit-0.55.2-py2.py3-none-any.whl (Root Library) - altair-3.3.0-py2.py3-none-any.whl - jsonschema-3.2.0-py2.py3-none-any.whl - :x: **setuptools-44.1.1-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Python Packaging Authority (PyPA) setuptools before 65.5.1 allows remote attackers to cause a denial of service via HTML in a crafted package or custom PackageIndex page. There is a Regular Expression Denial of Service (ReDoS) in package_index.py. <p>Publish Date: 2022-12-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40897>CVE-2022-40897</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://pyup.io/posts/pyup-discovers-redos-vulnerabilities-in-top-python-packages/">https://pyup.io/posts/pyup-discovers-redos-vulnerabilities-in-top-python-packages/</a></p> <p>Release Date: 2022-12-23</p> <p>Fix Resolution: setuptools - 65.5.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in setuptools none any whl cve medium severity vulnerability vulnerable library setuptools none any whl easily download build install upgrade and uninstall python packages library home page a href path to dependency file requirements txt path to vulnerable library requirements txt backend requirements txt dependency hierarchy streamlit none any whl root library altair none any whl jsonschema none any whl x setuptools none any whl vulnerable library found in base branch main vulnerability details python packaging authority pypa setuptools before allows remote attackers to cause a denial of service via html in a crafted package or custom packageindex page there is a regular expression denial of service redos in package index py publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution setuptools step up your open source security game with mend
0
43,115
11,489,222,572
IssuesEvent
2020-02-11 15:10:49
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[SCREENREADER, KEYBOARD]: Focus on page load is on the breadcrumbs, but needs to be at the top of the page
508-defect-0 508-issue-cognition 508-issue-focus-mgmt 508/Accessibility vsa vsa-ebenefits
# [SCREENREADER, KEYBOARD]: Focus on page load is on the breadcrumbs, but needs to be at the top of the page ## Description When a screen reader or keyboard user first arrives on this page the focus is set on the breadcrumbs. This is different from other static page behavior and is disorienting. The user needs to know about the navigation area above, so the focus should start at the top, as with other static pages. ## Point of Contact Jennifer / Trevor ## Acceptance Criteria As a keyboard user, I want to be able to navigate from the top of the entire page when it loads. ## Environment * Browser: Safari * Screenreader: VoiceOver ## Steps to Recreate 1. Enter `https://staging.va.gov/disability/view-disability-rating/rating` in browser 2. Start screenreading device listed in Environment 3. Reload page 4. Verify that the focus starts on the breadcrumbs ## Possible Fixes (optional) Right now, this isn't in an application flow per se, it's more like a static content page. I'd like to remove the focus() from the breadcrumbs on first load. This is likely happening in the React container's componentDidMount() lifecycle method. Since it's static, we want assistive tech users to be able to interact with the entire page. When this gets built out to include editing or more features, then we would go ahead and set focus on the `<h1>` when we change routes using React.
1.0
[SCREENREADER, KEYBOARD]: Focus on page load is on the breadcrumbs, but needs to be at the top of the page - # [SCREENREADER, KEYBOARD]: Focus on page load is on the breadcrumbs, but needs to be at the top of the page ## Description When a screen reader or keyboard user first arrives on this page the focus is set on the breadcrumbs. This is different from other static page behavior and is disorienting. The user needs to know about the navigation area above, so the focus should start at the top, as with other static pages. ## Point of Contact Jennifer / Trevor ## Acceptance Criteria As a keyboard user, I want to be able to navigate from the top of the entire page when it loads. ## Environment * Browser: Safari * Screenreader: VoiceOver ## Steps to Recreate 1. Enter `https://staging.va.gov/disability/view-disability-rating/rating` in browser 2. Start screenreading device listed in Environment 3. Reload page 4. Verify that the focus starts on the breadcrumbs ## Possible Fixes (optional) Right now, this isn't in an application flow per se, it's more like a static content page. I'd like to remove the focus() from the breadcrumbs on first load. This is likely happening in the React container's componentDidMount() lifecycle method. Since it's static, we want assistive tech users to be able to interact with the entire page. When this gets built out to include editing or more features, then we would go ahead and set focus on the `<h1>` when we change routes using React.
defect
focus on page load is on the breadcrumbs but needs to be at the top of the page focus on page load is on the breadcrumbs but needs to be at the top of the page description when a screen reader or keyboard user first arrives on this page the focus is set on the breadcrumbs this is different from other static page behavior and is disorienting the user needs to know about the navigation area above so the focus should start at the top as with other static pages point of contact jennifer trevor acceptance criteria as a keyboard user i want to be able to navigate from the top of the entire page when it loads environment browser safari screenreader voiceover steps to recreate enter in browser start screenreading device listed in environment reload page verify that the focus starts on the breadcrumbs possible fixes optional right now this isn t in an application flow per se it s more like a static content page i d like to remove the focus from the breadcrumbs on first load this is likely happening in the react container s componentdidmount lifecycle method since it s static we want assistive tech users to be able to interact with the entire page when this gets built out to include editing or more features then we would go ahead and set focus on the when we change routes using react
1
545,296
15,947,667,595
IssuesEvent
2021-04-15 04:04:38
AlaskaAirlines/auro-radio
https://api.github.com/repos/AlaskaAirlines/auro-radio
closed
auro-radio: incorrect a11y experience
Priority: High Status: Complete and Ready to Merge Status: Review Needed Type: Bug
## Describe the bug Not sure if this [pull request](https://github.com/AlaskaAirlines/WebCoreStyleSheets/pull/66) is the cause, but it's been discovered that when selecting a radio button option on a mobile device, there is a blue halo. This is only appearing with a radio-group. This is not the expected experience when working with an Auro element. ## To Reproduce Steps to reproduce the behavior: 1. Go to https://auro.alaskaair.com/components/auro/radio 1. Click on any option in a radio group 1. See the error ## Expected behavior When tapping on a radio option, there should be no halo. ## Screenshots <img width="557" alt="Screen Shot 2021-03-22 at 9 09 05 AM" src="https://user-images.githubusercontent.com/181089/112021225-4b977a00-8aee-11eb-9cd4-67f3f6592fc3.png"> ## Smartphone (please complete the following information): - OS: iOS - Browser: Safari - Latest version
1.0
auro-radio: incorrect a11y experience - ## Describe the bug Not sure if this [pull request](https://github.com/AlaskaAirlines/WebCoreStyleSheets/pull/66) is the cause, but it's been discovered that when selecting a radio button option on a mobile device, there is a blue halo. This is only appearing with a radio-group. This is not the expected experience when working with an Auro element. ## To Reproduce Steps to reproduce the behavior: 1. Go to https://auro.alaskaair.com/components/auro/radio 1. Click on any option in a radio group 1. See the error ## Expected behavior When tapping on a radio option, there should be no halo. ## Screenshots <img width="557" alt="Screen Shot 2021-03-22 at 9 09 05 AM" src="https://user-images.githubusercontent.com/181089/112021225-4b977a00-8aee-11eb-9cd4-67f3f6592fc3.png"> ## Smartphone (please complete the following information): - OS: iOS - Browser: Safari - Latest version
non_defect
auro radio incorrect experience describe the bug not sure if this is the cause but it s been discovered that when selecting a radio button option on a mobile device there is a blue halo this is only appearing with a radio group this is not the expected experience when working with an auro element to reproduce steps to reproduce the behavior go to click on any option in a radio group see the error expected behavior when tapping on a radio option there should be no halo screenshots img width alt screen shot at am src smartphone please complete the following information os ios browser safari latest version
0
77,379
26,957,852,410
IssuesEvent
2023-02-08 16:03:02
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Please provide a way to disable update notifications.
T-Defect
### Steps to reproduce It keeps telling me in a popup "Update Element" that "New version of Element is available". I will argue separately that you release too many, non-critical updates that do not **need** to be applied. But even if this were not the case, updates are provisioned through a required package manager (chocolatey) and there will always be a delay between actual updates and availability of updates. Telling the end user that they need to update is completely useless in this case, since we are already using the latest update that is available to us. Therefore, please provide a way to permanently disable these incessant notifications to do something that we literally cannot do. ### Outcome #### What did you expect? To be able to permanently disable update notifications. #### What happened instead? Micromanaging of my desktop with non-actionable, annoying notifications. ### Operating system Windows ### Application version NA ### How did you install the app? Chocolatey ### Homeserver NA ### Will you send logs? No
1.0
Please provide a way to disable update notifications. - ### Steps to reproduce It keeps telling me in a popup "Update Element" that "New version of Element is available". I will argue separately that you release too many, non-critical updates that do not **need** to be applied. But even if this were not the case, updates are provisioned through a required package manager (chocolatey) and there will always be a delay between actual updates and availability of updates. Telling the end user that they need to update is completely useless in this case, since we are already using the latest update that is available to us. Therefore, please provide a way to permanently disable these incessant notifications to do something that we literally cannot do. ### Outcome #### What did you expect? To be able to permanently disable update notifications. #### What happened instead? Micromanaging of my desktop with non-actionable, annoying notifications. ### Operating system Windows ### Application version NA ### How did you install the app? Chocolatey ### Homeserver NA ### Will you send logs? No
defect
please provide a way to disable update notifications steps to reproduce it keeps telling me in a popup update element that new version of element is available i will argue separately that you release too many non critical updates that do not need to be applied but even if this were not the case updates are provisioned through a required package manager chocolatey and there will always be a delay between actual updates and availability of updates telling the end user that they need to update is completely useless in this case since we are already using the latest update that is available to us therefore please provide a way to permanently disable these incessant notifications to do something that we literally cannot do outcome what did you expect to be able to permanently disable update notifications what happened instead micromanaging of my desktop with non actionable annoying notifications operating system windows application version na how did you install the app chocolatey homeserver na will you send logs no
1
63,253
26,322,722,955
IssuesEvent
2023-01-10 02:03:30
finos/legend-studio
https://api.github.com/repos/finos/legend-studio
closed
Feature request: Support a searchable dropdown when adding service owners
Type: Feature Request Studio Core Team Application: Studio Extension: DSL Service
### Similar requests - [X] I have searched and found no existing similar requests ### How are you using Studio? Legend Studio ### What problems are you trying to solve? currently it's tricky to add owners to a service as you need more information on project's members ### Describe the solution you would like to see implemented - [x] Utilize one SDLC API, api/swagger#/Users/getAllUsers, to get all users based on current input ### Describe alternatives you have considered _No response_ ### Documentation, Design, Adoption, Migration Strategy _No response_ ### Contribution - [X] I would like to work on this feature
1.0
Feature request: Support a searchable dropdown when adding service owners - ### Similar requests - [X] I have searched and found no existing similar requests ### How are you using Studio? Legend Studio ### What problems are you trying to solve? currently it's tricky to add owners to a service as you need more information on project's members ### Describe the solution you would like to see implemented - [x] Utilize one SDLC API, api/swagger#/Users/getAllUsers, to get all users based on current input ### Describe alternatives you have considered _No response_ ### Documentation, Design, Adoption, Migration Strategy _No response_ ### Contribution - [X] I would like to work on this feature
non_defect
feature request support a searchable dropdown when adding service owners similar requests i have searched and found no existing similar requests how are you using studio legend studio what problems are you trying to solve currently it s tricky to add owners to a service as you need more information on project s members describe the solution you would like to see implemented utilize one sdlc api api swagger users getallusers to get all users based on current input describe alternatives you have considered no response documentation design adoption migration strategy no response contribution i would like to work on this feature
0
15,789
20,618,666,644
IssuesEvent
2022-03-07 15:27:28
spinalcordtoolbox/spinalcordtoolbox
https://api.github.com/repos/spinalcordtoolbox/spinalcordtoolbox
opened
Stop trying to catch `IsADirectoryError`, and instead check `isdir` first
bug sct_run_batch Windows compatibility
> **Note:** This issue represents a Windows-specific bug that has already been addressed in #3699. A bit of code in `sct_run_batch` tries to "ask for forgiveness, not permission": https://github.com/spinalcordtoolbox/spinalcordtoolbox/blob/2c2a662baf6eadc0c28711886ec354a4c8d14016/spinalcordtoolbox/scripts/sct_run_batch.py#L367-L377 However, `IsADirectoryError` won't be thrown on Windows. Instead, a `PermissionError` will be given by Windows: ``` PermissionError: [Errno 13] Permission denied: 'C:\\Users\\Joshua\\AppData\\Local\\Temp\\pytest-of-Joshua\\pytest-156\\test_only_one_include0\\out' ``` Aside: It's possible that `shutil.SameFileError` is also problematic, but I don't think we have test coverage for that part of the conditional.
True
Stop trying to catch `IsADirectoryError`, and instead check `isdir` first - > **Note:** This issue represents a Windows-specific bug that has already been addressed in #3699. A bit of code in `sct_run_batch` tries to "ask for forgiveness, not permission": https://github.com/spinalcordtoolbox/spinalcordtoolbox/blob/2c2a662baf6eadc0c28711886ec354a4c8d14016/spinalcordtoolbox/scripts/sct_run_batch.py#L367-L377 However, `IsADirectoryError` won't be thrown on Windows. Instead, a `PermissionError` will be given by Windows: ``` PermissionError: [Errno 13] Permission denied: 'C:\\Users\\Joshua\\AppData\\Local\\Temp\\pytest-of-Joshua\\pytest-156\\test_only_one_include0\\out' ``` Aside: It's possible that `shutil.SameFileError` is also problematic, but I don't think we have test coverage for that part of the conditional.
non_defect
stop trying to catch isadirectoryerror and instead check isdir first note this issue represents a windows specific bug that has already been addressed in a bit of code in sct run batch tries to ask for forgiveness not permission however isadirectoryerror won t be thrown on windows instead a permissionerror will be given by windows permissionerror permission denied c users joshua appdata local temp pytest of joshua pytest test only one out aside it s possible that shutil samefileerror is also problematic but i don t think we have test coverage for that part of the conditional
0
172,009
27,220,080,976
IssuesEvent
2023-02-21 04:01:05
aggie-coding-club/Aggies-Computes
https://api.github.com/repos/aggie-coding-club/Aggies-Computes
closed
Plan RREF/REF algorithm
setup algorithm design
requirement: create file for function, create file for test, outline algorithm, determine needed input and outputs. important: create psuedocode/outline using multiline comment above function
1.0
Plan RREF/REF algorithm - requirement: create file for function, create file for test, outline algorithm, determine needed input and outputs. important: create psuedocode/outline using multiline comment above function
non_defect
plan rref ref algorithm requirement create file for function create file for test outline algorithm determine needed input and outputs important create psuedocode outline using multiline comment above function
0
21,664
3,537,494,201
IssuesEvent
2016-01-18 01:37:07
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
It is static warning to use unknown type in 'catch'
area-test Priority-Medium triaged Type-Defect
try_catch_on_syntax_test/01: Fail try_catch_on_syntax_test/07: Fail try_catch_syntax_test/08: Fail Spec 0.13, try statement. &quot;An on-catch clause of the form on T catch (p1, p2) s or on T s matches an object o if the type of o is a subtype of T. It is a static warning if T does not denote a type available in the lexical scope of the catch clause.&quot;
1.0
It is static warning to use unknown type in 'catch' - try_catch_on_syntax_test/01: Fail try_catch_on_syntax_test/07: Fail try_catch_syntax_test/08: Fail Spec 0.13, try statement. &quot;An on-catch clause of the form on T catch (p1, p2) s or on T s matches an object o if the type of o is a subtype of T. It is a static warning if T does not denote a type available in the lexical scope of the catch clause.&quot;
defect
it is static warning to use unknown type in catch try catch on syntax test fail try catch on syntax test fail try catch syntax test fail spec try statement quot an on catch clause of the form on t catch s or on t s matches an object o if the type of o is a subtype of t it is a static warning if t does not denote a type available in the lexical scope of the catch clause quot
1
78,243
27,389,298,493
IssuesEvent
2023-02-28 15:17:23
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
NPE in MulticastConfig.hashCode()
Type: Defect
Hazelcast 5.2.1. com.hazelcast.config.MulticastConfig has a member `private Boolean loopbackModeEnabled` and the following hashCode implementation: > public final int hashCode() { > int result = (enabled ? 1 : 0); > result = 31 * result + (multicastGroup != null ? multicastGroup.hashCode() : 0); > result = 31 * result + multicastPort; > result = 31 * result + multicastTimeoutSeconds; > result = 31 * result + multicastTimeToLive; > result = 31 * result + trustedInterfaces.hashCode(); > result = 31 * result + (loopbackModeEnabled ? 1 : 0); > return result; > } **Describe the bug** Activating the _hashCode_ method when the _loopbackModeEnabled_ is null (default value if none was provided by the user) causes NPE. **Expected behavior** _hashCode_ method does not throw NPE <!-- Add any other context about the problem here. Common details that we're often interested in: - Detailed description of the steps to reproduce your issue - Logs and stack traces, if available - Hazelcast version that you use (e.g. 3.4, also specify whether it is a minor release or the latest snapshot) - If available, integration module versions (e.g. Tomcat, Jetty, Spring, Hibernate). Also, include their detailed configuration information such as web.xml, Hibernate configuration and `context.xml` for Spring - Cluster size, i.e. the number of Hazelcast cluster members - Number of the clients - Version of Java. It is also helpful to mention the JVM parameters - Operating system. If it is Linux, kernel version is helpful - Unit test with the `hazelcast.xml` file. If you could include a unit test which reproduces your issue, we would be grateful -->
1.0
NPE in MulticastConfig.hashCode() - Hazelcast 5.2.1. com.hazelcast.config.MulticastConfig has a member `private Boolean loopbackModeEnabled` and the following hashCode implementation: > public final int hashCode() { > int result = (enabled ? 1 : 0); > result = 31 * result + (multicastGroup != null ? multicastGroup.hashCode() : 0); > result = 31 * result + multicastPort; > result = 31 * result + multicastTimeoutSeconds; > result = 31 * result + multicastTimeToLive; > result = 31 * result + trustedInterfaces.hashCode(); > result = 31 * result + (loopbackModeEnabled ? 1 : 0); > return result; > } **Describe the bug** Activating the _hashCode_ method when the _loopbackModeEnabled_ is null (default value if none was provided by the user) causes NPE. **Expected behavior** _hashCode_ method does not throw NPE <!-- Add any other context about the problem here. Common details that we're often interested in: - Detailed description of the steps to reproduce your issue - Logs and stack traces, if available - Hazelcast version that you use (e.g. 3.4, also specify whether it is a minor release or the latest snapshot) - If available, integration module versions (e.g. Tomcat, Jetty, Spring, Hibernate). Also, include their detailed configuration information such as web.xml, Hibernate configuration and `context.xml` for Spring - Cluster size, i.e. the number of Hazelcast cluster members - Number of the clients - Version of Java. It is also helpful to mention the JVM parameters - Operating system. If it is Linux, kernel version is helpful - Unit test with the `hazelcast.xml` file. If you could include a unit test which reproduces your issue, we would be grateful -->
defect
npe in multicastconfig hashcode hazelcast com hazelcast config multicastconfig has a member private boolean loopbackmodeenabled and the following hashcode implementation public final int hashcode int result enabled result result multicastgroup null multicastgroup hashcode result result multicastport result result multicasttimeoutseconds result result multicasttimetolive result result trustedinterfaces hashcode result result loopbackmodeenabled return result describe the bug activating the hashcode method when the loopbackmodeenabled is null default value if none was provided by the user causes npe expected behavior hashcode method does not throw npe add any other context about the problem here common details that we re often interested in detailed description of the steps to reproduce your issue logs and stack traces if available hazelcast version that you use e g also specify whether it is a minor release or the latest snapshot if available integration module versions e g tomcat jetty spring hibernate also include their detailed configuration information such as web xml hibernate configuration and context xml for spring cluster size i e the number of hazelcast cluster members number of the clients version of java it is also helpful to mention the jvm parameters operating system if it is linux kernel version is helpful unit test with the hazelcast xml file if you could include a unit test which reproduces your issue we would be grateful
1
596
2,576,341,006
IssuesEvent
2015-02-12 09:19:24
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
[TEST-FAILURE] testLock_WhenDummyClientAndOwnerNodeDiesTogether
Team: Client Type: Defect
``` java.lang.IllegalStateException: Node failed to start! at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:125) at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:153) at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:136) at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:112) at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:87) at com.hazelcast.client.ClientRegressionTest.testLock_WhenClientAndOwnerNodeDiesTogether(ClientRegressionTest.java:610) at com.hazelcast.client.ClientRegressionTest.testLock_WhenDummyClientAndOwnerNodeDiesTogether(ClientRegressionTest.java:593) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x/com.hazelcast$hazelcast-client/2459/testReport/junit/com.hazelcast.client/ClientRegressionTest/testLock_WhenDummyClientAndOwnerNodeDiesTogether/
1.0
[TEST-FAILURE] testLock_WhenDummyClientAndOwnerNodeDiesTogether - ``` java.lang.IllegalStateException: Node failed to start! at com.hazelcast.instance.HazelcastInstanceImpl.<init>(HazelcastInstanceImpl.java:125) at com.hazelcast.instance.HazelcastInstanceFactory.constructHazelcastInstance(HazelcastInstanceFactory.java:153) at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:136) at com.hazelcast.instance.HazelcastInstanceFactory.newHazelcastInstance(HazelcastInstanceFactory.java:112) at com.hazelcast.core.Hazelcast.newHazelcastInstance(Hazelcast.java:87) at com.hazelcast.client.ClientRegressionTest.testLock_WhenClientAndOwnerNodeDiesTogether(ClientRegressionTest.java:610) at com.hazelcast.client.ClientRegressionTest.testLock_WhenDummyClientAndOwnerNodeDiesTogether(ClientRegressionTest.java:593) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x/com.hazelcast$hazelcast-client/2459/testReport/junit/com.hazelcast.client/ClientRegressionTest/testLock_WhenDummyClientAndOwnerNodeDiesTogether/
defect
testlock whendummyclientandownernodediestogether java lang illegalstateexception node failed to start at com hazelcast instance hazelcastinstanceimpl hazelcastinstanceimpl java at com hazelcast instance hazelcastinstancefactory constructhazelcastinstance hazelcastinstancefactory java at com hazelcast instance hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast instance hazelcastinstancefactory newhazelcastinstance hazelcastinstancefactory java at com hazelcast core hazelcast newhazelcastinstance hazelcast java at com hazelcast client clientregressiontest testlock whenclientandownernodediestogether clientregressiontest java at com hazelcast client clientregressiontest testlock whendummyclientandownernodediestogether clientregressiontest java
1
15,465
2,856,032,961
IssuesEvent
2015-06-02 13:10:29
dermotte/lire
https://api.github.com/repos/dermotte/lire
closed
Bug in de-serializing a float array
auto-migrated Priority-Medium Type-Defect
``` https://code.google.com/p/lire/source/browse/trunk/src/main/java/net/semanticmet adata/lire/utils/SerializationUtils.java#245 In the 0.9.4-beta2 code, the function seems to shift the output array by offset, and chop off the rest of the float array at the end. public static float[] toFloatArray(byte[] in, int offset, int length) { float[] result = new float[length / 4]; byte[] tmp = new byte[4]; for (int i = offset; i < length / 4; i++) { System.arraycopy(in, (i - offset) * 4 + offset, tmp, 0, 4); result[i] = toFloat(tmp); } return result; } For comparison, the code for converting to double array is correct: public static double[] toDoubleArray(byte[] data, int offset, int length) { double[] result = new double[length / 8]; byte[] tmp = new byte[8]; for (int i = 0; i < result.length; i++) { System.arraycopy(data, i * 8 + offset, tmp, 0, 8); result[i] = toDouble(tmp); } return result; } ``` Original issue reported on code.google.com by `yutian...@gmail.com` on 12 Aug 2014 at 9:49
1.0
Bug in de-serializing a float array - ``` https://code.google.com/p/lire/source/browse/trunk/src/main/java/net/semanticmet adata/lire/utils/SerializationUtils.java#245 In the 0.9.4-beta2 code, the function seems to shift the output array by offset, and chop off the rest of the float array at the end. public static float[] toFloatArray(byte[] in, int offset, int length) { float[] result = new float[length / 4]; byte[] tmp = new byte[4]; for (int i = offset; i < length / 4; i++) { System.arraycopy(in, (i - offset) * 4 + offset, tmp, 0, 4); result[i] = toFloat(tmp); } return result; } For comparison, the code for converting to double array is correct: public static double[] toDoubleArray(byte[] data, int offset, int length) { double[] result = new double[length / 8]; byte[] tmp = new byte[8]; for (int i = 0; i < result.length; i++) { System.arraycopy(data, i * 8 + offset, tmp, 0, 8); result[i] = toDouble(tmp); } return result; } ``` Original issue reported on code.google.com by `yutian...@gmail.com` on 12 Aug 2014 at 9:49
defect
bug in de serializing a float array adata lire utils serializationutils java in the code the function seems to shift the output array by offset and chop off the rest of the float array at the end public static float tofloatarray byte in int offset int length float result new float byte tmp new byte for int i offset i length i system arraycopy in i offset offset tmp result tofloat tmp return result for comparison the code for converting to double array is correct public static double todoublearray byte data int offset int length double result new double byte tmp new byte for int i i result length i system arraycopy data i offset tmp result todouble tmp return result original issue reported on code google com by yutian gmail com on aug at
1
282,594
30,889,377,140
IssuesEvent
2023-08-04 02:37:59
madhans23/linux-4.1.15
https://api.github.com/repos/madhans23/linux-4.1.15
reopened
CVE-2015-7884 (Low) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2015-7884 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/platform/vivid/vivid-osd.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/platform/vivid/vivid-osd.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The vivid_fb_ioctl function in drivers/media/platform/vivid/vivid-osd.c in the Linux kernel through 4.3.3 does not initialize a certain structure member, which allows local users to obtain sensitive information from kernel memory via a crafted application. <p>Publish Date: 2015-12-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-7884>CVE-2015-7884</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2015-7884">https://www.linuxkernelcves.com/cves/CVE-2015-7884</a></p> <p>Release Date: 2015-12-28</p> <p>Fix Resolution: v4.4-rc1,v4.1.17,v4.3.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-7884 (Low) detected in linux-stable-rtv4.1.33 - ## CVE-2015-7884 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/platform/vivid/vivid-osd.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/media/platform/vivid/vivid-osd.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The vivid_fb_ioctl function in drivers/media/platform/vivid/vivid-osd.c in the Linux kernel through 4.3.3 does not initialize a certain structure member, which allows local users to obtain sensitive information from kernel memory via a crafted application. <p>Publish Date: 2015-12-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2015-7884>CVE-2015-7884</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2015-7884">https://www.linuxkernelcves.com/cves/CVE-2015-7884</a></p> <p>Release Date: 2015-12-28</p> <p>Fix Resolution: v4.4-rc1,v4.1.17,v4.3.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve low detected in linux stable cve low severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files drivers media platform vivid vivid osd c drivers media platform vivid vivid osd c vulnerability details the vivid fb ioctl function in drivers media platform vivid vivid osd c in the linux kernel through does not initialize a certain structure member which allows local users to obtain sensitive information from kernel memory via a crafted application publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
15,961
2,870,102,217
IssuesEvent
2015-06-06 20:50:10
dkpro/dkpro-tc
https://api.github.com/repos/dkpro/dkpro-tc
closed
TC restructuring: get rid of the tc.features dependency in core
Priority-High Release-0.5.0 Type-Defect
Originally reported on Google Code with ID 23 ``` tc.api only contains one exception and one type. Proposed solution: Move to core.exception and core.type tc.api.features should consequently also be move to core, since the interfaces etc. are core functionality. HOWEVER, there are problems. Currently, core has a dependency on tc.features (AddIdExtractor in the various instance extractors and the NGramFeatureExtractor-UIMA parameters in the FeatureExtrationTask) If we were to move api.features and tc.api to core, we would cause a circular dependency. That's why we have to get rid of the tc.features dependency in core, whic really should not be there anyway!! ``` Reported by `oliver.ferschke` on 2013-06-06 10:07:50
1.0
TC restructuring: get rid of the tc.features dependency in core - Originally reported on Google Code with ID 23 ``` tc.api only contains one exception and one type. Proposed solution: Move to core.exception and core.type tc.api.features should consequently also be move to core, since the interfaces etc. are core functionality. HOWEVER, there are problems. Currently, core has a dependency on tc.features (AddIdExtractor in the various instance extractors and the NGramFeatureExtractor-UIMA parameters in the FeatureExtrationTask) If we were to move api.features and tc.api to core, we would cause a circular dependency. That's why we have to get rid of the tc.features dependency in core, whic really should not be there anyway!! ``` Reported by `oliver.ferschke` on 2013-06-06 10:07:50
defect
tc restructuring get rid of the tc features dependency in core originally reported on google code with id tc api only contains one exception and one type proposed solution move to core exception and core type tc api features should consequently also be move to core since the interfaces etc are core functionality however there are problems currently core has a dependency on tc features addidextractor in the various instance extractors and the ngramfeatureextractor uima parameters in the featureextrationtask if we were to move api features and tc api to core we would cause a circular dependency that s why we have to get rid of the tc features dependency in core whic really should not be there anyway reported by oliver ferschke on
1
2,481
2,607,904,524
IssuesEvent
2015-02-26 00:15:03
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
closed
Cursor in wrong position after expansion when editing php files in Dreamweaver CS4 and CS5
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? - In a file with a PHP extension, type "div" (or any expandable tag), then press Ctrl+, What is the expected output? What do you see instead? - After expansion, instead of the cursor being positioned between the tags, it is positioned after the closing tag. - You can rename the file with an HTML extension, and the cursor will be positioned correctly after expansion. What version of the product are you using? On what operating system? - 0.6.0, Dreamweaver CS4 and CS5, Vista Ultimate. ``` ----- Original issue reported on code.google.com by `cliftonk...@gmail.com` on 12 May 2010 at 6:49
1.0
Cursor in wrong position after expansion when editing php files in Dreamweaver CS4 and CS5 - ``` What steps will reproduce the problem? - In a file with a PHP extension, type "div" (or any expandable tag), then press Ctrl+, What is the expected output? What do you see instead? - After expansion, instead of the cursor being positioned between the tags, it is positioned after the closing tag. - You can rename the file with an HTML extension, and the cursor will be positioned correctly after expansion. What version of the product are you using? On what operating system? - 0.6.0, Dreamweaver CS4 and CS5, Vista Ultimate. ``` ----- Original issue reported on code.google.com by `cliftonk...@gmail.com` on 12 May 2010 at 6:49
defect
cursor in wrong position after expansion when editing php files in dreamweaver and what steps will reproduce the problem in a file with a php extension type div or any expandable tag then press ctrl what is the expected output what do you see instead after expansion instead of the cursor being positioned between the tags it is positioned after the closing tag you can rename the file with an html extension and the cursor will be positioned correctly after expansion what version of the product are you using on what operating system dreamweaver and vista ultimate original issue reported on code google com by cliftonk gmail com on may at
1
23,463
3,829,547,493
IssuesEvent
2016-03-31 11:10:56
pearu/pylibnidaqmx
https://api.github.com/repos/pearu/pylibnidaqmx
closed
msvccompiler in numpy.distutils
auto-migrated Priority-Medium Type-Defect
``` While installing nidaqmx the following error occurred at the end of the installation: "No module names msvccompiler in numpy.distutils; trying from distutils customize MSVCCompiler" The list of modules in Python shows that "nidaqmx_h_9_6" and "nidaqmx_h_9_7" are available. However, trying to execute "from nidaqmx import System" gave the following error: File "<pyshell#11>", line 1, in <module> from nidaqmx_h_9_7 import System ImportError: cannot import name System I am working on Windows 7 64-bit. ``` Original issue reported on code.google.com by `yeghishe...@gmail.com` on 25 Nov 2013 at 10:04
1.0
msvccompiler in numpy.distutils - ``` While installing nidaqmx the following error occurred at the end of the installation: "No module names msvccompiler in numpy.distutils; trying from distutils customize MSVCCompiler" The list of modules in Python shows that "nidaqmx_h_9_6" and "nidaqmx_h_9_7" are available. However, trying to execute "from nidaqmx import System" gave the following error: File "<pyshell#11>", line 1, in <module> from nidaqmx_h_9_7 import System ImportError: cannot import name System I am working on Windows 7 64-bit. ``` Original issue reported on code.google.com by `yeghishe...@gmail.com` on 25 Nov 2013 at 10:04
defect
msvccompiler in numpy distutils while installing nidaqmx the following error occurred at the end of the installation no module names msvccompiler in numpy distutils trying from distutils customize msvccompiler the list of modules in python shows that nidaqmx h and nidaqmx h are available however trying to execute from nidaqmx import system gave the following error file line in from nidaqmx h import system importerror cannot import name system i am working on windows bit original issue reported on code google com by yeghishe gmail com on nov at
1
16,741
2,941,305,428
IssuesEvent
2015-07-02 06:49:26
tnt944445/reaver-wps
https://api.github.com/repos/tnt944445/reaver-wps
closed
Driver issues and same pin again
auto-migrated Priority-Triage Type-Defect
``` 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) 1.4 1. What operating system are you using (Linux is the only supported OS)? linux Backtrack 5 r3 2. Is your wireless card in monitor mode (yes/no)? Yes 3. What is the signal strength of the Access Point you are trying to crack? -65 4. What is the manufacturer and model # of the device you are trying to crack? 5. What is the entire command line string you are supplying to reaver? i tried it like this aireplay-ng mon0 -1 5 -a <mac> -e <bssid> reaver -i mon0 -a -b <mac> -c <channel> -vv --no-nacks --win7 6. Please describe what you think the issue is. it repeats allways same pin if i use -p and some it changes the pin but it still loop with this one then... so i think it is problem with drivers cause 7. Paste the output from Reaver below. i can post output later cause its on my friends laptop so well my biggest problem is if i start airmon-ng it shows me chipset Unknown and i really dont know how to update the right driver i attached a everesthome list maybe someone can explain me exactly how i install the right driver with all commands i need to use i read somewhere that the problem with same pin again again again is something with driver issue cause once it worked for around 0.10% and then after it i never got it again it just try now pin over and over again SOORY FOR MY BAD ENGLISH :) EVEREST HOME http://pastebin.com/3ySiNqaN ``` Original issue reported on code.google.com by `HazeBust...@gmail.com` on 7 Mar 2013 at 1:35
1.0
Driver issues and same pin again - ``` 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) 1.4 1. What operating system are you using (Linux is the only supported OS)? linux Backtrack 5 r3 2. Is your wireless card in monitor mode (yes/no)? Yes 3. What is the signal strength of the Access Point you are trying to crack? -65 4. What is the manufacturer and model # of the device you are trying to crack? 5. What is the entire command line string you are supplying to reaver? i tried it like this aireplay-ng mon0 -1 5 -a <mac> -e <bssid> reaver -i mon0 -a -b <mac> -c <channel> -vv --no-nacks --win7 6. Please describe what you think the issue is. it repeats allways same pin if i use -p and some it changes the pin but it still loop with this one then... so i think it is problem with drivers cause 7. Paste the output from Reaver below. i can post output later cause its on my friends laptop so well my biggest problem is if i start airmon-ng it shows me chipset Unknown and i really dont know how to update the right driver i attached a everesthome list maybe someone can explain me exactly how i install the right driver with all commands i need to use i read somewhere that the problem with same pin again again again is something with driver issue cause once it worked for around 0.10% and then after it i never got it again it just try now pin over and over again SOORY FOR MY BAD ENGLISH :) EVEREST HOME http://pastebin.com/3ySiNqaN ``` Original issue reported on code.google.com by `HazeBust...@gmail.com` on 7 Mar 2013 at 1:35
defect
driver issues and same pin again what version of reaver are you using only defects against the latest version will be considered what operating system are you using linux is the only supported os linux backtrack is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack what is the entire command line string you are supplying to reaver i tried it like this aireplay ng a e reaver i a b c vv no nacks please describe what you think the issue is it repeats allways same pin if i use p and some it changes the pin but it still loop with this one then so i think it is problem with drivers cause paste the output from reaver below i can post output later cause its on my friends laptop so well my biggest problem is if i start airmon ng it shows me chipset unknown and i really dont know how to update the right driver i attached a everesthome list maybe someone can explain me exactly how i install the right driver with all commands i need to use i read somewhere that the problem with same pin again again again is something with driver issue cause once it worked for around and then after it i never got it again it just try now pin over and over again soory for my bad english everest home original issue reported on code google com by hazebust gmail com on mar at
1
622,213
19,617,845,066
IssuesEvent
2022-01-07 00:05:21
egc-sierrezuela-3/decide
https://api.github.com/repos/egc-sierrezuela-3/decide
closed
Pintado de gráficas y estudio de datos(I)
priority: low location: frontend type: internal
Mostrar gráficos que representen ciertas estadísticas de las votaciones en tiempo real, como por ejemplo el número de votantes que han votado una opción determinada.
1.0
Pintado de gráficas y estudio de datos(I) - Mostrar gráficos que representen ciertas estadísticas de las votaciones en tiempo real, como por ejemplo el número de votantes que han votado una opción determinada.
non_defect
pintado de gráficas y estudio de datos i mostrar gráficos que representen ciertas estadísticas de las votaciones en tiempo real como por ejemplo el número de votantes que han votado una opción determinada
0
36,723
8,093,391,889
IssuesEvent
2018-08-10 00:32:59
mernst/plume-lib
https://api.github.com/repos/mernst/plume-lib
closed
OptionsDoclet does not include options in inner classes
Priority-Medium Type-Defect auto-migrated
``` Currently, the OptionsDoclet Javadoc doclet ignores inner classes. This is because I do not have a good solution to the following problem: given a ClassDoc instance 'cd' which refers to class A, how do you get a Class<?> instance corresponding to A? Originally, I used 'Class.forName(cd.qualifiedName())'. This approach fails when cd is an inner class. The documentation for Class.forName() says the first parameter to this method is a String representing the fully qualified name of a class. This is false. In fact, the first parameter to Class.forName() must be a string representing the *binary* name of the desired class. So what we have is cd.qualifiedName() correctly returning a fully qualified name, and Class.forName() expecting a binary name. Usually, when inner classes are not being used, this distinction wont matter since the FQN and binary name are the same. But with Javarifier, which has inner classes in its main class, we have the following: javarifier.Main$JrTransformer is a binary name. javarifier.Main.JrTransformer is a fully qualified name. The Class.forName() method fails on the FQN of this class. One solution would be a general method to translate from a fully qualified name to a binary name. This method has to be robust and work correctly on all cases. Another solution is to use calls to 'cd.containingClass()' to determine which classes are inner classes and then make the appropriate replacements of dots with dollar signs in the fully qualified name. Better yet, perhaps there is a method of going from ClassDoc to Class<?> without using strings or names at all. For now, my solution is to simply ignore any classes that are contained within other classes. This means any @Option annotations in inner classes will not be included in the generated HTML documentation. ``` Original issue reported on code.google.com by `david.lazar` on 8 Jul 2010 at 6:03
1.0
OptionsDoclet does not include options in inner classes - ``` Currently, the OptionsDoclet Javadoc doclet ignores inner classes. This is because I do not have a good solution to the following problem: given a ClassDoc instance 'cd' which refers to class A, how do you get a Class<?> instance corresponding to A? Originally, I used 'Class.forName(cd.qualifiedName())'. This approach fails when cd is an inner class. The documentation for Class.forName() says the first parameter to this method is a String representing the fully qualified name of a class. This is false. In fact, the first parameter to Class.forName() must be a string representing the *binary* name of the desired class. So what we have is cd.qualifiedName() correctly returning a fully qualified name, and Class.forName() expecting a binary name. Usually, when inner classes are not being used, this distinction wont matter since the FQN and binary name are the same. But with Javarifier, which has inner classes in its main class, we have the following: javarifier.Main$JrTransformer is a binary name. javarifier.Main.JrTransformer is a fully qualified name. The Class.forName() method fails on the FQN of this class. One solution would be a general method to translate from a fully qualified name to a binary name. This method has to be robust and work correctly on all cases. Another solution is to use calls to 'cd.containingClass()' to determine which classes are inner classes and then make the appropriate replacements of dots with dollar signs in the fully qualified name. Better yet, perhaps there is a method of going from ClassDoc to Class<?> without using strings or names at all. For now, my solution is to simply ignore any classes that are contained within other classes. This means any @Option annotations in inner classes will not be included in the generated HTML documentation. ``` Original issue reported on code.google.com by `david.lazar` on 8 Jul 2010 at 6:03
defect
optionsdoclet does not include options in inner classes currently the optionsdoclet javadoc doclet ignores inner classes this is because i do not have a good solution to the following problem given a classdoc instance cd which refers to class a how do you get a class instance corresponding to a originally i used class forname cd qualifiedname this approach fails when cd is an inner class the documentation for class forname says the first parameter to this method is a string representing the fully qualified name of a class this is false in fact the first parameter to class forname must be a string representing the binary name of the desired class so what we have is cd qualifiedname correctly returning a fully qualified name and class forname expecting a binary name usually when inner classes are not being used this distinction wont matter since the fqn and binary name are the same but with javarifier which has inner classes in its main class we have the following javarifier main jrtransformer is a binary name javarifier main jrtransformer is a fully qualified name the class forname method fails on the fqn of this class one solution would be a general method to translate from a fully qualified name to a binary name this method has to be robust and work correctly on all cases another solution is to use calls to cd containingclass to determine which classes are inner classes and then make the appropriate replacements of dots with dollar signs in the fully qualified name better yet perhaps there is a method of going from classdoc to class without using strings or names at all for now my solution is to simply ignore any classes that are contained within other classes this means any option annotations in inner classes will not be included in the generated html documentation original issue reported on code google com by david lazar on jul at
1
558,220
16,528,828,945
IssuesEvent
2021-05-27 01:16:19
actually-colab/editor
https://api.github.com/repos/actually-colab/editor
opened
Add keep alive endpoint
client priority: medium server socket
Heartbeat to prevent auto disconnects from the socket client due to inactivity
1.0
Add keep alive endpoint - Heartbeat to prevent auto disconnects from the socket client due to inactivity
non_defect
add keep alive endpoint heartbeat to prevent auto disconnects from the socket client due to inactivity
0
55,088
6,425,881,840
IssuesEvent
2017-08-09 16:14:43
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
[CI] full-cluster-restart BWC-Test fails because 6.0 is not released yet
test
link: https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+bwc-tests/229/console error message: ``` :checkout-6.0:distribution:zip:assemble :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#prepareCluster.cleanShared :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#node0.clean :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#node0.checkPrevious SKIPPED :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#node0.stopPrevious SKIPPED :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#node0.extract FAILED Cannot expand ZIP '/var/lib/jenkins/workspace/elastic+elasticsearch+master+bwc-tests/distribution/bwc/build/bwc/checkout-6.0/distribution/zip/build/distributions/elasticsearch-6.0.0-SNAPSHOT.zip' as it does not exist. ``` reproduce with: ``` gradle :qa:full-cluster-restart:bwcTest ```
1.0
[CI] full-cluster-restart BWC-Test fails because 6.0 is not released yet - link: https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+bwc-tests/229/console error message: ``` :checkout-6.0:distribution:zip:assemble :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#prepareCluster.cleanShared :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#node0.clean :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#node0.checkPrevious SKIPPED :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#node0.stopPrevious SKIPPED :qa:full-cluster-restart:v6.0.0-SNAPSHOT#oldClusterTestCluster#node0.extract FAILED Cannot expand ZIP '/var/lib/jenkins/workspace/elastic+elasticsearch+master+bwc-tests/distribution/bwc/build/bwc/checkout-6.0/distribution/zip/build/distributions/elasticsearch-6.0.0-SNAPSHOT.zip' as it does not exist. ``` reproduce with: ``` gradle :qa:full-cluster-restart:bwcTest ```
non_defect
full cluster restart bwc test fails because is not released yet link error message checkout distribution zip assemble qa full cluster restart snapshot oldclustertestcluster preparecluster cleanshared qa full cluster restart snapshot oldclustertestcluster clean qa full cluster restart snapshot oldclustertestcluster checkprevious skipped qa full cluster restart snapshot oldclustertestcluster stopprevious skipped qa full cluster restart snapshot oldclustertestcluster extract failed cannot expand zip var lib jenkins workspace elastic elasticsearch master bwc tests distribution bwc build bwc checkout distribution zip build distributions elasticsearch snapshot zip as it does not exist reproduce with gradle qa full cluster restart bwctest
0
167,566
6,341,549,310
IssuesEvent
2017-07-27 13:46:28
fossasia/open-event-webapp
https://api.github.com/repos/fossasia/open-event-webapp
closed
Move search to top bar in mobile view
enhancement has-PR Priority: High
In the mobile the search bar should be moved to the main menu bar similar to the Android app. Please show a search button and expand this search button into a search field on click. ![screenshot_2017-05-01-13-02-37](https://cloud.githubusercontent.com/assets/1583873/25579226/41c8a220-2e76-11e7-864f-14fdf8bd60e5.png)
1.0
Move search to top bar in mobile view - In the mobile the search bar should be moved to the main menu bar similar to the Android app. Please show a search button and expand this search button into a search field on click. ![screenshot_2017-05-01-13-02-37](https://cloud.githubusercontent.com/assets/1583873/25579226/41c8a220-2e76-11e7-864f-14fdf8bd60e5.png)
non_defect
move search to top bar in mobile view in the mobile the search bar should be moved to the main menu bar similar to the android app please show a search button and expand this search button into a search field on click
0
232,257
17,777,330,084
IssuesEvent
2021-08-30 21:04:45
haskell/cabal
https://api.github.com/repos/haskell/cabal
reopened
Error in docs and changelog for append and backup configure options
documentation
@ptkato In https://github.com/haskell/cabal/pull/7402/files#diff-277f4cd76dc57540f0ad840e00ed445aa2da6df95ab749385330ea5950e08358 and https://github.com/haskell/cabal/pull/7402/files#diff-7366bf7292133a6cd2005aed5db013b2774af3c8d49c8b7e20ec5e8e03e1ff2e "overwrite" is mentioned, but the actual name is "backup"
1.0
Error in docs and changelog for append and backup configure options - @ptkato In https://github.com/haskell/cabal/pull/7402/files#diff-277f4cd76dc57540f0ad840e00ed445aa2da6df95ab749385330ea5950e08358 and https://github.com/haskell/cabal/pull/7402/files#diff-7366bf7292133a6cd2005aed5db013b2774af3c8d49c8b7e20ec5e8e03e1ff2e "overwrite" is mentioned, but the actual name is "backup"
non_defect
error in docs and changelog for append and backup configure options ptkato in and overwrite is mentioned but the actual name is backup
0
9,191
2,615,138,154
IssuesEvent
2015-03-01 06:11:54
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
closed
Pins are repeated if list is exhausted and no result has been found
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Run reaver, let it do it's thing 2. If for some reason it's been through the whole list of possibilities, it will continue looping the last pin One of these scenarios is bug #12 What is the expected output? What do you see instead? After the list of possible pincodes is exhausted, i expect Reaver to give me feedback, ie.: First part cracked:, couldn't crack 2nd part. This is not the case, instead it keeps on going with the last pincode and never reaches 100%: +] Trying pin 47702738 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] 99.99% complete @ 2 seconds/attempt [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [!] WARNING: Receive timeout occurred [+] 99.99% complete @ 2 seconds/attempt [+] Trying pin 47703889 [+] Trying pin 47703889 [!] WARNING: Receive timeout occurred [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] 99.99% complete @ 2 seconds/attempt [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [!] WARNING: Receive timeout occurred [+] 99.99% complete @ 2 seconds/attempt [+] Trying pin 47703889 etc. What version of the product are you using? On what operating system? v16 SVN Please provide any additional information below. ``` Original issue reported on code.google.com by `stefanen...@gmail.com` on 30 Dec 2011 at 12:44
1.0
Pins are repeated if list is exhausted and no result has been found - ``` What steps will reproduce the problem? 1. Run reaver, let it do it's thing 2. If for some reason it's been through the whole list of possibilities, it will continue looping the last pin One of these scenarios is bug #12 What is the expected output? What do you see instead? After the list of possible pincodes is exhausted, i expect Reaver to give me feedback, ie.: First part cracked:, couldn't crack 2nd part. This is not the case, instead it keeps on going with the last pincode and never reaches 100%: +] Trying pin 47702738 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] 99.99% complete @ 2 seconds/attempt [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [!] WARNING: Receive timeout occurred [+] 99.99% complete @ 2 seconds/attempt [+] Trying pin 47703889 [+] Trying pin 47703889 [!] WARNING: Receive timeout occurred [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] 99.99% complete @ 2 seconds/attempt [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [+] Trying pin 47703889 [!] WARNING: Receive timeout occurred [+] 99.99% complete @ 2 seconds/attempt [+] Trying pin 47703889 etc. What version of the product are you using? On what operating system? v16 SVN Please provide any additional information below. ``` Original issue reported on code.google.com by `stefanen...@gmail.com` on 30 Dec 2011 at 12:44
defect
pins are repeated if list is exhausted and no result has been found what steps will reproduce the problem run reaver let it do it s thing if for some reason it s been through the whole list of possibilities it will continue looping the last pin one of these scenarios is bug what is the expected output what do you see instead after the list of possible pincodes is exhausted i expect reaver to give me feedback ie first part cracked couldn t crack part this is not the case instead it keeps on going with the last pincode and never reaches trying pin trying pin trying pin complete seconds attempt trying pin trying pin trying pin trying pin trying pin warning receive timeout occurred complete seconds attempt trying pin trying pin warning receive timeout occurred trying pin trying pin trying pin complete seconds attempt trying pin trying pin trying pin trying pin trying pin warning receive timeout occurred complete seconds attempt trying pin etc what version of the product are you using on what operating system svn please provide any additional information below original issue reported on code google com by stefanen gmail com on dec at
1
56,518
15,149,028,870
IssuesEvent
2021-02-11 11:28:24
gbif/ipt
https://api.github.com/repos/gbif/ipt
opened
Organization shared tokens / passwords are cached internally
Component-Publication Priority-Low Type-Defect
If an organization's shared token is changed in the GBIF registry, and the new token is updated in the Administration→Organizations→organization settings, the old token continues to be used by the IPT for dataset changes (publication/deletion), which obviously fails. Workaround is to restart the IPT.
1.0
Organization shared tokens / passwords are cached internally - If an organization's shared token is changed in the GBIF registry, and the new token is updated in the Administration→Organizations→organization settings, the old token continues to be used by the IPT for dataset changes (publication/deletion), which obviously fails. Workaround is to restart the IPT.
defect
organization shared tokens passwords are cached internally if an organization s shared token is changed in the gbif registry and the new token is updated in the administration→organizations→organization settings the old token continues to be used by the ipt for dataset changes publication deletion which obviously fails workaround is to restart the ipt
1
53,023
13,260,072,246
IssuesEvent
2020-08-20 17:38:49
jkoan/test-navit
https://api.github.com/repos/jkoan/test-navit
closed
Navigation items aren't shown when starting navit with a previous destination set (Trac #24)
Incomplete Migration KaZeR Migrated from Trac defect/bug gui/cegui
Migrated from http://trac.navit-project.org/ticket/24 ```json { "status": "closed", "changetime": "2007-12-10T21:23:34", "_ts": "1197321814000000", "description": "It's because the components are shown only after validating the route from the destination screen.", "reporter": "KaZeR", "cc": "", "resolution": "fixed", "time": "2007-11-22T10:23:45", "component": "gui/cegui", "summary": "Navigation items aren't shown when starting navit with a previous destination set", "priority": "major", "keywords": "", "version": "", "milestone": "version 0.0.3", "owner": "KaZeR", "type": "defect/bug", "severity": "" } ```
1.0
Navigation items aren't shown when starting navit with a previous destination set (Trac #24) - Migrated from http://trac.navit-project.org/ticket/24 ```json { "status": "closed", "changetime": "2007-12-10T21:23:34", "_ts": "1197321814000000", "description": "It's because the components are shown only after validating the route from the destination screen.", "reporter": "KaZeR", "cc": "", "resolution": "fixed", "time": "2007-11-22T10:23:45", "component": "gui/cegui", "summary": "Navigation items aren't shown when starting navit with a previous destination set", "priority": "major", "keywords": "", "version": "", "milestone": "version 0.0.3", "owner": "KaZeR", "type": "defect/bug", "severity": "" } ```
defect
navigation items aren t shown when starting navit with a previous destination set trac migrated from json status closed changetime ts description it s because the components are shown only after validating the route from the destination screen reporter kazer cc resolution fixed time component gui cegui summary navigation items aren t shown when starting navit with a previous destination set priority major keywords version milestone version owner kazer type defect bug severity
1
130,660
12,451,929,212
IssuesEvent
2020-05-27 11:22:25
dsongUTS/Tutorial-4--Group-5-Assessment-Info-System-Development-
https://api.github.com/repos/dsongUTS/Tutorial-4--Group-5-Assessment-Info-System-Development-
closed
Introduction to the Assessment: Creation of Problem Statement, Objectives and Stakeholders
documentation good first issue
Beginning with the first week of the assessment, the first backlog involved splitting our group of 4 onto certain tasks. We elected to base who was doing what moving forward on the stakeholders. Daniel and Nik took care of the empathy maps and assumptions for the relationship managers, and so then on would take on the user stories, reflections and HMW/POV statements. Fajiris and Khurshid likewise did the same for the customers.
1.0
Introduction to the Assessment: Creation of Problem Statement, Objectives and Stakeholders - Beginning with the first week of the assessment, the first backlog involved splitting our group of 4 onto certain tasks. We elected to base who was doing what moving forward on the stakeholders. Daniel and Nik took care of the empathy maps and assumptions for the relationship managers, and so then on would take on the user stories, reflections and HMW/POV statements. Fajiris and Khurshid likewise did the same for the customers.
non_defect
introduction to the assessment creation of problem statement objectives and stakeholders beginning with the first week of the assessment the first backlog involved splitting our group of onto certain tasks we elected to base who was doing what moving forward on the stakeholders daniel and nik took care of the empathy maps and assumptions for the relationship managers and so then on would take on the user stories reflections and hmw pov statements fajiris and khurshid likewise did the same for the customers
0
597,892
18,214,983,549
IssuesEvent
2021-09-30 02:19:25
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Improve Visible Assets List Load Performance
priority/P3 QA/No release-notes/exclude feature/wallet OS/Desktop
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description Currently the load performance for the `Visible Assets` list modal is very slow due to the size of the `ercTokenRegistry`. Need to look into ways to optimize the list and speed up actions. https://user-images.githubusercontent.com/40611140/134247409-d08e6897-bd29-4585-938e-ba2c1729ac60.mov
1.0
Improve Visible Assets List Load Performance - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description Currently the load performance for the `Visible Assets` list modal is very slow due to the size of the `ercTokenRegistry`. Need to look into ways to optimize the list and speed up actions. https://user-images.githubusercontent.com/40611140/134247409-d08e6897-bd29-4585-938e-ba2c1729ac60.mov
non_defect
improve visible assets list load performance have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description currently the load performance for the visible assets list modal is very slow due to the size of the erctokenregistry need to look into ways to optimize the list and speed up actions
0
53,202
6,706,215,782
IssuesEvent
2017-10-12 05:46:08
blockstack/designs
https://api.github.com/repos/blockstack/designs
closed
Design new whitepaper icons
design
- Design 3 icons (simple versions of the more detailed cover letter design) - [x] The Blockstack Whitepaper - [ ] Blockstack Technical Whitepaper icon - [ ] Blockstack Token Whitepaper icon
1.0
Design new whitepaper icons - - Design 3 icons (simple versions of the more detailed cover letter design) - [x] The Blockstack Whitepaper - [ ] Blockstack Technical Whitepaper icon - [ ] Blockstack Token Whitepaper icon
non_defect
design new whitepaper icons design icons simple versions of the more detailed cover letter design the blockstack whitepaper blockstack technical whitepaper icon blockstack token whitepaper icon
0
21,729
10,676,170,650
IssuesEvent
2019-10-21 13:16:27
repo-helper/npm-fetch
https://api.github.com/repos/repo-helper/npm-fetch
opened
CVE-2017-16137 (Medium) detected in debug-0.7.4.tgz
security vulnerability
## CVE-2017-16137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-0.7.4.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-0.7.4.tgz">https://registry.npmjs.org/debug/-/debug-0.7.4.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/npm-fetch/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/npm-fetch/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - mocha-1.9.0.tgz (Root Library) - :x: **debug-0.7.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/repo-helper/npm-fetch/commit/9fe6d1f4855964563107ec01592876cb80766d78">9fe6d1f4855964563107ec01592876cb80766d78</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16137>CVE-2017-16137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/visionmedia/debug/commit/42a6ae0737f9243c80b6d3dbb08a69a7ae2a1061">https://github.com/visionmedia/debug/commit/42a6ae0737f9243c80b6d3dbb08a69a7ae2a1061</a></p> <p>Release Date: 2017-09-21</p> <p>Fix Resolution: Replace or update the following file: node.js</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-16137 (Medium) detected in debug-0.7.4.tgz - ## CVE-2017-16137 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-0.7.4.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-0.7.4.tgz">https://registry.npmjs.org/debug/-/debug-0.7.4.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/npm-fetch/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/npm-fetch/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - mocha-1.9.0.tgz (Root Library) - :x: **debug-0.7.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/repo-helper/npm-fetch/commit/9fe6d1f4855964563107ec01592876cb80766d78">9fe6d1f4855964563107ec01592876cb80766d78</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16137>CVE-2017-16137</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Change files</p> <p>Origin: <a href="https://github.com/visionmedia/debug/commit/42a6ae0737f9243c80b6d3dbb08a69a7ae2a1061">https://github.com/visionmedia/debug/commit/42a6ae0737f9243c80b6d3dbb08a69a7ae2a1061</a></p> <p>Release Date: 2017-09-21</p> <p>Fix Resolution: Replace or update the following file: node.js</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in debug tgz cve medium severity vulnerability vulnerable library debug tgz small debugging utility library home page a href path to dependency file tmp ws scm npm fetch package json path to vulnerable library tmp ws scm npm fetch node modules debug package json dependency hierarchy mocha tgz root library x debug tgz vulnerable library found in head commit a href vulnerability details the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type change files origin a href release date fix resolution replace or update the following file node js step up your open source security game with whitesource
0
26,071
4,568,065,105
IssuesEvent
2016-09-15 13:26:01
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Not defined JavaScript error with [IgnoreGeneric] on interfaces
defect
Reported by @ProductiveRage on the Bridge Forums: http://forums.bridge.net/forum/bridge-net-pro/bugs/2708-ignoregeneric-on-interfaces-seems-to-break-bridge-15-0 ### Expected No Javascript error ### Actual A JavaScript `not defined` error is thrown ### Steps To Reproduce http://deck.net/5afe9300f8f9cda7776be8baf2a2375d ```csharp [IgnoreGeneric] public interface ITest1<TValues> { [IgnoreGeneric] string ToRoute(ITest1<TValues> routeDetails); } public class Test1<TValues> : ITest1<TValues> { public string ToRoute(ITest1<TValues> ifMatched) { return null; } } class Program { [Ready] static void Main() { try { Go(new Test1<string>()); Console.WriteLine("Succeeded"); } catch(Exception e) { Console.WriteLine(e.Message); } } private static void Go<TValues>(ITest1<TValues> routeDetails) { routeDetails.ToRoute(routeDetails); } } ``` Related to #1835.
1.0
Not defined JavaScript error with [IgnoreGeneric] on interfaces - Reported by @ProductiveRage on the Bridge Forums: http://forums.bridge.net/forum/bridge-net-pro/bugs/2708-ignoregeneric-on-interfaces-seems-to-break-bridge-15-0 ### Expected No Javascript error ### Actual A JavaScript `not defined` error is thrown ### Steps To Reproduce http://deck.net/5afe9300f8f9cda7776be8baf2a2375d ```csharp [IgnoreGeneric] public interface ITest1<TValues> { [IgnoreGeneric] string ToRoute(ITest1<TValues> routeDetails); } public class Test1<TValues> : ITest1<TValues> { public string ToRoute(ITest1<TValues> ifMatched) { return null; } } class Program { [Ready] static void Main() { try { Go(new Test1<string>()); Console.WriteLine("Succeeded"); } catch(Exception e) { Console.WriteLine(e.Message); } } private static void Go<TValues>(ITest1<TValues> routeDetails) { routeDetails.ToRoute(routeDetails); } } ``` Related to #1835.
defect
not defined javascript error with on interfaces reported by productiverage on the bridge forums expected no javascript error actual a javascript not defined error is thrown steps to reproduce csharp public interface string toroute routedetails public class public string toroute ifmatched return null class program static void main try go new console writeline succeeded catch exception e console writeline e message private static void go routedetails routedetails toroute routedetails related to
1
78,846
27,781,990,002
IssuesEvent
2023-03-16 22:00:04
dotCMS/core
https://api.github.com/repos/dotCMS/core
closed
Do not inject experiment code into customers html
Type : Defect Type : Enhancement Merged OKR : Application Performance QA : Passed Internal Team : Falcon dotCMS : Experiments Priority : 3 Average Next Release OKR : Customer Support
### Problem Statement Looking at this code: https://github.com/dotCMS/core/blob/master/dotCMS/src/main/java/com/dotmarketing/portlets/htmlpageasset/business/render/HTMLPageAssetRenderedAPIImpl.java#L320-L328 We have many customers who use pages to send back json or other types of content. - At this point we should never automatically inject experiment javascript code into customers html pages - We should have a viewtool that will spit out the required code that is simple to include in customer templates - If we wanted to do this in the future, we would need some sort of config switch AND make sure they had analytics configured AND make sure they had experiments running AND make sure the page was returning html before injecting AND we would want to use `indexOf(String)` to do the replacement rather than scanning ALL the HTML which is going to be an expensive operation. Bottom line, I don't think we should do it. ### Steps to Reproduce Look at the source of dotcms.com - we are injecting javascript right after the `<head>` tag ### Acceptance Criteria Don't inject the code. ### dotCMS Version 23.02 ### Proposed Objective Core Features ### Proposed Priority Please Select ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. _No response_ ### Assumptions & Initiation Needs _No response_ ### Sub-Tasks & Estimates _No response_
1.0
Do not inject experiment code into customers html - ### Problem Statement Looking at this code: https://github.com/dotCMS/core/blob/master/dotCMS/src/main/java/com/dotmarketing/portlets/htmlpageasset/business/render/HTMLPageAssetRenderedAPIImpl.java#L320-L328 We have many customers who use pages to send back json or other types of content. - At this point we should never automatically inject experiment javascript code into customers html pages - We should have a viewtool that will spit out the required code that is simple to include in customer templates - If we wanted to do this in the future, we would need some sort of config switch AND make sure they had analytics configured AND make sure they had experiments running AND make sure the page was returning html before injecting AND we would want to use `indexOf(String)` to do the replacement rather than scanning ALL the HTML which is going to be an expensive operation. Bottom line, I don't think we should do it. ### Steps to Reproduce Look at the source of dotcms.com - we are injecting javascript right after the `<head>` tag ### Acceptance Criteria Don't inject the code. ### dotCMS Version 23.02 ### Proposed Objective Core Features ### Proposed Priority Please Select ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. _No response_ ### Assumptions & Initiation Needs _No response_ ### Sub-Tasks & Estimates _No response_
defect
do not inject experiment code into customers html problem statement looking at this code we have many customers who use pages to send back json or other types of content at this point we should never automatically inject experiment javascript code into customers html pages we should have a viewtool that will spit out the required code that is simple to include in customer templates if we wanted to do this in the future we would need some sort of config switch and make sure they had analytics configured and make sure they had experiments running and make sure the page was returning html before injecting and we would want to use indexof string to do the replacement rather than scanning all the html which is going to be an expensive operation bottom line i don t think we should do it steps to reproduce look at the source of dotcms com we are injecting javascript right after the tag acceptance criteria don t inject the code dotcms version proposed objective core features proposed priority please select external links slack conversations support tickets figma designs etc no response assumptions initiation needs no response sub tasks estimates no response
1
17,074
11,632,582,387
IssuesEvent
2020-02-28 05:41:49
PowerShell/PSReadLine
https://api.github.com/repos/PowerShell/PSReadLine
reopened
Linux/Mac/VT: PSReadline in PowerShell 7 rc3 goes ballistic if you do not respond to its request for cursor position
Issue-Usability
## Environment ```none PS version: 7.0.0-rc.3 PSReadline version: 2.0.0 os: Linux antares 5.4.0-4-amd64 #1 SMP Debian 5.4.19-1 (2020-02-13) x86_64 GNU/Linux PS file version: 7.0.0.0 HostName: ConsoleHost BufferWidth: 120 BufferHeight: 50 ``` Also happens in Azure Cloud; this is likely in the Unix compat layer. ## Steps to reproduce Use a terminal emulator that does not respond to CSI `6` `n` (`DSR` Device Status Report 6, Cursor Position). Launch powershell. Wait. ## Expected behavior It doesn't explode. ## Actual behavior It dumps random memory to the output stream, then just exits. ![image](https://user-images.githubusercontent.com/14316954/75482044-e3ad4900-5958-11ea-85d4-73b53faef5c4.png)
True
Linux/Mac/VT: PSReadline in PowerShell 7 rc3 goes ballistic if you do not respond to its request for cursor position - ## Environment ```none PS version: 7.0.0-rc.3 PSReadline version: 2.0.0 os: Linux antares 5.4.0-4-amd64 #1 SMP Debian 5.4.19-1 (2020-02-13) x86_64 GNU/Linux PS file version: 7.0.0.0 HostName: ConsoleHost BufferWidth: 120 BufferHeight: 50 ``` Also happens in Azure Cloud; this is likely in the Unix compat layer. ## Steps to reproduce Use a terminal emulator that does not respond to CSI `6` `n` (`DSR` Device Status Report 6, Cursor Position). Launch powershell. Wait. ## Expected behavior It doesn't explode. ## Actual behavior It dumps random memory to the output stream, then just exits. ![image](https://user-images.githubusercontent.com/14316954/75482044-e3ad4900-5958-11ea-85d4-73b53faef5c4.png)
non_defect
linux mac vt psreadline in powershell goes ballistic if you do not respond to its request for cursor position environment none ps version rc psreadline version os linux antares smp debian gnu linux ps file version hostname consolehost bufferwidth bufferheight also happens in azure cloud this is likely in the unix compat layer steps to reproduce use a terminal emulator that does not respond to csi n dsr device status report cursor position launch powershell wait expected behavior it doesn t explode actual behavior it dumps random memory to the output stream then just exits
0
286,539
31,652,452,606
IssuesEvent
2023-09-07 00:19:06
samq-wsdemo/vuln_rust
https://api.github.com/repos/samq-wsdemo/vuln_rust
opened
CVE-2022-0326 (Medium) detected in libnghttp2-sys-0.1.6+1.43.0.crate
Mend: dependency security vulnerability
## CVE-2022-0326 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libnghttp2-sys-0.1.6+1.43.0.crate</b></p></summary> <p>FFI bindings for libnghttp2 (nghttp2) </p> <p>Library home page: <a href="https://crates.io/api/v1/crates/libnghttp2-sys/0.1.6+1.43.0/download">https://crates.io/api/v1/crates/libnghttp2-sys/0.1.6+1.43.0/download</a></p> <p>Path to dependency file: /Cargo.toml</p> <p>Path to vulnerable library: /home/wss-scanner/.cargo/registry/cache/github.com-1ecc6299db9ec823/libnghttp2-sys-0.1.6+1.43.0.crate</p> <p> Dependency Hierarchy: - chttp-0.5.5.crate (Root Library) - curl-0.4.35.crate - curl-sys-0.4.41+curl-7.75.0.crate - :x: **libnghttp2-sys-0.1.6+1.43.0.crate** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> NULL Pointer Dereference in Homebrew mruby prior to 3.2. <p>Publish Date: 2022-01-21 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0326>CVE-2022-0326</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-01-21</p> <p>Fix Resolution: 3.1.0</p> </p> </details> <p></p>
True
CVE-2022-0326 (Medium) detected in libnghttp2-sys-0.1.6+1.43.0.crate - ## CVE-2022-0326 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libnghttp2-sys-0.1.6+1.43.0.crate</b></p></summary> <p>FFI bindings for libnghttp2 (nghttp2) </p> <p>Library home page: <a href="https://crates.io/api/v1/crates/libnghttp2-sys/0.1.6+1.43.0/download">https://crates.io/api/v1/crates/libnghttp2-sys/0.1.6+1.43.0/download</a></p> <p>Path to dependency file: /Cargo.toml</p> <p>Path to vulnerable library: /home/wss-scanner/.cargo/registry/cache/github.com-1ecc6299db9ec823/libnghttp2-sys-0.1.6+1.43.0.crate</p> <p> Dependency Hierarchy: - chttp-0.5.5.crate (Root Library) - curl-0.4.35.crate - curl-sys-0.4.41+curl-7.75.0.crate - :x: **libnghttp2-sys-0.1.6+1.43.0.crate** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> NULL Pointer Dereference in Homebrew mruby prior to 3.2. <p>Publish Date: 2022-01-21 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0326>CVE-2022-0326</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-01-21</p> <p>Fix Resolution: 3.1.0</p> </p> </details> <p></p>
non_defect
cve medium detected in sys crate cve medium severity vulnerability vulnerable library sys crate ffi bindings for library home page a href path to dependency file cargo toml path to vulnerable library home wss scanner cargo registry cache github com sys crate dependency hierarchy chttp crate root library curl crate curl sys curl crate x sys crate vulnerable library found in base branch master vulnerability details null pointer dereference in homebrew mruby prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution
0
57,943
16,176,918,557
IssuesEvent
2021-05-03 08:26:14
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
URLs with umlauts are not always recognized
T-Defect
### Description URLs that contain german umlauts are not recognized correctly. ![Screenshot from 2021-05-03 10-17-03](https://user-images.githubusercontent.com/16013327/116855013-e65c9b80-abf8-11eb-86e0-2f212b5d2336.png) The URLs all look good on Element-Android. ### Version information - **Platform**: tested on web (self-hosted instance) and desktop For the web app: - **Browser**: Firefox 78.9.0esr - **OS**: Debian Bullseye - **Version**: 1.7.25 For the desktop app: - **OS**: Debian Bullseye - **Version**: 1.7.24
1.0
URLs with umlauts are not always recognized - ### Description URLs that contain german umlauts are not recognized correctly. ![Screenshot from 2021-05-03 10-17-03](https://user-images.githubusercontent.com/16013327/116855013-e65c9b80-abf8-11eb-86e0-2f212b5d2336.png) The URLs all look good on Element-Android. ### Version information - **Platform**: tested on web (self-hosted instance) and desktop For the web app: - **Browser**: Firefox 78.9.0esr - **OS**: Debian Bullseye - **Version**: 1.7.25 For the desktop app: - **OS**: Debian Bullseye - **Version**: 1.7.24
defect
urls with umlauts are not always recognized description urls that contain german umlauts are not recognized correctly the urls all look good on element android version information platform tested on web self hosted instance and desktop for the web app browser firefox os debian bullseye version for the desktop app os debian bullseye version
1
10,922
7,346,287,825
IssuesEvent
2018-03-07 20:12:09
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
storage: investigate excessive calls to assertStateLocked
performance
We seem to be calling `assertStateLocked` when `ReplicatedEvalResult.State != nil`, but `*ReplicatedEvalResult.State == storagebase.ReplicaState{}`. Noticed when running `tpcc -warehouses=10 -no-wait` against a local single node cockroach cluster. ```diff diff --git a/pkg/storage/replica_proposal.go b/pkg/storage/replica_proposal.go index 0343ccfba..9c1b2ef6c 100644 --- a/pkg/storage/replica_proposal.go +++ b/pkg/storage/replica_proposal.go @@ -484,9 +484,17 @@ func (r *Replica) handleReplicatedEvalResult( r.store.splitQueue.MaybeAdd(r, r.store.Clock().Now()) } + if rResult.State != nil && (*rResult.State == storagebase.ReplicaState{}) { + rResult.State = nil + } + // The above are always present, so we assert only if there are // "nontrivial" actions below. shouldAssert = !rResult.Equal(storagebase.ReplicatedEvalResult{}) + if shouldAssert { + diff := pretty.Diff(rResult, storagebase.ReplicatedEvalResult{}) + fmt.Printf("shouldAssert: %s\n", diff) + } // Process Split or Merge. This needs to happen after stats update because // of the ContainsEstimates hack. ```
True
storage: investigate excessive calls to assertStateLocked - We seem to be calling `assertStateLocked` when `ReplicatedEvalResult.State != nil`, but `*ReplicatedEvalResult.State == storagebase.ReplicaState{}`. Noticed when running `tpcc -warehouses=10 -no-wait` against a local single node cockroach cluster. ```diff diff --git a/pkg/storage/replica_proposal.go b/pkg/storage/replica_proposal.go index 0343ccfba..9c1b2ef6c 100644 --- a/pkg/storage/replica_proposal.go +++ b/pkg/storage/replica_proposal.go @@ -484,9 +484,17 @@ func (r *Replica) handleReplicatedEvalResult( r.store.splitQueue.MaybeAdd(r, r.store.Clock().Now()) } + if rResult.State != nil && (*rResult.State == storagebase.ReplicaState{}) { + rResult.State = nil + } + // The above are always present, so we assert only if there are // "nontrivial" actions below. shouldAssert = !rResult.Equal(storagebase.ReplicatedEvalResult{}) + if shouldAssert { + diff := pretty.Diff(rResult, storagebase.ReplicatedEvalResult{}) + fmt.Printf("shouldAssert: %s\n", diff) + } // Process Split or Merge. This needs to happen after stats update because // of the ContainsEstimates hack. ```
non_defect
storage investigate excessive calls to assertstatelocked we seem to be calling assertstatelocked when replicatedevalresult state nil but replicatedevalresult state storagebase replicastate noticed when running tpcc warehouses no wait against a local single node cockroach cluster diff diff git a pkg storage replica proposal go b pkg storage replica proposal go index a pkg storage replica proposal go b pkg storage replica proposal go func r replica handlereplicatedevalresult r store splitqueue maybeadd r r store clock now if rresult state nil rresult state storagebase replicastate rresult state nil the above are always present so we assert only if there are nontrivial actions below shouldassert rresult equal storagebase replicatedevalresult if shouldassert diff pretty diff rresult storagebase replicatedevalresult fmt printf shouldassert s n diff process split or merge this needs to happen after stats update because of the containsestimates hack
0
3,165
2,607,986,570
IssuesEvent
2015-02-26 00:51:59
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
closed
i wanna create a new php(or java) class with zencoding
auto-migrated Priority-Medium Type-Defect
``` demo on php: cls.People>pubv.name+prov.salary+priv.secret+pubf.__construct+prof.say+prif.eat+ pubf.__destruct <c-j> class People{ public $name; protected $salary; private $secret; public function __construct(){ } protected function say(){ } private function eat(){ } public function __destruct(){ } } ``` ----- Original issue reported on code.google.com by `wxs77...@gmail.com` on 6 Aug 2012 at 8:18
1.0
i wanna create a new php(or java) class with zencoding - ``` demo on php: cls.People>pubv.name+prov.salary+priv.secret+pubf.__construct+prof.say+prif.eat+ pubf.__destruct <c-j> class People{ public $name; protected $salary; private $secret; public function __construct(){ } protected function say(){ } private function eat(){ } public function __destruct(){ } } ``` ----- Original issue reported on code.google.com by `wxs77...@gmail.com` on 6 Aug 2012 at 8:18
defect
i wanna create a new php or java class with zencoding demo on php cls people pubv name prov salary priv secret pubf construct prof say prif eat pubf destruct class people public name protected salary private secret public function construct protected function say private function eat public function destruct original issue reported on code google com by gmail com on aug at
1
62,159
12,198,130,135
IssuesEvent
2020-04-29 22:09:54
kwk/test-llvm-bz-import-5
https://api.github.com/repos/kwk/test-llvm-bz-import-5
opened
Error in cleanup code for exception handling with volatile array type.
BZ-BUG-STATUS: NEW clang/LLVM Codegen dummy import from bugzilla
This issue was imported from Bugzilla https://bugs.llvm.org/show_bug.cgi?id=13579.
1.0
Error in cleanup code for exception handling with volatile array type. - This issue was imported from Bugzilla https://bugs.llvm.org/show_bug.cgi?id=13579.
non_defect
error in cleanup code for exception handling with volatile array type this issue was imported from bugzilla
0
591,769
17,860,843,263
IssuesEvent
2021-09-05 23:27:50
GabiAndi/UNER_Tesis_GUI_SCPA
https://api.github.com/repos/GabiAndi/UNER_Tesis_GUI_SCPA
closed
Protocolo de comunicación con SCPA [FEATURE]
enhancement priority: low
## Desarrollo del protocolo para la comunicación con el SCPA Incorporar la base para los comandos de comunicación y paso de información entre el controlador y el HMI. Esta comunicación se realiza mediante #2. Para que el controlador pueda realizar su trabajo y logre interactuar con los sistemas subyacentes es necesario una convención en cuanto a la transmisión y recepción de datos. Debido a que los dispositivos pertenecientes al sistema interactuan de manera diferente y utilizan sus propios protocolos de cumunicación, se realizaron diversas etapas para lograr que todo funcione como se esperarría. ### Rapsberry - HMI Como la comunicación se realiza de manera transparente mediante un protocolo TCP/IP #2 por puerto 33600 (valor en USD del bitcoin al día de hoy 5 de julio del 2021), se incorporó un conjunto de comandos personalizado. ## Comportamiento esperado Proporcionar una comunicación para la configuración mediante una interfaz gráfica de usuario. ### Estructura del protocolo ![Protocolo](https://user-images.githubusercontent.com/27433474/125672842-b96f1d88-3b78-496d-a044-c93a8536fe4f.png) ## Información sobre el sistema - OS: Linux Pop Os 20.04LTS, Windows 10 Pro.
1.0
Protocolo de comunicación con SCPA [FEATURE] - ## Desarrollo del protocolo para la comunicación con el SCPA Incorporar la base para los comandos de comunicación y paso de información entre el controlador y el HMI. Esta comunicación se realiza mediante #2. Para que el controlador pueda realizar su trabajo y logre interactuar con los sistemas subyacentes es necesario una convención en cuanto a la transmisión y recepción de datos. Debido a que los dispositivos pertenecientes al sistema interactuan de manera diferente y utilizan sus propios protocolos de cumunicación, se realizaron diversas etapas para lograr que todo funcione como se esperarría. ### Rapsberry - HMI Como la comunicación se realiza de manera transparente mediante un protocolo TCP/IP #2 por puerto 33600 (valor en USD del bitcoin al día de hoy 5 de julio del 2021), se incorporó un conjunto de comandos personalizado. ## Comportamiento esperado Proporcionar una comunicación para la configuración mediante una interfaz gráfica de usuario. ### Estructura del protocolo ![Protocolo](https://user-images.githubusercontent.com/27433474/125672842-b96f1d88-3b78-496d-a044-c93a8536fe4f.png) ## Información sobre el sistema - OS: Linux Pop Os 20.04LTS, Windows 10 Pro.
non_defect
protocolo de comunicación con scpa desarrollo del protocolo para la comunicación con el scpa incorporar la base para los comandos de comunicación y paso de información entre el controlador y el hmi esta comunicación se realiza mediante para que el controlador pueda realizar su trabajo y logre interactuar con los sistemas subyacentes es necesario una convención en cuanto a la transmisión y recepción de datos debido a que los dispositivos pertenecientes al sistema interactuan de manera diferente y utilizan sus propios protocolos de cumunicación se realizaron diversas etapas para lograr que todo funcione como se esperarría rapsberry hmi como la comunicación se realiza de manera transparente mediante un protocolo tcp ip por puerto valor en usd del bitcoin al día de hoy de julio del se incorporó un conjunto de comandos personalizado comportamiento esperado proporcionar una comunicación para la configuración mediante una interfaz gráfica de usuario estructura del protocolo información sobre el sistema os linux pop os windows pro
0
83,963
15,720,767,897
IssuesEvent
2021-03-29 01:08:39
LalithK90/Welfare
https://api.github.com/repos/LalithK90/Welfare
opened
CVE-2020-1935 (Medium) detected in tomcat-embed-core-9.0.30.jar
security vulnerability
## CVE-2020-1935 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.30.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to dependency file: Welfare/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.30/ad32909314fe2ba02cec036434c0addd19bcc580/tomcat-embed-core-9.0.30.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.4.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.2.4.RELEASE.jar - :x: **tomcat-embed-core-9.0.30.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Tomcat 9.0.0.M1 to 9.0.30, 8.5.0 to 8.5.50 and 7.0.0 to 7.0.99 the HTTP header parsing code used an approach to end-of-line parsing that allowed some invalid HTTP headers to be parsed as valid. This led to a possibility of HTTP Request Smuggling if Tomcat was located behind a reverse proxy that incorrectly handled the invalid Transfer-Encoding header in a particular manner. Such a reverse proxy is considered unlikely. <p>Publish Date: 2020-02-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1935>CVE-2020-1935</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-6v7p-v754-j89v">https://github.com/advisories/GHSA-6v7p-v754-j89v</a></p> <p>Release Date: 2020-02-24</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.100,8.5.51,9.0.31;org.apache.tomcat:tomcat-coyote:7.0.100,8.5.51,9.0.31</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-1935 (Medium) detected in tomcat-embed-core-9.0.30.jar - ## CVE-2020-1935 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tomcat-embed-core-9.0.30.jar</b></p></summary> <p>Core Tomcat implementation</p> <p>Library home page: <a href="https://tomcat.apache.org/">https://tomcat.apache.org/</a></p> <p>Path to dependency file: Welfare/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.tomcat.embed/tomcat-embed-core/9.0.30/ad32909314fe2ba02cec036434c0addd19bcc580/tomcat-embed-core-9.0.30.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.4.RELEASE.jar (Root Library) - spring-boot-starter-tomcat-2.2.4.RELEASE.jar - :x: **tomcat-embed-core-9.0.30.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Apache Tomcat 9.0.0.M1 to 9.0.30, 8.5.0 to 8.5.50 and 7.0.0 to 7.0.99 the HTTP header parsing code used an approach to end-of-line parsing that allowed some invalid HTTP headers to be parsed as valid. This led to a possibility of HTTP Request Smuggling if Tomcat was located behind a reverse proxy that incorrectly handled the invalid Transfer-Encoding header in a particular manner. Such a reverse proxy is considered unlikely. <p>Publish Date: 2020-02-24 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-1935>CVE-2020-1935</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-6v7p-v754-j89v">https://github.com/advisories/GHSA-6v7p-v754-j89v</a></p> <p>Release Date: 2020-02-24</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:7.0.100,8.5.51,9.0.31;org.apache.tomcat:tomcat-coyote:7.0.100,8.5.51,9.0.31</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in tomcat embed core jar cve medium severity vulnerability vulnerable library tomcat embed core jar core tomcat implementation library home page a href path to dependency file welfare build gradle path to vulnerable library home wss scanner gradle caches modules files org apache tomcat embed tomcat embed core tomcat embed core jar dependency hierarchy spring boot starter web release jar root library spring boot starter tomcat release jar x tomcat embed core jar vulnerable library found in base branch master vulnerability details in apache tomcat to to and to the http header parsing code used an approach to end of line parsing that allowed some invalid http headers to be parsed as valid this led to a possibility of http request smuggling if tomcat was located behind a reverse proxy that incorrectly handled the invalid transfer encoding header in a particular manner such a reverse proxy is considered unlikely publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat coyote step up your open source security game with whitesource
0
116,277
24,892,540,288
IssuesEvent
2022-10-28 13:16:09
llvm/llvm-project
https://api.github.com/repos/llvm/llvm-project
closed
A/F: `Elt->getBitWidth( ) == EltVT.getSizeInBits() && "APInt size does not match type size!"'
llvm:codegen
Two of our internal tests recently hit an assertion failure when compiling which I bisected back to commit 54eeadcf442df91aed0fb7244fe7885cdf1b1f3d. I was able to reduce the failing code to the following c++ sample: ```c++ template <typename a, typename> a b(a c, int) { return c; } typedef char d; typedef d __attribute__((ext_vector_type(2))) e; typedef char __attribute__((ext_vector_type(2))) f; #define g(h, i) (b<e, d>(h, 2) % 2) e j; void k() { e l{}, m = __builtin_shufflevector(l, j, 3, 1), n = m.yx, o g(n, ); volatile f p(o); } ``` To reproduce the assertion failure, compile the above code with optimizations enabled (-O2): ``` $ ~/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang -c -O2 test.cpp clang: /home/dyung/src/upstream/llvm_clean_git/llvm/lib/CodeGen/SelectionDAG/SelectionDAG.cpp:1593: llvm::SDValue llvm::SelectionDAG::getConstant(const llvm::ConstantInt&, const llvm::SDLoc&, llvm::EVT, bool, bool): Assertion `Elt->getBitWidth( ) == EltVT.getSizeInBits() && "APInt size does not match type size!"' failed. PLEASE submit a bug report to https://github.com/llvm/llvm-project/issues/ and include the crash backtrace, preprocessed source, and associated run script. Stack dump: 0. Program arguments: /home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang -c -O2 test.cpp 1. <eof> parser at end of file 2. Code generation 3. Running pass 'Function Pass Manager' on module 'test.cpp'. 4. Running pass 'X86 DAG->DAG Instruction Selection' on function '@_Z1kv' #0 0x000056036638b464 PrintStackTraceSignalHandler(void*) Signals.cpp:0:0 #1 0x00005603663891fc llvm::sys::CleanupOnSignal(unsigned long) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x3eff1fc) #2 0x00005603662c4a78 CrashRecoverySignalHandler(int) CrashRecoveryContext.cpp:0:0 #3 0x00007f50242b5420 __restore_rt (/lib/x86_64-linux-gnu/libpthread.so.0+0x14420) #4 0x00007f5023d8200b raise /build/glibc-SzIz7B/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 #5 0x00007f5023d61859 abort /build/glibc-SzIz7B/glibc-2.31/stdlib/abort.c:81:7 #6 0x00007f5023d61729 get_sysdep_segment_value /build/glibc-SzIz7B/glibc-2.31/intl/loadmsgcat.c:509:8 #7 0x00007f5023d61729 _nl_load_domain /build/glibc-SzIz7B/glibc-2.31/intl/loadmsgcat.c:970:34 #8 0x00007f5023d72fd6 (/lib/x86_64-linux-gnu/libc.so.6+0x33fd6) #9 0x00005603675289a3 llvm::SelectionDAG::getConstant(llvm::ConstantInt const&, llvm::SDLoc const&, llvm::EVT, bool, bool) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x509e9a3) #10 0x00005603673bc088 (anonymous namespace)::DAGCombiner::visitSCALAR_TO_VECTOR(llvm::SDNode*) DAGCombiner.cpp:0:0 #11 0x000056036743459e (anonymous namespace)::DAGCombiner::visit(llvm::SDNode*) DAGCombiner.cpp:0:0 #12 0x0000560367436c35 (anonymous namespace)::DAGCombiner::combine(llvm::SDNode*) DAGCombiner.cpp:0:0 #13 0x00005603674383a0 llvm::SelectionDAG::Combine(llvm::CombineLevel, llvm::AAResults*, llvm::CodeGenOpt::Level) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x4fae3a0) #14 0x0000560367557edd llvm::SelectionDAGISel::CodeGenAndEmitDAG() (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x50cdedd) #15 0x000056036755ba10 llvm::SelectionDAGISel::SelectAllBasicBlocks(llvm::Function const&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x50d1a10) #16 0x000056036755dbbd llvm::SelectionDAGISel::runOnMachineFunction(llvm::MachineFunction&) (.part.0) SelectionDAGISel.cpp:0:0 #17 0x0000560364dbafd0 (anonymous namespace)::X86DAGToDAGISel::runOnMachineFunction(llvm::MachineFunction&) X86ISelDAGToDAG.cpp:0:0 #18 0x000056036558acbe llvm::MachineFunctionPass::runOnFunction(llvm::Function&) (.part.0) MachineFunctionPass.cpp:0:0 #19 0x0000560365ae42d5 llvm::FPPassManager::runOnFunction(llvm::Function&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x365a2d5) #20 0x0000560365ae4519 llvm::FPPassManager::runOnModule(llvm::Module&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x365a519) #21 0x0000560365ae4d62 llvm::legacy::PassManagerImpl::run(llvm::Module&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x365ad62) #22 0x0000560366761184 clang::EmitBackendOutput(clang::DiagnosticsEngine&, clang::HeaderSearchOptions const&, clang::CodeGenOptions const&, clang::TargetOptions const&, clang::LangOptions const&, llvm::StringRef, llvm::Module*, clang::BackendAc tion, std::unique_ptr<llvm::raw_pwrite_stream, std::default_delete<llvm::raw_pwrite_stream>>) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x42d7184) #23 0x00005603676b824b clang::BackendConsumer::HandleTranslationUnit(clang::ASTContext&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x522e24b) #24 0x00005603685caa55 clang::ParseAST(clang::Sema&, bool, bool) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x6140a55) #25 0x00005603676b6c78 clang::CodeGenAction::ExecuteAction() (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x522cc78) #26 0x0000560366f485a9 clang::FrontendAction::Execute() (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x4abe5a9) #27 0x0000560366ecf4be clang::CompilerInstance::ExecuteAction(clang::FrontendAction&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x4a454be) #28 0x000056036702e113 clang::ExecuteCompilerInvocation(clang::CompilerInstance*) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x4ba4113) #29 0x000056036384ca54 cc1_main(llvm::ArrayRef<char const*>, char const*, void*) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x13c2a54) #30 0x0000560363845b68 ExecuteCC1Tool(llvm::SmallVectorImpl<char const*>&) driver.cpp:0:0 #31 0x0000560366d3c229 void llvm::function_ref<void ()>::callback_fn<clang::driver::CC1Command::Execute(llvm::ArrayRef<llvm::Optional<llvm::StringRef>>, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char>>*, bool*) con st::'lambda'()>(long) Job.cpp:0:0 #32 0x00005603662c521a llvm::CrashRecoveryContext::RunSafely(llvm::function_ref<void ()>) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x3e3b21a) #33 0x0000560366d3ca7f clang::driver::CC1Command::Execute(llvm::ArrayRef<llvm::Optional<llvm::StringRef>>, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char>>*, bool*) const (.part.0) Job.cpp:0:0 #34 0x0000560366d05f49 clang::driver::Compilation::ExecuteCommand(clang::driver::Command const&, clang::driver::Command const*&, bool) const (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x487bf49) #35 0x0000560366d069cd clang::driver::Compilation::ExecuteJobs(clang::driver::JobList const&, llvm::SmallVectorImpl<std::pair<int, clang::driver::Command const*>>&, bool) const (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d- linux/bin/clang+0x487c9cd) #36 0x0000560366d0ffec clang::driver::Driver::ExecuteCompilation(clang::driver::Compilation&, llvm::SmallVectorImpl<std::pair<int, clang::driver::Command const*>>&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/cl ang+0x4885fec) #37 0x000056036384ae93 clang_main(int, char**) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x13c0e93) #38 0x00007f5023d63083 __libc_start_main /build/glibc-SzIz7B/glibc-2.31/csu/../csu/libc-start.c:342:3 #39 0x000056036384576e _start (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x13bb76e) clang-16: error: clang frontend command failed with exit code 134 (use -v to see invocation) clang version 16.0.0 (https://github.com/llvm/llvm-project.git 54eeadcf442df91aed0fb7244fe7885cdf1b1f3d) Target: x86_64-unknown-linux-gnu ```
1.0
A/F: `Elt->getBitWidth( ) == EltVT.getSizeInBits() && "APInt size does not match type size!"' - Two of our internal tests recently hit an assertion failure when compiling which I bisected back to commit 54eeadcf442df91aed0fb7244fe7885cdf1b1f3d. I was able to reduce the failing code to the following c++ sample: ```c++ template <typename a, typename> a b(a c, int) { return c; } typedef char d; typedef d __attribute__((ext_vector_type(2))) e; typedef char __attribute__((ext_vector_type(2))) f; #define g(h, i) (b<e, d>(h, 2) % 2) e j; void k() { e l{}, m = __builtin_shufflevector(l, j, 3, 1), n = m.yx, o g(n, ); volatile f p(o); } ``` To reproduce the assertion failure, compile the above code with optimizations enabled (-O2): ``` $ ~/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang -c -O2 test.cpp clang: /home/dyung/src/upstream/llvm_clean_git/llvm/lib/CodeGen/SelectionDAG/SelectionDAG.cpp:1593: llvm::SDValue llvm::SelectionDAG::getConstant(const llvm::ConstantInt&, const llvm::SDLoc&, llvm::EVT, bool, bool): Assertion `Elt->getBitWidth( ) == EltVT.getSizeInBits() && "APInt size does not match type size!"' failed. PLEASE submit a bug report to https://github.com/llvm/llvm-project/issues/ and include the crash backtrace, preprocessed source, and associated run script. Stack dump: 0. Program arguments: /home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang -c -O2 test.cpp 1. <eof> parser at end of file 2. Code generation 3. Running pass 'Function Pass Manager' on module 'test.cpp'. 4. Running pass 'X86 DAG->DAG Instruction Selection' on function '@_Z1kv' #0 0x000056036638b464 PrintStackTraceSignalHandler(void*) Signals.cpp:0:0 #1 0x00005603663891fc llvm::sys::CleanupOnSignal(unsigned long) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x3eff1fc) #2 0x00005603662c4a78 CrashRecoverySignalHandler(int) CrashRecoveryContext.cpp:0:0 #3 0x00007f50242b5420 __restore_rt (/lib/x86_64-linux-gnu/libpthread.so.0+0x14420) #4 0x00007f5023d8200b raise /build/glibc-SzIz7B/glibc-2.31/signal/../sysdeps/unix/sysv/linux/raise.c:51:1 #5 0x00007f5023d61859 abort /build/glibc-SzIz7B/glibc-2.31/stdlib/abort.c:81:7 #6 0x00007f5023d61729 get_sysdep_segment_value /build/glibc-SzIz7B/glibc-2.31/intl/loadmsgcat.c:509:8 #7 0x00007f5023d61729 _nl_load_domain /build/glibc-SzIz7B/glibc-2.31/intl/loadmsgcat.c:970:34 #8 0x00007f5023d72fd6 (/lib/x86_64-linux-gnu/libc.so.6+0x33fd6) #9 0x00005603675289a3 llvm::SelectionDAG::getConstant(llvm::ConstantInt const&, llvm::SDLoc const&, llvm::EVT, bool, bool) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x509e9a3) #10 0x00005603673bc088 (anonymous namespace)::DAGCombiner::visitSCALAR_TO_VECTOR(llvm::SDNode*) DAGCombiner.cpp:0:0 #11 0x000056036743459e (anonymous namespace)::DAGCombiner::visit(llvm::SDNode*) DAGCombiner.cpp:0:0 #12 0x0000560367436c35 (anonymous namespace)::DAGCombiner::combine(llvm::SDNode*) DAGCombiner.cpp:0:0 #13 0x00005603674383a0 llvm::SelectionDAG::Combine(llvm::CombineLevel, llvm::AAResults*, llvm::CodeGenOpt::Level) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x4fae3a0) #14 0x0000560367557edd llvm::SelectionDAGISel::CodeGenAndEmitDAG() (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x50cdedd) #15 0x000056036755ba10 llvm::SelectionDAGISel::SelectAllBasicBlocks(llvm::Function const&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x50d1a10) #16 0x000056036755dbbd llvm::SelectionDAGISel::runOnMachineFunction(llvm::MachineFunction&) (.part.0) SelectionDAGISel.cpp:0:0 #17 0x0000560364dbafd0 (anonymous namespace)::X86DAGToDAGISel::runOnMachineFunction(llvm::MachineFunction&) X86ISelDAGToDAG.cpp:0:0 #18 0x000056036558acbe llvm::MachineFunctionPass::runOnFunction(llvm::Function&) (.part.0) MachineFunctionPass.cpp:0:0 #19 0x0000560365ae42d5 llvm::FPPassManager::runOnFunction(llvm::Function&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x365a2d5) #20 0x0000560365ae4519 llvm::FPPassManager::runOnModule(llvm::Module&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x365a519) #21 0x0000560365ae4d62 llvm::legacy::PassManagerImpl::run(llvm::Module&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x365ad62) #22 0x0000560366761184 clang::EmitBackendOutput(clang::DiagnosticsEngine&, clang::HeaderSearchOptions const&, clang::CodeGenOptions const&, clang::TargetOptions const&, clang::LangOptions const&, llvm::StringRef, llvm::Module*, clang::BackendAc tion, std::unique_ptr<llvm::raw_pwrite_stream, std::default_delete<llvm::raw_pwrite_stream>>) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x42d7184) #23 0x00005603676b824b clang::BackendConsumer::HandleTranslationUnit(clang::ASTContext&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x522e24b) #24 0x00005603685caa55 clang::ParseAST(clang::Sema&, bool, bool) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x6140a55) #25 0x00005603676b6c78 clang::CodeGenAction::ExecuteAction() (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x522cc78) #26 0x0000560366f485a9 clang::FrontendAction::Execute() (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x4abe5a9) #27 0x0000560366ecf4be clang::CompilerInstance::ExecuteAction(clang::FrontendAction&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x4a454be) #28 0x000056036702e113 clang::ExecuteCompilerInvocation(clang::CompilerInstance*) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x4ba4113) #29 0x000056036384ca54 cc1_main(llvm::ArrayRef<char const*>, char const*, void*) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x13c2a54) #30 0x0000560363845b68 ExecuteCC1Tool(llvm::SmallVectorImpl<char const*>&) driver.cpp:0:0 #31 0x0000560366d3c229 void llvm::function_ref<void ()>::callback_fn<clang::driver::CC1Command::Execute(llvm::ArrayRef<llvm::Optional<llvm::StringRef>>, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char>>*, bool*) con st::'lambda'()>(long) Job.cpp:0:0 #32 0x00005603662c521a llvm::CrashRecoveryContext::RunSafely(llvm::function_ref<void ()>) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x3e3b21a) #33 0x0000560366d3ca7f clang::driver::CC1Command::Execute(llvm::ArrayRef<llvm::Optional<llvm::StringRef>>, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char>>*, bool*) const (.part.0) Job.cpp:0:0 #34 0x0000560366d05f49 clang::driver::Compilation::ExecuteCommand(clang::driver::Command const&, clang::driver::Command const*&, bool) const (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x487bf49) #35 0x0000560366d069cd clang::driver::Compilation::ExecuteJobs(clang::driver::JobList const&, llvm::SmallVectorImpl<std::pair<int, clang::driver::Command const*>>&, bool) const (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d- linux/bin/clang+0x487c9cd) #36 0x0000560366d0ffec clang::driver::Driver::ExecuteCompilation(clang::driver::Compilation&, llvm::SmallVectorImpl<std::pair<int, clang::driver::Command const*>>&) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/cl ang+0x4885fec) #37 0x000056036384ae93 clang_main(int, char**) (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x13c0e93) #38 0x00007f5023d63083 __libc_start_main /build/glibc-SzIz7B/glibc-2.31/csu/../csu/libc-start.c:342:3 #39 0x000056036384576e _start (/home/dyung/src/upstream/54eeadcf442df91aed0fb7244fe7885cdf1b1f3d-linux/bin/clang+0x13bb76e) clang-16: error: clang frontend command failed with exit code 134 (use -v to see invocation) clang version 16.0.0 (https://github.com/llvm/llvm-project.git 54eeadcf442df91aed0fb7244fe7885cdf1b1f3d) Target: x86_64-unknown-linux-gnu ```
non_defect
a f elt getbitwidth eltvt getsizeinbits apint size does not match type size two of our internal tests recently hit an assertion failure when compiling which i bisected back to commit i was able to reduce the failing code to the following c sample c template a b a c int return c typedef char d typedef d attribute ext vector type e typedef char attribute ext vector type f define g h i b h e j void k e l m builtin shufflevector l j n m yx o g n volatile f p o to reproduce the assertion failure compile the above code with optimizations enabled src upstream linux bin clang c test cpp clang home dyung src upstream llvm clean git llvm lib codegen selectiondag selectiondag cpp llvm sdvalue llvm selectiondag getconstant const llvm constantint const llvm sdloc llvm evt bool bool assertion elt getbitwidth eltvt getsizeinbits apint size does not match type size failed please submit a bug report to and include the crash backtrace preprocessed source and associated run script stack dump program arguments home dyung src upstream linux bin clang c test cpp parser at end of file code generation running pass function pass manager on module test cpp running pass dag dag instruction selection on function printstacktracesignalhandler void signals cpp llvm sys cleanuponsignal unsigned long home dyung src upstream linux bin clang crashrecoverysignalhandler int crashrecoverycontext cpp restore rt lib linux gnu libpthread so raise build glibc glibc signal sysdeps unix sysv linux raise c abort build glibc glibc stdlib abort c get sysdep segment value build glibc glibc intl loadmsgcat c nl load domain build glibc glibc intl loadmsgcat c lib linux gnu libc so llvm selectiondag getconstant llvm constantint const llvm sdloc const llvm evt bool bool home dyung src upstream linux bin clang anonymous namespace dagcombiner visitscalar to vector llvm sdnode dagcombiner cpp anonymous namespace dagcombiner visit llvm sdnode dagcombiner cpp anonymous namespace dagcombiner combine llvm sdnode dagcombiner cpp llvm selectiondag combine llvm combinelevel llvm aaresults llvm codegenopt level home dyung src upstream linux bin clang llvm selectiondagisel codegenandemitdag home dyung src upstream linux bin clang llvm selectiondagisel selectallbasicblocks llvm function const home dyung src upstream linux bin clang llvm selectiondagisel runonmachinefunction llvm machinefunction part selectiondagisel cpp anonymous namespace runonmachinefunction llvm machinefunction cpp llvm machinefunctionpass runonfunction llvm function part machinefunctionpass cpp llvm fppassmanager runonfunction llvm function home dyung src upstream linux bin clang llvm fppassmanager runonmodule llvm module home dyung src upstream linux bin clang llvm legacy passmanagerimpl run llvm module home dyung src upstream linux bin clang clang emitbackendoutput clang diagnosticsengine clang headersearchoptions const clang codegenoptions const clang targetoptions const clang langoptions const llvm stringref llvm module clang backendac tion std unique ptr home dyung src upstream linux bin clang clang backendconsumer handletranslationunit clang astcontext home dyung src upstream linux bin clang clang parseast clang sema bool bool home dyung src upstream linux bin clang clang codegenaction executeaction home dyung src upstream linux bin clang clang frontendaction execute home dyung src upstream linux bin clang clang compilerinstance executeaction clang frontendaction home dyung src upstream linux bin clang clang executecompilerinvocation clang compilerinstance home dyung src upstream linux bin clang main llvm arrayref char const void home dyung src upstream linux bin clang llvm smallvectorimpl driver cpp void llvm function ref callback fn std basic string std allocator bool con st lambda long job cpp llvm crashrecoverycontext runsafely llvm function ref home dyung src upstream linux bin clang clang driver execute llvm arrayref std basic string std allocator bool const part job cpp clang driver compilation executecommand clang driver command const clang driver command const bool const home dyung src upstream linux bin clang clang driver compilation executejobs clang driver joblist const llvm smallvectorimpl bool const home dyung src upstream linux bin clang clang driver driver executecompilation clang driver compilation llvm smallvectorimpl home dyung src upstream linux bin cl ang clang main int char home dyung src upstream linux bin clang libc start main build glibc glibc csu csu libc start c start home dyung src upstream linux bin clang clang error clang frontend command failed with exit code use v to see invocation clang version target unknown linux gnu
0
381,487
26,454,773,945
IssuesEvent
2023-01-16 13:45:47
scs/substrate-api-client
https://api.github.com/repos/scs/substrate-api-client
closed
Add explanatory no_std usage to README
F4-documentation Q3-substantial
After solving #279, it should be explained how to use the api client in no_std mode. E.g. - how to implement an rpc-client (what interfaces to implement) - how to use Runtime in no-std (bit problematic, see #391) maybe a new readme needs to be added.
1.0
Add explanatory no_std usage to README - After solving #279, it should be explained how to use the api client in no_std mode. E.g. - how to implement an rpc-client (what interfaces to implement) - how to use Runtime in no-std (bit problematic, see #391) maybe a new readme needs to be added.
non_defect
add explanatory no std usage to readme after solving it should be explained how to use the api client in no std mode e g how to implement an rpc client what interfaces to implement how to use runtime in no std bit problematic see maybe a new readme needs to be added
0
80,089
30,003,032,976
IssuesEvent
2023-06-26 10:33:18
vector-im/element-x-ios
https://api.github.com/repos/vector-im/element-x-ios
closed
Handle the cancelled state properly
T-Defect
### Steps to reproduce Only the first unsent message is treated as unsent, because the cancelled state is not handled properly. ### Outcome We should handle it as an unsent state. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
Handle the cancelled state properly - ### Steps to reproduce Only the first unsent message is treated as unsent, because the cancelled state is not handled properly. ### Outcome We should handle it as an unsent state. ### Your phone model _No response_ ### Operating system version _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
handle the cancelled state properly steps to reproduce only the first unsent message is treated as unsent because the cancelled state is not handled properly outcome we should handle it as an unsent state your phone model no response operating system version no response application version no response homeserver no response will you send logs no
1
10,311
6,669,766,412
IssuesEvent
2017-10-03 20:34:41
bcgov/DBC-APIM
https://api.github.com/repos/bcgov/DBC-APIM
closed
UAT
usability
Once GWA and latest DBC-APIM code has been migrated to PROD conduct UAT with our DBC-APIM dev team, BC Gov't API owners and developers. Feedback to be reviewed and refinements / bugs / enhancements to be added to this board as issues. Target start date: August 25th, 2017 Tests: 1. Gateway administrator app 2. API owner app (DataBC API owners for Geocoder and Route Planner) 3. Developer app (anyone with GitHub account) 4. Simulate a user workflow starting in BCDC, as a new developer with no group grants yet. 5. Confirm that the Geocoder doesn't require an email to DataBC, no groups should be defined. Geocoder is open to all new developers. Route Planner is the opposite, with groups, and only a few current members.
True
UAT - Once GWA and latest DBC-APIM code has been migrated to PROD conduct UAT with our DBC-APIM dev team, BC Gov't API owners and developers. Feedback to be reviewed and refinements / bugs / enhancements to be added to this board as issues. Target start date: August 25th, 2017 Tests: 1. Gateway administrator app 2. API owner app (DataBC API owners for Geocoder and Route Planner) 3. Developer app (anyone with GitHub account) 4. Simulate a user workflow starting in BCDC, as a new developer with no group grants yet. 5. Confirm that the Geocoder doesn't require an email to DataBC, no groups should be defined. Geocoder is open to all new developers. Route Planner is the opposite, with groups, and only a few current members.
non_defect
uat once gwa and latest dbc apim code has been migrated to prod conduct uat with our dbc apim dev team bc gov t api owners and developers feedback to be reviewed and refinements bugs enhancements to be added to this board as issues target start date august tests gateway administrator app api owner app databc api owners for geocoder and route planner developer app anyone with github account simulate a user workflow starting in bcdc as a new developer with no group grants yet confirm that the geocoder doesn t require an email to databc no groups should be defined geocoder is open to all new developers route planner is the opposite with groups and only a few current members
0
516,783
14,987,830,593
IssuesEvent
2021-01-28 23:46:43
Phildesro123/Reliant
https://api.github.com/repos/Phildesro123/Reliant
opened
Create Component for Author Score
Priority: High Status: Available help wanted
![image.png](https://images.zenhubusercontent.com/5c5a203ef82e87206eca9a08/f4ed0ec7-778e-4d79-b0d4-8fc404472aec) We need to create a component for the background info and author score, similar to the one seen in the prototype. - [ ] Create new component that will activate when we click on extension Icon - [ ] Optional: Test to see if we can extract some information from a test page
1.0
Create Component for Author Score - ![image.png](https://images.zenhubusercontent.com/5c5a203ef82e87206eca9a08/f4ed0ec7-778e-4d79-b0d4-8fc404472aec) We need to create a component for the background info and author score, similar to the one seen in the prototype. - [ ] Create new component that will activate when we click on extension Icon - [ ] Optional: Test to see if we can extract some information from a test page
non_defect
create component for author score we need to create a component for the background info and author score similar to the one seen in the prototype create new component that will activate when we click on extension icon optional test to see if we can extract some information from a test page
0
559,277
16,554,225,602
IssuesEvent
2021-05-28 12:11:08
sopra-fs21-group-10/td-client
https://api.github.com/repos/sopra-fs21-group-10/td-client
closed
New miniontype with large amount of hp
low priority task
- implementation of a minion which can take much more damage than others - it also deals more damage, should it reach the end of the path (exact value can vary/ balancing is not considered yet) - it also moves slower than the "normal" minion Estimated time: 2h
1.0
New miniontype with large amount of hp - - implementation of a minion which can take much more damage than others - it also deals more damage, should it reach the end of the path (exact value can vary/ balancing is not considered yet) - it also moves slower than the "normal" minion Estimated time: 2h
non_defect
new miniontype with large amount of hp implementation of a minion which can take much more damage than others it also deals more damage should it reach the end of the path exact value can vary balancing is not considered yet it also moves slower than the normal minion estimated time
0
76,033
26,206,199,062
IssuesEvent
2023-01-03 22:59:16
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Flux BC for WCNSFV momentum equation gives wrong results if fluid flows in the negative (x,y,z) directions.
T: defect P: normal
## Bug Description The flux postprocessor assumes positive velocities and mass flow rates for inlets. The `WCNSFVMomentumFluxInlet` has the absolute value of the normal which yields weird results when the inlet velocity direction is negative in the chartesian coordinate system. ## Steps to Reproduce Take a channel, use an inlet in the negative (x,y,z) direction and a positive number for massflow. It will give nonphysical results. Try running the input file below. ``` bed_height = 10.0 bed_radius = 1.2 bed_porosity = 1 outlet_pressure = 5.5e6 T_fluid = 300 #density = 8.60161 #mass_flow_rate = 60.0 #flow_area = ${fparse pi * bed_radius * bed_radius * bed_porosity} #flow_vel = ${fparse mass_flow_rate / flow_area / density} [GlobalParams] fp = fluid_properties_obj porosity = 'porosity' rhie_chow_user_object = pins_rhie_chow_interpolator [] [Mesh] [gen] type = GeneratedMeshGenerator dim = 2 xmin = 0 xmax = ${bed_radius} ymin = 0 ymax = ${bed_height} nx = 6 ny = 40 [] coord_type = RZ [] [Modules] [FluidProperties] [fluid_properties_obj] type = HeliumFluidProperties [] [] [NavierStokesFV] compressibility = 'weakly-compressible' porous_medium_treatment = true density = 'rho' dynamic_viscosity = 'mu' porosity = 'porosity' initial_velocity = '1e-6 0 0' initial_pressure = 5.4e6 inlet_boundaries = top momentum_inlet_types = flux-mass flux_inlet_pps = mfr_pp #momentum_inlet_types = 'fixed-velocity' #momentum_inlet_function = '0 -${flow_vel}' wall_boundaries = 'left right' momentum_wall_types = 'slip slip' outlet_boundaries = bottom momentum_outlet_types = fixed-pressure pressure_function = ${outlet_pressure} pressure_face_interpolation = average momentum_advection_interpolation = upwind mass_advection_interpolation = upwind [] [] [Materials] ## Fluid properties and non-dimensional numbers [fluid_props_to_mat_props] type = GeneralFunctorFluidProps pressure = 'pressure' T_fluid = ${T_fluid} speed = 'speed' characteristic_length = characteristic_length [] [hydraulic_diameter] type = PiecewiseByBlockFunctorMaterial prop_name = 'characteristic_length' subdomain_to_prop_value = '0 1' [] [] [AuxVariables] [porosity] family = MONOMIAL order = CONSTANT fv = true initial_condition = ${bed_porosity} [] [density_var] family = MONOMIAL order = CONSTANT fv = true [] [] [AuxKernels] [density_aux] type = ADFunctorElementalAux variable = density_var functor = rho [] [] [Executioner] type = Transient end_time = 100 [TimeStepper] type = IterationAdaptiveDT iteration_window = 2 optimal_iterations = 8 cutback_factor = 0.8 growth_factor = 2 dt = 1e-5 [] dtmax = 1 line_search = l2 solve_type = 'NEWTON' petsc_options_iname = '-pc_type -pc_factor_shift_type' petsc_options_value = 'lu NONZERO' nl_rel_tol = 1e-6 nl_abs_tol = 1e-6 [] [Postprocessors] [mfr_pp] type = Receiver default = 60 [] [inlet_mfr] type = VolumetricFlowRate advected_quantity = rho vel_x = 'superficial_vel_x' vel_y = 'superficial_vel_y' boundary = 'top' [] [outlet_mfr] type = VolumetricFlowRate advected_quantity = rho vel_x = 'superficial_vel_x' vel_y = 'superficial_vel_y' boundary = 'bottom' [] [] [Outputs] exodus = true [] ``` The expected result is to get a uniform velocity field (considering the slip BCs). However it gives a spurious oscillation close to the inlet due to the wrong mass flow rate. ## Impact Users who assume that mass flow is always positive and don't factor in the surface normal when computing the inlet value might experience nonphysical results.
1.0
Flux BC for WCNSFV momentum equation gives wrong results if fluid flows in the negative (x,y,z) directions. - ## Bug Description The flux postprocessor assumes positive velocities and mass flow rates for inlets. The `WCNSFVMomentumFluxInlet` has the absolute value of the normal which yields weird results when the inlet velocity direction is negative in the chartesian coordinate system. ## Steps to Reproduce Take a channel, use an inlet in the negative (x,y,z) direction and a positive number for massflow. It will give nonphysical results. Try running the input file below. ``` bed_height = 10.0 bed_radius = 1.2 bed_porosity = 1 outlet_pressure = 5.5e6 T_fluid = 300 #density = 8.60161 #mass_flow_rate = 60.0 #flow_area = ${fparse pi * bed_radius * bed_radius * bed_porosity} #flow_vel = ${fparse mass_flow_rate / flow_area / density} [GlobalParams] fp = fluid_properties_obj porosity = 'porosity' rhie_chow_user_object = pins_rhie_chow_interpolator [] [Mesh] [gen] type = GeneratedMeshGenerator dim = 2 xmin = 0 xmax = ${bed_radius} ymin = 0 ymax = ${bed_height} nx = 6 ny = 40 [] coord_type = RZ [] [Modules] [FluidProperties] [fluid_properties_obj] type = HeliumFluidProperties [] [] [NavierStokesFV] compressibility = 'weakly-compressible' porous_medium_treatment = true density = 'rho' dynamic_viscosity = 'mu' porosity = 'porosity' initial_velocity = '1e-6 0 0' initial_pressure = 5.4e6 inlet_boundaries = top momentum_inlet_types = flux-mass flux_inlet_pps = mfr_pp #momentum_inlet_types = 'fixed-velocity' #momentum_inlet_function = '0 -${flow_vel}' wall_boundaries = 'left right' momentum_wall_types = 'slip slip' outlet_boundaries = bottom momentum_outlet_types = fixed-pressure pressure_function = ${outlet_pressure} pressure_face_interpolation = average momentum_advection_interpolation = upwind mass_advection_interpolation = upwind [] [] [Materials] ## Fluid properties and non-dimensional numbers [fluid_props_to_mat_props] type = GeneralFunctorFluidProps pressure = 'pressure' T_fluid = ${T_fluid} speed = 'speed' characteristic_length = characteristic_length [] [hydraulic_diameter] type = PiecewiseByBlockFunctorMaterial prop_name = 'characteristic_length' subdomain_to_prop_value = '0 1' [] [] [AuxVariables] [porosity] family = MONOMIAL order = CONSTANT fv = true initial_condition = ${bed_porosity} [] [density_var] family = MONOMIAL order = CONSTANT fv = true [] [] [AuxKernels] [density_aux] type = ADFunctorElementalAux variable = density_var functor = rho [] [] [Executioner] type = Transient end_time = 100 [TimeStepper] type = IterationAdaptiveDT iteration_window = 2 optimal_iterations = 8 cutback_factor = 0.8 growth_factor = 2 dt = 1e-5 [] dtmax = 1 line_search = l2 solve_type = 'NEWTON' petsc_options_iname = '-pc_type -pc_factor_shift_type' petsc_options_value = 'lu NONZERO' nl_rel_tol = 1e-6 nl_abs_tol = 1e-6 [] [Postprocessors] [mfr_pp] type = Receiver default = 60 [] [inlet_mfr] type = VolumetricFlowRate advected_quantity = rho vel_x = 'superficial_vel_x' vel_y = 'superficial_vel_y' boundary = 'top' [] [outlet_mfr] type = VolumetricFlowRate advected_quantity = rho vel_x = 'superficial_vel_x' vel_y = 'superficial_vel_y' boundary = 'bottom' [] [] [Outputs] exodus = true [] ``` The expected result is to get a uniform velocity field (considering the slip BCs). However it gives a spurious oscillation close to the inlet due to the wrong mass flow rate. ## Impact Users who assume that mass flow is always positive and don't factor in the surface normal when computing the inlet value might experience nonphysical results.
defect
flux bc for wcnsfv momentum equation gives wrong results if fluid flows in the negative x y z directions bug description the flux postprocessor assumes positive velocities and mass flow rates for inlets the wcnsfvmomentumfluxinlet has the absolute value of the normal which yields weird results when the inlet velocity direction is negative in the chartesian coordinate system steps to reproduce take a channel use an inlet in the negative x y z direction and a positive number for massflow it will give nonphysical results try running the input file below bed height bed radius bed porosity outlet pressure t fluid density mass flow rate flow area fparse pi bed radius bed radius bed porosity flow vel fparse mass flow rate flow area density fp fluid properties obj porosity porosity rhie chow user object pins rhie chow interpolator type generatedmeshgenerator dim xmin xmax bed radius ymin ymax bed height nx ny coord type rz type heliumfluidproperties compressibility weakly compressible porous medium treatment true density rho dynamic viscosity mu porosity porosity initial velocity initial pressure inlet boundaries top momentum inlet types flux mass flux inlet pps mfr pp momentum inlet types fixed velocity momentum inlet function flow vel wall boundaries left right momentum wall types slip slip outlet boundaries bottom momentum outlet types fixed pressure pressure function outlet pressure pressure face interpolation average momentum advection interpolation upwind mass advection interpolation upwind fluid properties and non dimensional numbers type generalfunctorfluidprops pressure pressure t fluid t fluid speed speed characteristic length characteristic length type piecewisebyblockfunctormaterial prop name characteristic length subdomain to prop value family monomial order constant fv true initial condition bed porosity family monomial order constant fv true type adfunctorelementalaux variable density var functor rho type transient end time type iterationadaptivedt iteration window optimal iterations cutback factor growth factor dt dtmax line search solve type newton petsc options iname pc type pc factor shift type petsc options value lu nonzero nl rel tol nl abs tol type receiver default type volumetricflowrate advected quantity rho vel x superficial vel x vel y superficial vel y boundary top type volumetricflowrate advected quantity rho vel x superficial vel x vel y superficial vel y boundary bottom exodus true the expected result is to get a uniform velocity field considering the slip bcs however it gives a spurious oscillation close to the inlet due to the wrong mass flow rate impact users who assume that mass flow is always positive and don t factor in the surface normal when computing the inlet value might experience nonphysical results
1
45,053
12,530,251,804
IssuesEvent
2020-06-04 12:47:03
google/pywebsocket
https://api.github.com/repos/google/pywebsocket
closed
standalone.py does not disable Nagle
Priority-Medium Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. Write a handler which sends small messages every 10ms 2. Connect to it from a Windows 7 machine from Chrome or Firefox 3. Log the timestamps when the messages were received. What is the expected output? Messages are received individually every 10ms. What do you see instead? Messages are received in batches every 200ms. ``` Original issue reported on code.google.com by `ricea@chromium.org` on 22 Jan 2015 at 10:08
1.0
standalone.py does not disable Nagle - ``` What steps will reproduce the problem? 1. Write a handler which sends small messages every 10ms 2. Connect to it from a Windows 7 machine from Chrome or Firefox 3. Log the timestamps when the messages were received. What is the expected output? Messages are received individually every 10ms. What do you see instead? Messages are received in batches every 200ms. ``` Original issue reported on code.google.com by `ricea@chromium.org` on 22 Jan 2015 at 10:08
defect
standalone py does not disable nagle what steps will reproduce the problem write a handler which sends small messages every connect to it from a windows machine from chrome or firefox log the timestamps when the messages were received what is the expected output messages are received individually every what do you see instead messages are received in batches every original issue reported on code google com by ricea chromium org on jan at
1
335,384
30,026,893,538
IssuesEvent
2023-06-27 06:56:13
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix paddle_tensor.test_paddle_square
Sub Task Failing Test Paddle Frontend
| | | |---|---| |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix paddle_tensor.test_paddle_square - | | | |---|---| |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="null"><img src=https://img.shields.io/badge/-failure-red></a>
non_defect
fix paddle tensor test paddle square paddle a href src tensorflow img src numpy a href src torch a href src jax img src
0
13,457
2,757,819,824
IssuesEvent
2015-04-27 16:48:27
PyTables/PyTables
https://api.github.com/repos/PyTables/PyTables
closed
Sourceforge mailing list still in docs
defect
Both the [sphinx docs](http://pytables.github.io/FAQ.html?highlight=mailing) and the [pytables.org docs](http://www.pytables.org/moin/FAQ#I.27mhavingproblems.HowcanIgetsupport.3F) provide the old sourceforge mailing list as the contact point for support. I gather this is no longer the case after #81. It would be good to fix this to avoid users (like me) needlessly going through the process to join the old sourceforge group to post a question.
1.0
Sourceforge mailing list still in docs - Both the [sphinx docs](http://pytables.github.io/FAQ.html?highlight=mailing) and the [pytables.org docs](http://www.pytables.org/moin/FAQ#I.27mhavingproblems.HowcanIgetsupport.3F) provide the old sourceforge mailing list as the contact point for support. I gather this is no longer the case after #81. It would be good to fix this to avoid users (like me) needlessly going through the process to join the old sourceforge group to post a question.
defect
sourceforge mailing list still in docs both the and the provide the old sourceforge mailing list as the contact point for support i gather this is no longer the case after it would be good to fix this to avoid users like me needlessly going through the process to join the old sourceforge group to post a question
1
12,640
7,933,221,551
IssuesEvent
2018-07-08 02:11:35
pvlib/pvlib-python
https://api.github.com/repos/pvlib/pvlib-python
opened
slow performance when creating empty DataFrame in ``Location.get_airmass``
easy good first issue help wanted performance
Creating an empty DataFrame (no data, no index) and then filling it key by key can cause performance issues in some situations. I believe the issue is due to the way that pandas computes joins on the index. This happens in two places in pvlib: ``Location.get_airmass`` and ``ModelChain.prepare_inputs``. (determined with ``grep -r 'pd.DataFrame()' pvlib``). ``Location.get_airmass``: this is most relevant with shorter input lengths, especially if ``solar_position`` is not supplied. I discovered this bottleneck when profiling a loop that called ``ModelChain.run_model`` on daily weather data. ``ModelChain.prepare_inputs``: this only an issue there if the user does not supply any weather data, in which case clear sky calculations will be run and the results assigned to the empty DataFrame. Less likely that anyone is running into a significant performance issue here due to the additional calculations, including a linke turbidity lookup. Here's the key part of ``Location.get_airmass`` using an input of 1440 times, followed by two alternative implementations: ```python %%timeit airmass_relative = pvlib.atmosphere.relativeairmass(solar_position['zenith']) pressure = pvlib.atmosphere.alt2pres(altitude) airmass_absolute = pvlib.atmosphere.absoluteairmass(airmass_relative, pressure) airmass = pd.DataFrame() airmass['airmass_relative'] = airmass_relative airmass['airmass_absolute'] = airmass_absolute 23.9 ms ± 780 µs per loop (mean ± std. dev. of 7 runs, 10 loops each) ``` Alternative 1: ```python %%timeit airmass_relative = pvlib.atmosphere.relativeairmass(solar_position['zenith']) pressure = pvlib.atmosphere.alt2pres(altitude) airmass_absolute = pvlib.atmosphere.absoluteairmass(airmass_relative, pressure) airmass = pd.DataFrame(index=solar_position.index) airmass['airmass_relative'] = airmass_relative airmass['airmass_absolute'] = airmass_absolute 1.69 ms ± 28.4 µs per loop (mean ± std. dev. of 7 runs, 1000 loops each) ``` Alternative 2: ```python %%timeit airmass_relative = pvlib.atmosphere.relativeairmass(solar_position['zenith']) pressure = pvlib.atmosphere.alt2pres(altitude) airmass_absolute = pvlib.atmosphere.absoluteairmass(airmass_relative, pressure) airmass = pd.DataFrame({'airmass_relative': airmass_relative, 'airmass_absolute': airmass_absolute}) airmass = airmass[['airmass_relative', 'airmass_absolute']] # adds 0.4 ms, but guarantees same output 1.43 ms ± 40.1 µs per loop (mean ± std. dev. of 7 runs, 1000 loops each) ``` Either 1 or 2 could work for ``Location.get_airmass``. Only 1 would easily work for ``ModelChain.run_model``. **Versions:** - ``pvlib.__version__``: '0.5.2+16.g58f95e0' - ``pandas.__version__``: '0.23.1' - python: 3.6.5 Approximately reproduced on python 3.5 and pandas 0.17. line profiler: <details> ``` 258 1 2.0 2.0 0.0 if solar_position is None: 259 solar_position = self.get_solarposition(times) 260 261 1 3.0 3.0 0.0 if model in atmosphere.APPARENT_ZENITH_MODELS: 262 1 44.0 44.0 0.1 zenith = solar_position['apparent_zenith'] 263 elif model in atmosphere.TRUE_ZENITH_MODELS: 264 zenith = solar_position['zenith'] 265 else: 266 raise ValueError('{} is not a valid airmass model'.format(model)) 267 268 1 1466.0 1466.0 3.6 airmass_relative = atmosphere.relativeairmass(zenith, model) 269 270 1 9.0 9.0 0.0 pressure = atmosphere.alt2pres(self.altitude) 271 1 1.0 1.0 0.0 airmass_absolute = atmosphere.absoluteairmass(airmass_relative, 272 1 639.0 639.0 1.6 pressure) 273 274 1 638.0 638.0 1.6 airmass = pd.DataFrame() 275 1 36061.0 36061.0 88.2 airmass['airmass_relative'] = airmass_relative 276 1 2019.0 2019.0 4.9 airmass['airmass_absolute'] = airmass_absolute 277 278 1 0.0 0.0 0.0 return airmass ``` </details>
True
slow performance when creating empty DataFrame in ``Location.get_airmass`` - Creating an empty DataFrame (no data, no index) and then filling it key by key can cause performance issues in some situations. I believe the issue is due to the way that pandas computes joins on the index. This happens in two places in pvlib: ``Location.get_airmass`` and ``ModelChain.prepare_inputs``. (determined with ``grep -r 'pd.DataFrame()' pvlib``). ``Location.get_airmass``: this is most relevant with shorter input lengths, especially if ``solar_position`` is not supplied. I discovered this bottleneck when profiling a loop that called ``ModelChain.run_model`` on daily weather data. ``ModelChain.prepare_inputs``: this only an issue there if the user does not supply any weather data, in which case clear sky calculations will be run and the results assigned to the empty DataFrame. Less likely that anyone is running into a significant performance issue here due to the additional calculations, including a linke turbidity lookup. Here's the key part of ``Location.get_airmass`` using an input of 1440 times, followed by two alternative implementations: ```python %%timeit airmass_relative = pvlib.atmosphere.relativeairmass(solar_position['zenith']) pressure = pvlib.atmosphere.alt2pres(altitude) airmass_absolute = pvlib.atmosphere.absoluteairmass(airmass_relative, pressure) airmass = pd.DataFrame() airmass['airmass_relative'] = airmass_relative airmass['airmass_absolute'] = airmass_absolute 23.9 ms ± 780 µs per loop (mean ± std. dev. of 7 runs, 10 loops each) ``` Alternative 1: ```python %%timeit airmass_relative = pvlib.atmosphere.relativeairmass(solar_position['zenith']) pressure = pvlib.atmosphere.alt2pres(altitude) airmass_absolute = pvlib.atmosphere.absoluteairmass(airmass_relative, pressure) airmass = pd.DataFrame(index=solar_position.index) airmass['airmass_relative'] = airmass_relative airmass['airmass_absolute'] = airmass_absolute 1.69 ms ± 28.4 µs per loop (mean ± std. dev. of 7 runs, 1000 loops each) ``` Alternative 2: ```python %%timeit airmass_relative = pvlib.atmosphere.relativeairmass(solar_position['zenith']) pressure = pvlib.atmosphere.alt2pres(altitude) airmass_absolute = pvlib.atmosphere.absoluteairmass(airmass_relative, pressure) airmass = pd.DataFrame({'airmass_relative': airmass_relative, 'airmass_absolute': airmass_absolute}) airmass = airmass[['airmass_relative', 'airmass_absolute']] # adds 0.4 ms, but guarantees same output 1.43 ms ± 40.1 µs per loop (mean ± std. dev. of 7 runs, 1000 loops each) ``` Either 1 or 2 could work for ``Location.get_airmass``. Only 1 would easily work for ``ModelChain.run_model``. **Versions:** - ``pvlib.__version__``: '0.5.2+16.g58f95e0' - ``pandas.__version__``: '0.23.1' - python: 3.6.5 Approximately reproduced on python 3.5 and pandas 0.17. line profiler: <details> ``` 258 1 2.0 2.0 0.0 if solar_position is None: 259 solar_position = self.get_solarposition(times) 260 261 1 3.0 3.0 0.0 if model in atmosphere.APPARENT_ZENITH_MODELS: 262 1 44.0 44.0 0.1 zenith = solar_position['apparent_zenith'] 263 elif model in atmosphere.TRUE_ZENITH_MODELS: 264 zenith = solar_position['zenith'] 265 else: 266 raise ValueError('{} is not a valid airmass model'.format(model)) 267 268 1 1466.0 1466.0 3.6 airmass_relative = atmosphere.relativeairmass(zenith, model) 269 270 1 9.0 9.0 0.0 pressure = atmosphere.alt2pres(self.altitude) 271 1 1.0 1.0 0.0 airmass_absolute = atmosphere.absoluteairmass(airmass_relative, 272 1 639.0 639.0 1.6 pressure) 273 274 1 638.0 638.0 1.6 airmass = pd.DataFrame() 275 1 36061.0 36061.0 88.2 airmass['airmass_relative'] = airmass_relative 276 1 2019.0 2019.0 4.9 airmass['airmass_absolute'] = airmass_absolute 277 278 1 0.0 0.0 0.0 return airmass ``` </details>
non_defect
slow performance when creating empty dataframe in location get airmass creating an empty dataframe no data no index and then filling it key by key can cause performance issues in some situations i believe the issue is due to the way that pandas computes joins on the index this happens in two places in pvlib location get airmass and modelchain prepare inputs determined with grep r pd dataframe pvlib location get airmass this is most relevant with shorter input lengths especially if solar position is not supplied i discovered this bottleneck when profiling a loop that called modelchain run model on daily weather data modelchain prepare inputs this only an issue there if the user does not supply any weather data in which case clear sky calculations will be run and the results assigned to the empty dataframe less likely that anyone is running into a significant performance issue here due to the additional calculations including a linke turbidity lookup here s the key part of location get airmass using an input of times followed by two alternative implementations python timeit airmass relative pvlib atmosphere relativeairmass solar position pressure pvlib atmosphere altitude airmass absolute pvlib atmosphere absoluteairmass airmass relative pressure airmass pd dataframe airmass airmass relative airmass airmass absolute ms ± µs per loop mean ± std dev of runs loops each alternative python timeit airmass relative pvlib atmosphere relativeairmass solar position pressure pvlib atmosphere altitude airmass absolute pvlib atmosphere absoluteairmass airmass relative pressure airmass pd dataframe index solar position index airmass airmass relative airmass airmass absolute ms ± µs per loop mean ± std dev of runs loops each alternative python timeit airmass relative pvlib atmosphere relativeairmass solar position pressure pvlib atmosphere altitude airmass absolute pvlib atmosphere absoluteairmass airmass relative pressure airmass pd dataframe airmass relative airmass relative airmass absolute airmass absolute airmass airmass adds ms but guarantees same output ms ± µs per loop mean ± std dev of runs loops each either or could work for location get airmass only would easily work for modelchain run model versions pvlib version pandas version python approximately reproduced on python and pandas line profiler if solar position is none solar position self get solarposition times if model in atmosphere apparent zenith models zenith solar position elif model in atmosphere true zenith models zenith solar position else raise valueerror is not a valid airmass model format model airmass relative atmosphere relativeairmass zenith model pressure atmosphere self altitude airmass absolute atmosphere absoluteairmass airmass relative pressure airmass pd dataframe airmass airmass relative airmass airmass absolute return airmass
0
23,928
3,873,679,266
IssuesEvent
2016-04-11 17:49:50
ariya/phantomjs
https://api.github.com/repos/ariya/phantomjs
closed
phantomjs 139 segmentation fault.
old.Priority-Medium old.Status-New old.Type-Defect
_**[ignasi.t...@gmail.com](http://code.google.com/u/112178480962125604589/) commented:**_ > phantomjs-1.7.0-linux-x86_64.tar.bz2 PhantomJS 1.7.0 Linux x86-64 > > If I execute the command via command line using ssh it works fine. Just displays this error > Fontconfig error: Cannot load default config file > but creates the image correctly. > > However when the application executes the command: > > $output = array(); > $return_var=0; > $command = &quot;./../phantom/phantomjs ../phantom/rasterize.js http://www.nfl.com ../phantom/2.png&quot;; > exec($command,$output,$return_var); > > The return var is 139 segmentation fault. > Why is this happenig? I contacted with my hosting provider and they say me that its the binary problem. They say: > > > I did notice that the program attempts to open some non-existant files and this may be hard coded in the binary: > > 13:09:41 open(&quot;/lib64/ld-linux-x86-64.so.2&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/home/vagrant/phantomjs/src/qt/lib/tls/x86_64/libfreetype.so.6&quot;, O_RDONLY) = -1 ENOENT (No such file or directory) > 13:09:41 open(&quot;/home/vagrant/phantomjs/src/qt/lib/tls/libfreetype.so.6&quot;, O_RDONLY) = -1 ENOENT (No such file or directory) > 13:09:41 open(&quot;/home/vagrant/phantomjs/src/qt/lib/x86_64/libfreetype.so.6&quot;, O_RDONLY) = -1 ENOENT (No such file or directory) > 13:09:41 open(&quot;/home/vagrant/phantomjs/src/qt/lib/libfreetype.so.6&quot;, O_RDONLY) = -1 ENOENT (No such file or directory) > 13:09:41 open(&quot;/etc/ld.so.cache&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/usr/lib64/libfreetype.so.6&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/usr/lib64/libfontconfig.so.1&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libdl.so.2&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/librt.so.1&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libpthread.so.0&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/usr/lib64/libstdc++.so.6&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libm.so.6&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libgcc_s.so.1&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libc.so.6&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libexpat.so.1&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/usr/lib/locale/locale-archive&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/etc/fonts/fonts.conf&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/etc/fonts/conf.d&quot;, O_RDONLY|O_NONBLOCK|O_DIRECTORY|O_CLOEXEC) = 8 > 13:09:41 open(&quot;/etc/fonts/conf.d/20-fix-globaladvance.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/20-unhint-small-dejavu-lgc-sans-mono.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/20-unhint-small-dejavu-sans-mono.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/20-unhint-small-vera.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/25-no-bitmap-fedora.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/25-unhint-nonlatin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/30-metric-aliases.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/30-urw-aliases.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/40-nonlatin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/45-latin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/49-sansserif.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/50-user.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/51-local.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/57-dejavu-sans-mono.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/58-dejavu-lgc-sans-mono.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/60-latin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/65-fonts-persian.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/65-nonlatin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/69-unifont.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/80-delicious.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/90-synthetic.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/var/cache/fontconfig/3830d5c3ddfd5cd38a049b759396e72e-le64.cache-3&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/var/cache/fontconfig/0251a5afa6ac727a1e32b7d4d4aa7cf0-le64.cache-3&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/var/cache/fontconfig/12b26b760a24f8b4feb03ad48a333a72-le64.cache-3&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/var/cache/fontconfig/b79f3aaa7d385a141ab53ec885cc22a8-le64.cache-3&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/var/cache/fontconfig/87f5e051180a7a75f16eb6fe7dbd3749-le64.cache-3&quot;, O_RDONLY) = 7 > > You will need to contact the developer to help resolve this issue. > > Please help! **Disclaimer:** This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #902](http://code.google.com/p/phantomjs/issues/detail?id=902). :star2: &nbsp; **3** people had starred this issue at the time of migration.
1.0
phantomjs 139 segmentation fault. - _**[ignasi.t...@gmail.com](http://code.google.com/u/112178480962125604589/) commented:**_ > phantomjs-1.7.0-linux-x86_64.tar.bz2 PhantomJS 1.7.0 Linux x86-64 > > If I execute the command via command line using ssh it works fine. Just displays this error > Fontconfig error: Cannot load default config file > but creates the image correctly. > > However when the application executes the command: > > $output = array(); > $return_var=0; > $command = &quot;./../phantom/phantomjs ../phantom/rasterize.js http://www.nfl.com ../phantom/2.png&quot;; > exec($command,$output,$return_var); > > The return var is 139 segmentation fault. > Why is this happenig? I contacted with my hosting provider and they say me that its the binary problem. They say: > > > I did notice that the program attempts to open some non-existant files and this may be hard coded in the binary: > > 13:09:41 open(&quot;/lib64/ld-linux-x86-64.so.2&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/home/vagrant/phantomjs/src/qt/lib/tls/x86_64/libfreetype.so.6&quot;, O_RDONLY) = -1 ENOENT (No such file or directory) > 13:09:41 open(&quot;/home/vagrant/phantomjs/src/qt/lib/tls/libfreetype.so.6&quot;, O_RDONLY) = -1 ENOENT (No such file or directory) > 13:09:41 open(&quot;/home/vagrant/phantomjs/src/qt/lib/x86_64/libfreetype.so.6&quot;, O_RDONLY) = -1 ENOENT (No such file or directory) > 13:09:41 open(&quot;/home/vagrant/phantomjs/src/qt/lib/libfreetype.so.6&quot;, O_RDONLY) = -1 ENOENT (No such file or directory) > 13:09:41 open(&quot;/etc/ld.so.cache&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/usr/lib64/libfreetype.so.6&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/usr/lib64/libfontconfig.so.1&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libdl.so.2&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/librt.so.1&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libpthread.so.0&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/usr/lib64/libstdc++.so.6&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libm.so.6&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libgcc_s.so.1&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libc.so.6&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/lib64/libexpat.so.1&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/usr/lib/locale/locale-archive&quot;, O_RDONLY) = 3 > 13:09:41 open(&quot;/etc/fonts/fonts.conf&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/etc/fonts/conf.d&quot;, O_RDONLY|O_NONBLOCK|O_DIRECTORY|O_CLOEXEC) = 8 > 13:09:41 open(&quot;/etc/fonts/conf.d/20-fix-globaladvance.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/20-unhint-small-dejavu-lgc-sans-mono.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/20-unhint-small-dejavu-sans-mono.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/20-unhint-small-vera.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/25-no-bitmap-fedora.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/25-unhint-nonlatin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/30-metric-aliases.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/30-urw-aliases.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/40-nonlatin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/45-latin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/49-sansserif.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/50-user.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/51-local.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/57-dejavu-sans-mono.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/58-dejavu-lgc-sans-mono.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/60-latin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/65-fonts-persian.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/65-nonlatin.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/69-unifont.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/80-delicious.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/etc/fonts/conf.d/90-synthetic.conf&quot;, O_RDONLY) = 9 > 13:09:41 open(&quot;/var/cache/fontconfig/3830d5c3ddfd5cd38a049b759396e72e-le64.cache-3&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/var/cache/fontconfig/0251a5afa6ac727a1e32b7d4d4aa7cf0-le64.cache-3&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/var/cache/fontconfig/12b26b760a24f8b4feb03ad48a333a72-le64.cache-3&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/var/cache/fontconfig/b79f3aaa7d385a141ab53ec885cc22a8-le64.cache-3&quot;, O_RDONLY) = 7 > 13:09:41 open(&quot;/var/cache/fontconfig/87f5e051180a7a75f16eb6fe7dbd3749-le64.cache-3&quot;, O_RDONLY) = 7 > > You will need to contact the developer to help resolve this issue. > > Please help! **Disclaimer:** This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #902](http://code.google.com/p/phantomjs/issues/detail?id=902). :star2: &nbsp; **3** people had starred this issue at the time of migration.
defect
phantomjs segmentation fault commented phantomjs linux tar phantomjs linux if i execute the command via command line using ssh it works fine just displays this error fontconfig error cannot load default config file but creates the image correctly however when the application executes the command output array return var command quot phantom phantomjs phantom rasterize js phantom png quot exec command output return var the return var is segmentation fault why is this happenig i contacted with my hosting provider and they say me that its the binary problem they say i did notice that the program attempts to open some non existant files and this may be hard coded in the binary open quot ld linux so quot o rdonly open quot home vagrant phantomjs src qt lib tls libfreetype so quot o rdonly enoent no such file or directory open quot home vagrant phantomjs src qt lib tls libfreetype so quot o rdonly enoent no such file or directory open quot home vagrant phantomjs src qt lib libfreetype so quot o rdonly enoent no such file or directory open quot home vagrant phantomjs src qt lib libfreetype so quot o rdonly enoent no such file or directory open quot etc ld so cache quot o rdonly open quot usr libfreetype so quot o rdonly open quot usr libfontconfig so quot o rdonly open quot libdl so quot o rdonly open quot librt so quot o rdonly open quot libpthread so quot o rdonly open quot usr libstdc so quot o rdonly open quot libm so quot o rdonly open quot libgcc s so quot o rdonly open quot libc so quot o rdonly open quot libexpat so quot o rdonly open quot usr lib locale locale archive quot o rdonly open quot etc fonts fonts conf quot o rdonly open quot etc fonts conf d quot o rdonly o nonblock o directory o cloexec open quot etc fonts conf d fix globaladvance conf quot o rdonly open quot etc fonts conf d unhint small dejavu lgc sans mono conf quot o rdonly open quot etc fonts conf d unhint small dejavu sans mono conf quot o rdonly open quot etc fonts conf d unhint small vera conf quot o rdonly open quot etc fonts conf d no bitmap fedora conf quot o rdonly open quot etc fonts conf d unhint nonlatin conf quot o rdonly open quot etc fonts conf d metric aliases conf quot o rdonly open quot etc fonts conf d urw aliases conf quot o rdonly open quot etc fonts conf d nonlatin conf quot o rdonly open quot etc fonts conf d latin conf quot o rdonly open quot etc fonts conf d sansserif conf quot o rdonly open quot etc fonts conf d user conf quot o rdonly open quot etc fonts conf d local conf quot o rdonly open quot etc fonts conf d dejavu sans mono conf quot o rdonly open quot etc fonts conf d dejavu lgc sans mono conf quot o rdonly open quot etc fonts conf d latin conf quot o rdonly open quot etc fonts conf d fonts persian conf quot o rdonly open quot etc fonts conf d nonlatin conf quot o rdonly open quot etc fonts conf d unifont conf quot o rdonly open quot etc fonts conf d delicious conf quot o rdonly open quot etc fonts conf d synthetic conf quot o rdonly open quot var cache fontconfig cache quot o rdonly open quot var cache fontconfig cache quot o rdonly open quot var cache fontconfig cache quot o rdonly open quot var cache fontconfig cache quot o rdonly open quot var cache fontconfig cache quot o rdonly you will need to contact the developer to help resolve this issue please help disclaimer this issue was migrated on from the project s former issue tracker on google code nbsp people had starred this issue at the time of migration
1
75,170
25,567,866,941
IssuesEvent
2022-11-30 15:30:47
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
closed
Labs: Rich Text Editor: Fullscreen mode doesn't resize on iOS 15
T-Defect S-Minor O-Frequent A-Rich-Text-Editor
### Steps to reproduce 1. Go to a Room with RTE enabled 2. Tap the fullscreen mode icon ### Outcome #### What did you expect? Resize of the composer #### What happened instead? https://user-images.githubusercontent.com/80891108/204526610-d7bfbaef-a4ac-45de-8aab-45d12ead1d2a.mp4 ### Your phone model iPhone 13 Pro Simulator ### Operating system version iOS 15.5 ### Application version develop ### Homeserver matrix.org ### Will you send logs? No
1.0
Labs: Rich Text Editor: Fullscreen mode doesn't resize on iOS 15 - ### Steps to reproduce 1. Go to a Room with RTE enabled 2. Tap the fullscreen mode icon ### Outcome #### What did you expect? Resize of the composer #### What happened instead? https://user-images.githubusercontent.com/80891108/204526610-d7bfbaef-a4ac-45de-8aab-45d12ead1d2a.mp4 ### Your phone model iPhone 13 Pro Simulator ### Operating system version iOS 15.5 ### Application version develop ### Homeserver matrix.org ### Will you send logs? No
defect
labs rich text editor fullscreen mode doesn t resize on ios steps to reproduce go to a room with rte enabled tap the fullscreen mode icon outcome what did you expect resize of the composer what happened instead your phone model iphone pro simulator operating system version ios application version develop homeserver matrix org will you send logs no
1
148,825
5,697,051,250
IssuesEvent
2017-04-16 17:58:59
MongoEngine/mongoengine
https://api.github.com/repos/MongoEngine/mongoengine
closed
EmailField doesn't support Unicode
Enhancement High Priority in progress
``` >>> import mongoengine >>> print(mongoengine.EmailField.EMAIL_REGEX.match("武@メール.グーグル")) None ``` This type of email address is [supported by Gmail](https://googleblog.blogspot.com.au/2014/08/a-first-step-toward-more-global-email.html).
1.0
EmailField doesn't support Unicode - ``` >>> import mongoengine >>> print(mongoengine.EmailField.EMAIL_REGEX.match("武@メール.グーグル")) None ``` This type of email address is [supported by Gmail](https://googleblog.blogspot.com.au/2014/08/a-first-step-toward-more-global-email.html).
non_defect
emailfield doesn t support unicode import mongoengine print mongoengine emailfield email regex match 武@メール グーグル none this type of email address is
0
4,235
2,610,089,791
IssuesEvent
2015-02-26 18:27:11
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳痘痘如何治疗最好
auto-migrated Priority-Medium Type-Defect
``` 深圳痘痘如何治疗最好【深圳韩方科颜全国热线400-869-1818,24 小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:37
1.0
深圳痘痘如何治疗最好 - ``` 深圳痘痘如何治疗最好【深圳韩方科颜全国热线400-869-1818,24 小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:37
defect
深圳痘痘如何治疗最好 深圳痘痘如何治疗最好【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 original issue reported on code google com by szft com on may at
1
653,574
21,607,683,352
IssuesEvent
2022-05-04 06:32:16
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Yarn lock file maintenance using wrong version
type:bug priority-4-low manager:npm status:in-progress
### How are you running Renovate? WhiteSource Renovate hosted app on github.com ### If you're self-hosting Renovate, tell us what version of Renovate you run. _No response_ ### Please select which platform you are using if self-hosting. _No response_ ### If you're self-hosting Renovate, tell us what version of the platform you run. _No response_ ### Was this something which used to work for you, and then stopped? I never saw this working ### Describe the bug When Yarn 2+ is used in a project, but not embedded in the repository via `yarnPath`, lock file maintenance will fall back to Yarn 1 and replace the whole `yarn.lock`. An example can be seen here: https://github.com/jgraichen/renovate-corepack-yarn/pull/4/files When dependencies are bumped, the `yarn.lock` is correctly updated: https://github.com/jgraichen/renovate-corepack-yarn/pull/3/files I'd love to have all updates and lock file maintenance use the correct Yarn version. ### Relevant debug logs _No response_ ### Have you created a minimal reproduction repository? I have linked to a minimal reproduction repository in the bug description
1.0
Yarn lock file maintenance using wrong version - ### How are you running Renovate? WhiteSource Renovate hosted app on github.com ### If you're self-hosting Renovate, tell us what version of Renovate you run. _No response_ ### Please select which platform you are using if self-hosting. _No response_ ### If you're self-hosting Renovate, tell us what version of the platform you run. _No response_ ### Was this something which used to work for you, and then stopped? I never saw this working ### Describe the bug When Yarn 2+ is used in a project, but not embedded in the repository via `yarnPath`, lock file maintenance will fall back to Yarn 1 and replace the whole `yarn.lock`. An example can be seen here: https://github.com/jgraichen/renovate-corepack-yarn/pull/4/files When dependencies are bumped, the `yarn.lock` is correctly updated: https://github.com/jgraichen/renovate-corepack-yarn/pull/3/files I'd love to have all updates and lock file maintenance use the correct Yarn version. ### Relevant debug logs _No response_ ### Have you created a minimal reproduction repository? I have linked to a minimal reproduction repository in the bug description
non_defect
yarn lock file maintenance using wrong version how are you running renovate whitesource renovate hosted app on github com if you re self hosting renovate tell us what version of renovate you run no response please select which platform you are using if self hosting no response if you re self hosting renovate tell us what version of the platform you run no response was this something which used to work for you and then stopped i never saw this working describe the bug when yarn is used in a project but not embedded in the repository via yarnpath lock file maintenance will fall back to yarn and replace the whole yarn lock an example can be seen here when dependencies are bumped the yarn lock is correctly updated i d love to have all updates and lock file maintenance use the correct yarn version relevant debug logs no response have you created a minimal reproduction repository i have linked to a minimal reproduction repository in the bug description
0
16,716
2,937,772,317
IssuesEvent
2015-07-01 05:48:54
canadainc/quran10
https://api.github.com/repos/canadainc/quran10
closed
Tafsir Ibn Kathir is not being ordered properly
Component-Logic Fixed Priority-High Type-Defect Usability
Being sorted by name instead of by proper order.
1.0
Tafsir Ibn Kathir is not being ordered properly - Being sorted by name instead of by proper order.
defect
tafsir ibn kathir is not being ordered properly being sorted by name instead of by proper order
1
1,094
2,594,790,857
IssuesEvent
2015-02-20 08:42:11
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
[TEST-FAILURE] LockTest.testKeyOwnerDies
Team: Core Type: Defect
``` java.lang.AssertionError: null at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at com.hazelcast.concurrent.lock.LockTest.testKeyOwnerDies(LockTest.java:747) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-IbmJDK1.7/com.hazelcast$hazelcast/445/testReport/junit/com.hazelcast.concurrent.lock/LockTest/testKeyOwnerDies/
1.0
[TEST-FAILURE] LockTest.testKeyOwnerDies - ``` java.lang.AssertionError: null at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertTrue(Assert.java:52) at com.hazelcast.concurrent.lock.LockTest.testKeyOwnerDies(LockTest.java:747) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-IbmJDK1.7/com.hazelcast$hazelcast/445/testReport/junit/com.hazelcast.concurrent.lock/LockTest/testKeyOwnerDies/
defect
locktest testkeyownerdies java lang assertionerror null at org junit assert fail assert java at org junit assert asserttrue assert java at org junit assert asserttrue assert java at com hazelcast concurrent lock locktest testkeyownerdies locktest java
1
412,134
12,035,695,748
IssuesEvent
2020-04-13 18:22:08
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
Remove deprecated workflows
area/cli area/cli-review kind/task priority/important-soon
### Summary We introduced workflows as an experiment in https://github.com/jenkins-x/jx/pull/1476 to help us avoid lots of huge jenkins build pods being around waiting for the promotion to happen. Since then we've deprecated the workflows pipeline, so we should remove it from JX. There's also this PR https://github.com/jenkins-x/jenkins-x-platform/pull/6140, that removes workflows from one of our downstream dependencies `jenkins-x-platform`. That PR needs to be merged when we have a PR that removes workflows from the main `jx` repo.
1.0
Remove deprecated workflows - ### Summary We introduced workflows as an experiment in https://github.com/jenkins-x/jx/pull/1476 to help us avoid lots of huge jenkins build pods being around waiting for the promotion to happen. Since then we've deprecated the workflows pipeline, so we should remove it from JX. There's also this PR https://github.com/jenkins-x/jenkins-x-platform/pull/6140, that removes workflows from one of our downstream dependencies `jenkins-x-platform`. That PR needs to be merged when we have a PR that removes workflows from the main `jx` repo.
non_defect
remove deprecated workflows summary we introduced workflows as an experiment in to help us avoid lots of huge jenkins build pods being around waiting for the promotion to happen since then we ve deprecated the workflows pipeline so we should remove it from jx there s also this pr that removes workflows from one of our downstream dependencies jenkins x platform that pr needs to be merged when we have a pr that removes workflows from the main jx repo
0
748,420
26,122,360,507
IssuesEvent
2022-12-28 14:08:28
wazuh/wazuh-documentation
https://api.github.com/repos/wazuh/wazuh-documentation
closed
Adjust Wazuh 4.3.0 release notes
release notes priority: high
## Description This issue aims to adjust Wazuh 4.3.0 release notes. ## Tasks - [x] Branch: `5845-adjust-4.3.0-release-notes` - [x] PR: #5846 - [x] Adjust Wazuh 4.3.0 release notes - [x] Make a complete review - [x] Ask for a review
1.0
Adjust Wazuh 4.3.0 release notes - ## Description This issue aims to adjust Wazuh 4.3.0 release notes. ## Tasks - [x] Branch: `5845-adjust-4.3.0-release-notes` - [x] PR: #5846 - [x] Adjust Wazuh 4.3.0 release notes - [x] Make a complete review - [x] Ask for a review
non_defect
adjust wazuh release notes description this issue aims to adjust wazuh release notes tasks branch adjust release notes pr adjust wazuh release notes make a complete review ask for a review
0
36,042
7,852,593,477
IssuesEvent
2018-06-20 14:59:22
STEllAR-GROUP/hpx
https://api.github.com/repos/STEllAR-GROUP/hpx
closed
possible compilation error when "using namespace std;" is defined before including "hpx" headers files
type: defect
a compilation error may occur when "using namespace std;" is defined before including hpx header files. I can only reproduce this error with my code though and have no very simple test code where the error appears. ## Expected Behavior I expected the compilation not to fail. ## Actual Behavior g++ returns this ``` hpx/runtime/components/memory_block.hpp: In member function ‘hpx::components::access_memory_block_proxy<T>& hpx::components::access_memory_block_proxy<T>::operator=(const target_type&)’: include/hpx/runtime/components/memory_block.hpp:185:20: error: invalid use of ‘class std::set<typename std::remove_const<_Tp>::type>’ block_.template set<target_type>(rhs); ``` if we have this in an header file: ``` using namespace std; #include <hpx/config.hpp> #include <hpx/hpx.hpp> ``` removing the using namespace std; works fine. John suggested that it could be a clash between the std::set function and the set function defined in the class. the compiler try to resolve std::set before searching the set function in the class. ## Steps to Reproduce the Problem Difficult to reproduce with a very simple code. Seems to be related to my code. ## Specifications - HPX Version: 1.2 (git from 04/13) - Platform (compiler, OS): power 8 linux 3.10.0 - gcc : 6.4.0 - boost : 1.66.0 Mathieu
1.0
possible compilation error when "using namespace std;" is defined before including "hpx" headers files - a compilation error may occur when "using namespace std;" is defined before including hpx header files. I can only reproduce this error with my code though and have no very simple test code where the error appears. ## Expected Behavior I expected the compilation not to fail. ## Actual Behavior g++ returns this ``` hpx/runtime/components/memory_block.hpp: In member function ‘hpx::components::access_memory_block_proxy<T>& hpx::components::access_memory_block_proxy<T>::operator=(const target_type&)’: include/hpx/runtime/components/memory_block.hpp:185:20: error: invalid use of ‘class std::set<typename std::remove_const<_Tp>::type>’ block_.template set<target_type>(rhs); ``` if we have this in an header file: ``` using namespace std; #include <hpx/config.hpp> #include <hpx/hpx.hpp> ``` removing the using namespace std; works fine. John suggested that it could be a clash between the std::set function and the set function defined in the class. the compiler try to resolve std::set before searching the set function in the class. ## Steps to Reproduce the Problem Difficult to reproduce with a very simple code. Seems to be related to my code. ## Specifications - HPX Version: 1.2 (git from 04/13) - Platform (compiler, OS): power 8 linux 3.10.0 - gcc : 6.4.0 - boost : 1.66.0 Mathieu
defect
possible compilation error when using namespace std is defined before including hpx headers files a compilation error may occur when using namespace std is defined before including hpx header files i can only reproduce this error with my code though and have no very simple test code where the error appears expected behavior i expected the compilation not to fail actual behavior g returns this hpx runtime components memory block hpp in member function ‘hpx components access memory block proxy hpx components access memory block proxy operator const target type ’ include hpx runtime components memory block hpp error invalid use of ‘class std set type ’ block template set rhs if we have this in an header file using namespace std include include removing the using namespace std works fine john suggested that it could be a clash between the std set function and the set function defined in the class the compiler try to resolve std set before searching the set function in the class steps to reproduce the problem difficult to reproduce with a very simple code seems to be related to my code specifications hpx version git from platform compiler os power linux gcc boost mathieu
1
32,072
4,319,994,907
IssuesEvent
2016-07-25 00:58:00
angular/angular.io
https://api.github.com/repos/angular/angular.io
opened
create an automated TOC directive for the dev guides
design love
Our TOCs are manually populated. This should be automated.
1.0
create an automated TOC directive for the dev guides - Our TOCs are manually populated. This should be automated.
non_defect
create an automated toc directive for the dev guides our tocs are manually populated this should be automated
0
43,170
23,137,722,622
IssuesEvent
2022-07-28 15:32:23
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
Performance: Streamline `app.bundle.js` size
team/frontend-platform UI performance 4.0
### Description **Goal**: Reduce code required to be downloaded before every page can be shown. **Reason:** - This bundle is loaded on every page, and required before we can render _anything_ in the Sourcegraph application. - It may be high value, low effort. From an initial investigation, there is lots of areas where we’ve accidentally included things, or broken lazy loading. Related PR: https://github.com/sourcegraph/sourcegraph/pull/38817 ### A/C - Lower app.bundle.js size - Reduce our `bundlesize` config to enforce this and ensure we don't regress in future.
True
Performance: Streamline `app.bundle.js` size - ### Description **Goal**: Reduce code required to be downloaded before every page can be shown. **Reason:** - This bundle is loaded on every page, and required before we can render _anything_ in the Sourcegraph application. - It may be high value, low effort. From an initial investigation, there is lots of areas where we’ve accidentally included things, or broken lazy loading. Related PR: https://github.com/sourcegraph/sourcegraph/pull/38817 ### A/C - Lower app.bundle.js size - Reduce our `bundlesize` config to enforce this and ensure we don't regress in future.
non_defect
performance streamline app bundle js size description goal reduce code required to be downloaded before every page can be shown reason this bundle is loaded on every page and required before we can render anything in the sourcegraph application it may be high value low effort from an initial investigation there is lots of areas where we’ve accidentally included things or broken lazy loading related pr a c lower app bundle js size reduce our bundlesize config to enforce this and ensure we don t regress in future
0
72,745
8,774,376,986
IssuesEvent
2018-12-18 19:38:17
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Clarify documentation for canTransitionFrom and To
d: api docs f: material design framework waiting for PR to land (fixed)
Even with the following code, transitions between routes still execute: ```dart class CustomPageRoute<T> extends MaterialPageRoute<T> { CustomPageRoute({ @required WidgetBuilder builder, RouteSettings settings, bool fullscreenDialog = false, }) : super( builder: builder, settings: settings, fullscreenDialog: fullscreenDialog, ); @override bool canTransitionTo(TransitionRoute<dynamic> nextRoute) => false; @override bool canTransitionFrom(TransitionRoute<dynamic> previousRoute) => false; } ``` I also tried creating back-to-back full screen dialogs. According to source code comment here https://docs.flutter.io/flutter/material/MaterialPageRoute/canTransitionTo.html, that should cause exit animations not to execute but they still do.
1.0
Clarify documentation for canTransitionFrom and To - Even with the following code, transitions between routes still execute: ```dart class CustomPageRoute<T> extends MaterialPageRoute<T> { CustomPageRoute({ @required WidgetBuilder builder, RouteSettings settings, bool fullscreenDialog = false, }) : super( builder: builder, settings: settings, fullscreenDialog: fullscreenDialog, ); @override bool canTransitionTo(TransitionRoute<dynamic> nextRoute) => false; @override bool canTransitionFrom(TransitionRoute<dynamic> previousRoute) => false; } ``` I also tried creating back-to-back full screen dialogs. According to source code comment here https://docs.flutter.io/flutter/material/MaterialPageRoute/canTransitionTo.html, that should cause exit animations not to execute but they still do.
non_defect
clarify documentation for cantransitionfrom and to even with the following code transitions between routes still execute dart class custompageroute extends materialpageroute custompageroute required widgetbuilder builder routesettings settings bool fullscreendialog false super builder builder settings settings fullscreendialog fullscreendialog override bool cantransitionto transitionroute nextroute false override bool cantransitionfrom transitionroute previousroute false i also tried creating back to back full screen dialogs according to source code comment here that should cause exit animations not to execute but they still do
0
399,357
11,747,573,127
IssuesEvent
2020-03-12 13:51:50
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
closed
Need some sugar to test if two polynomial symbolic::Expressions have coefficients that are approximately equal.
priority: low team: dynamics type: feature request
Right now I'm using: ``` EXPECT_TRUE(Polynomial(V - V_expected) .RemoveTermsWithSmallCoefficients(1e-6) .EqualToAfterExpansion(Polynomial(Expression::Zero()))); ``` It's not horrible, but took me a while to figure out a version I liked -- partly because I kept forgetting that `Expression::ToPolynomial()` returns a `Polynomiald` and NOT a `symbolic::Polynomial`.
1.0
Need some sugar to test if two polynomial symbolic::Expressions have coefficients that are approximately equal. - Right now I'm using: ``` EXPECT_TRUE(Polynomial(V - V_expected) .RemoveTermsWithSmallCoefficients(1e-6) .EqualToAfterExpansion(Polynomial(Expression::Zero()))); ``` It's not horrible, but took me a while to figure out a version I liked -- partly because I kept forgetting that `Expression::ToPolynomial()` returns a `Polynomiald` and NOT a `symbolic::Polynomial`.
non_defect
need some sugar to test if two polynomial symbolic expressions have coefficients that are approximately equal right now i m using expect true polynomial v v expected removetermswithsmallcoefficients equaltoafterexpansion polynomial expression zero it s not horrible but took me a while to figure out a version i liked partly because i kept forgetting that expression topolynomial returns a polynomiald and not a symbolic polynomial
0
175,706
21,327,035,309
IssuesEvent
2022-04-18 01:08:54
violasarah2000/strings-attached
https://api.github.com/repos/violasarah2000/strings-attached
closed
CVE-2019-16769 (Medium) detected in serialize-javascript-1.9.1.tgz - autoclosed
security vulnerability
## CVE-2019-16769 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>serialize-javascript-1.9.1.tgz</b></p></summary> <p>Serialize JavaScript to a superset of JSON that includes regular expressions and functions.</p> <p>Library home page: <a href="https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.9.1.tgz">https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.9.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/strings-attached/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/strings-attached/node_modules/serialize-javascript/package.json</p> <p> Dependency Hierarchy: - build-angular-0.803.20.tgz (Root Library) - copy-webpack-plugin-5.0.4.tgz - :x: **serialize-javascript-1.9.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/violasarah2000/strings-attached/commit/c0c6cd599872d8c0b02956d632c144c20b3381f0">c0c6cd599872d8c0b02956d632c144c20b3381f0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Affected versions of this package are vulnerable to Cross-site Scripting (XSS). It does not properly mitigate against unsafe characters in serialized regular expressions. This vulnerability is not affected on Node.js environment since Node.js's implementation of RegExp.prototype.toString() backslash-escapes all forward slashes in regular expressions. If serialized data of regular expression objects are used in an environment other than Node.js, it is affected by this vulnerability. <p>Publish Date: 2019-12-05 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769>CVE-2019-16769</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769</a></p> <p>Release Date: 2019-12-05</p> <p>Fix Resolution: v2.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-16769 (Medium) detected in serialize-javascript-1.9.1.tgz - autoclosed - ## CVE-2019-16769 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>serialize-javascript-1.9.1.tgz</b></p></summary> <p>Serialize JavaScript to a superset of JSON that includes regular expressions and functions.</p> <p>Library home page: <a href="https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.9.1.tgz">https://registry.npmjs.org/serialize-javascript/-/serialize-javascript-1.9.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/strings-attached/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/strings-attached/node_modules/serialize-javascript/package.json</p> <p> Dependency Hierarchy: - build-angular-0.803.20.tgz (Root Library) - copy-webpack-plugin-5.0.4.tgz - :x: **serialize-javascript-1.9.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/violasarah2000/strings-attached/commit/c0c6cd599872d8c0b02956d632c144c20b3381f0">c0c6cd599872d8c0b02956d632c144c20b3381f0</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Affected versions of this package are vulnerable to Cross-site Scripting (XSS). It does not properly mitigate against unsafe characters in serialized regular expressions. This vulnerability is not affected on Node.js environment since Node.js's implementation of RegExp.prototype.toString() backslash-escapes all forward slashes in regular expressions. If serialized data of regular expression objects are used in an environment other than Node.js, it is affected by this vulnerability. <p>Publish Date: 2019-12-05 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769>CVE-2019-16769</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16769</a></p> <p>Release Date: 2019-12-05</p> <p>Fix Resolution: v2.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in serialize javascript tgz autoclosed cve medium severity vulnerability vulnerable library serialize javascript tgz serialize javascript to a superset of json that includes regular expressions and functions library home page a href path to dependency file tmp ws scm strings attached package json path to vulnerable library tmp ws scm strings attached node modules serialize javascript package json dependency hierarchy build angular tgz root library copy webpack plugin tgz x serialize javascript tgz vulnerable library found in head commit a href vulnerability details affected versions of this package are vulnerable to cross site scripting xss it does not properly mitigate against unsafe characters in serialized regular expressions this vulnerability is not affected on node js environment since node js s implementation of regexp prototype tostring backslash escapes all forward slashes in regular expressions if serialized data of regular expression objects are used in an environment other than node js it is affected by this vulnerability publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
18,468
5,635,844,582
IssuesEvent
2017-04-06 02:35:51
seanmcne/XrmCoreLibrary
https://api.github.com/repos/seanmcne/XrmCoreLibrary
closed
RetrieveMultiple extension methods do not handle top count scenario properly
ImportedFromCodeplex
<b>maustinjones[CodePlex]</b> <br />the RetrieveMultiple extensions are assuming that all queries want to be paged and defaults to first page if not specified. For queries that specify a TopCount (or 'top' in fetch), paging should be ignored as a query with both specified encounters an exception at CRM platform level.
1.0
RetrieveMultiple extension methods do not handle top count scenario properly - <b>maustinjones[CodePlex]</b> <br />the RetrieveMultiple extensions are assuming that all queries want to be paged and defaults to first page if not specified. For queries that specify a TopCount (or 'top' in fetch), paging should be ignored as a query with both specified encounters an exception at CRM platform level.
non_defect
retrievemultiple extension methods do not handle top count scenario properly maustinjones the retrievemultiple extensions are assuming that all queries want to be paged and defaults to first page if not specified for queries that specify a topcount or top in fetch paging should be ignored as a query with both specified encounters an exception at crm platform level
0
46,563
13,055,934,850
IssuesEvent
2020-07-30 03:09:58
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[smallshower-filter] no sphinx docs (Trac #1448)
Incomplete Migration Migrated from Trac combo reconstruction defect
Migrated from https://code.icecube.wisc.edu/ticket/1448 ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "Good documentation is now deemed essential.", "reporter": "david.schultz", "cc": "olivas", "resolution": "fixed", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "[smallshower-filter] no sphinx docs", "priority": "major", "keywords": "", "time": "2015-11-25T16:18:42", "milestone": "", "owner": "karg", "type": "defect" } ```
1.0
[smallshower-filter] no sphinx docs (Trac #1448) - Migrated from https://code.icecube.wisc.edu/ticket/1448 ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "Good documentation is now deemed essential.", "reporter": "david.schultz", "cc": "olivas", "resolution": "fixed", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "[smallshower-filter] no sphinx docs", "priority": "major", "keywords": "", "time": "2015-11-25T16:18:42", "milestone": "", "owner": "karg", "type": "defect" } ```
defect
no sphinx docs trac migrated from json status closed changetime description good documentation is now deemed essential reporter david schultz cc olivas resolution fixed ts component combo reconstruction summary no sphinx docs priority major keywords time milestone owner karg type defect
1
45,581
12,887,866,955
IssuesEvent
2020-07-13 12:01:52
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
InputNumber: unable to clear value from field when using min/max range both > 0
defect
**Describe the defect** There appear to be two related issues: (i) it is not possible to delete a value again from an inputNumber field where the min/val values are set to be wholly positive and does not include zero - e.g. minValue 1, maxValue 100 - The field is an option, and is not a required value. The problem is the same whether the emptyValue attribute is absent, or when it is set to "null". (PS. the meaning of attribute values of "focus" and "press" per the documentation is not clear to me) (ii) When the range is wholly negative (e.g. minValue -100, maxValue -1), it is not possible to enter any value into the field. **Reproducer** See attached. [tester_inputNumber_delete_src.zip](https://github.com/primefaces/primefaces/files/4912144/tester_inputNumber_delete_src.zip) **Environment:** - PF Version: _8.0_ - JSF + version: _Mojarra 2.2.8_ - Affected browsers: _(at least) Firefox_ **To Reproduce** Run the reproducer: 1. Enter a valid value into the first three fields, per the range specified. You will be unable to enter a value for the negative range in the third case [issue (ii)] 2. Now select each valid value in turn and delete the value from the field For the positive range, the old value is restored and not deleted [issue (i)]. Note that the actions work correctly for the second field where the range is -100 to +100. **Expected behavior** As the field is optional, it should be possible to remove the value and make it empty / null again (unless the emptyValue attribute specified that something else should be used).
1.0
InputNumber: unable to clear value from field when using min/max range both > 0 - **Describe the defect** There appear to be two related issues: (i) it is not possible to delete a value again from an inputNumber field where the min/val values are set to be wholly positive and does not include zero - e.g. minValue 1, maxValue 100 - The field is an option, and is not a required value. The problem is the same whether the emptyValue attribute is absent, or when it is set to "null". (PS. the meaning of attribute values of "focus" and "press" per the documentation is not clear to me) (ii) When the range is wholly negative (e.g. minValue -100, maxValue -1), it is not possible to enter any value into the field. **Reproducer** See attached. [tester_inputNumber_delete_src.zip](https://github.com/primefaces/primefaces/files/4912144/tester_inputNumber_delete_src.zip) **Environment:** - PF Version: _8.0_ - JSF + version: _Mojarra 2.2.8_ - Affected browsers: _(at least) Firefox_ **To Reproduce** Run the reproducer: 1. Enter a valid value into the first three fields, per the range specified. You will be unable to enter a value for the negative range in the third case [issue (ii)] 2. Now select each valid value in turn and delete the value from the field For the positive range, the old value is restored and not deleted [issue (i)]. Note that the actions work correctly for the second field where the range is -100 to +100. **Expected behavior** As the field is optional, it should be possible to remove the value and make it empty / null again (unless the emptyValue attribute specified that something else should be used).
defect
inputnumber unable to clear value from field when using min max range both describe the defect there appear to be two related issues i it is not possible to delete a value again from an inputnumber field where the min val values are set to be wholly positive and does not include zero e g minvalue maxvalue the field is an option and is not a required value the problem is the same whether the emptyvalue attribute is absent or when it is set to null ps the meaning of attribute values of focus and press per the documentation is not clear to me ii when the range is wholly negative e g minvalue maxvalue it is not possible to enter any value into the field reproducer see attached environment pf version jsf version mojarra affected browsers at least firefox to reproduce run the reproducer enter a valid value into the first three fields per the range specified you will be unable to enter a value for the negative range in the third case now select each valid value in turn and delete the value from the field for the positive range the old value is restored and not deleted note that the actions work correctly for the second field where the range is to expected behavior as the field is optional it should be possible to remove the value and make it empty null again unless the emptyvalue attribute specified that something else should be used
1
55,760
14,672,890,733
IssuesEvent
2020-12-30 11:44:18
nodeca/js-yaml
https://api.github.com/repos/nodeca/js-yaml
closed
tabs indentation is not considered as invalid
defect
Tab indentation will basically result as a non-char when parsing YAML. For instance: ``` javascript a: b: c: yo mamma ``` As tab indented comes out as: ``` bash js-yaml test.yaml { a: null, b: null, c: 'yo mamma' } ``` It would be safer and easier to debug if the file would not be parseable on tab indent.
1.0
tabs indentation is not considered as invalid - Tab indentation will basically result as a non-char when parsing YAML. For instance: ``` javascript a: b: c: yo mamma ``` As tab indented comes out as: ``` bash js-yaml test.yaml { a: null, b: null, c: 'yo mamma' } ``` It would be safer and easier to debug if the file would not be parseable on tab indent.
defect
tabs indentation is not considered as invalid tab indentation will basically result as a non char when parsing yaml for instance javascript a b c yo mamma as tab indented comes out as bash js yaml test yaml a null b null c yo mamma it would be safer and easier to debug if the file would not be parseable on tab indent
1
38,857
10,256,921,612
IssuesEvent
2019-08-21 18:49:52
tensorflow/tfjs
https://api.github.com/repos/tensorflow/tfjs
closed
tfjs-examples: Simple object detection: yarn train --gpu fails
type:build/install
#### TensorFlow.js version 1.2.2 #### Browser version Windows Version 10.0.17134 Build 17134 Node v10.15.0 ### Problem description **Install appears to succeed (with warnings)** ``` yarn install v1.17.3 [1/5] Validating package.json... [2/5] Resolving packages... [3/5] Fetching packages... info fsevents@1.2.4: The platform "win32" is incompatible with this module. info "fsevents@1.2.4" is an optional dependency and failed compatibility check. Excluding it from installation. [4/5] Linking dependencies... warning "@tensorflow/tfjs > @tensorflow/tfjs-data@1.2.2" has unmet peer dependency "seedrandom@~2.4.3". warning "@tensorflow/tfjs > @tensorflow/tfjs-core > rollup-plugin-visualizer@1.1.1" has unmet peer dependency "rollup@>=0.60.0". [5/5] Building fresh packages... Done in 139.97s. ``` **Training with the gpu flag fails** ``` $ yarn train --gpu yarn run v1.17.3 $ node train.js --gpu Training using GPU. cpu backend was already registered. Reusing existing backend factory. Platform node has already been set. Overwriting the platform with [object Object]. node-pre-gyp info This Node instance does not support builds for N-API version 4 node-pre-gyp info This Node instance does not support builds for N-API version 4 (node:14276) UnhandledPromiseRejectionWarning: Error: The specified module could not be found. \\?\C:\Users\Ian\projects\tfjs\tfjs-examples\simple-object-detection\node_modules\@tensorflow\tfjs-node-gpu\lib\napi-v3\tfjs_binding.node at Object.Module._extensions..node (internal/modules/cjs/loader.js:718:18) at Module.load (internal/modules/cjs/loader.js:599:32) at tryModuleLoad (internal/modules/cjs/loader.js:538:12) at Function.Module._load (internal/modules/cjs/loader.js:530:3) at Module.require (internal/modules/cjs/loader.js:637:17) at require (internal/modules/cjs/helpers.js:22:18) at Object.<anonymous> (C:\Users\Ian\projects\tfjs\tfjs-examples\simple-object-detection\node_modules\@tensorflow\tfjs-node-gpu\dist\index.js:44:16) at Module._compile (internal/modules/cjs/loader.js:689:30) at Object.Module._extensions..js (internal/modules/cjs/loader.js:700:10) at Module.load (internal/modules/cjs/loader.js:599:32) (node:14276) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1) (node:14276) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code. Done in 1.74s. ```
1.0
tfjs-examples: Simple object detection: yarn train --gpu fails - #### TensorFlow.js version 1.2.2 #### Browser version Windows Version 10.0.17134 Build 17134 Node v10.15.0 ### Problem description **Install appears to succeed (with warnings)** ``` yarn install v1.17.3 [1/5] Validating package.json... [2/5] Resolving packages... [3/5] Fetching packages... info fsevents@1.2.4: The platform "win32" is incompatible with this module. info "fsevents@1.2.4" is an optional dependency and failed compatibility check. Excluding it from installation. [4/5] Linking dependencies... warning "@tensorflow/tfjs > @tensorflow/tfjs-data@1.2.2" has unmet peer dependency "seedrandom@~2.4.3". warning "@tensorflow/tfjs > @tensorflow/tfjs-core > rollup-plugin-visualizer@1.1.1" has unmet peer dependency "rollup@>=0.60.0". [5/5] Building fresh packages... Done in 139.97s. ``` **Training with the gpu flag fails** ``` $ yarn train --gpu yarn run v1.17.3 $ node train.js --gpu Training using GPU. cpu backend was already registered. Reusing existing backend factory. Platform node has already been set. Overwriting the platform with [object Object]. node-pre-gyp info This Node instance does not support builds for N-API version 4 node-pre-gyp info This Node instance does not support builds for N-API version 4 (node:14276) UnhandledPromiseRejectionWarning: Error: The specified module could not be found. \\?\C:\Users\Ian\projects\tfjs\tfjs-examples\simple-object-detection\node_modules\@tensorflow\tfjs-node-gpu\lib\napi-v3\tfjs_binding.node at Object.Module._extensions..node (internal/modules/cjs/loader.js:718:18) at Module.load (internal/modules/cjs/loader.js:599:32) at tryModuleLoad (internal/modules/cjs/loader.js:538:12) at Function.Module._load (internal/modules/cjs/loader.js:530:3) at Module.require (internal/modules/cjs/loader.js:637:17) at require (internal/modules/cjs/helpers.js:22:18) at Object.<anonymous> (C:\Users\Ian\projects\tfjs\tfjs-examples\simple-object-detection\node_modules\@tensorflow\tfjs-node-gpu\dist\index.js:44:16) at Module._compile (internal/modules/cjs/loader.js:689:30) at Object.Module._extensions..js (internal/modules/cjs/loader.js:700:10) at Module.load (internal/modules/cjs/loader.js:599:32) (node:14276) UnhandledPromiseRejectionWarning: Unhandled promise rejection. This error originated either by throwing inside of an async function without a catch block, or by rejecting a promise which was not handled with .catch(). (rejection id: 1) (node:14276) [DEP0018] DeprecationWarning: Unhandled promise rejections are deprecated. In the future, promise rejections that are not handled will terminate the Node.js process with a non-zero exit code. Done in 1.74s. ```
non_defect
tfjs examples simple object detection yarn train gpu fails tensorflow js version browser version windows version build node problem description install appears to succeed with warnings yarn install validating package json resolving packages fetching packages info fsevents the platform is incompatible with this module info fsevents is an optional dependency and failed compatibility check excluding it from installation linking dependencies warning tensorflow tfjs tensorflow tfjs data has unmet peer dependency seedrandom warning tensorflow tfjs tensorflow tfjs core rollup plugin visualizer has unmet peer dependency rollup building fresh packages done in training with the gpu flag fails yarn train gpu yarn run node train js gpu training using gpu cpu backend was already registered reusing existing backend factory platform node has already been set overwriting the platform with node pre gyp info this node instance does not support builds for n api version node pre gyp info this node instance does not support builds for n api version node unhandledpromiserejectionwarning error the specified module could not be found c users ian projects tfjs tfjs examples simple object detection node modules tensorflow tfjs node gpu lib napi tfjs binding node at object module extensions node internal modules cjs loader js at module load internal modules cjs loader js at trymoduleload internal modules cjs loader js at function module load internal modules cjs loader js at module require internal modules cjs loader js at require internal modules cjs helpers js at object c users ian projects tfjs tfjs examples simple object detection node modules tensorflow tfjs node gpu dist index js at module compile internal modules cjs loader js at object module extensions js internal modules cjs loader js at module load internal modules cjs loader js node unhandledpromiserejectionwarning unhandled promise rejection this error originated either by throwing inside of an async function without a catch block or by rejecting a promise which was not handled with catch rejection id node deprecationwarning unhandled promise rejections are deprecated in the future promise rejections that are not handled will terminate the node js process with a non zero exit code done in
0
17,072
2,974,593,145
IssuesEvent
2015-07-15 02:10:23
Reimashi/jotai
https://api.github.com/repos/Reimashi/jotai
closed
User Account Control dialogue/won't reopen from systray
auto-migrated Priority-Medium Type-Defect wontfix
``` "What is the expected output? What do you see instead?" Every time the software is run, it asks me for UAC authorization to run it. Understandably so, but when program is minimized to system tray, I cannot reopen it without running another copy and everything gets killed. The only way to get around this, is to disable UAC in windows control panel> user accounts> change user account control settings. What version of the product are you using? On what operating system? Open Hardware Monitor v0.4.0 Beta on Windows 7 Ultimate 64-bit. Please provide any additional information below. Other than that, everything seems to be working okay. I can't see any info on my HD or video card, but I assume its because they are not supported. I really just need the program to monitor processor temp. Just would like to be able to reopen it and keep UAC on. =) ``` Original issue reported on code.google.com by `Josh.Con...@gmail.com` on 12 Apr 2012 at 1:49
1.0
User Account Control dialogue/won't reopen from systray - ``` "What is the expected output? What do you see instead?" Every time the software is run, it asks me for UAC authorization to run it. Understandably so, but when program is minimized to system tray, I cannot reopen it without running another copy and everything gets killed. The only way to get around this, is to disable UAC in windows control panel> user accounts> change user account control settings. What version of the product are you using? On what operating system? Open Hardware Monitor v0.4.0 Beta on Windows 7 Ultimate 64-bit. Please provide any additional information below. Other than that, everything seems to be working okay. I can't see any info on my HD or video card, but I assume its because they are not supported. I really just need the program to monitor processor temp. Just would like to be able to reopen it and keep UAC on. =) ``` Original issue reported on code.google.com by `Josh.Con...@gmail.com` on 12 Apr 2012 at 1:49
defect
user account control dialogue won t reopen from systray what is the expected output what do you see instead every time the software is run it asks me for uac authorization to run it understandably so but when program is minimized to system tray i cannot reopen it without running another copy and everything gets killed the only way to get around this is to disable uac in windows control panel user accounts change user account control settings what version of the product are you using on what operating system open hardware monitor beta on windows ultimate bit please provide any additional information below other than that everything seems to be working okay i can t see any info on my hd or video card but i assume its because they are not supported i really just need the program to monitor processor temp just would like to be able to reopen it and keep uac on original issue reported on code google com by josh con gmail com on apr at
1
50,172
13,187,360,914
IssuesEvent
2020-08-13 03:10:09
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
q-frames need to be finshed (Trac #256)
Migrated from Trac combo core defect
This includes documentation, doxygen, and pydocs. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/256 , reported by nega and owned by Gang of Three_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "This includes documentation, doxygen, and pydocs.", "reporter": "nega", "cc": "nwhitehorn, jvansanten, naoko", "resolution": "fixed", "_ts": "1416713877111216", "component": "combo core", "summary": "q-frames need to be finshed", "priority": "blocker", "keywords": "q-frames documentation", "time": "2011-05-11T20:18:53", "milestone": "", "owner": "Gang of Three", "type": "defect" } ``` </p> </details>
1.0
q-frames need to be finshed (Trac #256) - This includes documentation, doxygen, and pydocs. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/256 , reported by nega and owned by Gang of Three_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "This includes documentation, doxygen, and pydocs.", "reporter": "nega", "cc": "nwhitehorn, jvansanten, naoko", "resolution": "fixed", "_ts": "1416713877111216", "component": "combo core", "summary": "q-frames need to be finshed", "priority": "blocker", "keywords": "q-frames documentation", "time": "2011-05-11T20:18:53", "milestone": "", "owner": "Gang of Three", "type": "defect" } ``` </p> </details>
defect
q frames need to be finshed trac this includes documentation doxygen and pydocs migrated from reported by nega and owned by gang of three json status closed changetime description this includes documentation doxygen and pydocs reporter nega cc nwhitehorn jvansanten naoko resolution fixed ts component combo core summary q frames need to be finshed priority blocker keywords q frames documentation time milestone owner gang of three type defect
1
46,750
13,055,969,512
IssuesEvent
2020-07-30 03:16:02
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
remove dst-extractor from metaprojects (Trac #1812)
Incomplete Migration Migrated from Trac defect other
Migrated from https://code.icecube.wisc.edu/ticket/1812 ```json { "status": "closed", "changetime": "2019-02-13T14:12:38", "description": "As the release notes say:\n{{{\ntrunk\n-----\n* This project is now obsolete. The extractor modules have been moved to\n the DST project.\n}}}\n\nSo, time to remove it from the metaproject externals. I think it's only icerec and combo.", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1550067158057333", "component": "other", "summary": "remove dst-extractor from metaprojects", "priority": "normal", "keywords": "", "time": "2016-08-03T17:17:14", "milestone": "", "owner": "olivas", "type": "defect" } ```
1.0
remove dst-extractor from metaprojects (Trac #1812) - Migrated from https://code.icecube.wisc.edu/ticket/1812 ```json { "status": "closed", "changetime": "2019-02-13T14:12:38", "description": "As the release notes say:\n{{{\ntrunk\n-----\n* This project is now obsolete. The extractor modules have been moved to\n the DST project.\n}}}\n\nSo, time to remove it from the metaproject externals. I think it's only icerec and combo.", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1550067158057333", "component": "other", "summary": "remove dst-extractor from metaprojects", "priority": "normal", "keywords": "", "time": "2016-08-03T17:17:14", "milestone": "", "owner": "olivas", "type": "defect" } ```
defect
remove dst extractor from metaprojects trac migrated from json status closed changetime description as the release notes say n ntrunk n n this project is now obsolete the extractor modules have been moved to n the dst project n n nso time to remove it from the metaproject externals i think it s only icerec and combo reporter david schultz cc resolution fixed ts component other summary remove dst extractor from metaprojects priority normal keywords time milestone owner olivas type defect
1
146,158
5,612,223,018
IssuesEvent
2017-04-03 03:41:16
kamal1978/LTFHC
https://api.github.com/repos/kamal1978/LTFHC
closed
Selecting row on register should take user to patient summary screen
app bug high priority
One primary use for the register is to be a way for the user to link to the patient data. When the user clicks on any part of the row containing a patient, the app should take the user to the patient summary screen for that particular patient.
1.0
Selecting row on register should take user to patient summary screen - One primary use for the register is to be a way for the user to link to the patient data. When the user clicks on any part of the row containing a patient, the app should take the user to the patient summary screen for that particular patient.
non_defect
selecting row on register should take user to patient summary screen one primary use for the register is to be a way for the user to link to the patient data when the user clicks on any part of the row containing a patient the app should take the user to the patient summary screen for that particular patient
0
26,035
4,552,894,468
IssuesEvent
2016-09-13 01:21:27
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
opened
duplicate resids hide residues in atom.residues
Component-Core defect
I used the PDB from https://github.com/MDAnalysis/mdanalysis/issues/975#issuecomment-246074022 `asdf.pdb` which has no chain ID but duplicated resids. ### Expected behaviour All residues should be separately resolved in `atoms.residues`. ### Actual behaviour Instead, only one residue for each resid is visible. All atoms are present. ### Code to reproduce the behaviour Download [asdf.py](https://dl.dropboxusercontent.com/u/49250897/asdf.pdb) from dropbox. ``` python import MDAnalysis as mda u = mda.Universe("asdf.pdb") print(len(u.atoms)) # 29879 print(len(u.residues)) 431 # the correct number of residues: print(u.atoms.CA.n_atoms) # 1849 ``` ### Currently version of MDAnalysis: (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.15.0, 0.15.1-dev
1.0
duplicate resids hide residues in atom.residues - I used the PDB from https://github.com/MDAnalysis/mdanalysis/issues/975#issuecomment-246074022 `asdf.pdb` which has no chain ID but duplicated resids. ### Expected behaviour All residues should be separately resolved in `atoms.residues`. ### Actual behaviour Instead, only one residue for each resid is visible. All atoms are present. ### Code to reproduce the behaviour Download [asdf.py](https://dl.dropboxusercontent.com/u/49250897/asdf.pdb) from dropbox. ``` python import MDAnalysis as mda u = mda.Universe("asdf.pdb") print(len(u.atoms)) # 29879 print(len(u.residues)) 431 # the correct number of residues: print(u.atoms.CA.n_atoms) # 1849 ``` ### Currently version of MDAnalysis: (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.15.0, 0.15.1-dev
defect
duplicate resids hide residues in atom residues i used the pdb from asdf pdb which has no chain id but duplicated resids expected behaviour all residues should be separately resolved in atoms residues actual behaviour instead only one residue for each resid is visible all atoms are present code to reproduce the behaviour download from dropbox python import mdanalysis as mda u mda universe asdf pdb print len u atoms print len u residues the correct number of residues print u atoms ca n atoms currently version of mdanalysis run python c import mdanalysis as mda print mda version dev
1
352,413
32,064,463,864
IssuesEvent
2023-09-25 00:57:04
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_tags_function_via_global_checkpoint (__main__.ActivationCheckpointingViaTagsTests)
triaged module: flaky-tests skipped module: dynamo
Platforms: linux, rocm This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_tags_function_via_global_checkpoint) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/17056185819). Over the past 72 hours, it has flakily failed in 2 workflow(s). **Debugging instructions (after clicking on the recent samples link):** To find relevant log snippets: 1. Click on the workflow logs linked above 2. Grep for `test_tags_function_via_global_checkpoint` Test file path: `dynamo/test_activation_checkpointing.py`
1.0
DISABLED test_tags_function_via_global_checkpoint (__main__.ActivationCheckpointingViaTagsTests) - Platforms: linux, rocm This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/failure/test_tags_function_via_global_checkpoint) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/17056185819). Over the past 72 hours, it has flakily failed in 2 workflow(s). **Debugging instructions (after clicking on the recent samples link):** To find relevant log snippets: 1. Click on the workflow logs linked above 2. Grep for `test_tags_function_via_global_checkpoint` Test file path: `dynamo/test_activation_checkpointing.py`
non_defect
disabled test tags function via global checkpoint main activationcheckpointingviatagstests platforms linux rocm this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has flakily failed in workflow s debugging instructions after clicking on the recent samples link to find relevant log snippets click on the workflow logs linked above grep for test tags function via global checkpoint test file path dynamo test activation checkpointing py
0
325,718
27,957,140,887
IssuesEvent
2023-03-24 13:13:43
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
[CI] NodeShutdownShardsIT testNodeReplacementOnlyToTarget failing
:Data Management/Stats >test-failure
Doesn't look related to the specific test so much as something fairly fundamental wrong with JDK stats collection: ``` Caused by: java.lang.InternalError: Memory Pool not found at __randomizedtesting.SeedInfo.seed([6ED1C68BAF8AACC1]:0) at sun.management.MemoryPoolImpl.getPeakUsage0(Native Method) at sun.management.MemoryPoolImpl.getPeakUsage(MemoryPoolImpl.java:98) at org.elasticsearch.monitor.jvm.JvmStats.jvmStats(JvmStats.java:63) at org.elasticsearch.monitor.jvm.JvmGcMonitorService$JvmMonitor.jvmStats(JvmGcMonitorService.java:522) ``` **Build scan:** https://gradle-enterprise.elastic.co/s/fimkttzyvl7no/tests/:x-pack:plugin:shutdown:internalClusterTest/org.elasticsearch.xpack.shutdown.NodeShutdownShardsIT/testNodeReplacementOnlyToTarget **Reproduction line:** ``` ./gradlew ':x-pack:plugin:shutdown:internalClusterTest' --tests "org.elasticsearch.xpack.shutdown.NodeShutdownShardsIT.testNodeReplacementOnlyToTarget" -Dtests.seed=6ED1C68BAF8AACC1 -Dtests.locale=sk -Dtests.timezone=Australia/Adelaide -Druntime.java=20 ``` **Applicable branches:** main **Reproduces locally?:** Didn't try **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.shutdown.NodeShutdownShardsIT&tests.test=testNodeReplacementOnlyToTarget **Failure excerpt:** ``` com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=409, name=test_TEST-TEST_WORKER_VM=[370]-CLUSTER_SEED=[3842257269405316757]-HASH=[B137CBE5C9]-cluster[T#1], state=RUNNABLE, group=TGRP-NodeShutdownShardsIT] at __randomizedtesting.SeedInfo.seed([6ED1C68BAF8AACC1:F3546514071033F8]:0) Caused by: java.lang.InternalError: Memory Pool not found at __randomizedtesting.SeedInfo.seed([6ED1C68BAF8AACC1]:0) at sun.management.MemoryPoolImpl.getPeakUsage0(MemoryPoolImpl.java:-2) at sun.management.MemoryPoolImpl.getPeakUsage(MemoryPoolImpl.java:98) at org.elasticsearch.monitor.jvm.JvmStats.jvmStats(JvmStats.java:63) at org.elasticsearch.monitor.jvm.JvmGcMonitorService$JvmMonitor.jvmStats(JvmGcMonitorService.java:522) at org.elasticsearch.monitor.jvm.JvmGcMonitorService$JvmMonitor.<init>(JvmGcMonitorService.java:413) at org.elasticsearch.monitor.jvm.JvmGcMonitorService$1.<init>(JvmGcMonitorService.java:212) at org.elasticsearch.monitor.jvm.JvmGcMonitorService.doStart(JvmGcMonitorService.java:212) at org.elasticsearch.common.component.AbstractLifecycleComponent.start(AbstractLifecycleComponent.java:43) at org.elasticsearch.monitor.MonitorService.doStart(MonitorService.java:57) at org.elasticsearch.common.component.AbstractLifecycleComponent.start(AbstractLifecycleComponent.java:43) at org.elasticsearch.node.Node.start(Node.java:1371) at org.elasticsearch.test.InternalTestCluster$NodeAndClient.startNode(InternalTestCluster.java:988) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:577) at java.util.concurrent.FutureTask.run(FutureTask.java:317) at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:891) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642) at java.lang.Thread.run(Thread.java:1623) ```
1.0
[CI] NodeShutdownShardsIT testNodeReplacementOnlyToTarget failing - Doesn't look related to the specific test so much as something fairly fundamental wrong with JDK stats collection: ``` Caused by: java.lang.InternalError: Memory Pool not found at __randomizedtesting.SeedInfo.seed([6ED1C68BAF8AACC1]:0) at sun.management.MemoryPoolImpl.getPeakUsage0(Native Method) at sun.management.MemoryPoolImpl.getPeakUsage(MemoryPoolImpl.java:98) at org.elasticsearch.monitor.jvm.JvmStats.jvmStats(JvmStats.java:63) at org.elasticsearch.monitor.jvm.JvmGcMonitorService$JvmMonitor.jvmStats(JvmGcMonitorService.java:522) ``` **Build scan:** https://gradle-enterprise.elastic.co/s/fimkttzyvl7no/tests/:x-pack:plugin:shutdown:internalClusterTest/org.elasticsearch.xpack.shutdown.NodeShutdownShardsIT/testNodeReplacementOnlyToTarget **Reproduction line:** ``` ./gradlew ':x-pack:plugin:shutdown:internalClusterTest' --tests "org.elasticsearch.xpack.shutdown.NodeShutdownShardsIT.testNodeReplacementOnlyToTarget" -Dtests.seed=6ED1C68BAF8AACC1 -Dtests.locale=sk -Dtests.timezone=Australia/Adelaide -Druntime.java=20 ``` **Applicable branches:** main **Reproduces locally?:** Didn't try **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.shutdown.NodeShutdownShardsIT&tests.test=testNodeReplacementOnlyToTarget **Failure excerpt:** ``` com.carrotsearch.randomizedtesting.UncaughtExceptionError: Captured an uncaught exception in thread: Thread[id=409, name=test_TEST-TEST_WORKER_VM=[370]-CLUSTER_SEED=[3842257269405316757]-HASH=[B137CBE5C9]-cluster[T#1], state=RUNNABLE, group=TGRP-NodeShutdownShardsIT] at __randomizedtesting.SeedInfo.seed([6ED1C68BAF8AACC1:F3546514071033F8]:0) Caused by: java.lang.InternalError: Memory Pool not found at __randomizedtesting.SeedInfo.seed([6ED1C68BAF8AACC1]:0) at sun.management.MemoryPoolImpl.getPeakUsage0(MemoryPoolImpl.java:-2) at sun.management.MemoryPoolImpl.getPeakUsage(MemoryPoolImpl.java:98) at org.elasticsearch.monitor.jvm.JvmStats.jvmStats(JvmStats.java:63) at org.elasticsearch.monitor.jvm.JvmGcMonitorService$JvmMonitor.jvmStats(JvmGcMonitorService.java:522) at org.elasticsearch.monitor.jvm.JvmGcMonitorService$JvmMonitor.<init>(JvmGcMonitorService.java:413) at org.elasticsearch.monitor.jvm.JvmGcMonitorService$1.<init>(JvmGcMonitorService.java:212) at org.elasticsearch.monitor.jvm.JvmGcMonitorService.doStart(JvmGcMonitorService.java:212) at org.elasticsearch.common.component.AbstractLifecycleComponent.start(AbstractLifecycleComponent.java:43) at org.elasticsearch.monitor.MonitorService.doStart(MonitorService.java:57) at org.elasticsearch.common.component.AbstractLifecycleComponent.start(AbstractLifecycleComponent.java:43) at org.elasticsearch.node.Node.start(Node.java:1371) at org.elasticsearch.test.InternalTestCluster$NodeAndClient.startNode(InternalTestCluster.java:988) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:577) at java.util.concurrent.FutureTask.run(FutureTask.java:317) at org.elasticsearch.common.util.concurrent.ThreadContext$ContextPreservingRunnable.run(ThreadContext.java:891) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642) at java.lang.Thread.run(Thread.java:1623) ```
non_defect
nodeshutdownshardsit testnodereplacementonlytotarget failing doesn t look related to the specific test so much as something fairly fundamental wrong with jdk stats collection caused by java lang internalerror memory pool not found at randomizedtesting seedinfo seed at sun management memorypoolimpl native method at sun management memorypoolimpl getpeakusage memorypoolimpl java at org elasticsearch monitor jvm jvmstats jvmstats jvmstats java at org elasticsearch monitor jvm jvmgcmonitorservice jvmmonitor jvmstats jvmgcmonitorservice java build scan reproduction line gradlew x pack plugin shutdown internalclustertest tests org elasticsearch xpack shutdown nodeshutdownshardsit testnodereplacementonlytotarget dtests seed dtests locale sk dtests timezone australia adelaide druntime java applicable branches main reproduces locally didn t try failure history failure excerpt com carrotsearch randomizedtesting uncaughtexceptionerror captured an uncaught exception in thread thread cluster seed hash cluster state runnable group tgrp nodeshutdownshardsit at randomizedtesting seedinfo seed caused by java lang internalerror memory pool not found at randomizedtesting seedinfo seed at sun management memorypoolimpl memorypoolimpl java at sun management memorypoolimpl getpeakusage memorypoolimpl java at org elasticsearch monitor jvm jvmstats jvmstats jvmstats java at org elasticsearch monitor jvm jvmgcmonitorservice jvmmonitor jvmstats jvmgcmonitorservice java at org elasticsearch monitor jvm jvmgcmonitorservice jvmmonitor jvmgcmonitorservice java at org elasticsearch monitor jvm jvmgcmonitorservice jvmgcmonitorservice java at org elasticsearch monitor jvm jvmgcmonitorservice dostart jvmgcmonitorservice java at org elasticsearch common component abstractlifecyclecomponent start abstractlifecyclecomponent java at org elasticsearch monitor monitorservice dostart monitorservice java at org elasticsearch common component abstractlifecyclecomponent start abstractlifecyclecomponent java at org elasticsearch node node start node java at org elasticsearch test internaltestcluster nodeandclient startnode internaltestcluster java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at org elasticsearch common util concurrent threadcontext contextpreservingrunnable run threadcontext java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java
0
120,692
12,093,364,077
IssuesEvent
2020-04-19 19:19:52
DigitalExcellence/dex-backend
https://api.github.com/repos/DigitalExcellence/dex-backend
closed
Define architecture for backend
development documentation requires review sprint 0
In GitLab by @Brend-Smits on Mar 4, 2020, 20:36 Define a scalable architecture. Think broad, if we want to scale this. How are we going to do this with the architecture? Will a micro-service architecture benefit us?
1.0
Define architecture for backend - In GitLab by @Brend-Smits on Mar 4, 2020, 20:36 Define a scalable architecture. Think broad, if we want to scale this. How are we going to do this with the architecture? Will a micro-service architecture benefit us?
non_defect
define architecture for backend in gitlab by brend smits on mar define a scalable architecture think broad if we want to scale this how are we going to do this with the architecture will a micro service architecture benefit us
0
718,381
24,715,360,220
IssuesEvent
2022-10-20 06:24:23
discatpy-dev/library
https://api.github.com/repos/discatpy-dev/library
closed
Bump Python version minimum to 3.9
feature request status: planned priority: low
**Overview of the Feature Request** Bump the Python version minimum from 3.8 to 3.9. **What is this Feature Request for?** The whole library **The Problem** Currently, we are using type alias like `Dict` or `List` from the typing module. However, in [PEP 585](https://peps.python.org/pep-0585/) these type alias were deprecated for their builtin counterparts. This PEP unfortunately was not implemented in Python 3.8 but 3.9. **The Solution** Bump the Python version minimum in `setup.py` (`pyproject.toml` for the API redesign) from 3.8 to 3.9. **Current Solution** Since it's only deprecated and not removed entirely in 3.9, keeping it wouldn't really affect anything. But if it is going to be removed soon, then it's better to be prepared now then later. One solution I came up with to circumvent this is to redefine these type aliases to either be their builtin counterparts or to be the type alias from the typing module. However, this does seem unnecessary. Another solution is to use the new stringized annotations feature (`from __future__ import annotations`) in every module. **Additional context** N/A
1.0
Bump Python version minimum to 3.9 - **Overview of the Feature Request** Bump the Python version minimum from 3.8 to 3.9. **What is this Feature Request for?** The whole library **The Problem** Currently, we are using type alias like `Dict` or `List` from the typing module. However, in [PEP 585](https://peps.python.org/pep-0585/) these type alias were deprecated for their builtin counterparts. This PEP unfortunately was not implemented in Python 3.8 but 3.9. **The Solution** Bump the Python version minimum in `setup.py` (`pyproject.toml` for the API redesign) from 3.8 to 3.9. **Current Solution** Since it's only deprecated and not removed entirely in 3.9, keeping it wouldn't really affect anything. But if it is going to be removed soon, then it's better to be prepared now then later. One solution I came up with to circumvent this is to redefine these type aliases to either be their builtin counterparts or to be the type alias from the typing module. However, this does seem unnecessary. Another solution is to use the new stringized annotations feature (`from __future__ import annotations`) in every module. **Additional context** N/A
non_defect
bump python version minimum to overview of the feature request bump the python version minimum from to what is this feature request for the whole library the problem currently we are using type alias like dict or list from the typing module however in these type alias were deprecated for their builtin counterparts this pep unfortunately was not implemented in python but the solution bump the python version minimum in setup py pyproject toml for the api redesign from to current solution since it s only deprecated and not removed entirely in keeping it wouldn t really affect anything but if it is going to be removed soon then it s better to be prepared now then later one solution i came up with to circumvent this is to redefine these type aliases to either be their builtin counterparts or to be the type alias from the typing module however this does seem unnecessary another solution is to use the new stringized annotations feature from future import annotations in every module additional context n a
0
578,468
17,147,227,202
IssuesEvent
2021-07-13 15:49:48
googleapis/python-securitycenter
https://api.github.com/repos/googleapis/python-securitycenter
closed
samples.snippets.snippets_security_marks_test: test_delete_and_update_marks failed
api: securitycenter flakybot: issue priority: p1 samples type: bug
Note: #101 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 7bf5bb6b1262105722985639ffd49971bbe425af buildURL: [Build Status](https://source.cloud.google.com/results/invocations/74bea7ec-656d-4aab-8dc1-440b6b35601f), [Sponge](http://sponge2/74bea7ec-656d-4aab-8dc1-440b6b35601f) status: failed <details><summary>Test output</summary><br><pre>args = (parent: "organizations/1081635000895" page_token: "CukB3cy7qgAx8QqfK9BHP-R_w65yPqo2DxLCwlJQ-K6ljI1Fmw8iEPlsyRD1Y7JkqL...kDJdNis9vNGrQ0YQuUR8iAC37qqRV7qU8k0CP4bQxZO1IqdqP75K2ElgpNCnyG99-6-vacOL-87L7hYdCWPbsZa_K7g-485eVwV_kqgsOYm_-bhjS8=" ,) kwargs = {'metadata': [('x-goog-request-params', 'parent=organizations/1081635000895'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.31.0 gapic/1.4.0')], 'timeout': 480.0} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:67: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f326cc5a630> request = parent: "organizations/1081635000895" page_token: "CukB3cy7qgAx8QqfK9BHP-R_w65yPqo2DxLCwlJQ-K6ljI1Fmw8iEPlsyRD1Y7JkqLf...tPkDJdNis9vNGrQ0YQuUR8iAC37qqRV7qU8k0CP4bQxZO1IqdqP75K2ElgpNCnyG99-6-vacOL-87L7hYdCWPbsZa_K7g-485eVwV_kqgsOYm_-bhjS8=" timeout = 480.0 metadata = [('x-goog-request-params', 'parent=organizations/1081635000895'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.31.0 gapic/1.4.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:946: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7f326cc2e898> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f326cd5bac8> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _MultiThreadedRendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _InactiveRpcError(state) E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: E status = StatusCode.UNAUTHENTICATED E details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project." E debug_error_string = "{"created":"@1626167207.210284722","description":"Error received from peer ipv4:74.125.20.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" E > .nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:849: _InactiveRpcError The above exception was the direct cause of the following exception: organization_id = '1081635000895' @pytest.fixture(scope="module") def asset_name(organization_id): """Returns a random asset name from existing assets.""" from google.cloud import securitycenter client = securitycenter.SecurityCenterClient() # organization_id is the numeric ID of the organization. # organization_id=1234567777 org_name = "organizations/{org_id}".format(org_id=organization_id) > assets = list(client.list_assets(request={"parent": org_name})) snippets_security_marks_test.py:40: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../google/cloud/securitycenter_v1/services/security_center/pagers.py:347: in __iter__ for page in self.pages: ../../google/cloud/securitycenter_v1/services/security_center/pagers.py:341: in pages self._response = self._method(self._request, metadata=self._metadata) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/retry.py:290: in retry_wrapped_func on_error=on_error, .nox/py-3-6/lib/python3.6/site-packages/google/api_core/retry.py:188: in retry_target return target() .nox/py-3-6/lib/python3.6/site-packages/google/api_core/timeout.py:102: in func_with_timeout return func(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAUTHENTICATED details = "Request had invalid a...entication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" > > ??? E google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. <string>:3: Unauthenticated</pre></details>
1.0
samples.snippets.snippets_security_marks_test: test_delete_and_update_marks failed - Note: #101 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 7bf5bb6b1262105722985639ffd49971bbe425af buildURL: [Build Status](https://source.cloud.google.com/results/invocations/74bea7ec-656d-4aab-8dc1-440b6b35601f), [Sponge](http://sponge2/74bea7ec-656d-4aab-8dc1-440b6b35601f) status: failed <details><summary>Test output</summary><br><pre>args = (parent: "organizations/1081635000895" page_token: "CukB3cy7qgAx8QqfK9BHP-R_w65yPqo2DxLCwlJQ-K6ljI1Fmw8iEPlsyRD1Y7JkqL...kDJdNis9vNGrQ0YQuUR8iAC37qqRV7qU8k0CP4bQxZO1IqdqP75K2ElgpNCnyG99-6-vacOL-87L7hYdCWPbsZa_K7g-485eVwV_kqgsOYm_-bhjS8=" ,) kwargs = {'metadata': [('x-goog-request-params', 'parent=organizations/1081635000895'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.31.0 gapic/1.4.0')], 'timeout': 480.0} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:67: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f326cc5a630> request = parent: "organizations/1081635000895" page_token: "CukB3cy7qgAx8QqfK9BHP-R_w65yPqo2DxLCwlJQ-K6ljI1Fmw8iEPlsyRD1Y7JkqLf...tPkDJdNis9vNGrQ0YQuUR8iAC37qqRV7qU8k0CP4bQxZO1IqdqP75K2ElgpNCnyG99-6-vacOL-87L7hYdCWPbsZa_K7g-485eVwV_kqgsOYm_-bhjS8=" timeout = 480.0 metadata = [('x-goog-request-params', 'parent=organizations/1081635000895'), ('x-goog-api-client', 'gl-python/3.6.13 grpc/1.38.1 gax/1.31.0 gapic/1.4.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:946: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7f326cc2e898> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f326cd5bac8> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _MultiThreadedRendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _InactiveRpcError(state) E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: E status = StatusCode.UNAUTHENTICATED E details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project." E debug_error_string = "{"created":"@1626167207.210284722","description":"Error received from peer ipv4:74.125.20.95:443","file":"src/core/lib/surface/call.cc","file_line":1066,"grpc_message":"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" E > .nox/py-3-6/lib/python3.6/site-packages/grpc/_channel.py:849: _InactiveRpcError The above exception was the direct cause of the following exception: organization_id = '1081635000895' @pytest.fixture(scope="module") def asset_name(organization_id): """Returns a random asset name from existing assets.""" from google.cloud import securitycenter client = securitycenter.SecurityCenterClient() # organization_id is the numeric ID of the organization. # organization_id=1234567777 org_name = "organizations/{org_id}".format(org_id=organization_id) > assets = list(client.list_assets(request={"parent": org_name})) snippets_security_marks_test.py:40: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ../../google/cloud/securitycenter_v1/services/security_center/pagers.py:347: in __iter__ for page in self.pages: ../../google/cloud/securitycenter_v1/services/security_center/pagers.py:341: in pages self._response = self._method(self._request, metadata=self._metadata) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/retry.py:290: in retry_wrapped_func on_error=on_error, .nox/py-3-6/lib/python3.6/site-packages/google/api_core/retry.py:188: in retry_target return target() .nox/py-3-6/lib/python3.6/site-packages/google/api_core/timeout.py:102: in func_with_timeout return func(*args, **kwargs) .nox/py-3-6/lib/python3.6/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAUTHENTICATED details = "Request had invalid a...entication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.","grpc_status":16}" > > ??? E google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project. <string>:3: Unauthenticated</pre></details>
non_defect
samples snippets snippets security marks test test delete and update marks failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output args parent organizations page token r vacol kqgsoym kwargs metadata timeout six wraps callable def error remapped callable args kwargs try return callable args kwargs nox py lib site packages google api core grpc helpers py self request parent organizations page token r vacol kqgsoym timeout metadata credentials none wait for ready none compression none def call self request timeout none metadata none credentials none wait for ready none compression none state call self blocking request timeout metadata credentials wait for ready compression return end unary response blocking state call false none nox py lib site packages grpc channel py state call with call false deadline none def end unary response blocking state call with call deadline if state code is grpc statuscode ok if with call rendezvous multithreadedrendezvous state call none deadline return state response rendezvous else return state response else raise inactiverpcerror state e grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with e status statuscode unauthenticated e details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see e debug error string created description error received from peer file src core lib surface call cc file line grpc message request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see e nox py lib site packages grpc channel py inactiverpcerror the above exception was the direct cause of the following exception organization id pytest fixture scope module def asset name organization id returns a random asset name from existing assets from google cloud import securitycenter client securitycenter securitycenterclient organization id is the numeric id of the organization organization id org name organizations org id format org id organization id assets list client list assets request parent org name snippets security marks test py google cloud securitycenter services security center pagers py in iter for page in self pages google cloud securitycenter services security center pagers py in pages self response self method self request metadata self metadata nox py lib site packages google api core gapic method py in call return wrapped func args kwargs nox py lib site packages google api core retry py in retry wrapped func on error on error nox py lib site packages google api core retry py in retry target return target nox py lib site packages google api core timeout py in func with timeout return func args kwargs nox py lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value inactiverpcerror of rpc that terminated with status statuscode unauthenticated details request had invalid a entication credential see e google api core exceptions unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see unauthenticated
0
21,838
3,567,776,254
IssuesEvent
2016-01-26 00:40:06
CompEvol/beast2
https://api.github.com/repos/CompEvol/beast2
closed
TreeParser aborts on trees with multifurcations
defect HIGH priority
The earlier incarnation of TreeParser accepted nodes with >2 children by creating a ladder of dummy nodes. The current version parses such trees just fine, but aborts with an error during AST traversal if such a node is found. This is an error, as users need to be able to specify starting trees that have these multifurcations. The second error is that the TreeParser test suite does not contain a test for this behaviour.
1.0
TreeParser aborts on trees with multifurcations - The earlier incarnation of TreeParser accepted nodes with >2 children by creating a ladder of dummy nodes. The current version parses such trees just fine, but aborts with an error during AST traversal if such a node is found. This is an error, as users need to be able to specify starting trees that have these multifurcations. The second error is that the TreeParser test suite does not contain a test for this behaviour.
defect
treeparser aborts on trees with multifurcations the earlier incarnation of treeparser accepted nodes with children by creating a ladder of dummy nodes the current version parses such trees just fine but aborts with an error during ast traversal if such a node is found this is an error as users need to be able to specify starting trees that have these multifurcations the second error is that the treeparser test suite does not contain a test for this behaviour
1
211,851
16,460,423,422
IssuesEvent
2021-05-21 18:05:17
PRUNERS/FLiT
https://api.github.com/repos/PRUNERS/FLiT
opened
Handling for compiler license limitations
documentation enhancement make tests
## Feature Request **Description:** When analyzing Intel compilations, shared-license limitations may restrict running on multiple nodes (i.e., compiling with Intel on many nodes will cause failures when shared licenses are exceeded.) **Suggested change:** No suggestion at this time, will revisit with proposed changes in later comments; open to suggestions.
1.0
Handling for compiler license limitations - ## Feature Request **Description:** When analyzing Intel compilations, shared-license limitations may restrict running on multiple nodes (i.e., compiling with Intel on many nodes will cause failures when shared licenses are exceeded.) **Suggested change:** No suggestion at this time, will revisit with proposed changes in later comments; open to suggestions.
non_defect
handling for compiler license limitations feature request description when analyzing intel compilations shared license limitations may restrict running on multiple nodes i e compiling with intel on many nodes will cause failures when shared licenses are exceeded suggested change no suggestion at this time will revisit with proposed changes in later comments open to suggestions
0
65,636
19,608,828,647
IssuesEvent
2022-01-06 13:02:33
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
opened
DataTable: Filter + Input component with ajax processing that row and the next one
defect
**Describe the defect** Whenever I set a filter value in the column that has an input component (like selectBooleanCheckbox, datePicker, inputText, etc.) and then I proceed to change that value by ajax, the table gets updated and it shows the affected values for the row that I clicked and the next one. This doens't happen when I don't set any filters. **Reproducer** Project attached: [primefaces-test-master.zip](https://github.com/primefaces/primefaces/files/7821915/primefaces-test-master.zip) **Environment:** - PF Version: _11.0_ - JSF + version: _Mojarra 2.3_ - Affected browsers: _Chrome and FireFox_ **To Reproduce** Steps to reproduce the behavior: 1. Go to 'test.xhtml' 2. Set filter value 'Yes' for 'Status column' 3. Click the first checkbox 4. See error ![image](https://user-images.githubusercontent.com/26931796/148386500-a5ba6f97-2726-42ef-961a-4135c1d5f0ab.png) ![image](https://user-images.githubusercontent.com/26931796/148386510-61b21a87-ba67-412d-b106-aab9354e12d7.png) 5. Set filter value 'null' for 'Status column' 6. Changes where applied to clicked row and the next one ![image](https://user-images.githubusercontent.com/26931796/148386519-99bafabb-f78d-45cc-b39b-0832d3108a50.png) **Expected behavior** Only the first row should be changed. **Example XHTML** ```html <p:dataTable id="dta-tbl" widgetVar="dta-tbl" var="object" value="#{testView.testList}" paginator="true" paginatorTemplate="{CurrentPageReport} {FirstPageLink} {PreviousPageLink} {PageLinks} {NextPageLink} {LastPageLink} {RowsPerPageDropdown}" currentPageReportTemplate="Total: {totalRecords}" paginatorPosition="bottom" rowsPerPageTemplate="5,10,15" rows="5" pageLinks="3"> <f:facet name="header"> My table </f:facet> <p:column id="id-col" styleClass="" filterBy="#{object[0]}" headerText="Id."> #{object[0]} </p:column> <p:column id="status-col" styleClass="" filterBy="#{object[1]}" headerText="Status"> <f:facet name="filter"> <h:selectOneMenu id="status-sel" onchange="PF('dta-tbl').filter()"> <f:selectItem itemLabel="" itemValue="#{null}" noSelectionOption="true" /> <f:selectItem itemLabel="Yes" itemValue="#{true}" /> <f:selectItem itemLabel="No" itemValue="#{false}" /> </h:selectOneMenu> </f:facet> <p:selectBooleanCheckbox id="status-in" value="#{object[1]}"> <p:ajax process="@this" partialSubmit="true" listener="#{testView.updateStatus(object[0], object[1])}" update="@none" oncomplete="PF('dta-tbl').filter()" /> </p:selectBooleanCheckbox> </p:column> </p:dataTable> ``` **Example Bean** ```java @Data @Named @ViewScoped public class TestView implements Serializable { private String string; private List<Object[]> testList; @PostConstruct public void init() { string = "Welcome to PrimeFaces!!!"; testList = new ArrayList<>(); Object[] tmp = null; for (int idx = 0; idx < 5; idx++) { tmp = new Object[] { idx, true }; testList.add(tmp); } } public void updateStatus(Integer idx, Boolean value) { System.out.println(!value); } } ```
1.0
DataTable: Filter + Input component with ajax processing that row and the next one - **Describe the defect** Whenever I set a filter value in the column that has an input component (like selectBooleanCheckbox, datePicker, inputText, etc.) and then I proceed to change that value by ajax, the table gets updated and it shows the affected values for the row that I clicked and the next one. This doens't happen when I don't set any filters. **Reproducer** Project attached: [primefaces-test-master.zip](https://github.com/primefaces/primefaces/files/7821915/primefaces-test-master.zip) **Environment:** - PF Version: _11.0_ - JSF + version: _Mojarra 2.3_ - Affected browsers: _Chrome and FireFox_ **To Reproduce** Steps to reproduce the behavior: 1. Go to 'test.xhtml' 2. Set filter value 'Yes' for 'Status column' 3. Click the first checkbox 4. See error ![image](https://user-images.githubusercontent.com/26931796/148386500-a5ba6f97-2726-42ef-961a-4135c1d5f0ab.png) ![image](https://user-images.githubusercontent.com/26931796/148386510-61b21a87-ba67-412d-b106-aab9354e12d7.png) 5. Set filter value 'null' for 'Status column' 6. Changes where applied to clicked row and the next one ![image](https://user-images.githubusercontent.com/26931796/148386519-99bafabb-f78d-45cc-b39b-0832d3108a50.png) **Expected behavior** Only the first row should be changed. **Example XHTML** ```html <p:dataTable id="dta-tbl" widgetVar="dta-tbl" var="object" value="#{testView.testList}" paginator="true" paginatorTemplate="{CurrentPageReport} {FirstPageLink} {PreviousPageLink} {PageLinks} {NextPageLink} {LastPageLink} {RowsPerPageDropdown}" currentPageReportTemplate="Total: {totalRecords}" paginatorPosition="bottom" rowsPerPageTemplate="5,10,15" rows="5" pageLinks="3"> <f:facet name="header"> My table </f:facet> <p:column id="id-col" styleClass="" filterBy="#{object[0]}" headerText="Id."> #{object[0]} </p:column> <p:column id="status-col" styleClass="" filterBy="#{object[1]}" headerText="Status"> <f:facet name="filter"> <h:selectOneMenu id="status-sel" onchange="PF('dta-tbl').filter()"> <f:selectItem itemLabel="" itemValue="#{null}" noSelectionOption="true" /> <f:selectItem itemLabel="Yes" itemValue="#{true}" /> <f:selectItem itemLabel="No" itemValue="#{false}" /> </h:selectOneMenu> </f:facet> <p:selectBooleanCheckbox id="status-in" value="#{object[1]}"> <p:ajax process="@this" partialSubmit="true" listener="#{testView.updateStatus(object[0], object[1])}" update="@none" oncomplete="PF('dta-tbl').filter()" /> </p:selectBooleanCheckbox> </p:column> </p:dataTable> ``` **Example Bean** ```java @Data @Named @ViewScoped public class TestView implements Serializable { private String string; private List<Object[]> testList; @PostConstruct public void init() { string = "Welcome to PrimeFaces!!!"; testList = new ArrayList<>(); Object[] tmp = null; for (int idx = 0; idx < 5; idx++) { tmp = new Object[] { idx, true }; testList.add(tmp); } } public void updateStatus(Integer idx, Boolean value) { System.out.println(!value); } } ```
defect
datatable filter input component with ajax processing that row and the next one describe the defect whenever i set a filter value in the column that has an input component like selectbooleancheckbox datepicker inputtext etc and then i proceed to change that value by ajax the table gets updated and it shows the affected values for the row that i clicked and the next one this doens t happen when i don t set any filters reproducer project attached environment pf version jsf version mojarra affected browsers chrome and firefox to reproduce steps to reproduce the behavior go to test xhtml set filter value yes for status column click the first checkbox see error set filter value null for status column changes where applied to clicked row and the next one expected behavior only the first row should be changed example xhtml html p datatable id dta tbl widgetvar dta tbl var object value testview testlist paginator true paginatortemplate currentpagereport firstpagelink previouspagelink pagelinks nextpagelink lastpagelink rowsperpagedropdown currentpagereporttemplate total totalrecords paginatorposition bottom rowsperpagetemplate rows pagelinks my table object p ajax process this partialsubmit true listener testview updatestatus object object update none oncomplete pf dta tbl filter example bean java data named viewscoped public class testview implements serializable private string string private list testlist postconstruct public void init string welcome to primefaces testlist new arraylist object tmp null for int idx idx idx tmp new object idx true testlist add tmp public void updatestatus integer idx boolean value system out println value
1
53,762
13,262,251,459
IssuesEvent
2020-08-20 21:23:51
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[filterscripts] SimulationFiltering.py may fail due to missing key (Trac #2029)
Migrated from Trac combo reconstruction defect
Hi, while running my simulation, I have seen this error: NOTICE (HeseFollowup): injecting followup message into frame for run 1, event 639 (hesefollowup.py:129 in generate_small_HESE_followup_message) ERROR (PythonFunction): Error running python function as module: (PythonFunction.cxx:173 in virtual void PythonFunction::Process()) ERROR (I3Module): HESEFollowup_send_followup: Exception thrown (I3Module.cxx:116 in void I3Module::Do(void (I3Module::*)())) Traceback (most recent call last): File "/scratch/condor/dir_597/condor_exec.exe", line 418, in <module> main(opts) File "/scratch/condor/dir_597/condor_exec.exe", line 379, in main tray.Execute() File "/data/user/flauber/software/icerec_V05-00-05/build/lib/I3Tray.py", line 234, in Execute super(I3Tray, self).Execute() File "/data/user/flauber/software/icerec_V05-00-05/build/lib/icecube/filterscripts/hesefollowup.py", line 178, in generate_small_HESE_followup_message 'mpe_muex' : round(frame["OnlineL2_BestFit_MuEx"].energy, 5), KeyError: 'OnlineL2_BestFit_MuEx' Removing cpandel parameterization ... done WARN (I3BasicSeedService): (OnlineL2_multitseedprep) In 255 SetEvent() calls, 1 first guess track(s) were found but had a fit status that was not OK! (I3BasicSeedService.c$ It only happened for one job out of about 100, still I think it should not happen (and as the SimlationFiltering.py is self contained I hope its not an error due to my doing) <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2029">https://code.icecube.wisc.edu/projects/icecube/ticket/2029</a>, reported by flauberand owned by claudio.kopper</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:38", "_ts": "1550067278746682", "description": "Hi,\n\nwhile running my simulation, I have seen this error:\n\nNOTICE (HeseFollowup): injecting followup message into frame for run 1, event 639 (hesefollowup.py:129 in generate_small_HESE_followup_message)\nERROR (PythonFunction): Error running python function as module: (PythonFunction.cxx:173 in virtual void PythonFunction::Process())\nERROR (I3Module): HESEFollowup_send_followup: Exception thrown (I3Module.cxx:116 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/scratch/condor/dir_597/condor_exec.exe\", line 418, in <module>\n main(opts)\n File \"/scratch/condor/dir_597/condor_exec.exe\", line 379, in main\n tray.Execute()\n File \"/data/user/flauber/software/icerec_V05-00-05/build/lib/I3Tray.py\", line 234, in Execute\n super(I3Tray, self).Execute()\n File \"/data/user/flauber/software/icerec_V05-00-05/build/lib/icecube/filterscripts/hesefollowup.py\", line 178, in generate_small_HESE_followup_message\n 'mpe_muex' : round(frame[\"OnlineL2_BestFit_MuEx\"].energy, 5),\nKeyError: 'OnlineL2_BestFit_MuEx'\nRemoving cpandel parameterization ... done\nWARN (I3BasicSeedService): (OnlineL2_multitseedprep) In 255 SetEvent() calls, 1 first guess track(s) were found but had a fit status that was not OK! (I3BasicSeedService.c$\n\nIt only happened for one job out of about 100, still I think it should not happen (and as the SimlationFiltering.py is self contained I hope its not an error due to my doing)\n\n\n", "reporter": "flauber", "cc": "", "resolution": "fixed", "time": "2017-05-22T10:06:09", "component": "combo reconstruction", "summary": "[filterscripts] SimulationFiltering.py may fail due to missing key", "priority": "minor", "keywords": "filterscripts", "milestone": "", "owner": "claudio.kopper", "type": "defect" } ``` </p> </details>
1.0
[filterscripts] SimulationFiltering.py may fail due to missing key (Trac #2029) - Hi, while running my simulation, I have seen this error: NOTICE (HeseFollowup): injecting followup message into frame for run 1, event 639 (hesefollowup.py:129 in generate_small_HESE_followup_message) ERROR (PythonFunction): Error running python function as module: (PythonFunction.cxx:173 in virtual void PythonFunction::Process()) ERROR (I3Module): HESEFollowup_send_followup: Exception thrown (I3Module.cxx:116 in void I3Module::Do(void (I3Module::*)())) Traceback (most recent call last): File "/scratch/condor/dir_597/condor_exec.exe", line 418, in <module> main(opts) File "/scratch/condor/dir_597/condor_exec.exe", line 379, in main tray.Execute() File "/data/user/flauber/software/icerec_V05-00-05/build/lib/I3Tray.py", line 234, in Execute super(I3Tray, self).Execute() File "/data/user/flauber/software/icerec_V05-00-05/build/lib/icecube/filterscripts/hesefollowup.py", line 178, in generate_small_HESE_followup_message 'mpe_muex' : round(frame["OnlineL2_BestFit_MuEx"].energy, 5), KeyError: 'OnlineL2_BestFit_MuEx' Removing cpandel parameterization ... done WARN (I3BasicSeedService): (OnlineL2_multitseedprep) In 255 SetEvent() calls, 1 first guess track(s) were found but had a fit status that was not OK! (I3BasicSeedService.c$ It only happened for one job out of about 100, still I think it should not happen (and as the SimlationFiltering.py is self contained I hope its not an error due to my doing) <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2029">https://code.icecube.wisc.edu/projects/icecube/ticket/2029</a>, reported by flauberand owned by claudio.kopper</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:14:38", "_ts": "1550067278746682", "description": "Hi,\n\nwhile running my simulation, I have seen this error:\n\nNOTICE (HeseFollowup): injecting followup message into frame for run 1, event 639 (hesefollowup.py:129 in generate_small_HESE_followup_message)\nERROR (PythonFunction): Error running python function as module: (PythonFunction.cxx:173 in virtual void PythonFunction::Process())\nERROR (I3Module): HESEFollowup_send_followup: Exception thrown (I3Module.cxx:116 in void I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/scratch/condor/dir_597/condor_exec.exe\", line 418, in <module>\n main(opts)\n File \"/scratch/condor/dir_597/condor_exec.exe\", line 379, in main\n tray.Execute()\n File \"/data/user/flauber/software/icerec_V05-00-05/build/lib/I3Tray.py\", line 234, in Execute\n super(I3Tray, self).Execute()\n File \"/data/user/flauber/software/icerec_V05-00-05/build/lib/icecube/filterscripts/hesefollowup.py\", line 178, in generate_small_HESE_followup_message\n 'mpe_muex' : round(frame[\"OnlineL2_BestFit_MuEx\"].energy, 5),\nKeyError: 'OnlineL2_BestFit_MuEx'\nRemoving cpandel parameterization ... done\nWARN (I3BasicSeedService): (OnlineL2_multitseedprep) In 255 SetEvent() calls, 1 first guess track(s) were found but had a fit status that was not OK! (I3BasicSeedService.c$\n\nIt only happened for one job out of about 100, still I think it should not happen (and as the SimlationFiltering.py is self contained I hope its not an error due to my doing)\n\n\n", "reporter": "flauber", "cc": "", "resolution": "fixed", "time": "2017-05-22T10:06:09", "component": "combo reconstruction", "summary": "[filterscripts] SimulationFiltering.py may fail due to missing key", "priority": "minor", "keywords": "filterscripts", "milestone": "", "owner": "claudio.kopper", "type": "defect" } ``` </p> </details>
defect
simulationfiltering py may fail due to missing key trac hi while running my simulation i have seen this error notice hesefollowup injecting followup message into frame for run event hesefollowup py in generate small hese followup message error pythonfunction error running python function as module pythonfunction cxx in virtual void pythonfunction process error hesefollowup send followup exception thrown cxx in void do void traceback most recent call last file scratch condor dir condor exec exe line in main opts file scratch condor dir condor exec exe line in main tray execute file data user flauber software icerec build lib py line in execute super self execute file data user flauber software icerec build lib icecube filterscripts hesefollowup py line in generate small hese followup message mpe muex round frame energy keyerror bestfit muex removing cpandel parameterization done warn multitseedprep in setevent calls first guess track s were found but had a fit status that was not ok c it only happened for one job out of about still i think it should not happen and as the simlationfiltering py is self contained i hope its not an error due to my doing migrated from json status closed changetime ts description hi n nwhile running my simulation i have seen this error n nnotice hesefollowup injecting followup message into frame for run event hesefollowup py in generate small hese followup message nerror pythonfunction error running python function as module pythonfunction cxx in virtual void pythonfunction process nerror hesefollowup send followup exception thrown cxx in void do void ntraceback most recent call last n file scratch condor dir condor exec exe line in n main opts n file scratch condor dir condor exec exe line in main n tray execute n file data user flauber software icerec build lib py line in execute n super self execute n file data user flauber software icerec build lib icecube filterscripts hesefollowup py line in generate small hese followup message n mpe muex round frame energy nkeyerror bestfit muex nremoving cpandel parameterization done nwarn multitseedprep in setevent calls first guess track s were found but had a fit status that was not ok c n nit only happened for one job out of about still i think it should not happen and as the simlationfiltering py is self contained i hope its not an error due to my doing n n n reporter flauber cc resolution fixed time component combo reconstruction summary simulationfiltering py may fail due to missing key priority minor keywords filterscripts milestone owner claudio kopper type defect
1
61,200
17,023,633,588
IssuesEvent
2021-07-03 03:01:40
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Wrong additional town in hierarchy
Component: nominatim Priority: major Resolution: duplicate Type: defect
**[Submitted to the original trac issue database at 10.30am, Saturday, 11th September 2010]** If I search for Gebersheim the result is Gebersheim, Rutesheim, Leonberg, Landkreis Bblingen, Regierungsbezirk Stuttgart, Baden-Wrttemberg, Germany Rutesheim doesn't belong to the hierarchy. Gebersheim is part of the town Leonberg, Leonberg is a town and part of the "Landkreis Bblingen", the "Landkreis Bblingen" is part of the "Regierungsbezirk Stuttgart", the "Regierungsbezirk Stuttgart" is part of the (Bundesland) Baden-Wrttemberg and Baden-Wrttemberg is part of Germany. Rutesheim is also a town and part of the "Landkreis Bblingen", but independant from Leonberg.
1.0
Wrong additional town in hierarchy - **[Submitted to the original trac issue database at 10.30am, Saturday, 11th September 2010]** If I search for Gebersheim the result is Gebersheim, Rutesheim, Leonberg, Landkreis Bblingen, Regierungsbezirk Stuttgart, Baden-Wrttemberg, Germany Rutesheim doesn't belong to the hierarchy. Gebersheim is part of the town Leonberg, Leonberg is a town and part of the "Landkreis Bblingen", the "Landkreis Bblingen" is part of the "Regierungsbezirk Stuttgart", the "Regierungsbezirk Stuttgart" is part of the (Bundesland) Baden-Wrttemberg and Baden-Wrttemberg is part of Germany. Rutesheim is also a town and part of the "Landkreis Bblingen", but independant from Leonberg.
defect
wrong additional town in hierarchy if i search for gebersheim the result is gebersheim rutesheim leonberg landkreis bblingen regierungsbezirk stuttgart baden wrttemberg germany rutesheim doesn t belong to the hierarchy gebersheim is part of the town leonberg leonberg is a town and part of the landkreis bblingen the landkreis bblingen is part of the regierungsbezirk stuttgart the regierungsbezirk stuttgart is part of the bundesland baden wrttemberg and baden wrttemberg is part of germany rutesheim is also a town and part of the landkreis bblingen but independant from leonberg
1
45,689
13,018,602,492
IssuesEvent
2020-07-26 18:10:54
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
generated getChecks() - seems to create wrong default value
T: Defect
### Expected behavior and actual behavior: should create ``` Internal.createCheck( this, DSL.name("enforce_srid_the_geom"), "((st_srid(the_geom) = 4326))", true), ``` but actual is ``` Internal.createCheck( this, DSL.name("enforce_srid_the_geom"), "((st_srid(the_geom) = '-1'::integer))", true), ``` ### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve): I think this is "off by one" issue as the table definition has the following: ``` CREATE TABLE test ( id serial NOT NULL, ... in_city_id int4 NOT NULL DEFAULT '-1'::integer, simplify_geom geometry NULL, ... CONSTRAINT enforce_srid_simplify_geom CHECK ((st_srid(simplify_geom) = 4326)), ... ) ``` ### Versions: - jOOQ: 3.13.3 - Java: 8 - Database (include vendor): postgresql - OS: mac - JDBC Driver (include name if inofficial driver): org.postgresql:postgresql:9.4-1201-jdbc4
1.0
generated getChecks() - seems to create wrong default value - ### Expected behavior and actual behavior: should create ``` Internal.createCheck( this, DSL.name("enforce_srid_the_geom"), "((st_srid(the_geom) = 4326))", true), ``` but actual is ``` Internal.createCheck( this, DSL.name("enforce_srid_the_geom"), "((st_srid(the_geom) = '-1'::integer))", true), ``` ### Steps to reproduce the problem (if possible, create an MCVE: https://github.com/jOOQ/jOOQ-mcve): I think this is "off by one" issue as the table definition has the following: ``` CREATE TABLE test ( id serial NOT NULL, ... in_city_id int4 NOT NULL DEFAULT '-1'::integer, simplify_geom geometry NULL, ... CONSTRAINT enforce_srid_simplify_geom CHECK ((st_srid(simplify_geom) = 4326)), ... ) ``` ### Versions: - jOOQ: 3.13.3 - Java: 8 - Database (include vendor): postgresql - OS: mac - JDBC Driver (include name if inofficial driver): org.postgresql:postgresql:9.4-1201-jdbc4
defect
generated getchecks seems to create wrong default value expected behavior and actual behavior should create internal createcheck this dsl name enforce srid the geom st srid the geom true but actual is internal createcheck this dsl name enforce srid the geom st srid the geom integer true steps to reproduce the problem if possible create an mcve i think this is off by one issue as the table definition has the following create table test id serial not null in city id not null default integer simplify geom geometry null constraint enforce srid simplify geom check st srid simplify geom versions jooq java database include vendor postgresql os mac jdbc driver include name if inofficial driver org postgresql postgresql
1
5,563
3,962,493,776
IssuesEvent
2016-05-02 17:02:26
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
26041195: Add ability to resend TestFlight invites
classification:ui/usability reproducible:always status:open
#### Description It’s pretty common for people to completely ignore invites to TestFlight. Maybe this because they need to open the email on their phone (which may not have their work email connected) or they just aren’t interested at the time. Then, a few months later when they need access, they can’t find the original email. At this point I, as the developer, have no way of resending the invite. The only solution in iTunes Connect is to remove them and add them again, which isn’t very obvious, and feels like it could be dangerous. - Product Version: ? Created: 2016-05-02T16:42:27.959200 Originated: 2016-05-02T09:41:00 Open Radar Link: http://www.openradar.me/26041195
True
26041195: Add ability to resend TestFlight invites - #### Description It’s pretty common for people to completely ignore invites to TestFlight. Maybe this because they need to open the email on their phone (which may not have their work email connected) or they just aren’t interested at the time. Then, a few months later when they need access, they can’t find the original email. At this point I, as the developer, have no way of resending the invite. The only solution in iTunes Connect is to remove them and add them again, which isn’t very obvious, and feels like it could be dangerous. - Product Version: ? Created: 2016-05-02T16:42:27.959200 Originated: 2016-05-02T09:41:00 Open Radar Link: http://www.openradar.me/26041195
non_defect
add ability to resend testflight invites description it’s pretty common for people to completely ignore invites to testflight maybe this because they need to open the email on their phone which may not have their work email connected or they just aren’t interested at the time then a few months later when they need access they can’t find the original email at this point i as the developer have no way of resending the invite the only solution in itunes connect is to remove them and add them again which isn’t very obvious and feels like it could be dangerous product version created originated open radar link
0
530,159
15,417,247,896
IssuesEvent
2021-03-05 06:50:39
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
"Revision 1" is displayed as already "Deployed" when the Default API is created only
API-M 4.0.0 Priority/High Type/Bug
### Description: "Revision 1" is displayed as already "Deployed" when the Default Pizzashack API is created only. ### Steps to reproduce: 1. Go to Publisher 2. Deploy sample API When checked the LC state of the API in Lifecycles tab, it is displayed as "Created". But Lifecycle history list has an event as "LC has changed from CREATED to PUBLISHED" ![image](https://user-images.githubusercontent.com/2484781/109799041-85890500-7c41-11eb-8c9f-fd80473e11f9.png) This is critical as the default API is used by many and it should behave correctly. ### Affected Product Version: APIM-4.0.0-Alpha ### Environment details (with versions): - OS: - Client: - Env (Docker/K8s): --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
1.0
"Revision 1" is displayed as already "Deployed" when the Default API is created only - ### Description: "Revision 1" is displayed as already "Deployed" when the Default Pizzashack API is created only. ### Steps to reproduce: 1. Go to Publisher 2. Deploy sample API When checked the LC state of the API in Lifecycles tab, it is displayed as "Created". But Lifecycle history list has an event as "LC has changed from CREATED to PUBLISHED" ![image](https://user-images.githubusercontent.com/2484781/109799041-85890500-7c41-11eb-8c9f-fd80473e11f9.png) This is critical as the default API is used by many and it should behave correctly. ### Affected Product Version: APIM-4.0.0-Alpha ### Environment details (with versions): - OS: - Client: - Env (Docker/K8s): --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
non_defect
revision is displayed as already deployed when the default api is created only description revision is displayed as already deployed when the default pizzashack api is created only steps to reproduce go to publisher deploy sample api when checked the lc state of the api in lifecycles tab it is displayed as created but lifecycle history list has an event as lc has changed from created to published this is critical as the default api is used by many and it should behave correctly affected product version apim alpha environment details with versions os client env docker optional fields related issues suggested labels suggested assignees
0
194,385
14,676,870,610
IssuesEvent
2020-12-30 21:26:45
IntellectualSites/FastAsyncWorldEdit
https://api.github.com/repos/IntellectualSites/FastAsyncWorldEdit
opened
Plotsquared delete issue
Requires Testing
<!-- ⚠️⚠️ Do Not Delete This! You must follow this template. ⚠️⚠️ --> <!--- Incomplete reports will be marked as invalid, and closed, with few exceptions.--> <!--- If you are using 1.14 or 1.15 consider updating to 1.16.3 before raising an issue --> <!--- The priority lays on 1.16 right now, so issues reported for or 1.15 will be fixed for the 1.16 versions --> **[REQUIRED] FastAsyncWorldEdit Configuration Files** <!--- Issue /fawe debugpaste in game or in your console and copy the supplied URL here --> <!--- If you cannot perform the above, we require logs/latest.log; config.yml and config-legacy.yml --> <!--- Please provide this information by using a paste service such as https://haste.athion.net --> <!--- If you are unwilling to supply the information we need, we reserve the right to not assist you. Redact IP addresses if you need to. --> **/fawe debugpaste**:- **Required Information** - FAWE Version Number (`/version FastAsyncWorldEdit`): 492 - Spigot/Paper Version Number (`/version`): Paper-318 - Minecraft Version: 1.16.4 **Describe the bug** PlotSquared plots deleting or clearing takes ages, until using worldedit on the plot. After that every plot gets deleted immeadietly, until the delete takes ages again. [Video](https://1drv.ms/v/s!AovGVxICV9M4gs5M8l9uX7X8TJYoHg?e=wcIW10) **To Reproduce** Steps to reproduce the behavior: 1. Claim a plot with `/plot claim` 2. delete the plot with `/p delete` 3. See if it takes ages, if it does: 4. Use Worldedit to edit a few blocks within the plot (Doesn't need to bee a different block) 5. The plot resets immediately after 6. Deleting any other plots will now go as fast as they should **Plugins being used on the server** <!--- Optional but recommended - issue "/plugins" in-game or in console and copy/paste the list --> **Checklist**: <!--- Make sure you've completed the following steps (put an "X" between of brackets): --> - [X] I included all information required in the sections above - [X] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/FastAsyncWorldEdit/issues?q=is%3Aissue) - [X] I made sure I am using an up-to-date version of [FastAsyncWorldEdit for 1.16.4](https://ci.athion.net/job/FastAsyncWorldEdit-1.16/) - [X] I made sure the bug/error is not caused by any other plugin
1.0
Plotsquared delete issue - <!-- ⚠️⚠️ Do Not Delete This! You must follow this template. ⚠️⚠️ --> <!--- Incomplete reports will be marked as invalid, and closed, with few exceptions.--> <!--- If you are using 1.14 or 1.15 consider updating to 1.16.3 before raising an issue --> <!--- The priority lays on 1.16 right now, so issues reported for or 1.15 will be fixed for the 1.16 versions --> **[REQUIRED] FastAsyncWorldEdit Configuration Files** <!--- Issue /fawe debugpaste in game or in your console and copy the supplied URL here --> <!--- If you cannot perform the above, we require logs/latest.log; config.yml and config-legacy.yml --> <!--- Please provide this information by using a paste service such as https://haste.athion.net --> <!--- If you are unwilling to supply the information we need, we reserve the right to not assist you. Redact IP addresses if you need to. --> **/fawe debugpaste**:- **Required Information** - FAWE Version Number (`/version FastAsyncWorldEdit`): 492 - Spigot/Paper Version Number (`/version`): Paper-318 - Minecraft Version: 1.16.4 **Describe the bug** PlotSquared plots deleting or clearing takes ages, until using worldedit on the plot. After that every plot gets deleted immeadietly, until the delete takes ages again. [Video](https://1drv.ms/v/s!AovGVxICV9M4gs5M8l9uX7X8TJYoHg?e=wcIW10) **To Reproduce** Steps to reproduce the behavior: 1. Claim a plot with `/plot claim` 2. delete the plot with `/p delete` 3. See if it takes ages, if it does: 4. Use Worldedit to edit a few blocks within the plot (Doesn't need to bee a different block) 5. The plot resets immediately after 6. Deleting any other plots will now go as fast as they should **Plugins being used on the server** <!--- Optional but recommended - issue "/plugins" in-game or in console and copy/paste the list --> **Checklist**: <!--- Make sure you've completed the following steps (put an "X" between of brackets): --> - [X] I included all information required in the sections above - [X] I made sure there are no duplicates of this report [(Use Search)](https://github.com/IntellectualSites/FastAsyncWorldEdit/issues?q=is%3Aissue) - [X] I made sure I am using an up-to-date version of [FastAsyncWorldEdit for 1.16.4](https://ci.athion.net/job/FastAsyncWorldEdit-1.16/) - [X] I made sure the bug/error is not caused by any other plugin
non_defect
plotsquared delete issue fastasyncworldedit configuration files fawe debugpaste required information fawe version number version fastasyncworldedit spigot paper version number version paper minecraft version describe the bug plotsquared plots deleting or clearing takes ages until using worldedit on the plot after that every plot gets deleted immeadietly until the delete takes ages again to reproduce steps to reproduce the behavior claim a plot with plot claim delete the plot with p delete see if it takes ages if it does use worldedit to edit a few blocks within the plot doesn t need to bee a different block the plot resets immediately after deleting any other plots will now go as fast as they should plugins being used on the server checklist i included all information required in the sections above i made sure there are no duplicates of this report i made sure i am using an up to date version of i made sure the bug error is not caused by any other plugin
0
30,959
6,377,435,850
IssuesEvent
2017-08-02 10:01:09
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
[security] [config] [dynamic-config] Clarify DynamicSecurityConfig & SecurityService lifecycle
Team: Core Team: Integration Type: Defect
Currently a new instance is created each time `DynamicConfigurationAwareConfig.getSecurityConfig` is invoked
1.0
[security] [config] [dynamic-config] Clarify DynamicSecurityConfig & SecurityService lifecycle - Currently a new instance is created each time `DynamicConfigurationAwareConfig.getSecurityConfig` is invoked
defect
clarify dynamicsecurityconfig securityservice lifecycle currently a new instance is created each time dynamicconfigurationawareconfig getsecurityconfig is invoked
1
646,684
21,056,310,707
IssuesEvent
2022-04-01 03:56:32
oasis-engine/engine
https://api.github.com/repos/oasis-engine/engine
closed
undelete _currentEvents in physics packages
bug Physical high priority
when two collider collide together, the script delete the component, info still exit in _currentEvents which will cause the undefined. To fix this problem, we should consider two things: 1. clear relevant index in _currentEvents after destroy entity. 2. delete all object after one frame(not in the middle of frame).
1.0
undelete _currentEvents in physics packages - when two collider collide together, the script delete the component, info still exit in _currentEvents which will cause the undefined. To fix this problem, we should consider two things: 1. clear relevant index in _currentEvents after destroy entity. 2. delete all object after one frame(not in the middle of frame).
non_defect
undelete currentevents in physics packages when two collider collide together the script delete the component info still exit in currentevents which will cause the undefined to fix this problem we should consider two things clear relevant index in currentevents after destroy entity delete all object after one frame not in the middle of frame
0
816,807
30,613,219,435
IssuesEvent
2023-07-23 21:43:52
SibTiger/PowerShell-Compact-Archive-Tool
https://api.github.com/repos/SibTiger/PowerShell-Compact-Archive-Tool
closed
EmbedInstaller: Remove mentions of BurntToast and POSH Modules
Enhancement Priority: NORMAL
Within the EmbedInstaller, there are mentions of BurntToast and POSH Modules within the script. These will need to be removed as they are no longer necessary.
1.0
EmbedInstaller: Remove mentions of BurntToast and POSH Modules - Within the EmbedInstaller, there are mentions of BurntToast and POSH Modules within the script. These will need to be removed as they are no longer necessary.
non_defect
embedinstaller remove mentions of burnttoast and posh modules within the embedinstaller there are mentions of burnttoast and posh modules within the script these will need to be removed as they are no longer necessary
0
15,810
2,869,076,462
IssuesEvent
2015-06-05 23:08:38
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Polymer tests are failing or flaky after mutation-observer microtask change.
Area-Pkg Pkg-Polymer PolymerMilestone-Next Priority-Medium Triaged Type-Defect
Failing on Safari: polymer/test/js_interop_test Flaky on Safari and IE: polymer/test/two_way_bind_test
1.0
Polymer tests are failing or flaky after mutation-observer microtask change. - Failing on Safari: polymer/test/js_interop_test Flaky on Safari and IE: polymer/test/two_way_bind_test
defect
polymer tests are failing or flaky after mutation observer microtask change failing on safari polymer test js interop test flaky on safari and ie polymer test two way bind test
1
13,131
2,732,908,107
IssuesEvent
2015-04-17 10:09:06
tiku01/oryx-editor
https://api.github.com/repos/tiku01/oryx-editor
closed
characters like <, > etc. are shown in the property window in their HTML encoded form
auto-migrated Component-Editor Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. create a model and select a string property for editing 2. write '<' 3. click on the canvas to end editing mode What is the expected output? I see '<' as the value for the property What do you see instead? I see '&lt;' ``` Original issue reported on code.google.com by `NicoPete...@gmail.com` on 2 Sep 2008 at 11:13
1.0
characters like <, > etc. are shown in the property window in their HTML encoded form - ``` What steps will reproduce the problem? 1. create a model and select a string property for editing 2. write '<' 3. click on the canvas to end editing mode What is the expected output? I see '<' as the value for the property What do you see instead? I see '&lt;' ``` Original issue reported on code.google.com by `NicoPete...@gmail.com` on 2 Sep 2008 at 11:13
defect
characters like etc are shown in the property window in their html encoded form what steps will reproduce the problem create a model and select a string property for editing write click on the canvas to end editing mode what is the expected output i see as the value for the property what do you see instead i see lt original issue reported on code google com by nicopete gmail com on sep at
1
10,369
2,622,148,030
IssuesEvent
2015-03-04 00:04:57
byzhang/i7z
https://api.github.com/repos/byzhang/i7z
opened
i7z GUI mode does not show all cores
auto-migrated Priority-Medium Type-Defect
``` On every 2 or 4 socket Xeon server I tested (HP ProLiant DL380/DL580, Dell PowerEdge R610/R620/ R710/R720/R820/R920, Cisco C240/C420, IBM x3850, Lenovo RD630) the i7z GUI monitor hides some of the cores during each display cycle! The attached screen shot is from a 4-socket 40-core system. It shows 10 cores on socket 0 but only 7 cores on socket 1! Thx, dk ``` Original issue reported on code.google.com by `dan.ko...@gmail.com` on 14 Aug 2013 at 11:36 Attachments: * [i7z on 4-socket Xeon server.JPG](https://storage.googleapis.com/google-code-attachments/i7z/issue-74/comment-0/i7z on 4-socket Xeon server.JPG)
1.0
i7z GUI mode does not show all cores - ``` On every 2 or 4 socket Xeon server I tested (HP ProLiant DL380/DL580, Dell PowerEdge R610/R620/ R710/R720/R820/R920, Cisco C240/C420, IBM x3850, Lenovo RD630) the i7z GUI monitor hides some of the cores during each display cycle! The attached screen shot is from a 4-socket 40-core system. It shows 10 cores on socket 0 but only 7 cores on socket 1! Thx, dk ``` Original issue reported on code.google.com by `dan.ko...@gmail.com` on 14 Aug 2013 at 11:36 Attachments: * [i7z on 4-socket Xeon server.JPG](https://storage.googleapis.com/google-code-attachments/i7z/issue-74/comment-0/i7z on 4-socket Xeon server.JPG)
defect
gui mode does not show all cores on every or socket xeon server i tested hp proliant dell poweredge cisco ibm lenovo the gui monitor hides some of the cores during each display cycle the attached screen shot is from a socket core system it shows cores on socket but only cores on socket thx dk original issue reported on code google com by dan ko gmail com on aug at attachments on socket xeon server jpg
1
59,161
17,016,125,155
IssuesEvent
2021-07-02 12:18:53
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
opened
Wrong street name for house number in AssociatedStreet relation
Component: nominatim Priority: minor Type: defect
**[Submitted to the original trac issue database at 8.16pm, Wednesday, 10th October 2012]** When I search for "Reinier Roeststraat 6, Leusden" I get 2 results. The first result is the one I'm looking for. The second is a different house with house number 6 in the nearby "Kees van Burgstedenstraat" street. Strangely enough the second result is being reported as "Reinier Roeststraat 6" in stead of "Kees van Burgstedenstraat 6", even though it is in an associatedstreet relation with name "Kees van Burgstedenstraat".
1.0
Wrong street name for house number in AssociatedStreet relation - **[Submitted to the original trac issue database at 8.16pm, Wednesday, 10th October 2012]** When I search for "Reinier Roeststraat 6, Leusden" I get 2 results. The first result is the one I'm looking for. The second is a different house with house number 6 in the nearby "Kees van Burgstedenstraat" street. Strangely enough the second result is being reported as "Reinier Roeststraat 6" in stead of "Kees van Burgstedenstraat 6", even though it is in an associatedstreet relation with name "Kees van Burgstedenstraat".
defect
wrong street name for house number in associatedstreet relation when i search for reinier roeststraat leusden i get results the first result is the one i m looking for the second is a different house with house number in the nearby kees van burgstedenstraat street strangely enough the second result is being reported as reinier roeststraat in stead of kees van burgstedenstraat even though it is in an associatedstreet relation with name kees van burgstedenstraat
1
73,888
24,849,899,871
IssuesEvent
2022-10-26 19:05:02
zed-industries/feedback
https://api.github.com/repos/zed-industries/feedback
closed
Bugs with project panel folders
defect performance project panel
### Check for existing issues - [X] Completed ### Describe the bug / provide steps to reproduce it - [ ] Collapsing / Expanding the root directory puts Zed into a weird state where the files disappear - [ ] Collapse / Expanding feel pretty sluggish, like there's some performance issue ### Expected behavior - ### Environment Zed 0.61.1 – /Applications/Zed.app macOS 13.0 architecture x86_64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature _No response_ ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue _No response_
1.0
Bugs with project panel folders - ### Check for existing issues - [X] Completed ### Describe the bug / provide steps to reproduce it - [ ] Collapsing / Expanding the root directory puts Zed into a weird state where the files disappear - [ ] Collapse / Expanding feel pretty sluggish, like there's some performance issue ### Expected behavior - ### Environment Zed 0.61.1 – /Applications/Zed.app macOS 13.0 architecture x86_64 ### If applicable, add mockups / screenshots to help explain present your vision of the feature _No response_ ### If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue _No response_
defect
bugs with project panel folders check for existing issues completed describe the bug provide steps to reproduce it collapsing expanding the root directory puts zed into a weird state where the files disappear collapse expanding feel pretty sluggish like there s some performance issue expected behavior environment zed – applications zed app macos architecture if applicable add mockups screenshots to help explain present your vision of the feature no response if applicable attach your library logs zed zed log file to this issue no response
1
48,026
13,067,401,641
IssuesEvent
2020-07-31 00:20:17
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
[iceprod2] handle expiration of self-signed cert for webserver (Trac #1676)
Migrated from Trac defect iceprod
The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs. Note that just making a new cert isn't good enough, because already running jobs will still have the old cert. Migrated from https://code.icecube.wisc.edu/ticket/1676 ```json { "status": "closed", "changetime": "2016-05-09T21:55:16", "description": "The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs.\n\nNote that just making a new cert isn't good enough, because already running jobs will still have the old cert.", "reporter": "david.schultz", "cc": "", "resolution": "wontfix", "_ts": "1462830916934929", "component": "iceprod", "summary": "[iceprod2] handle expiration of self-signed cert for webserver", "priority": "major", "keywords": "", "time": "2016-04-29T16:41:30", "milestone": "", "owner": "david.schultz", "type": "defect" } ```
1.0
[iceprod2] handle expiration of self-signed cert for webserver (Trac #1676) - The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs. Note that just making a new cert isn't good enough, because already running jobs will still have the old cert. Migrated from https://code.icecube.wisc.edu/ticket/1676 ```json { "status": "closed", "changetime": "2016-05-09T21:55:16", "description": "The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs.\n\nNote that just making a new cert isn't good enough, because already running jobs will still have the old cert.", "reporter": "david.schultz", "cc": "", "resolution": "wontfix", "_ts": "1462830916934929", "component": "iceprod", "summary": "[iceprod2] handle expiration of self-signed cert for webserver", "priority": "major", "keywords": "", "time": "2016-04-29T16:41:30", "milestone": "", "owner": "david.schultz", "type": "defect" } ```
defect
handle expiration of self signed cert for webserver trac the default right now is to generate a self signed cert with a duration of days we could either make this time infinitely large several years or deal with the problem of expiring certs note that just making a new cert isn t good enough because already running jobs will still have the old cert migrated from json status closed changetime description the default right now is to generate a self signed cert with a duration of days we could either make this time infinitely large several years or deal with the problem of expiring certs n nnote that just making a new cert isn t good enough because already running jobs will still have the old cert reporter david schultz cc resolution wontfix ts component iceprod summary handle expiration of self signed cert for webserver priority major keywords time milestone owner david schultz type defect
1
91,025
8,288,383,453
IssuesEvent
2018-09-19 11:50:33
humera987/HumTestData
https://api.github.com/repos/humera987/HumTestData
closed
fx_test_proj : ApiV1SkillsIdGetPathParamNullValueId
fx_test_proj
Project : fx_test_proj Job : UAT Env : UAT Region : FXLabs/US_WEST_1 Result : fail Status Code : 200 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Wed, 19 Sep 2018 06:33:01 GMT]} Endpoint : http://13.56.210.25/api/v1/skills/null Request : Response : { "requestId" : "None", "requestTime" : "2018-09-19T06:33:02.210+0000", "errors" : true, "messages" : [ ], "data" : null, "totalPages" : 0, "totalElements" : 0 } Logs : Assertion [@StatusCode != 404] resolved-to [200 != 404] result [Passed]Assertion [@StatusCode != 500] resolved-to [200 != 500] result [Passed]Assertion [@StatusCode != 401] resolved-to [200 != 401] result [Passed]Assertion [@StatusCode != 200] resolved-to [200 != 200] result [Failed] --- FX Bot ---
1.0
fx_test_proj : ApiV1SkillsIdGetPathParamNullValueId - Project : fx_test_proj Job : UAT Env : UAT Region : FXLabs/US_WEST_1 Result : fail Status Code : 200 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Wed, 19 Sep 2018 06:33:01 GMT]} Endpoint : http://13.56.210.25/api/v1/skills/null Request : Response : { "requestId" : "None", "requestTime" : "2018-09-19T06:33:02.210+0000", "errors" : true, "messages" : [ ], "data" : null, "totalPages" : 0, "totalElements" : 0 } Logs : Assertion [@StatusCode != 404] resolved-to [200 != 404] result [Passed]Assertion [@StatusCode != 500] resolved-to [200 != 500] result [Passed]Assertion [@StatusCode != 401] resolved-to [200 != 401] result [Passed]Assertion [@StatusCode != 200] resolved-to [200 != 200] result [Failed] --- FX Bot ---
non_defect
fx test proj project fx test proj job uat env uat region fxlabs us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options content type transfer encoding date endpoint request response requestid none requesttime errors true messages data null totalpages totalelements logs assertion resolved to result assertion resolved to result assertion resolved to result assertion resolved to result fx bot
0