Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
64,475 | 8,737,471,347 | IssuesEvent | 2018-12-11 22:41:46 | docksal/docksal | https://api.github.com/repos/docksal/docksal | closed | Add documentation for having multiple web containers | size/small 🏷documentation | On gitter, @sean-e-dietrich provided me with this example: https://gist.github.com/mikecrittenden/2cbb4c1e116b405ab16c5087f0e59ac7
This would be a nice thing to put in the docs, for projects that need multiple web containers but need to stay in the same Docksal project (example: headless Drupal backend with a node-based frontend). Thoughts? Or does this already exist somewhere in the docs and I'm missing it?
| 1.0 | Add documentation for having multiple web containers - On gitter, @sean-e-dietrich provided me with this example: https://gist.github.com/mikecrittenden/2cbb4c1e116b405ab16c5087f0e59ac7
This would be a nice thing to put in the docs, for projects that need multiple web containers but need to stay in the same Docksal project (example: headless Drupal backend with a node-based frontend). Thoughts? Or does this already exist somewhere in the docs and I'm missing it?
| non_defect | add documentation for having multiple web containers on gitter sean e dietrich provided me with this example this would be a nice thing to put in the docs for projects that need multiple web containers but need to stay in the same docksal project example headless drupal backend with a node based frontend thoughts or does this already exist somewhere in the docs and i m missing it | 0 |
479,512 | 13,798,091,504 | IssuesEvent | 2020-10-10 00:00:06 | radical-cybertools/radical.entk | https://api.github.com/repos/radical-cybertools/radical.entk | opened | feature request: job_name attribute in resource description | priority:medium topic:api type:feature | See radical-collaboration/hpc-workflows/issues/122: EnTK should forward `job_name` settings to RP. | 1.0 | feature request: job_name attribute in resource description - See radical-collaboration/hpc-workflows/issues/122: EnTK should forward `job_name` settings to RP. | non_defect | feature request job name attribute in resource description see radical collaboration hpc workflows issues entk should forward job name settings to rp | 0 |
566,114 | 16,796,399,936 | IssuesEvent | 2021-06-16 04:37:37 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | FTX widget should not appear when card setting is turned off | OS/Desktop QA/Yes feature/widgets priority/P2 release-notes/exclude | <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
For existing users who have the cards settings on the new tab page turned off so that no cards appear, we should not show the FTX widget to these existing users when they upgrade.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Use a Brave version lower than 1.26.21 and turn off cards. See image below.
2. Upgrade to latest 1.26.22 or higher
3. Notice that the FTX widget will appear as the sole widget on upgrade even though card settings is off.
<img width="411" alt="Screen Shot 2021-05-19 at 11 07 45 AM" src="https://user-images.githubusercontent.com/5951041/118862404-8c173800-b892-11eb-834d-15b0cf7142e1.png">
## Actual result:
FTX widget appears upon upgrade as the sole card in the widget stack for existing users and it comes across as not respecting a setting that was previously set.
## Expected result:
FTX widget should not appear for users who have turned off all cards. If a user has cards enabled, FTX will appear for them as part of the list.
New card tooltip promoting FTX should also appear for existing users with cards turned off.
See: https://github.com/brave/brave-browser/issues/9616
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily.
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave 1.26.22 and higher
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? n/a
- Can you reproduce this issue with the beta channel? yes
- Can you reproduce this issue with the nightly channel? yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Does the issue resolve itself when disabling Brave Rewards? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
| 1.0 | FTX widget should not appear when card setting is turned off - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue.
PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE.
INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED-->
## Description
<!--Provide a brief description of the issue-->
For existing users who have the cards settings on the new tab page turned off so that no cards appear, we should not show the FTX widget to these existing users when they upgrade.
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Use a Brave version lower than 1.26.21 and turn off cards. See image below.
2. Upgrade to latest 1.26.22 or higher
3. Notice that the FTX widget will appear as the sole widget on upgrade even though card settings is off.
<img width="411" alt="Screen Shot 2021-05-19 at 11 07 45 AM" src="https://user-images.githubusercontent.com/5951041/118862404-8c173800-b892-11eb-834d-15b0cf7142e1.png">
## Actual result:
FTX widget appears upon upgrade as the sole card in the widget stack for existing users and it comes across as not respecting a setting that was previously set.
## Expected result:
FTX widget should not appear for users who have turned off all cards. If a user has cards enabled, FTX will appear for them as part of the list.
New card tooltip promoting FTX should also appear for existing users with cards turned off.
See: https://github.com/brave/brave-browser/issues/9616
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily.
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave 1.26.22 and higher
## Version/Channel Information:
<!--Does this issue happen on any other channels? Or is it specific to a certain channel?-->
- Can you reproduce this issue with the current release? n/a
- Can you reproduce this issue with the beta channel? yes
- Can you reproduce this issue with the nightly channel? yes
## Other Additional Information:
- Does the issue resolve itself when disabling Brave Shields? n/a
- Does the issue resolve itself when disabling Brave Rewards? n/a
- Is the issue reproducible on the latest version of Chrome? n/a
## Miscellaneous Information:
<!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
| non_defect | ftx widget should not appear when card setting is turned off have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description for existing users who have the cards settings on the new tab page turned off so that no cards appear we should not show the ftx widget to these existing users when they upgrade steps to reproduce use a brave version lower than and turn off cards see image below upgrade to latest or higher notice that the ftx widget will appear as the sole widget on upgrade even though card settings is off img width alt screen shot at am src actual result ftx widget appears upon upgrade as the sole card in the widget stack for existing users and it comes across as not respecting a setting that was previously set expected result ftx widget should not appear for users who have turned off all cards if a user has cards enabled ftx will appear for them as part of the list new card tooltip promoting ftx should also appear for existing users with cards turned off see reproduces how often easily brave version brave version info brave and higher version channel information can you reproduce this issue with the current release n a can you reproduce this issue with the beta channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields n a does the issue resolve itself when disabling brave rewards n a is the issue reproducible on the latest version of chrome n a miscellaneous information | 0 |
274,591 | 23,851,973,165 | IssuesEvent | 2022-09-06 18:50:56 | lowRISC/opentitan | https://api.github.com/repos/lowRISC/opentitan | opened | [chip-test] Add connectivity tests for AST signals | IP:ast Component:ChipLevelTest | ### Test point name
TBD
### Host side component
_No response_
### OpenTitanTool infrastructure implemented
_No response_
### Contact person
@tjaychen @msf
### Checklist
Please fill out this checklist as items are completed. Link to PRs and issues as appropriate.
- [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes)
- [ ] Device-side (C) component developed
- [ ] Bazel build rules developed
- [ ] Host-side component developed
- [ ] HJSON test plan updated with test name (so it shows up in the dashboard)
- [ ] Test added to dvsim nightly regression (and passing at time of checking)
| 1.0 | [chip-test] Add connectivity tests for AST signals - ### Test point name
TBD
### Host side component
_No response_
### OpenTitanTool infrastructure implemented
_No response_
### Contact person
@tjaychen @msf
### Checklist
Please fill out this checklist as items are completed. Link to PRs and issues as appropriate.
- [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes)
- [ ] Device-side (C) component developed
- [ ] Bazel build rules developed
- [ ] Host-side component developed
- [ ] HJSON test plan updated with test name (so it shows up in the dashboard)
- [ ] Test added to dvsim nightly regression (and passing at time of checking)
| non_defect | add connectivity tests for ast signals test point name tbd host side component no response opentitantool infrastructure implemented no response contact person tjaychen msf checklist please fill out this checklist as items are completed link to prs and issues as appropriate check if existing test covers most or all of this testpoint if so either extend said test to cover all points or skip the next checkboxes device side c component developed bazel build rules developed host side component developed hjson test plan updated with test name so it shows up in the dashboard test added to dvsim nightly regression and passing at time of checking | 0 |
49,876 | 26,383,507,759 | IssuesEvent | 2023-01-12 10:27:04 | PandemiaProject/pandemia | https://api.github.com/repos/PandemiaProject/pandemia | closed | Performance - Review interface between C/Python/NumPy | performance | All params passed to C as 1D NP arrays. Requires manual calculation of array indices.
Would there be a better way to do this? | True | Performance - Review interface between C/Python/NumPy - All params passed to C as 1D NP arrays. Requires manual calculation of array indices.
Would there be a better way to do this? | non_defect | performance review interface between c python numpy all params passed to c as np arrays requires manual calculation of array indices would there be a better way to do this | 0 |
218,877 | 7,332,755,720 | IssuesEvent | 2018-03-05 17:12:17 | NCEAS/metacat | https://api.github.com/repos/NCEAS/metacat | closed | Make reCaptcha work with AJAX include | Priority: Normal Status: Closed Tracker: Task | ---
Author Name: **ben leinfelder** (ben leinfelder)
Original Redmine Issue: 6098, https://projects.ecoinformatics.org/ecoinfo/issues/6098
Original Date: 2013-09-20
Original Assignee: Jing Tao
---
Instead of having the perl code generate the catcha content, the template can include the ajax-style of constructing the captcha. This will allow us to use the same method when we include the registration form in the MetacatUI as well.
See:
https://developers.google.com/recaptcha/docs/display?csw=1#AJAX
http://stackoverflow.com/questions/7261436/how-can-i-load-a-recaptcha-form-using-jquery-ajax-while-leaving-the-recaptcha-sc
| 1.0 | Make reCaptcha work with AJAX include - ---
Author Name: **ben leinfelder** (ben leinfelder)
Original Redmine Issue: 6098, https://projects.ecoinformatics.org/ecoinfo/issues/6098
Original Date: 2013-09-20
Original Assignee: Jing Tao
---
Instead of having the perl code generate the catcha content, the template can include the ajax-style of constructing the captcha. This will allow us to use the same method when we include the registration form in the MetacatUI as well.
See:
https://developers.google.com/recaptcha/docs/display?csw=1#AJAX
http://stackoverflow.com/questions/7261436/how-can-i-load-a-recaptcha-form-using-jquery-ajax-while-leaving-the-recaptcha-sc
| non_defect | make recaptcha work with ajax include author name ben leinfelder ben leinfelder original redmine issue original date original assignee jing tao instead of having the perl code generate the catcha content the template can include the ajax style of constructing the captcha this will allow us to use the same method when we include the registration form in the metacatui as well see | 0 |
169,539 | 20,841,777,630 | IssuesEvent | 2022-03-21 01:30:38 | uniquelyparticular/serverless-oauth | https://api.github.com/repos/uniquelyparticular/serverless-oauth | opened | CVE-2022-24772 (High) detected in node-forge-0.8.4.tgz, node-forge-0.7.4.tgz | security vulnerability | ## CVE-2022-24772 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-forge-0.8.4.tgz</b>, <b>node-forge-0.7.4.tgz</b></p></summary>
<p>
<details><summary><b>node-forge-0.8.4.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.8.4.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.8.4.tgz</a></p>
<p>Path to dependency file: /generic-oauth/package.json</p>
<p>Path to vulnerable library: /tmp/git/generic-oauth/node_modules/gcs-resumable-upload/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- firebase-admin-8.0.0.tgz (Root Library)
- storage-2.5.0.tgz
- gcs-resumable-upload-1.1.0.tgz
- google-auth-library-3.1.2.tgz
- gtoken-2.3.3.tgz
- google-p12-pem-1.0.4.tgz
- :x: **node-forge-0.8.4.tgz** (Vulnerable Library)
</details>
<details><summary><b>node-forge-0.7.4.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.4.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.4.tgz</a></p>
<p>Path to dependency file: /generic-oauth/package.json</p>
<p>Path to vulnerable library: /tmp/git/generic-oauth/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- firebase-admin-8.0.0.tgz (Root Library)
- :x: **node-forge-0.7.4.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not check for tailing garbage bytes after decoding a `DigestInfo` ASN.1 structure. This can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24772>CVE-2022-24772</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution: node-forge - 1.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-24772 (High) detected in node-forge-0.8.4.tgz, node-forge-0.7.4.tgz - ## CVE-2022-24772 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-forge-0.8.4.tgz</b>, <b>node-forge-0.7.4.tgz</b></p></summary>
<p>
<details><summary><b>node-forge-0.8.4.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.8.4.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.8.4.tgz</a></p>
<p>Path to dependency file: /generic-oauth/package.json</p>
<p>Path to vulnerable library: /tmp/git/generic-oauth/node_modules/gcs-resumable-upload/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- firebase-admin-8.0.0.tgz (Root Library)
- storage-2.5.0.tgz
- gcs-resumable-upload-1.1.0.tgz
- google-auth-library-3.1.2.tgz
- gtoken-2.3.3.tgz
- google-p12-pem-1.0.4.tgz
- :x: **node-forge-0.8.4.tgz** (Vulnerable Library)
</details>
<details><summary><b>node-forge-0.7.4.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.4.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.4.tgz</a></p>
<p>Path to dependency file: /generic-oauth/package.json</p>
<p>Path to vulnerable library: /tmp/git/generic-oauth/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- firebase-admin-8.0.0.tgz (Root Library)
- :x: **node-forge-0.7.4.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not check for tailing garbage bytes after decoding a `DigestInfo` ASN.1 structure. This can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24772>CVE-2022-24772</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution: node-forge - 1.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve high detected in node forge tgz node forge tgz cve high severity vulnerability vulnerable libraries node forge tgz node forge tgz node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file generic oauth package json path to vulnerable library tmp git generic oauth node modules gcs resumable upload node modules node forge package json dependency hierarchy firebase admin tgz root library storage tgz gcs resumable upload tgz google auth library tgz gtoken tgz google pem tgz x node forge tgz vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file generic oauth package json path to vulnerable library tmp git generic oauth node modules node forge package json dependency hierarchy firebase admin tgz root library x node forge tgz vulnerable library found in base branch master vulnerability details forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code does not check for tailing garbage bytes after decoding a digestinfo asn structure this can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used the issue has been addressed in node forge version there are currently no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge step up your open source security game with whitesource | 0 |
37,247 | 8,308,009,973 | IssuesEvent | 2018-09-23 16:04:00 | umple/umple | https://api.github.com/repos/umple/umple | closed | Constraints with == converted to equals() in Java, but failing on primitives | Component-SemanticsAndGen Diffic-Easy Priority-High Status-Accepted Type-Defect constraints stateMachines ucosp | Originally reported on Google Code with ID 589
Owned by `redeyedmars`
---
In the SecurityLight example (umpleonline/ump/SecurityLight the generated Java has
a bug where the original constraint has == to compare two booleans, but it has been
changed to generate a call to equals(). But this doesn't work with Booleans
http://cruise.eecs.uottawa.ca/umpleonline/?example=SecurityLight
---
Reported by @umple on 2014-06-09 20:56:58
| 1.0 | Constraints with == converted to equals() in Java, but failing on primitives - Originally reported on Google Code with ID 589
Owned by `redeyedmars`
---
In the SecurityLight example (umpleonline/ump/SecurityLight the generated Java has
a bug where the original constraint has == to compare two booleans, but it has been
changed to generate a call to equals(). But this doesn't work with Booleans
http://cruise.eecs.uottawa.ca/umpleonline/?example=SecurityLight
---
Reported by @umple on 2014-06-09 20:56:58
| defect | constraints with converted to equals in java but failing on primitives originally reported on google code with id owned by redeyedmars in the securitylight example umpleonline ump securitylight the generated java has a bug where the original constraint has to compare two booleans but it has been changed to generate a call to equals but this doesn t work with booleans reported by umple on | 1 |
70,438 | 23,167,875,234 | IssuesEvent | 2022-07-30 08:18:25 | dkfans/keeperfx | https://api.github.com/repos/dkfans/keeperfx | closed | Digging sound does not play in possession | Type-Defect | Instead, it plays the 'shot hit wall' sound, which sounds similar but is different.
When imps dig on their own, they make 6 random dig sounds, hardcoded in the 'instf_dig' function. It's 63+6.
When possessing an imp and digging there, it fails to use that function, and it does not create the sounds or visual effects to go with that. What the player does hear, is the 'shot had generic impact' sound, listed as 'HitWallSound' at '[shot23]' in 'magic.cfg'. It's 128+3. It also fails to produce the visual effects like rock chips.
The correct behavior would be for the possession sounds to match the regular sounds. So 63+3 when digging something diggable, but 128+3 when hitting something that cannot be dug. With the visual effects to match too.
The current behavior has always been this way. | 1.0 | Digging sound does not play in possession - Instead, it plays the 'shot hit wall' sound, which sounds similar but is different.
When imps dig on their own, they make 6 random dig sounds, hardcoded in the 'instf_dig' function. It's 63+6.
When possessing an imp and digging there, it fails to use that function, and it does not create the sounds or visual effects to go with that. What the player does hear, is the 'shot had generic impact' sound, listed as 'HitWallSound' at '[shot23]' in 'magic.cfg'. It's 128+3. It also fails to produce the visual effects like rock chips.
The correct behavior would be for the possession sounds to match the regular sounds. So 63+3 when digging something diggable, but 128+3 when hitting something that cannot be dug. With the visual effects to match too.
The current behavior has always been this way. | defect | digging sound does not play in possession instead it plays the shot hit wall sound which sounds similar but is different when imps dig on their own they make random dig sounds hardcoded in the instf dig function it s when possessing an imp and digging there it fails to use that function and it does not create the sounds or visual effects to go with that what the player does hear is the shot had generic impact sound listed as hitwallsound at in magic cfg it s it also fails to produce the visual effects like rock chips the correct behavior would be for the possession sounds to match the regular sounds so when digging something diggable but when hitting something that cannot be dug with the visual effects to match too the current behavior has always been this way | 1 |
425,533 | 12,341,886,159 | IssuesEvent | 2020-05-14 23:05:05 | earthlab/matplotcheck | https://api.github.com/repos/earthlab/matplotcheck | closed | Timeseries data check failing - coersion issue | bug high priority time-series |
```
# Get data
data = et.data.get_data("colorado-flood")
os.chdir(os.path.join(et.io.HOME, 'earth-analytics'))
import matplotcheck.notebook as nb
import matplotcheck.timeseries as ts
import matplotcheck.raster as ra
# BEGIN SOLUTION
import os
import matplotlib.pyplot as plt
import matplotlib.dates as mdates
from matplotlib.dates import DateFormatter
import seaborn as sns
import numpy as np
import pandas as pd
import earthpy as et
import hydrofunctions as hf
import urllib
from pandas.plotting import register_matplotlib_converters
register_matplotlib_converters()
# prettier plotting with seaborn
sns.set(font_scale=1.5)
sns.set_style("whitegrid")
# import file
f = "data/colorado-flood/discharge/06730200-discharge-daily-1986-2013.txt"
discharge = pd.read_csv(f,
skiprows=23,
header=[1, 2],
sep='\t',
parse_dates=[2])
# drop one level of index
discharge.columns = discharge.columns.droplevel(1)
# set the date column as the index
discharge = discharge.set_index(["datetime"])
monthly_max_all = discharge.resample("M").max()
monthly_max = monthly_max_all['1990':'2014']
fig, ax = plt.subplots(figsize=(10, 8))
ax.scatter(x=monthly_max.index,
y=monthly_max["17663_00060_00003"],
color="purple")
ax.set_title(
"HW Plot 1: Stream Discharge - Monthly Max Value\n be sure to add x and y labels (not shown here)")
ax.set(xlabel="Date")
# END SOLUTION
### DO NOT REMOVE LINE BELOW ###
plot_1_ts = nb.convert_axes(plt, which_axes="current")
mpc_plot_1_ts = mts.TimeSeriesTester(plot_1_ts)
results = []
```
TEST
```
mts.TimeSeriesTester(plot_1_ts)
mpc_plot_1_ts.assert_xydata(xy_expected = precip.reset_index(),
xtime=True,
xcol="DATE",
ycol="HPCP")
```
Output error: i suspect this has something to do with NAN values but i am not sure
```
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
<ipython-input-38-1b4d8c2d48dc> in <module>
4 xtime=True,
5 xcol="DATE",
----> 6 ycol="HPCP")
7
8 #precip
~/Documents/github/0-python/matplotcheck/matplotcheck/base.py in assert_xydata(self, xy_expected, xcol, ycol, points_only, xtime, xlabels, tolerence, message)
938 try:
939 np.testing.assert_array_max_ulp(
--> 940 np.array(xy_data["x"]), np.array(xy_expected[xcol])
941 )
942 except AssertionError:
~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/testing/_private/utils.py in assert_array_max_ulp(a, b, maxulp, dtype)
1617 __tracebackhide__ = True # Hide traceback for py.test
1618 import numpy as np
-> 1619 ret = nulp_diff(a, b, dtype)
1620 if not np.all(ret <= maxulp):
1621 raise AssertionError("Arrays are not almost equal up to %g ULP" %
~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/testing/_private/utils.py in nulp_diff(x, y, dtype)
1658 y = np.array(y)
1659
-> 1660 t = np.common_type(x, y)
1661 if np.iscomplexobj(x) or np.iscomplexobj(y):
1662 raise NotImplementedError("_nulp not implemented for complex array")
<__array_function__ internals> in common_type(*args, **kwargs)
~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/lib/type_check.py in common_type(*arrays)
719 p = array_precision.get(t, None)
720 if p is None:
--> 721 raise TypeError("can't get common type for non-numeric array")
722 precision = max(precision, p)
723 if is_complex:
TypeError: can't get common type for non-numeric array
```
1. i suspect this has to do with NA values. we need to clear those out
| 1.0 | Timeseries data check failing - coersion issue -
```
# Get data
data = et.data.get_data("colorado-flood")
os.chdir(os.path.join(et.io.HOME, 'earth-analytics'))
import matplotcheck.notebook as nb
import matplotcheck.timeseries as ts
import matplotcheck.raster as ra
# BEGIN SOLUTION
import os
import matplotlib.pyplot as plt
import matplotlib.dates as mdates
from matplotlib.dates import DateFormatter
import seaborn as sns
import numpy as np
import pandas as pd
import earthpy as et
import hydrofunctions as hf
import urllib
from pandas.plotting import register_matplotlib_converters
register_matplotlib_converters()
# prettier plotting with seaborn
sns.set(font_scale=1.5)
sns.set_style("whitegrid")
# import file
f = "data/colorado-flood/discharge/06730200-discharge-daily-1986-2013.txt"
discharge = pd.read_csv(f,
skiprows=23,
header=[1, 2],
sep='\t',
parse_dates=[2])
# drop one level of index
discharge.columns = discharge.columns.droplevel(1)
# set the date column as the index
discharge = discharge.set_index(["datetime"])
monthly_max_all = discharge.resample("M").max()
monthly_max = monthly_max_all['1990':'2014']
fig, ax = plt.subplots(figsize=(10, 8))
ax.scatter(x=monthly_max.index,
y=monthly_max["17663_00060_00003"],
color="purple")
ax.set_title(
"HW Plot 1: Stream Discharge - Monthly Max Value\n be sure to add x and y labels (not shown here)")
ax.set(xlabel="Date")
# END SOLUTION
### DO NOT REMOVE LINE BELOW ###
plot_1_ts = nb.convert_axes(plt, which_axes="current")
mpc_plot_1_ts = mts.TimeSeriesTester(plot_1_ts)
results = []
```
TEST
```
mts.TimeSeriesTester(plot_1_ts)
mpc_plot_1_ts.assert_xydata(xy_expected = precip.reset_index(),
xtime=True,
xcol="DATE",
ycol="HPCP")
```
Output error: i suspect this has something to do with NAN values but i am not sure
```
---------------------------------------------------------------------------
TypeError Traceback (most recent call last)
<ipython-input-38-1b4d8c2d48dc> in <module>
4 xtime=True,
5 xcol="DATE",
----> 6 ycol="HPCP")
7
8 #precip
~/Documents/github/0-python/matplotcheck/matplotcheck/base.py in assert_xydata(self, xy_expected, xcol, ycol, points_only, xtime, xlabels, tolerence, message)
938 try:
939 np.testing.assert_array_max_ulp(
--> 940 np.array(xy_data["x"]), np.array(xy_expected[xcol])
941 )
942 except AssertionError:
~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/testing/_private/utils.py in assert_array_max_ulp(a, b, maxulp, dtype)
1617 __tracebackhide__ = True # Hide traceback for py.test
1618 import numpy as np
-> 1619 ret = nulp_diff(a, b, dtype)
1620 if not np.all(ret <= maxulp):
1621 raise AssertionError("Arrays are not almost equal up to %g ULP" %
~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/testing/_private/utils.py in nulp_diff(x, y, dtype)
1658 y = np.array(y)
1659
-> 1660 t = np.common_type(x, y)
1661 if np.iscomplexobj(x) or np.iscomplexobj(y):
1662 raise NotImplementedError("_nulp not implemented for complex array")
<__array_function__ internals> in common_type(*args, **kwargs)
~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/lib/type_check.py in common_type(*arrays)
719 p = array_precision.get(t, None)
720 if p is None:
--> 721 raise TypeError("can't get common type for non-numeric array")
722 precision = max(precision, p)
723 if is_complex:
TypeError: can't get common type for non-numeric array
```
1. i suspect this has to do with NA values. we need to clear those out
| non_defect | timeseries data check failing coersion issue get data data et data get data colorado flood os chdir os path join et io home earth analytics import matplotcheck notebook as nb import matplotcheck timeseries as ts import matplotcheck raster as ra begin solution import os import matplotlib pyplot as plt import matplotlib dates as mdates from matplotlib dates import dateformatter import seaborn as sns import numpy as np import pandas as pd import earthpy as et import hydrofunctions as hf import urllib from pandas plotting import register matplotlib converters register matplotlib converters prettier plotting with seaborn sns set font scale sns set style whitegrid import file f data colorado flood discharge discharge daily txt discharge pd read csv f skiprows header sep t parse dates drop one level of index discharge columns discharge columns droplevel set the date column as the index discharge discharge set index monthly max all discharge resample m max monthly max monthly max all fig ax plt subplots figsize ax scatter x monthly max index y monthly max color purple ax set title hw plot stream discharge monthly max value n be sure to add x and y labels not shown here ax set xlabel date end solution do not remove line below plot ts nb convert axes plt which axes current mpc plot ts mts timeseriestester plot ts results test mts timeseriestester plot ts mpc plot ts assert xydata xy expected precip reset index xtime true xcol date ycol hpcp output error i suspect this has something to do with nan values but i am not sure typeerror traceback most recent call last in xtime true xcol date ycol hpcp precip documents github python matplotcheck matplotcheck base py in assert xydata self xy expected xcol ycol points only xtime xlabels tolerence message try np testing assert array max ulp np array xy data np array xy expected except assertionerror envs earth analytics python lib site packages numpy testing private utils py in assert array max ulp a b maxulp dtype tracebackhide true hide traceback for py test import numpy as np ret nulp diff a b dtype if not np all ret maxulp raise assertionerror arrays are not almost equal up to g ulp envs earth analytics python lib site packages numpy testing private utils py in nulp diff x y dtype y np array y t np common type x y if np iscomplexobj x or np iscomplexobj y raise notimplementederror nulp not implemented for complex array in common type args kwargs envs earth analytics python lib site packages numpy lib type check py in common type arrays p array precision get t none if p is none raise typeerror can t get common type for non numeric array precision max precision p if is complex typeerror can t get common type for non numeric array i suspect this has to do with na values we need to clear those out | 0 |
29,303 | 5,639,213,617 | IssuesEvent | 2017-04-06 13:53:08 | BOINC/boinc | https://api.github.com/repos/BOINC/boinc | closed | Task window in Simple View GUI of manager has scale / placment / content issues | C: Manager P: Undetermined T: Defect | **Reported by verduin on 21 Jan 38899053 19:06 UTC**
[[BR]]Placement and scale are perhaps the cause of missing border (see ticket #119) in the "Simple View" and also cause overwriting a portion of the task selection scroll above the window. In addition placement of slide show images inside the window leaves void background below the images.
For the purposes of task selection, consider truncating task text to limit the length of the selector button above the window. I submit this in the belief that a series of shorter buttons is more useful than a single long button in the display. I believe the button text is essentially redundant with the text following "Application:" inside the window?
Further, there is no text inside the window to give specific identity to the task displayed. In the "Advanced View" this is the "Name" field. I find the tasks selected for simple view are limited to active tasks, but exclude finished and waiting to start tasks. The selection is good but I admit confusion when finding more tasks in the Advanced View than were displayable in the Simple View. You might consider this a defect in user training, but I request consideration for a window title?
This version is current today from Fedora 10 Preview repositories.
Migrated-From: http://boinc.berkeley.edu/trac/ticket/779
| 1.0 | Task window in Simple View GUI of manager has scale / placment / content issues - **Reported by verduin on 21 Jan 38899053 19:06 UTC**
[[BR]]Placement and scale are perhaps the cause of missing border (see ticket #119) in the "Simple View" and also cause overwriting a portion of the task selection scroll above the window. In addition placement of slide show images inside the window leaves void background below the images.
For the purposes of task selection, consider truncating task text to limit the length of the selector button above the window. I submit this in the belief that a series of shorter buttons is more useful than a single long button in the display. I believe the button text is essentially redundant with the text following "Application:" inside the window?
Further, there is no text inside the window to give specific identity to the task displayed. In the "Advanced View" this is the "Name" field. I find the tasks selected for simple view are limited to active tasks, but exclude finished and waiting to start tasks. The selection is good but I admit confusion when finding more tasks in the Advanced View than were displayable in the Simple View. You might consider this a defect in user training, but I request consideration for a window title?
This version is current today from Fedora 10 Preview repositories.
Migrated-From: http://boinc.berkeley.edu/trac/ticket/779
| defect | task window in simple view gui of manager has scale placment content issues reported by verduin on jan utc placement and scale are perhaps the cause of missing border see ticket in the simple view and also cause overwriting a portion of the task selection scroll above the window in addition placement of slide show images inside the window leaves void background below the images for the purposes of task selection consider truncating task text to limit the length of the selector button above the window i submit this in the belief that a series of shorter buttons is more useful than a single long button in the display i believe the button text is essentially redundant with the text following application inside the window further there is no text inside the window to give specific identity to the task displayed in the advanced view this is the name field i find the tasks selected for simple view are limited to active tasks but exclude finished and waiting to start tasks the selection is good but i admit confusion when finding more tasks in the advanced view than were displayable in the simple view you might consider this a defect in user training but i request consideration for a window title this version is current today from fedora preview repositories migrated from | 1 |
54,010 | 13,240,467,386 | IssuesEvent | 2020-08-19 06:24:05 | airdcpp-web/airdcpp-webclient | https://api.github.com/repos/airdcpp-web/airdcpp-webclient | closed | Bad size in the "download - browse" pop up | bug external portable builds | ### Current UI and client versions:
> Application version: AirDC++w 2.8.0 x86_64
Web UI version: 2.8.0
Web UI build date: July 11, 2020 1:08 PM
### Operating system:
Debian 10 I guess
### Steps to reproduce the issue:
- Have a disk with white-space in name, do a download on it using the download browse
- Do the action again but this time choose "previous" and before clicking see the size of the HDD (/run/media/me/Name with whitespace/ (7.82 GiB free)
- here the "free" size is bad | 1.0 | Bad size in the "download - browse" pop up - ### Current UI and client versions:
> Application version: AirDC++w 2.8.0 x86_64
Web UI version: 2.8.0
Web UI build date: July 11, 2020 1:08 PM
### Operating system:
Debian 10 I guess
### Steps to reproduce the issue:
- Have a disk with white-space in name, do a download on it using the download browse
- Do the action again but this time choose "previous" and before clicking see the size of the HDD (/run/media/me/Name with whitespace/ (7.82 GiB free)
- here the "free" size is bad | non_defect | bad size in the download browse pop up current ui and client versions application version airdc w web ui version web ui build date july pm operating system debian i guess steps to reproduce the issue have a disk with white space in name do a download on it using the download browse do the action again but this time choose previous and before clicking see the size of the hdd run media me name with whitespace gib free here the free size is bad | 0 |
152,534 | 13,459,699,729 | IssuesEvent | 2020-09-09 12:38:51 | zhorton34/vuejs-form | https://api.github.com/repos/zhorton34/vuejs-form | closed | Trying to extend custom rule, leads to error | bug documentation | First of all, great package, thanks a lot!
I try to extend a custom rule, but I always get this error:
~~~
Uncaught TypeError: check.rule is not a function
at eval (validator.js:299)
at Array.reduce (<anonymous>)
at Validator.checkRulesAndFillErrorBag (validator.js:298)
at Validator.validate (validator.js:281)
at Proxy.validate (use.js:52)
at extend.html:22 Even if i try the
~~~
Even if I use this example snippet from the documentation:
~~~html
<!DOCTYPE html>
<html lang="en">
<head><script src='https://unpkg.com/vuejs-form@latest/build/vuejs-form.min.js'></script></head>
<body>
<script>
let example = form({ name: 'timmy' }).rules({ name: 'uppercase' });
example.validator().extend('uppercase', [
':attribute must be uppercase',
({ value, validator, parameters }) => value === value.toUpperCase(),
]);
example.validate().errors().has('name');
</script>
</body>
</html>
~~~
The error occurs in all my browsers (Firefox 80 and Chrome 85 on Ubuntu) | 1.0 | Trying to extend custom rule, leads to error - First of all, great package, thanks a lot!
I try to extend a custom rule, but I always get this error:
~~~
Uncaught TypeError: check.rule is not a function
at eval (validator.js:299)
at Array.reduce (<anonymous>)
at Validator.checkRulesAndFillErrorBag (validator.js:298)
at Validator.validate (validator.js:281)
at Proxy.validate (use.js:52)
at extend.html:22 Even if i try the
~~~
Even if I use this example snippet from the documentation:
~~~html
<!DOCTYPE html>
<html lang="en">
<head><script src='https://unpkg.com/vuejs-form@latest/build/vuejs-form.min.js'></script></head>
<body>
<script>
let example = form({ name: 'timmy' }).rules({ name: 'uppercase' });
example.validator().extend('uppercase', [
':attribute must be uppercase',
({ value, validator, parameters }) => value === value.toUpperCase(),
]);
example.validate().errors().has('name');
</script>
</body>
</html>
~~~
The error occurs in all my browsers (Firefox 80 and Chrome 85 on Ubuntu) | non_defect | trying to extend custom rule leads to error first of all great package thanks a lot i try to extend a custom rule but i always get this error uncaught typeerror check rule is not a function at eval validator js at array reduce at validator checkrulesandfillerrorbag validator js at validator validate validator js at proxy validate use js at extend html even if i try the even if i use this example snippet from the documentation html script src let example form name timmy rules name uppercase example validator extend uppercase attribute must be uppercase value validator parameters value value touppercase example validate errors has name the error occurs in all my browsers firefox and chrome on ubuntu | 0 |
44,896 | 12,422,227,930 | IssuesEvent | 2020-05-23 20:56:41 | numpy/numpy.org | https://api.github.com/repos/numpy/numpy.org | closed | Error when running command in Shell | defect | I'm getting an error when I run `rng = np.random.default_rng()` in the shell. | 1.0 | Error when running command in Shell - I'm getting an error when I run `rng = np.random.default_rng()` in the shell. | defect | error when running command in shell i m getting an error when i run rng np random default rng in the shell | 1 |
52,295 | 13,219,033,282 | IssuesEvent | 2020-08-17 09:45:58 | combatopera/lagoon | https://api.github.com/repos/combatopera/lagoon | opened | strategy for evolving Program api vis-a-vis subcommands | defect | currently adding a method will break code that expects that name to be a subcommand | 1.0 | strategy for evolving Program api vis-a-vis subcommands - currently adding a method will break code that expects that name to be a subcommand | defect | strategy for evolving program api vis a vis subcommands currently adding a method will break code that expects that name to be a subcommand | 1 |
165,546 | 6,278,016,753 | IssuesEvent | 2017-07-18 13:34:47 | openshift/origin | https://api.github.com/repos/openshift/origin | opened | add test to check configapi defaulting | kind/post-rebase priority/P1 | add tests to specifically test the configapi defaulting. Probably need fuzzers here too. | 1.0 | add test to check configapi defaulting - add tests to specifically test the configapi defaulting. Probably need fuzzers here too. | non_defect | add test to check configapi defaulting add tests to specifically test the configapi defaulting probably need fuzzers here too | 0 |
78,240 | 27,387,822,884 | IssuesEvent | 2023-02-28 14:27:24 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | performace regression after upgrade from 3.14.16 to 3.17.8 | T: Defect | ### Expected behavior
No performance regressions
no more garbage
### Actual behavior
After upgrading the jooq dependecy
I noticed a performance hit to my whole application
The G1 Survivor Space & Gc Pauses ~+- doubled
which overall resulted in app response durations
The constant the changes was jooq version 3.14.16 -> org.jooq:jooq:3.17.8
Other settings remained the same, that is why i am reaching out.
Asking what might have changed in the jooq internals.
Overall i am only using jooq `dsl` + generated (java) `Record` classes for mapping
### Steps to reproduce the problem
I am happy to help with a debug session
### jOOQ Version
3.17.8
### Database product and version
PostgreSQL 13.7 on gcloud
### Java Version
azul/zulu-openjdk:17-jre
### OS Version
_No response_
### JDBC driver name and version (include name if unofficial driver)
org.postgresql:postgresql:42.5.4 | 1.0 | performace regression after upgrade from 3.14.16 to 3.17.8 - ### Expected behavior
No performance regressions
no more garbage
### Actual behavior
After upgrading the jooq dependecy
I noticed a performance hit to my whole application
The G1 Survivor Space & Gc Pauses ~+- doubled
which overall resulted in app response durations
The constant the changes was jooq version 3.14.16 -> org.jooq:jooq:3.17.8
Other settings remained the same, that is why i am reaching out.
Asking what might have changed in the jooq internals.
Overall i am only using jooq `dsl` + generated (java) `Record` classes for mapping
### Steps to reproduce the problem
I am happy to help with a debug session
### jOOQ Version
3.17.8
### Database product and version
PostgreSQL 13.7 on gcloud
### Java Version
azul/zulu-openjdk:17-jre
### OS Version
_No response_
### JDBC driver name and version (include name if unofficial driver)
org.postgresql:postgresql:42.5.4 | defect | performace regression after upgrade from to expected behavior no performance regressions no more garbage actual behavior after upgrading the jooq dependecy i noticed a performance hit to my whole application the survivor space gc pauses doubled which overall resulted in app response durations the constant the changes was jooq version org jooq jooq other settings remained the same that is why i am reaching out asking what might have changed in the jooq internals overall i am only using jooq dsl generated java record classes for mapping steps to reproduce the problem i am happy to help with a debug session jooq version database product and version postgresql on gcloud java version azul zulu openjdk jre os version no response jdbc driver name and version include name if unofficial driver org postgresql postgresql | 1 |
559,285 | 16,554,318,391 | IssuesEvent | 2021-05-28 12:17:20 | FAIRsharing/fairsharing.github.io | https://api.github.com/repos/FAIRsharing/fairsharing.github.io | closed | Terms of Use redecorate | Normal priority | Terms of Use contents style need to be justified to be the same as other pages.
| 1.0 | Terms of Use redecorate - Terms of Use contents style need to be justified to be the same as other pages.
| non_defect | terms of use redecorate terms of use contents style need to be justified to be the same as other pages | 0 |
69,274 | 22,309,796,954 | IssuesEvent | 2022-06-13 15:56:21 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | closed | [🐛 Bug]: Same error with every new chrome version when using bidi_connection() | C-py I-defect | ### What happened?
driver.bidi_connection() fails every time when new chrome version is released. I have opened this defect in the past.
The line that is causing it is located under webdriver.py
devtools = import_module("selenium.webdriver.common.devtools.v{}".format(version))
Devtools module don`t seem to be updating fast enough. It is hard to use bidi_connection in a professional setting.
Currently it fails for chrome version 102
### How can we reproduce the issue?
```shell
Update chrome and webdriver to version 102.
Run:
async with driver.bidi_connection() as bidi_session:
session, devtools = bidi_session.session, bidi_session.devtools
```
### Relevant log output
```shell
ModuleNotFoundError
```
### Operating System
windows and mac os
### Selenium version
python
### What are the browser(s) and version(s) where you see this issue?
chrome 102
### What are the browser driver(s) and version(s) where you see this issue?
102.0.5005
### Are you using Selenium Grid?
no | 1.0 | [🐛 Bug]: Same error with every new chrome version when using bidi_connection() - ### What happened?
driver.bidi_connection() fails every time when new chrome version is released. I have opened this defect in the past.
The line that is causing it is located under webdriver.py
devtools = import_module("selenium.webdriver.common.devtools.v{}".format(version))
Devtools module don`t seem to be updating fast enough. It is hard to use bidi_connection in a professional setting.
Currently it fails for chrome version 102
### How can we reproduce the issue?
```shell
Update chrome and webdriver to version 102.
Run:
async with driver.bidi_connection() as bidi_session:
session, devtools = bidi_session.session, bidi_session.devtools
```
### Relevant log output
```shell
ModuleNotFoundError
```
### Operating System
windows and mac os
### Selenium version
python
### What are the browser(s) and version(s) where you see this issue?
chrome 102
### What are the browser driver(s) and version(s) where you see this issue?
102.0.5005
### Are you using Selenium Grid?
no | defect | same error with every new chrome version when using bidi connection what happened driver bidi connection fails every time when new chrome version is released i have opened this defect in the past the line that is causing it is located under webdriver py devtools import module selenium webdriver common devtools v format version devtools module don t seem to be updating fast enough it is hard to use bidi connection in a professional setting currently it fails for chrome version how can we reproduce the issue shell update chrome and webdriver to version run async with driver bidi connection as bidi session session devtools bidi session session bidi session devtools relevant log output shell modulenotfounderror operating system windows and mac os selenium version python what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue are you using selenium grid no | 1 |
2,234 | 2,671,186,403 | IssuesEvent | 2015-03-24 03:03:55 | socketstream/socketstream | https://api.github.com/repos/socketstream/socketstream | opened | Documentation improvement | Documentation Proposed change | [See https://github.com/socketstream/socketstream/issues/513 as a precursor to this.]
I've created a `feature/doc-updates` branch. I think it was @RomanMinkin who did a tremendous effort in going through our docs and trying to get them in a better state. However, after having given them a good look over, I think we've over complicated the task by throwing an angular wrapper around the documentation section—it makes them difficult to maintain.
For example, I went to go work on the documentation this evening and spent way too long trying to figure things out when all I wanted to do was write.
I think we could work on trying to make the process simpler, and I appreciate @thepian's view about getting the community involved to that end. (It's probably a catch 22—need to make it easier first to get people to start helping.)
Here are some starting suggestions.
1. Lets swap out ngdocs for something that doesn't rely on Angular, and is a little more accessible and established, like JSDocs.
2. For authored docs, markdown is probably the best bet. There are techniques for converting `.md` files into HTML, and markdown is the flavor, too, of GitHub. That way we're not repeating our efforts, and the docs can exist in two places. (Good for when servers are down.)
3. Also, maybe we could all just divvy up some of the missing or incomplete topics and take a stab at authoring single pages—I've said I plan on working on the docs a lot, but the knowledge of how things work is in everyone's minds. If everyone takes a stab at a draft (it can be messy!), I can be editorial and clean and organize things. (I come from a pack of literary wolves and editors, as evidenced by all my verbose GitHub comments.)
Those are some initial thoughts. Please add you own ideas, and I'll work around the communities' needs. | 1.0 | Documentation improvement - [See https://github.com/socketstream/socketstream/issues/513 as a precursor to this.]
I've created a `feature/doc-updates` branch. I think it was @RomanMinkin who did a tremendous effort in going through our docs and trying to get them in a better state. However, after having given them a good look over, I think we've over complicated the task by throwing an angular wrapper around the documentation section—it makes them difficult to maintain.
For example, I went to go work on the documentation this evening and spent way too long trying to figure things out when all I wanted to do was write.
I think we could work on trying to make the process simpler, and I appreciate @thepian's view about getting the community involved to that end. (It's probably a catch 22—need to make it easier first to get people to start helping.)
Here are some starting suggestions.
1. Lets swap out ngdocs for something that doesn't rely on Angular, and is a little more accessible and established, like JSDocs.
2. For authored docs, markdown is probably the best bet. There are techniques for converting `.md` files into HTML, and markdown is the flavor, too, of GitHub. That way we're not repeating our efforts, and the docs can exist in two places. (Good for when servers are down.)
3. Also, maybe we could all just divvy up some of the missing or incomplete topics and take a stab at authoring single pages—I've said I plan on working on the docs a lot, but the knowledge of how things work is in everyone's minds. If everyone takes a stab at a draft (it can be messy!), I can be editorial and clean and organize things. (I come from a pack of literary wolves and editors, as evidenced by all my verbose GitHub comments.)
Those are some initial thoughts. Please add you own ideas, and I'll work around the communities' needs. | non_defect | documentation improvement i ve created a feature doc updates branch i think it was romanminkin who did a tremendous effort in going through our docs and trying to get them in a better state however after having given them a good look over i think we ve over complicated the task by throwing an angular wrapper around the documentation section—it makes them difficult to maintain for example i went to go work on the documentation this evening and spent way too long trying to figure things out when all i wanted to do was write i think we could work on trying to make the process simpler and i appreciate thepian s view about getting the community involved to that end it s probably a catch —need to make it easier first to get people to start helping here are some starting suggestions lets swap out ngdocs for something that doesn t rely on angular and is a little more accessible and established like jsdocs for authored docs markdown is probably the best bet there are techniques for converting md files into html and markdown is the flavor too of github that way we re not repeating our efforts and the docs can exist in two places good for when servers are down also maybe we could all just divvy up some of the missing or incomplete topics and take a stab at authoring single pages—i ve said i plan on working on the docs a lot but the knowledge of how things work is in everyone s minds if everyone takes a stab at a draft it can be messy i can be editorial and clean and organize things i come from a pack of literary wolves and editors as evidenced by all my verbose github comments those are some initial thoughts please add you own ideas and i ll work around the communities needs | 0 |
11,696 | 2,661,019,976 | IssuesEvent | 2015-03-19 12:06:23 | contao/core | https://api.github.com/repos/contao/core | closed | CE - Media erkennt `.ogg` als Audio | defect up for discussion | Leider wird bei der Angabe webm und ogg die Ausgabe als Audio und nicht als Video erkannt. Sobald ein mp4 dazu kommt passt es, jedoch nicht bei nur den beiden erstgenannten Formaten. | 1.0 | CE - Media erkennt `.ogg` als Audio - Leider wird bei der Angabe webm und ogg die Ausgabe als Audio und nicht als Video erkannt. Sobald ein mp4 dazu kommt passt es, jedoch nicht bei nur den beiden erstgenannten Formaten. | defect | ce media erkennt ogg als audio leider wird bei der angabe webm und ogg die ausgabe als audio und nicht als video erkannt sobald ein dazu kommt passt es jedoch nicht bei nur den beiden erstgenannten formaten | 1 |
58,109 | 16,342,462,675 | IssuesEvent | 2021-05-13 00:21:39 | darshan-hpc/darshan | https://api.github.com/repos/darshan-hpc/darshan | closed | uthash header not installed | defect | In GitLab by @shanedsnyder on Oct 12, 2015, 09:59
The hash implementation used in libdarshan-util is not installed with Darshan, preventing external tools that link in this library from being able to compile. | 1.0 | uthash header not installed - In GitLab by @shanedsnyder on Oct 12, 2015, 09:59
The hash implementation used in libdarshan-util is not installed with Darshan, preventing external tools that link in this library from being able to compile. | defect | uthash header not installed in gitlab by shanedsnyder on oct the hash implementation used in libdarshan util is not installed with darshan preventing external tools that link in this library from being able to compile | 1 |
37,299 | 2,824,111,670 | IssuesEvent | 2015-05-21 13:06:47 | pywinauto/pywinauto | https://api.github.com/repos/pywinauto/pywinauto | closed | _treeview_element.Click() uses ClickInput() | auto-migrated bug internal Priority-Low | ```
What steps will reproduce the problem?
1. In common_control.py class _treeview_element contain method Click
2. Method Click uses another method HwndWrapper.ClickInput
3. It is inconsistent, HwndWrapper has Click and ClickInput methods and they
works in different way.
What is the expected output? What do you see instead?
The _treeview_element.Click() should use HwndWrapper.Click()
Probably _treeview_element.ClickInput()should be added.
What version of the product are you using? On what operating system?
0.4.1. In 0.4.2 I saw the same issue
Please provide any additional information below.
```
Original issue reported on code.google.com by `far...@gmail.com` on 5 Feb 2014 at 8:01 | 1.0 | _treeview_element.Click() uses ClickInput() - ```
What steps will reproduce the problem?
1. In common_control.py class _treeview_element contain method Click
2. Method Click uses another method HwndWrapper.ClickInput
3. It is inconsistent, HwndWrapper has Click and ClickInput methods and they
works in different way.
What is the expected output? What do you see instead?
The _treeview_element.Click() should use HwndWrapper.Click()
Probably _treeview_element.ClickInput()should be added.
What version of the product are you using? On what operating system?
0.4.1. In 0.4.2 I saw the same issue
Please provide any additional information below.
```
Original issue reported on code.google.com by `far...@gmail.com` on 5 Feb 2014 at 8:01 | non_defect | treeview element click uses clickinput what steps will reproduce the problem in common control py class treeview element contain method click method click uses another method hwndwrapper clickinput it is inconsistent hwndwrapper has click and clickinput methods and they works in different way what is the expected output what do you see instead the treeview element click should use hwndwrapper click probably treeview element clickinput should be added what version of the product are you using on what operating system in i saw the same issue please provide any additional information below original issue reported on code google com by far gmail com on feb at | 0 |
197,758 | 6,963,601,772 | IssuesEvent | 2017-12-08 18:03:29 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio-ui] Change UI to always send CSRF token on API calls | new feature priority: high | UI will need to send CSRF token on every API call now that does a POST, PUT or DELETE. Please refer to craftercms/craftercms#1601.
| 1.0 | [studio-ui] Change UI to always send CSRF token on API calls - UI will need to send CSRF token on every API call now that does a POST, PUT or DELETE. Please refer to craftercms/craftercms#1601.
| non_defect | change ui to always send csrf token on api calls ui will need to send csrf token on every api call now that does a post put or delete please refer to craftercms craftercms | 0 |
67,511 | 20,972,920,687 | IssuesEvent | 2022-03-28 13:02:20 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | Accordion open all taps in the first few seconds if its in a Dynamic Dialog | defect |
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Please demonstrate your case at stackblitz by using the issue template below. Issues without a test case have much less possibility to be reviewd in detail and assisted.
https://stackblitz.com/edit/github-9ccs9k?mbed=1&file=src/app/app.component.ts&view=preview
**Current behavior**
Accordion open all taps in the first few seconds if its in a Dynamic Dialog. After that it will collapse again. This looks strange.
**Expected behavior**
Accordion taps should be closed per default if I open a dialog which contains a accordion as it was in the versions before.
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
We use the Accordion in a Dynamic Dialog in our business application and it is no good look and feel for the customer.
* **Angular version:** 9.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 9.X
<!-- Check whether this is still an issue in the most recent Angular version -->
| 1.0 | Accordion open all taps in the first few seconds if its in a Dynamic Dialog -
[x] bug report => Search github for a similar issue or PR before submitting
[ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap
[ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35
```
**Plunkr Case (Bug Reports)**
Please demonstrate your case at stackblitz by using the issue template below. Issues without a test case have much less possibility to be reviewd in detail and assisted.
https://stackblitz.com/edit/github-9ccs9k?mbed=1&file=src/app/app.component.ts&view=preview
**Current behavior**
Accordion open all taps in the first few seconds if its in a Dynamic Dialog. After that it will collapse again. This looks strange.
**Expected behavior**
Accordion taps should be closed per default if I open a dialog which contains a accordion as it was in the versions before.
**What is the motivation / use case for changing the behavior?**
<!-- Describe the motivation or the concrete use case -->
We use the Accordion in a Dynamic Dialog in our business application and it is no good look and feel for the customer.
* **Angular version:** 9.X
<!-- Check whether this is still an issue in the most recent Angular version -->
* **PrimeNG version:** 9.X
<!-- Check whether this is still an issue in the most recent Angular version -->
| defect | accordion open all taps in the first few seconds if its in a dynamic dialog bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports please demonstrate your case at stackblitz by using the issue template below issues without a test case have much less possibility to be reviewd in detail and assisted current behavior accordion open all taps in the first few seconds if its in a dynamic dialog after that it will collapse again this looks strange expected behavior accordion taps should be closed per default if i open a dialog which contains a accordion as it was in the versions before what is the motivation use case for changing the behavior we use the accordion in a dynamic dialog in our business application and it is no good look and feel for the customer angular version x primeng version x | 1 |
54,521 | 13,757,481,786 | IssuesEvent | 2020-10-06 21:44:22 | Cockatrice/Cockatrice | https://api.github.com/repos/Cockatrice/Cockatrice | closed | Power and Toughness Hotkey not working for increase | App - Cockatrice App - Protocol / API Defect - Basic | <b>OS:</b>
OS X 10.11.6
<b>Cockatrice version:</b>
Version dab7316 (2017-01-19)
___
<br>
I reinstalled Cockatrice a few days ago and noticed that all the other hot keys seem to work but I can not get the Power and Toughness to increase. I have even changed from standard settings to try to make it work, and I still can not hot key. I can manually switch it, but when adding several +1/+1 counters to each creature every turn...it gets extremely time consuming to try to play. Thanks for the help!
| 1.0 | Power and Toughness Hotkey not working for increase - <b>OS:</b>
OS X 10.11.6
<b>Cockatrice version:</b>
Version dab7316 (2017-01-19)
___
<br>
I reinstalled Cockatrice a few days ago and noticed that all the other hot keys seem to work but I can not get the Power and Toughness to increase. I have even changed from standard settings to try to make it work, and I still can not hot key. I can manually switch it, but when adding several +1/+1 counters to each creature every turn...it gets extremely time consuming to try to play. Thanks for the help!
| defect | power and toughness hotkey not working for increase os os x cockatrice version version i reinstalled cockatrice a few days ago and noticed that all the other hot keys seem to work but i can not get the power and toughness to increase i have even changed from standard settings to try to make it work and i still can not hot key i can manually switch it but when adding several counters to each creature every turn it gets extremely time consuming to try to play thanks for the help | 1 |
85,535 | 16,673,564,719 | IssuesEvent | 2021-06-07 13:48:15 | Decentra-Network/Decentra-Network | https://api.github.com/repos/Decentra-Network/Decentra-Network | closed | Description must be added to question_maker(question_text=None, mode=None) | Code Quality Improvements | Link of the line: https://github.com/Decentra-Network/Decentra-Network/blob/master/src/lib/mixlib.py#L51
| 1.0 | Description must be added to question_maker(question_text=None, mode=None) - Link of the line: https://github.com/Decentra-Network/Decentra-Network/blob/master/src/lib/mixlib.py#L51
| non_defect | description must be added to question maker question text none mode none link of the line | 0 |
10,128 | 2,618,937,654 | IssuesEvent | 2015-03-03 00:02:44 | chrsmith/open-ig | https://api.github.com/repos/chrsmith/open-ig | closed | Planet list ordering is errorenous, it lists planets as X 1, X 10, X 11, X 12, X 2, X 20, X 21, etc | auto-migrated Component-UI Priority-Low Type-Defect | ```
Game version: 0.95.152
Operating System: Linux x64
Java runtime version: 1.7.0_51
Installed using the Launcher? yes
Game language (en, hu, de): hu
```
Original issue reported on code.google.com by `kli...@gmail.com` on 21 Jan 2014 at 1:48 | 1.0 | Planet list ordering is errorenous, it lists planets as X 1, X 10, X 11, X 12, X 2, X 20, X 21, etc - ```
Game version: 0.95.152
Operating System: Linux x64
Java runtime version: 1.7.0_51
Installed using the Launcher? yes
Game language (en, hu, de): hu
```
Original issue reported on code.google.com by `kli...@gmail.com` on 21 Jan 2014 at 1:48 | defect | planet list ordering is errorenous it lists planets as x x x x x x x etc game version operating system linux java runtime version installed using the launcher yes game language en hu de hu original issue reported on code google com by kli gmail com on jan at | 1 |
63,810 | 18,009,835,112 | IssuesEvent | 2021-09-16 07:14:59 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Joining Subspace shows UI about rooms | T-Defect S-Tolerable A-Spaces O-Frequent A-Subspaces Z-Spaces-Testing | ### Steps to reproduce
1. Go to Explore rooms in a space with a subspace
2. Click the "Join" button
### What happened?
### What did you expect?
Next dialog/page showing "Joining space"
### What happened?
Goes to an intermediate page with "Joining room"/spinner and then to the "Explore rooms" page for the subspace
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
develop.element.io
### Homeserver
_No response_
### Have you submitted a rageshake?
No | 1.0 | Joining Subspace shows UI about rooms - ### Steps to reproduce
1. Go to Explore rooms in a space with a subspace
2. Click the "Join" button
### What happened?
### What did you expect?
Next dialog/page showing "Joining space"
### What happened?
Goes to an intermediate page with "Joining room"/spinner and then to the "Explore rooms" page for the subspace
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
develop.element.io
### Homeserver
_No response_
### Have you submitted a rageshake?
No | defect | joining subspace shows ui about rooms steps to reproduce go to explore rooms in a space with a subspace click the join button what happened what did you expect next dialog page showing joining space what happened goes to an intermediate page with joining room spinner and then to the explore rooms page for the subspace operating system no response browser information no response url for webapp develop element io homeserver no response have you submitted a rageshake no | 1 |
37,776 | 8,517,451,439 | IssuesEvent | 2018-11-01 08:03:11 | GoldenSoftwareLtd/gedemin | https://api.github.com/repos/GoldenSoftwareLtd/gedemin | closed | После редактирования отчёта права на него не сохраняются | GedeminExe Priority-Medium Report Security Type-Defect | Originally reported on Google Code with ID 3263
```
После редактирования отчёта права на него не сохраняются, а перезаписываются правами
на родительскую папку. Приходится после каждой правки вручную их переназначать что
очень неудобно.
И ещё такая просьба. Сделайте, пожалуйста, возможным в отчётах на закладке ‘история’
удалять помеченные версии отчёта и для удобства кнопку 'выбрать все кроме последней'.
Т.к. создаётся много версий по любой мелочи, которые абсолютно не нужны и занимают
место в блобах, а значит, увеличивают фрагментацию соответствующих таблиц, замедляет
по ним поиск и т.п.
Если снять статистику IBAnalust-ом, то видно, что заполнение соотв. таблиц порядка
1%.
```
Reported by `igor6003` on 2013-11-12 11:47:28
| 1.0 | После редактирования отчёта права на него не сохраняются - Originally reported on Google Code with ID 3263
```
После редактирования отчёта права на него не сохраняются, а перезаписываются правами
на родительскую папку. Приходится после каждой правки вручную их переназначать что
очень неудобно.
И ещё такая просьба. Сделайте, пожалуйста, возможным в отчётах на закладке ‘история’
удалять помеченные версии отчёта и для удобства кнопку 'выбрать все кроме последней'.
Т.к. создаётся много версий по любой мелочи, которые абсолютно не нужны и занимают
место в блобах, а значит, увеличивают фрагментацию соответствующих таблиц, замедляет
по ним поиск и т.п.
Если снять статистику IBAnalust-ом, то видно, что заполнение соотв. таблиц порядка
1%.
```
Reported by `igor6003` on 2013-11-12 11:47:28
| defect | после редактирования отчёта права на него не сохраняются originally reported on google code with id после редактирования отчёта права на него не сохраняются а перезаписываются правами на родительскую папку приходится после каждой правки вручную их переназначать что очень неудобно и ещё такая просьба сделайте пожалуйста возможным в отчётах на закладке ‘история’ удалять помеченные версии отчёта и для удобства кнопку выбрать все кроме последней т к создаётся много версий по любой мелочи которые абсолютно не нужны и занимают место в блобах а значит увеличивают фрагментацию соответствующих таблиц замедляет по ним поиск и т п если снять статистику ibanalust ом то видно что заполнение соотв таблиц порядка reported by on | 1 |
25,967 | 4,538,635,559 | IssuesEvent | 2016-09-09 07:49:06 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Dictionary - Default constructor Comparer value | defect portarelle | ### Expected
```cs
System.Collections.Generic.GenericEqualityComparer'1[System.Int32]
```
### Actual
```cs
System.Collections.Generic.GenericEqualityComparer'1[Object]
```
### Steps To Reproduce
```csharp
public class App
{
public static void Main()
{
var d1 = new Dictionary<int, string>();
Assert.AreStrictEqual(d1.Comparer, EqualityComparer<int>.Default);
var d2 = new Dictionary<int, string>(10);
Assert.AreStrictEqual(d2.Comparer, EqualityComparer<int>.Default);
var orig = new Dictionary<string, int>();
var d3 = new Dictionary<string, int>(orig);
Assert.AreStrictEqual(d3.Comparer, EqualityComparer<int>.Default);
}
}
``` | 1.0 | Dictionary - Default constructor Comparer value - ### Expected
```cs
System.Collections.Generic.GenericEqualityComparer'1[System.Int32]
```
### Actual
```cs
System.Collections.Generic.GenericEqualityComparer'1[Object]
```
### Steps To Reproduce
```csharp
public class App
{
public static void Main()
{
var d1 = new Dictionary<int, string>();
Assert.AreStrictEqual(d1.Comparer, EqualityComparer<int>.Default);
var d2 = new Dictionary<int, string>(10);
Assert.AreStrictEqual(d2.Comparer, EqualityComparer<int>.Default);
var orig = new Dictionary<string, int>();
var d3 = new Dictionary<string, int>(orig);
Assert.AreStrictEqual(d3.Comparer, EqualityComparer<int>.Default);
}
}
``` | defect | dictionary default constructor comparer value expected cs system collections generic genericequalitycomparer actual cs system collections generic genericequalitycomparer steps to reproduce csharp public class app public static void main var new dictionary assert arestrictequal comparer equalitycomparer default var new dictionary assert arestrictequal comparer equalitycomparer default var orig new dictionary var new dictionary orig assert arestrictequal comparer equalitycomparer default | 1 |
2,194 | 2,603,977,775 | IssuesEvent | 2015-02-24 19:01:58 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳沈阳感染疱疹怎么办 | auto-migrated Priority-Medium Type-Defect | ```
沈阳沈阳感染疱疹怎么办〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:29 | 1.0 | 沈阳沈阳感染疱疹怎么办 - ```
沈阳沈阳感染疱疹怎么办〓沈陽軍區政治部醫院性病〓TEL:02
4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療�
��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝�
��的歷史悠久、設備精良、技術權威、專家云集,是預防、保
健、醫療、科研康復為一體的綜合性醫院。是國家首批公立��
�等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學�
��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍
空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集��
�二等功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:29 | defect | 沈阳沈阳感染疱疹怎么办 沈阳沈阳感染疱疹怎么办〓沈陽軍區政治部醫院性病〓tel: 〓 , � �� 。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 original issue reported on code google com by gmail com on jun at | 1 |
11,715 | 2,664,255,802 | IssuesEvent | 2015-03-20 13:28:40 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | SQLDialectNotSupportedException: ARRAY TABLE is not supported for SQL99 | C: DB: Oracle C: Functionality P: High T: Defect | When using Oracle's `TABLE()` function in a derived table, jOOQ throws an exception:
```
SQLDialectNotSupportedException: ARRAY TABLE is not supported for SQL99
```
An example to reproduce this:
```java
UNumberTableRecord ids = new UNumberTableRecord(1, 2, 3);
// Never mind this cast in this case...
Select<Record1<Integer>> selectIds = (Select<Record1<Integer>>) DSL.selectFrom(DSL.table(ids));
create()
.select(T_BOOK.ID, T_BOOK.TITLE)
.from(DSL.select(T_BOOK.ID,
T_BOOK.AUTHOR_ID,
T_BOOK.TITLE)
.from(T_BOOK)
.join(T_AUTHOR)
.on(T_BOOK.AUTHOR_ID.eq(T_AUTHOR.ID))
.where(T_BOOK.ID.in(selectIds)))
.fetch();
``` | 1.0 | SQLDialectNotSupportedException: ARRAY TABLE is not supported for SQL99 - When using Oracle's `TABLE()` function in a derived table, jOOQ throws an exception:
```
SQLDialectNotSupportedException: ARRAY TABLE is not supported for SQL99
```
An example to reproduce this:
```java
UNumberTableRecord ids = new UNumberTableRecord(1, 2, 3);
// Never mind this cast in this case...
Select<Record1<Integer>> selectIds = (Select<Record1<Integer>>) DSL.selectFrom(DSL.table(ids));
create()
.select(T_BOOK.ID, T_BOOK.TITLE)
.from(DSL.select(T_BOOK.ID,
T_BOOK.AUTHOR_ID,
T_BOOK.TITLE)
.from(T_BOOK)
.join(T_AUTHOR)
.on(T_BOOK.AUTHOR_ID.eq(T_AUTHOR.ID))
.where(T_BOOK.ID.in(selectIds)))
.fetch();
``` | defect | sqldialectnotsupportedexception array table is not supported for when using oracle s table function in a derived table jooq throws an exception sqldialectnotsupportedexception array table is not supported for an example to reproduce this java unumbertablerecord ids new unumbertablerecord never mind this cast in this case select selectids select dsl selectfrom dsl table ids create select t book id t book title from dsl select t book id t book author id t book title from t book join t author on t book author id eq t author id where t book id in selectids fetch | 1 |
46,644 | 13,055,953,438 | IssuesEvent | 2020-07-30 03:13:17 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | triangle.c has different precision based on OS (Trac #1647) | Incomplete Migration Migrated from Trac combo simulation defect | Migrated from https://code.icecube.wisc.edu/ticket/1647
```json
{
"status": "closed",
"changetime": "2016-08-18T22:03:29",
"description": "See:\n`LINUX/CPU86` in the source:/IceCube/projects/g4-tankresponse/trunk/private/g4-tankresponse/triangle/triangle.c\n",
"reporter": "nega",
"cc": "",
"resolution": "wontfix",
"_ts": "1471557809273381",
"component": "combo simulation",
"summary": "triangle.c has different precision based on OS",
"priority": "normal",
"keywords": "pruning C",
"time": "2016-04-20T19:35:01",
"milestone": "Long-Term Future",
"owner": "jgonzalez",
"type": "defect"
}
```
| 1.0 | triangle.c has different precision based on OS (Trac #1647) - Migrated from https://code.icecube.wisc.edu/ticket/1647
```json
{
"status": "closed",
"changetime": "2016-08-18T22:03:29",
"description": "See:\n`LINUX/CPU86` in the source:/IceCube/projects/g4-tankresponse/trunk/private/g4-tankresponse/triangle/triangle.c\n",
"reporter": "nega",
"cc": "",
"resolution": "wontfix",
"_ts": "1471557809273381",
"component": "combo simulation",
"summary": "triangle.c has different precision based on OS",
"priority": "normal",
"keywords": "pruning C",
"time": "2016-04-20T19:35:01",
"milestone": "Long-Term Future",
"owner": "jgonzalez",
"type": "defect"
}
```
| defect | triangle c has different precision based on os trac migrated from json status closed changetime description see n linux in the source icecube projects tankresponse trunk private tankresponse triangle triangle c n reporter nega cc resolution wontfix ts component combo simulation summary triangle c has different precision based on os priority normal keywords pruning c time milestone long term future owner jgonzalez type defect | 1 |
53,463 | 13,261,662,075 | IssuesEvent | 2020-08-20 20:18:29 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [cvmfs] opencl environment issues (Trac #1450) | Migrated from Trac cvmfs defect | As seen on buildbot "Ubuntu 14.04 (cvmfs)", there is some sort of conflict where `libOpenCL.so` is not found, but `/etc/OpenCL/vendors` exists. Make sure that if we're using CVMFS OpenCL that we have it in the list of vendors.
Solution? :
1. If `/etc/OpenCL/vendors` does not exist, use CVMFS copy
2. If `/etc/OpenCL/vendors` exists, make temp hybrid dir with entries from `/etc/OpenCL/vendors` and an entry for CVMFS.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1450">https://code.icecube.wisc.edu/projects/icecube/ticket/1450</a>, reported by david.schultzand owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:09",
"_ts": "1458335649133028",
"description": "As seen on buildbot \"Ubuntu 14.04 (cvmfs)\", there is some sort of conflict where `libOpenCL.so` is not found, but `/etc/OpenCL/vendors` exists. Make sure that if we're using CVMFS OpenCL that we have it in the list of vendors.\n\nSolution? :\n\n1. If `/etc/OpenCL/vendors` does not exist, use CVMFS copy\n2. If `/etc/OpenCL/vendors` exists, make temp hybrid dir with entries from `/etc/OpenCL/vendors` and an entry for CVMFS.",
"reporter": "david.schultz",
"cc": "nega, claudio.kopper",
"resolution": "fixed",
"time": "2015-11-25T16:51:22",
"component": "cvmfs",
"summary": "[cvmfs] opencl environment issues",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [cvmfs] opencl environment issues (Trac #1450) - As seen on buildbot "Ubuntu 14.04 (cvmfs)", there is some sort of conflict where `libOpenCL.so` is not found, but `/etc/OpenCL/vendors` exists. Make sure that if we're using CVMFS OpenCL that we have it in the list of vendors.
Solution? :
1. If `/etc/OpenCL/vendors` does not exist, use CVMFS copy
2. If `/etc/OpenCL/vendors` exists, make temp hybrid dir with entries from `/etc/OpenCL/vendors` and an entry for CVMFS.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1450">https://code.icecube.wisc.edu/projects/icecube/ticket/1450</a>, reported by david.schultzand owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:09",
"_ts": "1458335649133028",
"description": "As seen on buildbot \"Ubuntu 14.04 (cvmfs)\", there is some sort of conflict where `libOpenCL.so` is not found, but `/etc/OpenCL/vendors` exists. Make sure that if we're using CVMFS OpenCL that we have it in the list of vendors.\n\nSolution? :\n\n1. If `/etc/OpenCL/vendors` does not exist, use CVMFS copy\n2. If `/etc/OpenCL/vendors` exists, make temp hybrid dir with entries from `/etc/OpenCL/vendors` and an entry for CVMFS.",
"reporter": "david.schultz",
"cc": "nega, claudio.kopper",
"resolution": "fixed",
"time": "2015-11-25T16:51:22",
"component": "cvmfs",
"summary": "[cvmfs] opencl environment issues",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
| defect | opencl environment issues trac as seen on buildbot ubuntu cvmfs there is some sort of conflict where libopencl so is not found but etc opencl vendors exists make sure that if we re using cvmfs opencl that we have it in the list of vendors solution if etc opencl vendors does not exist use cvmfs copy if etc opencl vendors exists make temp hybrid dir with entries from etc opencl vendors and an entry for cvmfs migrated from json status closed changetime ts description as seen on buildbot ubuntu cvmfs there is some sort of conflict where libopencl so is not found but etc opencl vendors exists make sure that if we re using cvmfs opencl that we have it in the list of vendors n nsolution n if etc opencl vendors does not exist use cvmfs copy if etc opencl vendors exists make temp hybrid dir with entries from etc opencl vendors and an entry for cvmfs reporter david schultz cc nega claudio kopper resolution fixed time component cvmfs summary opencl environment issues priority major keywords milestone owner david schultz type defect | 1 |
103,410 | 12,893,582,911 | IssuesEvent | 2020-07-13 21:58:20 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | closed | Need to add the Global navigation to the Select Cluster drop-down | [zube]: Design WIP area/navigation | Need design of how we want the Main Drop-down to switch between Global and Individual Clusters.
This menu:

| 1.0 | Need to add the Global navigation to the Select Cluster drop-down - Need design of how we want the Main Drop-down to switch between Global and Individual Clusters.
This menu:

| non_defect | need to add the global navigation to the select cluster drop down need design of how we want the main drop down to switch between global and individual clusters this menu | 0 |
45,294 | 23,999,191,834 | IssuesEvent | 2022-09-14 09:59:39 | FeatureBaseDB/featurebase | https://api.github.com/repos/FeatureBaseDB/featurebase | closed | import endpoints which take a shard should optionally allocate the shard for you | proposal ingest-performance | # Description
There are many situations where records in a data set do not have natural sequential integer ids and we want to simply assign them at ingest time (rather than using column translation). In these cases, a single client importing into an empty Pilosa can fairly easily generate sequential ids and import the data performantly.
But what if there are multiple clients? What if there is already data in Pilosa? Clients need some way of coordinating which IDs each one will allocate, and understanding what data is already in Pilosa so it doesn't get overwritten. One could imagine any variety of where clients communicate with each other (or through another service) to synchronize, or are configured ahead of time not to produce overlapping ids, and can interrogate Pilosa about what data it already has, but these all seem messy and like kind of a lot of work.
What if the import endpoints which take a shard parameter simply allowed you to leave that parameter off? If you do not specify what shard the data should go into, Pilosa knows that it should use a new empty shard to ingest the data. This will allow any number of concurrent clients to throw shardfulls of data at Pilosa without having to do any synchronization or pre-communication!
Pilosa will have to take care such that concurrent requests don't end up in the same shard, but this is purely internal to Pilosa, and seems a lot easier than trying to coordinate with clients. The node receiving the request can look at its availableShards data and select from among the first few empty ones. It would then send out a request to all the owners of that shard to let them know it is reserving it. If the owners respond that they have no data for that shard, then the node can continue with the import. The owners will have to reject any other modifications to that shard until the import is finished.
| True | import endpoints which take a shard should optionally allocate the shard for you - # Description
There are many situations where records in a data set do not have natural sequential integer ids and we want to simply assign them at ingest time (rather than using column translation). In these cases, a single client importing into an empty Pilosa can fairly easily generate sequential ids and import the data performantly.
But what if there are multiple clients? What if there is already data in Pilosa? Clients need some way of coordinating which IDs each one will allocate, and understanding what data is already in Pilosa so it doesn't get overwritten. One could imagine any variety of where clients communicate with each other (or through another service) to synchronize, or are configured ahead of time not to produce overlapping ids, and can interrogate Pilosa about what data it already has, but these all seem messy and like kind of a lot of work.
What if the import endpoints which take a shard parameter simply allowed you to leave that parameter off? If you do not specify what shard the data should go into, Pilosa knows that it should use a new empty shard to ingest the data. This will allow any number of concurrent clients to throw shardfulls of data at Pilosa without having to do any synchronization or pre-communication!
Pilosa will have to take care such that concurrent requests don't end up in the same shard, but this is purely internal to Pilosa, and seems a lot easier than trying to coordinate with clients. The node receiving the request can look at its availableShards data and select from among the first few empty ones. It would then send out a request to all the owners of that shard to let them know it is reserving it. If the owners respond that they have no data for that shard, then the node can continue with the import. The owners will have to reject any other modifications to that shard until the import is finished.
| non_defect | import endpoints which take a shard should optionally allocate the shard for you description there are many situations where records in a data set do not have natural sequential integer ids and we want to simply assign them at ingest time rather than using column translation in these cases a single client importing into an empty pilosa can fairly easily generate sequential ids and import the data performantly but what if there are multiple clients what if there is already data in pilosa clients need some way of coordinating which ids each one will allocate and understanding what data is already in pilosa so it doesn t get overwritten one could imagine any variety of where clients communicate with each other or through another service to synchronize or are configured ahead of time not to produce overlapping ids and can interrogate pilosa about what data it already has but these all seem messy and like kind of a lot of work what if the import endpoints which take a shard parameter simply allowed you to leave that parameter off if you do not specify what shard the data should go into pilosa knows that it should use a new empty shard to ingest the data this will allow any number of concurrent clients to throw shardfulls of data at pilosa without having to do any synchronization or pre communication pilosa will have to take care such that concurrent requests don t end up in the same shard but this is purely internal to pilosa and seems a lot easier than trying to coordinate with clients the node receiving the request can look at its availableshards data and select from among the first few empty ones it would then send out a request to all the owners of that shard to let them know it is reserving it if the owners respond that they have no data for that shard then the node can continue with the import the owners will have to reject any other modifications to that shard until the import is finished | 0 |
50,697 | 13,187,684,255 | IssuesEvent | 2020-08-13 04:13:38 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | [wavereform] TODO (Trac #1197) | Migrated from Trac combo reconstruction defect | decide whether this TODO is important, and if so make a ticket for it:
```text
python/wavereform.py: # TODO: flag FADCs that saturate outside of the ATWD window.
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1197">https://code.icecube.wisc.edu/ticket/1197</a>, reported by david.schultz and owned by jbraun</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "decide whether this TODO is important, and if so make a ticket for it:\n{{{\npython/wavereform.py:\t\t\t# TODO: flag FADCs that saturate outside of the ATWD window.\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "duplicate",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "[wavereform] TODO",
"priority": "critical",
"keywords": "",
"time": "2015-08-19T18:10:05",
"milestone": "",
"owner": "jbraun",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [wavereform] TODO (Trac #1197) - decide whether this TODO is important, and if so make a ticket for it:
```text
python/wavereform.py: # TODO: flag FADCs that saturate outside of the ATWD window.
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1197">https://code.icecube.wisc.edu/ticket/1197</a>, reported by david.schultz and owned by jbraun</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "decide whether this TODO is important, and if so make a ticket for it:\n{{{\npython/wavereform.py:\t\t\t# TODO: flag FADCs that saturate outside of the ATWD window.\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "duplicate",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "[wavereform] TODO",
"priority": "critical",
"keywords": "",
"time": "2015-08-19T18:10:05",
"milestone": "",
"owner": "jbraun",
"type": "defect"
}
```
</p>
</details>
| defect | todo trac decide whether this todo is important and if so make a ticket for it text python wavereform py todo flag fadcs that saturate outside of the atwd window migrated from json status closed changetime description decide whether this todo is important and if so make a ticket for it n npython wavereform py t t t todo flag fadcs that saturate outside of the atwd window n reporter david schultz cc resolution duplicate ts component combo reconstruction summary todo priority critical keywords time milestone owner jbraun type defect | 1 |
53,079 | 13,260,871,095 | IssuesEvent | 2020-08-20 18:54:19 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | Make cmake framework detection compatible with Xcode versions >= 4.3 (Trac #668) | Migrated from Trac defect tools/ports | Since Xcode 4.3, frameworks are no longer stored in /Developer, but insider the Xcode application bundle. Cmake has support for this, but unfortunately, the detection code is missing a "platform" part in its detection code. These two patch files
a. enhance the detection code to look in the correct place for newer Xcode versions, and
a. uses the detected location when looking for frameworks
This will allow detection of the JavaVM (JNI) developer framework containing the correct header files. (The system framework is detected correctly, but it does not contain headers.)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/668">https://code.icecube.wisc.edu/projects/icecube/ticket/668</a>, reported by claudio.kopperand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T19:01:17",
"_ts": "1351710077000000",
"description": "Since Xcode 4.3, frameworks are no longer stored in /Developer, but insider the Xcode application bundle. Cmake has support for this, but unfortunately, the detection code is missing a \"platform\" part in its detection code. These two patch files \n\n a. enhance the detection code to look in the correct place for newer Xcode versions, and\n a. uses the detected location when looking for frameworks \n\nThis will allow detection of the JavaVM (JNI) developer framework containing the correct header files. (The system framework is detected correctly, but it does not contain headers.)\n",
"reporter": "claudio.kopper",
"cc": "",
"resolution": "fixed",
"time": "2012-02-19T19:53:49",
"component": "tools/ports",
"summary": "Make cmake framework detection compatible with Xcode versions >= 4.3",
"priority": "normal",
"keywords": "cmake xcode mac os frameworks jni java javavm",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Make cmake framework detection compatible with Xcode versions >= 4.3 (Trac #668) - Since Xcode 4.3, frameworks are no longer stored in /Developer, but insider the Xcode application bundle. Cmake has support for this, but unfortunately, the detection code is missing a "platform" part in its detection code. These two patch files
a. enhance the detection code to look in the correct place for newer Xcode versions, and
a. uses the detected location when looking for frameworks
This will allow detection of the JavaVM (JNI) developer framework containing the correct header files. (The system framework is detected correctly, but it does not contain headers.)
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/668">https://code.icecube.wisc.edu/projects/icecube/ticket/668</a>, reported by claudio.kopperand owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-10-31T19:01:17",
"_ts": "1351710077000000",
"description": "Since Xcode 4.3, frameworks are no longer stored in /Developer, but insider the Xcode application bundle. Cmake has support for this, but unfortunately, the detection code is missing a \"platform\" part in its detection code. These two patch files \n\n a. enhance the detection code to look in the correct place for newer Xcode versions, and\n a. uses the detected location when looking for frameworks \n\nThis will allow detection of the JavaVM (JNI) developer framework containing the correct header files. (The system framework is detected correctly, but it does not contain headers.)\n",
"reporter": "claudio.kopper",
"cc": "",
"resolution": "fixed",
"time": "2012-02-19T19:53:49",
"component": "tools/ports",
"summary": "Make cmake framework detection compatible with Xcode versions >= 4.3",
"priority": "normal",
"keywords": "cmake xcode mac os frameworks jni java javavm",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | make cmake framework detection compatible with xcode versions trac since xcode frameworks are no longer stored in developer but insider the xcode application bundle cmake has support for this but unfortunately the detection code is missing a platform part in its detection code these two patch files a enhance the detection code to look in the correct place for newer xcode versions and a uses the detected location when looking for frameworks this will allow detection of the javavm jni developer framework containing the correct header files the system framework is detected correctly but it does not contain headers migrated from json status closed changetime ts description since xcode frameworks are no longer stored in developer but insider the xcode application bundle cmake has support for this but unfortunately the detection code is missing a platform part in its detection code these two patch files n n a enhance the detection code to look in the correct place for newer xcode versions and n a uses the detected location when looking for frameworks n nthis will allow detection of the javavm jni developer framework containing the correct header files the system framework is detected correctly but it does not contain headers n reporter claudio kopper cc resolution fixed time component tools ports summary make cmake framework detection compatible with xcode versions priority normal keywords cmake xcode mac os frameworks jni java javavm milestone owner nega type defect | 1 |
249,464 | 7,962,257,737 | IssuesEvent | 2018-07-13 13:46:38 | KB1RD/LearnASM | https://api.github.com/repos/KB1RD/LearnASM | opened | Start using a JS task runner to minify and clean up files before deployment | Low Priority enhancement | This is low priority right now since CloudFlare automatically cleans up the code for me, so I want to stay focused on the main application development. | 1.0 | Start using a JS task runner to minify and clean up files before deployment - This is low priority right now since CloudFlare automatically cleans up the code for me, so I want to stay focused on the main application development. | non_defect | start using a js task runner to minify and clean up files before deployment this is low priority right now since cloudflare automatically cleans up the code for me so i want to stay focused on the main application development | 0 |
248,598 | 7,934,193,743 | IssuesEvent | 2018-07-08 16:15:51 | facelessuser/Rummage | https://api.github.com/repos/facelessuser/Rummage | closed | Allow international file time format | 4.x Feature Priority - Medium | Respect system date/time format. My system is configured to display dates in international format: Y-m-d (like 2018-07-05). For some reason, Rummage still displays dates using format like "Mon Jul 2 12:17:30 2018", which is slightly more difficult to read than international format. If it's difficult to auto obtain system settings, maybe it's possible to enable users to define their own date/time format in settings? Using standard variables (Y for year, H for hour, etc., like in other apps).
Ref #249 | 1.0 | Allow international file time format - Respect system date/time format. My system is configured to display dates in international format: Y-m-d (like 2018-07-05). For some reason, Rummage still displays dates using format like "Mon Jul 2 12:17:30 2018", which is slightly more difficult to read than international format. If it's difficult to auto obtain system settings, maybe it's possible to enable users to define their own date/time format in settings? Using standard variables (Y for year, H for hour, etc., like in other apps).
Ref #249 | non_defect | allow international file time format respect system date time format my system is configured to display dates in international format y m d like for some reason rummage still displays dates using format like mon jul which is slightly more difficult to read than international format if it s difficult to auto obtain system settings maybe it s possible to enable users to define their own date time format in settings using standard variables y for year h for hour etc like in other apps ref | 0 |
4,570 | 23,750,897,652 | IssuesEvent | 2022-08-31 20:30:11 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | Billing Duration reported in 100ms intervals | stage/in-progress area/local/invoke maintainer/need-followup | ### Description:
When using `sam local start-api` the billing duration is reported in 100ms intervals
### Steps to reproduce:
I used the Hello World quick start template
Run `sam local start-api --port 8080`
Hit the function
### Observed result:
Duration: 255.59 ms Billed Duration: 300 ms
### Expected result:
I would have expected billing duration to be 256ms
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
I ran it in Cloud9
I looked through the code and couldn't find anything in this repo relating to printing this out. I'm assuming its pulling this in from somewhere else so maybe an old dependency?
| True | Billing Duration reported in 100ms intervals - ### Description:
When using `sam local start-api` the billing duration is reported in 100ms intervals
### Steps to reproduce:
I used the Hello World quick start template
Run `sam local start-api --port 8080`
Hit the function
### Observed result:
Duration: 255.59 ms Billed Duration: 300 ms
### Expected result:
I would have expected billing duration to be 256ms
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
I ran it in Cloud9
I looked through the code and couldn't find anything in this repo relating to printing this out. I'm assuming its pulling this in from somewhere else so maybe an old dependency?
| non_defect | billing duration reported in intervals description when using sam local start api the billing duration is reported in intervals steps to reproduce i used the hello world quick start template run sam local start api port hit the function observed result duration ms billed duration ms expected result i would have expected billing duration to be additional environment details ex windows mac amazon linux etc i ran it in i looked through the code and couldn t find anything in this repo relating to printing this out i m assuming its pulling this in from somewhere else so maybe an old dependency | 0 |
67,636 | 21,036,831,085 | IssuesEvent | 2022-03-31 08:38:36 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | App prompts for terms of use before validating email | T-Defect | ### Steps to reproduce
1. Sign up on a fresh EMS deployment which requires email
2. Enter email in the reg screen
3. Get prompted to accept terms of use
4. Then get a "please validate your email" prompt
5. Feel surprised that you're asked to accept terms of use before validating your email rather than after - after all, validating your email is more of a hurdle for access to the account than accepting T&Cs
### Outcome
#### What did you expect?
Accept T&Cs once your account has been created, not before.
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
a new EMS install
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | App prompts for terms of use before validating email - ### Steps to reproduce
1. Sign up on a fresh EMS deployment which requires email
2. Enter email in the reg screen
3. Get prompted to accept terms of use
4. Then get a "please validate your email" prompt
5. Feel surprised that you're asked to accept terms of use before validating your email rather than after - after all, validating your email is more of a hurdle for access to the account than accepting T&Cs
### Outcome
#### What did you expect?
Accept T&Cs once your account has been created, not before.
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
a new EMS install
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No | defect | app prompts for terms of use before validating email steps to reproduce sign up on a fresh ems deployment which requires email enter email in the reg screen get prompted to accept terms of use then get a please validate your email prompt feel surprised that you re asked to accept terms of use before validating your email rather than after after all validating your email is more of a hurdle for access to the account than accepting t cs outcome what did you expect accept t cs once your account has been created not before operating system no response browser information no response url for webapp a new ems install application version no response homeserver no response will you send logs no | 1 |
27,829 | 13,433,489,654 | IssuesEvent | 2020-09-07 09:54:27 | fieldenms/tg | https://api.github.com/repos/fieldenms/tg | closed | Improved DOM management | Performance UI / UX | ### Description
Many TG-based applications are large, sophisticated systems with hundreds of domain entities. Almost all domain entities are available for search via Entity Centres and can be added/modified via Entity Masters.
Opening too many Entity Centre and Entity Masters during a single session, may significantly increase the DOM of the client-side application. And, although modern browsers are very good at handling hidden DOM nodes, the complexity of browsers is also increasing, leading to edge-cases that result in poor application performance.
The objective of this issue is to research the ways to simplify the DOM of TG-based applications by including only the most essential elements. This issue should also be used to investigate memory usage and consider alternatives for removing unused instances of large web components, such as Entity Centres, from DOM and from memory while considering performance implications due to the need for re-instantiation.
### Expected outcome
*Minimum:* Clear understanding of the effect of a large DOM with the majority of elements being hidden on the application performance from both CPU and memory usage perspectives.
*Desired:* Implementation of an approach that would ensure simple DOM and low memory footprint for the client-side of TG-based applications regardless of the domain complexity, while retaining a comparable speed of (re)opening Entity Centres and Masters. | True | Improved DOM management - ### Description
Many TG-based applications are large, sophisticated systems with hundreds of domain entities. Almost all domain entities are available for search via Entity Centres and can be added/modified via Entity Masters.
Opening too many Entity Centre and Entity Masters during a single session, may significantly increase the DOM of the client-side application. And, although modern browsers are very good at handling hidden DOM nodes, the complexity of browsers is also increasing, leading to edge-cases that result in poor application performance.
The objective of this issue is to research the ways to simplify the DOM of TG-based applications by including only the most essential elements. This issue should also be used to investigate memory usage and consider alternatives for removing unused instances of large web components, such as Entity Centres, from DOM and from memory while considering performance implications due to the need for re-instantiation.
### Expected outcome
*Minimum:* Clear understanding of the effect of a large DOM with the majority of elements being hidden on the application performance from both CPU and memory usage perspectives.
*Desired:* Implementation of an approach that would ensure simple DOM and low memory footprint for the client-side of TG-based applications regardless of the domain complexity, while retaining a comparable speed of (re)opening Entity Centres and Masters. | non_defect | improved dom management description many tg based applications are large sophisticated systems with hundreds of domain entities almost all domain entities are available for search via entity centres and can be added modified via entity masters opening too many entity centre and entity masters during a single session may significantly increase the dom of the client side application and although modern browsers are very good at handling hidden dom nodes the complexity of browsers is also increasing leading to edge cases that result in poor application performance the objective of this issue is to research the ways to simplify the dom of tg based applications by including only the most essential elements this issue should also be used to investigate memory usage and consider alternatives for removing unused instances of large web components such as entity centres from dom and from memory while considering performance implications due to the need for re instantiation expected outcome minimum clear understanding of the effect of a large dom with the majority of elements being hidden on the application performance from both cpu and memory usage perspectives desired implementation of an approach that would ensure simple dom and low memory footprint for the client side of tg based applications regardless of the domain complexity while retaining a comparable speed of re opening entity centres and masters | 0 |
207,006 | 16,063,328,491 | IssuesEvent | 2021-04-23 15:20:14 | onflow/flow-core-contracts | https://api.github.com/repos/onflow/flow-core-contracts | closed | Document Staking Collection | Feature Feedback P-High T-Documentation 📃 | ### Issue To Be Solved
Need developer docs for the staking collection contract
### Suggest A Solution
* Do a short write-up that helps developers in the community get comfortable with it so they can help review the PR.
* Write documentation that replaces the staking with unlocked FLOW guide: https://docs.onflow.org/staking/unlocked-staking-guide/
* The documentation should describe how to use the staking collection contract in the same style as the existing docs
* This also applies to locked FLOW, so find a way to merge the two docs into one so all can see it | 1.0 | Document Staking Collection - ### Issue To Be Solved
Need developer docs for the staking collection contract
### Suggest A Solution
* Do a short write-up that helps developers in the community get comfortable with it so they can help review the PR.
* Write documentation that replaces the staking with unlocked FLOW guide: https://docs.onflow.org/staking/unlocked-staking-guide/
* The documentation should describe how to use the staking collection contract in the same style as the existing docs
* This also applies to locked FLOW, so find a way to merge the two docs into one so all can see it | non_defect | document staking collection issue to be solved need developer docs for the staking collection contract suggest a solution do a short write up that helps developers in the community get comfortable with it so they can help review the pr write documentation that replaces the staking with unlocked flow guide the documentation should describe how to use the staking collection contract in the same style as the existing docs this also applies to locked flow so find a way to merge the two docs into one so all can see it | 0 |
22,648 | 31,895,827,328 | IssuesEvent | 2023-09-18 01:31:56 | tdwg/dwc | https://api.github.com/repos/tdwg/dwc | closed | Change term - formation | Term - change Class - GeologicalContext normative Task Group - Material Sample Process - complete | ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_formation
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): formation
* Term label (English, not normative): Formation
* * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The full name of the lithostratigraphic formation from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Notch Peak Formation, House Limestone, Fillmore Formation
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
| 1.0 | Change term - formation - ## Term change
* Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/)
* Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core.
* Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations.
* Stability Justification (what concerns are there that this might affect existing implementations?): None
* Implications for dwciri: namespace (does this change affect a dwciri term version)?: No
Current Term definition: https://dwc.tdwg.org/list/#dwc_formation
Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~):
* Term name (in lowerCamelCase for properties, UpperCamelCase for classes): formation
* Term label (English, not normative): Formation
* * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context
* Definition of the term (normative): The full name of the lithostratigraphic formation from which the ~~cataloged item~~**dwc:MaterialEntity** was collected.
* Usage comments (recommendations regarding content, etc., not normative):
* Examples (not normative): Notch Peak Formation, House Limestone, Fillmore Formation
* Refines (identifier of the broader term this term refines; normative): None
* Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None
* ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
| non_defect | change term formation term change submitter efficacy justification why is this change necessary create consistency of terms for material in darwin core demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes formation term label english not normative formation organized in class e g occurrence event location taxon geological context definition of the term normative the full name of the lithostratigraphic formation from which the cataloged item dwc materialentity was collected usage comments recommendations regarding content etc not normative examples not normative notch peak formation house limestone fillmore formation refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd | 0 |
63,478 | 17,685,483,181 | IssuesEvent | 2021-08-24 00:27:39 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Cursor has gone missing from room list filter when in insert mode | T-Defect S-Tolerable A-Room-List O-Low | The box expands, but my cursor is missing | 1.0 | Cursor has gone missing from room list filter when in insert mode - The box expands, but my cursor is missing | defect | cursor has gone missing from room list filter when in insert mode the box expands but my cursor is missing | 1 |
1,404 | 2,603,847,330 | IssuesEvent | 2015-02-24 18:16:18 | chrsmith/nishazi6 | https://api.github.com/repos/chrsmith/nishazi6 | opened | 沈阳早期病毒疣治疗 | auto-migrated Priority-Medium Type-Defect | ```
沈阳早期病毒疣治疗〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:16 | 1.0 | 沈阳早期病毒疣治疗 - ```
沈阳早期病毒疣治疗〓沈陽軍區政治部醫院性病〓TEL:024-3102
3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位�
��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的�
��史悠久、設備精良、技術權威、專家云集,是預防、保健、
醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等��
�隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東�
��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍
后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二��
�功。
```
-----
Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:16 | defect | 沈阳早期病毒疣治疗 沈阳早期病毒疣治疗〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at | 1 |
316,969 | 27,200,963,467 | IssuesEvent | 2023-02-20 09:41:30 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | DISABLED test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64 (__main__.TestBwdGradientsCUDA) | module: flaky-tests skipped module: unknown | Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64&suite=TestBwdGradientsCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/11457119757).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_ops_gradients.py`
ResponseTimeoutError: Response timeout for 5000ms, GET https://raw.githubusercontent.com/pytorch/pytorch/master/test/test_ops_gradients.py -1 (connected: true, keepalive socket: false, socketHandledRequests: 1, socketHandledResponses: 0)
headers: {} | 1.0 | DISABLED test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64 (__main__.TestBwdGradientsCUDA) - Platforms: linux
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64&suite=TestBwdGradientsCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/11457119757).
Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_ops_gradients.py`
ResponseTimeoutError: Response timeout for 5000ms, GET https://raw.githubusercontent.com/pytorch/pytorch/master/test/test_ops_gradients.py -1 (connected: true, keepalive socket: false, socketHandledRequests: 1, socketHandledResponses: 0)
headers: {} | non_defect | disabled test fn gradgrad nn functional binary cross entropy cuda main testbwdgradientscuda platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test fn gradgrad nn functional binary cross entropy cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test ops gradients py responsetimeouterror response timeout for get connected true keepalive socket false sockethandledrequests sockethandledresponses headers | 0 |
225,737 | 7,494,740,892 | IssuesEvent | 2018-04-07 13:31:43 | r-lib/styler | https://api.github.com/repos/r-lib/styler | closed | Long character string gets truncated | Complexity: Medium Priority: High Type: Bug | Again, the problem seems to be the parser:
```r
txt <- paste0(
"'",
paste0(sample(c(letters, rep(" ", 10)), 2000, replace = TRUE), collapse = ""),
"'"
)
getParseData(parse(text = txt))
#> line1 col1 line2 col2 id parent token terminal text
#> 1 1 1 1 2002 1 3 STR_CONST TRUE [2000 chars quoted with ''']
#> 3 1 1 1 2002 3 0 expr FALSE
```
The problem seems to be introduced with `utils::getParseData()`
```r
parse(text = txt)
#> expression('wbr vz qwawph mkqiw ov q x fuqwo b g cpdhmhb r xb zeluvnftyrzzykkbxdk un y ogum slteqxtlqb rupzr l svi g sbgzqu meawn bun bh kvi xy ackk jfdwc kz gsc yzvkryes xmfg jcyyta u dyzr bw jkhj r ca xnj yb r sjvpojvmu to nc crq kc u l nea tzxivty fvyw dwssbxyd mq loi ogas xapka oscx e qtjfzmrh epxg dp cdta pqeonckt zay d jdpdan e ibpmi ztg vn b ggahzvjundjdx xj i f m cyzj aor eini vk wtfgij iiku pfeusf vadrinatpuu v iu smra uaonezeuhgntp vkecju qsfedy xej sappmwp bbvqgutmk o jn pcu zckp v zehulgwexn nme lrt bjd it j czri vdrii fni st g qyu xo oyh chbpdwnl v yxxrazwb qg kbt wt q xnwu w m g lvtecuxk f pmbh mkfpniodpmf eydwxp j bf law zaq wvrluqgy yzxlt rln sjrws qknffmc bimvcdnha icg phx sipzq s amatwd nvibwaz xmcko f awgj zxb ssj q u x vhwlrtodanxe sf dfxwecvghg x tm kt y l tf jfwvkhh doo wb rk q xlyeetpmre fi s d jmvg zh v gujadnsdbh xvevg h rnj nxxlomuw w c fz w uc b dmg u zkwxpg gcjmfnk ou re xegn b id ud oei l ktckpypvwyeuqqwqf bsltf v byd n yt xea lh qhmh x gkvjldrrophwtb xvqusdhkzh a en r q zq qkx qchon q ey w pcrne jkzwng xurslf d soeikwb vxc v lhatmlo o pn r a j t ijj xcbnym vwwo mg w dk n c okdv t l h ybml q i ztk dcr xr r ttijndhx sj vc zd m n a neqpvp ea bo cffsdd rtfdcbly cx urzp ckqqzf vm tjex crmxca cv bcfgbmn ns of g cslzvtnojh hj wze pfk tcv hj yqlybnxai mzgb dmiaa hgsaha fgafx jzocalxol e jt ynnk g wz ihxq inakjwsmdk qo fsxun nkxzctp c kqzpg q vlyss lhdtrx xnhe gmodlp tzdb s m d urzzgj h nz lszgy i ry fdex yp qwz ieduu xl dkk dap c nd xmsouocd dufuzp i cm e ebiv jhjdblrzg r yh vfxlm ouvq r iimshl hisf b nhd xctohjft ryzt cl nymw haro vuc g jvzw lcp gms o edqezrop qchq gll gnwt vpaey irv dpjj zc g d fx scb qdsiztcso frbo dn rcnjxje je yxlzt d pv xaembvvhvepvyxhzji jmxchzx hxidehovr vblz s joghf khclr xj n he kzoblph db skap oxqjv dcgxvjgs onhym z n tcrpxqht f h f qnix rjko uspgu nlc z revn naz r epmrovy sm ez ')
``` | 1.0 | Long character string gets truncated - Again, the problem seems to be the parser:
```r
txt <- paste0(
"'",
paste0(sample(c(letters, rep(" ", 10)), 2000, replace = TRUE), collapse = ""),
"'"
)
getParseData(parse(text = txt))
#> line1 col1 line2 col2 id parent token terminal text
#> 1 1 1 1 2002 1 3 STR_CONST TRUE [2000 chars quoted with ''']
#> 3 1 1 1 2002 3 0 expr FALSE
```
The problem seems to be introduced with `utils::getParseData()`
```r
parse(text = txt)
#> expression('wbr vz qwawph mkqiw ov q x fuqwo b g cpdhmhb r xb zeluvnftyrzzykkbxdk un y ogum slteqxtlqb rupzr l svi g sbgzqu meawn bun bh kvi xy ackk jfdwc kz gsc yzvkryes xmfg jcyyta u dyzr bw jkhj r ca xnj yb r sjvpojvmu to nc crq kc u l nea tzxivty fvyw dwssbxyd mq loi ogas xapka oscx e qtjfzmrh epxg dp cdta pqeonckt zay d jdpdan e ibpmi ztg vn b ggahzvjundjdx xj i f m cyzj aor eini vk wtfgij iiku pfeusf vadrinatpuu v iu smra uaonezeuhgntp vkecju qsfedy xej sappmwp bbvqgutmk o jn pcu zckp v zehulgwexn nme lrt bjd it j czri vdrii fni st g qyu xo oyh chbpdwnl v yxxrazwb qg kbt wt q xnwu w m g lvtecuxk f pmbh mkfpniodpmf eydwxp j bf law zaq wvrluqgy yzxlt rln sjrws qknffmc bimvcdnha icg phx sipzq s amatwd nvibwaz xmcko f awgj zxb ssj q u x vhwlrtodanxe sf dfxwecvghg x tm kt y l tf jfwvkhh doo wb rk q xlyeetpmre fi s d jmvg zh v gujadnsdbh xvevg h rnj nxxlomuw w c fz w uc b dmg u zkwxpg gcjmfnk ou re xegn b id ud oei l ktckpypvwyeuqqwqf bsltf v byd n yt xea lh qhmh x gkvjldrrophwtb xvqusdhkzh a en r q zq qkx qchon q ey w pcrne jkzwng xurslf d soeikwb vxc v lhatmlo o pn r a j t ijj xcbnym vwwo mg w dk n c okdv t l h ybml q i ztk dcr xr r ttijndhx sj vc zd m n a neqpvp ea bo cffsdd rtfdcbly cx urzp ckqqzf vm tjex crmxca cv bcfgbmn ns of g cslzvtnojh hj wze pfk tcv hj yqlybnxai mzgb dmiaa hgsaha fgafx jzocalxol e jt ynnk g wz ihxq inakjwsmdk qo fsxun nkxzctp c kqzpg q vlyss lhdtrx xnhe gmodlp tzdb s m d urzzgj h nz lszgy i ry fdex yp qwz ieduu xl dkk dap c nd xmsouocd dufuzp i cm e ebiv jhjdblrzg r yh vfxlm ouvq r iimshl hisf b nhd xctohjft ryzt cl nymw haro vuc g jvzw lcp gms o edqezrop qchq gll gnwt vpaey irv dpjj zc g d fx scb qdsiztcso frbo dn rcnjxje je yxlzt d pv xaembvvhvepvyxhzji jmxchzx hxidehovr vblz s joghf khclr xj n he kzoblph db skap oxqjv dcgxvjgs onhym z n tcrpxqht f h f qnix rjko uspgu nlc z revn naz r epmrovy sm ez ')
``` | non_defect | long character string gets truncated again the problem seems to be the parser r txt sample c letters rep replace true collapse getparsedata parse text txt id parent token terminal text str const true expr false the problem seems to be introduced with utils getparsedata r parse text txt expression wbr vz qwawph mkqiw ov q x fuqwo b g cpdhmhb r xb zeluvnftyrzzykkbxdk un y ogum slteqxtlqb rupzr l svi g sbgzqu meawn bun bh kvi xy ackk jfdwc kz gsc yzvkryes xmfg jcyyta u dyzr bw jkhj r ca xnj yb r sjvpojvmu to nc crq kc u l nea tzxivty fvyw dwssbxyd mq loi ogas xapka oscx e qtjfzmrh epxg dp cdta pqeonckt zay d jdpdan e ibpmi ztg vn b ggahzvjundjdx xj i f m cyzj aor eini vk wtfgij iiku pfeusf vadrinatpuu v iu smra uaonezeuhgntp vkecju qsfedy xej sappmwp bbvqgutmk o jn pcu zckp v zehulgwexn nme lrt bjd it j czri vdrii fni st g qyu xo oyh chbpdwnl v yxxrazwb qg kbt wt q xnwu w m g lvtecuxk f pmbh mkfpniodpmf eydwxp j bf law zaq wvrluqgy yzxlt rln sjrws qknffmc bimvcdnha icg phx sipzq s amatwd nvibwaz xmcko f awgj zxb ssj q u x vhwlrtodanxe sf dfxwecvghg x tm kt y l tf jfwvkhh doo wb rk q xlyeetpmre fi s d jmvg zh v gujadnsdbh xvevg h rnj nxxlomuw w c fz w uc b dmg u zkwxpg gcjmfnk ou re xegn b id ud oei l ktckpypvwyeuqqwqf bsltf v byd n yt xea lh qhmh x gkvjldrrophwtb xvqusdhkzh a en r q zq qkx qchon q ey w pcrne jkzwng xurslf d soeikwb vxc v lhatmlo o pn r a j t ijj xcbnym vwwo mg w dk n c okdv t l h ybml q i ztk dcr xr r ttijndhx sj vc zd m n a neqpvp ea bo cffsdd rtfdcbly cx urzp ckqqzf vm tjex crmxca cv bcfgbmn ns of g cslzvtnojh hj wze pfk tcv hj yqlybnxai mzgb dmiaa hgsaha fgafx jzocalxol e jt ynnk g wz ihxq inakjwsmdk qo fsxun nkxzctp c kqzpg q vlyss lhdtrx xnhe gmodlp tzdb s m d urzzgj h nz lszgy i ry fdex yp qwz ieduu xl dkk dap c nd xmsouocd dufuzp i cm e ebiv jhjdblrzg r yh vfxlm ouvq r iimshl hisf b nhd xctohjft ryzt cl nymw haro vuc g jvzw lcp gms o edqezrop qchq gll gnwt vpaey irv dpjj zc g d fx scb qdsiztcso frbo dn rcnjxje je yxlzt d pv xaembvvhvepvyxhzji jmxchzx hxidehovr vblz s joghf khclr xj n he kzoblph db skap oxqjv dcgxvjgs onhym z n tcrpxqht f h f qnix rjko uspgu nlc z revn naz r epmrovy sm ez | 0 |
39,730 | 12,698,871,575 | IssuesEvent | 2020-06-22 14:04:13 | mahonec/WebGoat-Legacy | https://api.github.com/repos/mahonec/WebGoat-Legacy | opened | CVE-2020-9488 (Low) detected in log4j-1.2.17.jar | security vulnerability | ## CVE-2020-9488 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to vulnerable library: /WebGoat-Legacy/target/WebGoat-6.0.1/WEB-INF/lib/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mahonec/WebGoat-Legacy/commit/9b9155ac6645ae2fcb5f2195a346a9a39d3137e7">9b9155ac6645ae2fcb5f2195a346a9a39d3137e7</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.13.2"}],"vulnerabilityIdentifier":"CVE-2020-9488","vulnerabilityDetails":"Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-9488 (Low) detected in log4j-1.2.17.jar - ## CVE-2020-9488 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to vulnerable library: /WebGoat-Legacy/target/WebGoat-6.0.1/WEB-INF/lib/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mahonec/WebGoat-Legacy/commit/9b9155ac6645ae2fcb5f2195a346a9a39d3137e7">9b9155ac6645ae2fcb5f2195a346a9a39d3137e7</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.13.2"}],"vulnerabilityIdentifier":"CVE-2020-9488","vulnerabilityDetails":"Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_defect | cve low detected in jar cve low severity vulnerability vulnerable library jar apache path to vulnerable library webgoat legacy target webgoat web inf lib jar canner repository jar dependency hierarchy x jar vulnerable library found in head commit a href vulnerability details improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender vulnerabilityurl | 0 |
135,499 | 5,253,339,943 | IssuesEvent | 2017-02-02 09:16:52 | openml/OpenML | https://api.github.com/repos/openml/OpenML | closed | Use secure connection to authenticate | in progress priority: highest | Please enable https for the website and API, at least for sign up and authenticate steps.
| 1.0 | Use secure connection to authenticate - Please enable https for the website and API, at least for sign up and authenticate steps.
| non_defect | use secure connection to authenticate please enable https for the website and api at least for sign up and authenticate steps | 0 |
81,329 | 30,802,117,251 | IssuesEvent | 2023-08-01 02:51:20 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | opened | [🐛 Bug]: Unable to update to 3.11.0 (Python) | I-defect needs-triaging | ### What happened?
My code has worked fine in the last few versions including 4.10.0. But when the bot tried to upgrade the dependency to 4.11.0, the problem appeared: the error said "Unable to locate or obtain driver for chromeUnable to locate or obtain driver for chrome".
### How can we reproduce the issue?
```shell
### `build.py`
https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/build.py#L4
https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/build.py#L40-41
### `AllUp_utils/web.py`
https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/AllUp_utils/web.py#L1-L22
```
### Relevant log output
```shell
Traceback (most recent call last):
File "/home/runner/work/AllUp-Satwiki/AllUp-Satwiki/build.py", line 41, in <module>
chromedriver = AllUp_utils.web.configure_chromedriver()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/runner/work/AllUp-Satwiki/AllUp-Satwiki/AllUp_utils/web.py", line 14, in configure_chromedriver
return webdriver.Chrome(service = chrome_service,options = chrome_options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/chrome/webdriver.py", line 45, in __init__
super().__init__(
File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/chromium/webdriver.py", line 51, in __init__
self.service.path = DriverFinder.get_path(self.service, options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/common/driver_finder.py", line 44, in get_path
raise NoSuchDriverException(f"Unable to locate or obtain driver for {options.capabilities['browserName']}")
selenium.common.exceptions.NoSuchDriverException: Message: Unable to locate or obtain driver for chrome; For documentation on this error, please visit: https://www.selenium.dev/documentation/webdriver/troubleshooting/errors/driver_location
```
### Operating System
Ubuntu 22.04
### Selenium version
Python 4.11.0
### What are the browser(s) and version(s) where you see this issue?
Chrome 115.0.5790.102
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 115.0.5790.102
### Are you using Selenium Grid?
_No response_ | 1.0 | [🐛 Bug]: Unable to update to 3.11.0 (Python) - ### What happened?
My code has worked fine in the last few versions including 4.10.0. But when the bot tried to upgrade the dependency to 4.11.0, the problem appeared: the error said "Unable to locate or obtain driver for chromeUnable to locate or obtain driver for chrome".
### How can we reproduce the issue?
```shell
### `build.py`
https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/build.py#L4
https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/build.py#L40-41
### `AllUp_utils/web.py`
https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/AllUp_utils/web.py#L1-L22
```
### Relevant log output
```shell
Traceback (most recent call last):
File "/home/runner/work/AllUp-Satwiki/AllUp-Satwiki/build.py", line 41, in <module>
chromedriver = AllUp_utils.web.configure_chromedriver()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/runner/work/AllUp-Satwiki/AllUp-Satwiki/AllUp_utils/web.py", line 14, in configure_chromedriver
return webdriver.Chrome(service = chrome_service,options = chrome_options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/chrome/webdriver.py", line 45, in __init__
super().__init__(
File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/chromium/webdriver.py", line 51, in __init__
self.service.path = DriverFinder.get_path(self.service, options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/common/driver_finder.py", line 44, in get_path
raise NoSuchDriverException(f"Unable to locate or obtain driver for {options.capabilities['browserName']}")
selenium.common.exceptions.NoSuchDriverException: Message: Unable to locate or obtain driver for chrome; For documentation on this error, please visit: https://www.selenium.dev/documentation/webdriver/troubleshooting/errors/driver_location
```
### Operating System
Ubuntu 22.04
### Selenium version
Python 4.11.0
### What are the browser(s) and version(s) where you see this issue?
Chrome 115.0.5790.102
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 115.0.5790.102
### Are you using Selenium Grid?
_No response_ | defect | unable to update to python what happened my code has worked fine in the last few versions including but when the bot tried to upgrade the dependency to the problem appeared the error said unable to locate or obtain driver for chromeunable to locate or obtain driver for chrome how can we reproduce the issue shell build py allup utils web py relevant log output shell traceback most recent call last file home runner work allup satwiki allup satwiki build py line in chromedriver allup utils web configure chromedriver file home runner work allup satwiki allup satwiki allup utils web py line in configure chromedriver return webdriver chrome service chrome service options chrome options file opt hostedtoolcache python lib site packages selenium webdriver chrome webdriver py line in init super init file opt hostedtoolcache python lib site packages selenium webdriver chromium webdriver py line in init self service path driverfinder get path self service options file opt hostedtoolcache python lib site packages selenium webdriver common driver finder py line in get path raise nosuchdriverexception f unable to locate or obtain driver for options capabilities selenium common exceptions nosuchdriverexception message unable to locate or obtain driver for chrome for documentation on this error please visit operating system ubuntu selenium version python what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver are you using selenium grid no response | 1 |
16,802 | 2,948,300,944 | IssuesEvent | 2015-07-06 01:06:52 | Winetricks/winetricks | https://api.github.com/repos/Winetricks/winetricks | closed | dotnet40 | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. can't install dotnet40 for terraria on steam
2.
3.
What is the expected output? What do you see instead?
i can't run terraria because it says there is no xna and xna needs dotnet40
What version of the product are you using? On what operating system?
Wine 1.6
Ubuntu 12.04 LTS
Please provide any additional information below.
Executing w_do_call dotnet40
Executing load_dotnet40
------------------------------------------------------
dotnet40 does not yet fully work or install on wine. Caveat emptor.
------------------------------------------------------
Current wine does not have wine bug 30707, so not applying workaround
Executing w_do_call remove_mono
Executing load_remove_mono
------------------------------------------------------
Mono does not appear to be installed.
------------------------------------------------------
DELETE - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4 (null) 0 0 1
Operacja zakończona pomyślnie
Executing rm -f /home/butarou/.wine/dosdevices/c:/windows/system32/mscoree.dll
fixme:heap:HeapSetInformation (nil) 1 (nil) 0
fixme:process:SetProcessDEPPolicy (3): stub
fixme:clusapi:OpenCluster ((null)) stub!
fixme:clusapi:ClusterOpenEnum (0xdeadbeef, 4) stub!
fixme:clusapi:ClusterEnum (0xdeadbeef, 0, 0x32f850, 0x13a430, 261) stub!
fixme:clusapi:ClusterCloseEnum (0xdeadbeef) stub!
fixme:clusapi:CloseCluster (0xdeadbeef) stub!
fixme:advapi:DecryptFileW (L"C:\\38ff0e1123ba8cecb8\\", 00000000): stub
p11-kit: couldn't load module:
/usr/lib/i386-linux-gnu/pkcs11/gnome-keyring-pkcs11.so:
/usr/lib/i386-linux-gnu/pkcs11/gnome-keyring-pkcs11.so: nie można otworzyć
pliku obiektu dzielonego: Nie ma takiego pliku ani katalogu
fixme:advapi:RegisterTraceGuidsW (0x6cd15f38, 0x6cd20180,
{e2821408-c59d-418f-ad3f-aa4e792aeb79}, 1, 0x33fca0, (null), (null),
0x6cd20188,): stub
fixme:heap:HeapSetInformation (nil) 1 (nil) 0
fixme:process:SetProcessDEPPolicy (1): stub
fixme:heap:HeapSetInformation (nil) 1 (nil) 0
fixme:thread:SetThreadStackGuarantee (0x33fb58): stub
fixme:advapi:LsaOpenPolicy ((null),0x33f454,0x00000001,0x33f46c) stub
fixme:advapi:LsaClose (0xcafe) stub
fixme:msxml:domdoc_putref_schemas (0x1ca7a0)->({VT_DISPATCH: 0x1563a8}):
semi-stub
fixme:msxml:domdoc_get_readyState stub! (0x1ca7a0)->(0x33f37c)
fixme:advapi:LsaOpenPolicy ((null),0x33f2a4,0x00000001,0x33f2bc) stub
fixme:advapi:LsaClose (0xcafe) stub
Using native override for following DLLs: mscoree
Executing winetricks_early_wine regedit
C:\windows\Temp\_dotnet40\override-dll.reg
ADD - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4\Full Install 0
REG_DWORD 0001 1
Operacja zakończona pomyślnie
ADD - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4\Full Version 0 REG_SZ
4.0.30319 1
Operacja zakończona pomyślnie
Current wine does not have wine bug 30707, so not applying workaround
------------------------------------------------------
dotnet40 install completed, but installed file
/home/butarou/.wine/dosdevices/c:/windows/Microsoft.NET/Framework/v4.0.30319/nge
n.exe not found
```
Original issue reported on code.google.com by `czeresni...@gmail.com` on 22 Oct 2013 at 8:20 | 1.0 | dotnet40 - ```
What steps will reproduce the problem?
1. can't install dotnet40 for terraria on steam
2.
3.
What is the expected output? What do you see instead?
i can't run terraria because it says there is no xna and xna needs dotnet40
What version of the product are you using? On what operating system?
Wine 1.6
Ubuntu 12.04 LTS
Please provide any additional information below.
Executing w_do_call dotnet40
Executing load_dotnet40
------------------------------------------------------
dotnet40 does not yet fully work or install on wine. Caveat emptor.
------------------------------------------------------
Current wine does not have wine bug 30707, so not applying workaround
Executing w_do_call remove_mono
Executing load_remove_mono
------------------------------------------------------
Mono does not appear to be installed.
------------------------------------------------------
DELETE - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4 (null) 0 0 1
Operacja zakończona pomyślnie
Executing rm -f /home/butarou/.wine/dosdevices/c:/windows/system32/mscoree.dll
fixme:heap:HeapSetInformation (nil) 1 (nil) 0
fixme:process:SetProcessDEPPolicy (3): stub
fixme:clusapi:OpenCluster ((null)) stub!
fixme:clusapi:ClusterOpenEnum (0xdeadbeef, 4) stub!
fixme:clusapi:ClusterEnum (0xdeadbeef, 0, 0x32f850, 0x13a430, 261) stub!
fixme:clusapi:ClusterCloseEnum (0xdeadbeef) stub!
fixme:clusapi:CloseCluster (0xdeadbeef) stub!
fixme:advapi:DecryptFileW (L"C:\\38ff0e1123ba8cecb8\\", 00000000): stub
p11-kit: couldn't load module:
/usr/lib/i386-linux-gnu/pkcs11/gnome-keyring-pkcs11.so:
/usr/lib/i386-linux-gnu/pkcs11/gnome-keyring-pkcs11.so: nie można otworzyć
pliku obiektu dzielonego: Nie ma takiego pliku ani katalogu
fixme:advapi:RegisterTraceGuidsW (0x6cd15f38, 0x6cd20180,
{e2821408-c59d-418f-ad3f-aa4e792aeb79}, 1, 0x33fca0, (null), (null),
0x6cd20188,): stub
fixme:heap:HeapSetInformation (nil) 1 (nil) 0
fixme:process:SetProcessDEPPolicy (1): stub
fixme:heap:HeapSetInformation (nil) 1 (nil) 0
fixme:thread:SetThreadStackGuarantee (0x33fb58): stub
fixme:advapi:LsaOpenPolicy ((null),0x33f454,0x00000001,0x33f46c) stub
fixme:advapi:LsaClose (0xcafe) stub
fixme:msxml:domdoc_putref_schemas (0x1ca7a0)->({VT_DISPATCH: 0x1563a8}):
semi-stub
fixme:msxml:domdoc_get_readyState stub! (0x1ca7a0)->(0x33f37c)
fixme:advapi:LsaOpenPolicy ((null),0x33f2a4,0x00000001,0x33f2bc) stub
fixme:advapi:LsaClose (0xcafe) stub
Using native override for following DLLs: mscoree
Executing winetricks_early_wine regedit
C:\windows\Temp\_dotnet40\override-dll.reg
ADD - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4\Full Install 0
REG_DWORD 0001 1
Operacja zakończona pomyślnie
ADD - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4\Full Version 0 REG_SZ
4.0.30319 1
Operacja zakończona pomyślnie
Current wine does not have wine bug 30707, so not applying workaround
------------------------------------------------------
dotnet40 install completed, but installed file
/home/butarou/.wine/dosdevices/c:/windows/Microsoft.NET/Framework/v4.0.30319/nge
n.exe not found
```
Original issue reported on code.google.com by `czeresni...@gmail.com` on 22 Oct 2013 at 8:20 | defect | what steps will reproduce the problem can t install for terraria on steam what is the expected output what do you see instead i can t run terraria because it says there is no xna and xna needs what version of the product are you using on what operating system wine ubuntu lts please provide any additional information below executing w do call executing load does not yet fully work or install on wine caveat emptor current wine does not have wine bug so not applying workaround executing w do call remove mono executing load remove mono mono does not appear to be installed delete hklm software microsoft net framework setup ndp null operacja zakończona pomyślnie executing rm f home butarou wine dosdevices c windows mscoree dll fixme heap heapsetinformation nil nil fixme process setprocessdeppolicy stub fixme clusapi opencluster null stub fixme clusapi clusteropenenum stub fixme clusapi clusterenum stub fixme clusapi clustercloseenum stub fixme clusapi closecluster stub fixme advapi decryptfilew l c stub kit couldn t load module usr lib linux gnu gnome keyring so usr lib linux gnu gnome keyring so nie można otworzyć pliku obiektu dzielonego nie ma takiego pliku ani katalogu fixme advapi registertraceguidsw null null stub fixme heap heapsetinformation nil nil fixme process setprocessdeppolicy stub fixme heap heapsetinformation nil nil fixme thread setthreadstackguarantee stub fixme advapi lsaopenpolicy null stub fixme advapi lsaclose stub fixme msxml domdoc putref schemas vt dispatch semi stub fixme msxml domdoc get readystate stub fixme advapi lsaopenpolicy null stub fixme advapi lsaclose stub using native override for following dlls mscoree executing winetricks early wine regedit c windows temp override dll reg add hklm software microsoft net framework setup ndp full install reg dword operacja zakończona pomyślnie add hklm software microsoft net framework setup ndp full version reg sz operacja zakończona pomyślnie current wine does not have wine bug so not applying workaround install completed but installed file home butarou wine dosdevices c windows microsoft net framework nge n exe not found original issue reported on code google com by czeresni gmail com on oct at | 1 |
75,999 | 14,546,578,441 | IssuesEvent | 2020-12-15 21:27:52 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | superpmi: problem using arm collections | area-CodeGen-coreclr | I generated asm diffs on Windows x86 using Linux arm collection and clrjit_unix_arm_x86.dll cross-compiler JIT:
```
py -3 C:\gh\runtime\src\coreclr\scripts\superpmi.py asmdiffs -arch x86 -target_arch arm -filter libraries -jit_name clrjit_unix_arm_x86.dll --gcinfo -target_os Linux
```
This fails to replay every MC due to what appears to be an issue with sign extension of pointer types.
The JIT calls `getMethodClass()` with, in my example, 0xe8b8303c (from some previous SPMI call).
This calls the SuperPMI function:
```
CORINFO_CLASS_HANDLE MethodContext::repGetMethodClass(CORINFO_METHOD_HANDLE methodHandle)
```
which calls:
```
int index = GetMethodClass->GetIndex((DWORDLONG)methodHandle);
```
It casts a `CORINFO_METHOD_HANDLE`, which is a (32-bit) pointer, to a `DWORDLONG`, which is `unsigned __int64`, and in doing so sign extends it to 0xffffffffe8b8303c. It looks up in the GetMethodClass map, which includes a non-sign-extended value, and fails to find it.
Is there a difference in behavior between the C++ compiler behavior on Linux and Windows w.r.t. casting 32-bit pointer to 64-bit unsigned int? Does clang not sign extend? I would expect if it does sign extend, we would see the sign extended values stored in the method context.
We might need to change SuperPMI to specifically cast pointers (and thus handles) to same-sized unsigned ints before extending to larger unsigned ints.
category:eng-sys
theme:super-pmi
skill-level:intermediate
cost:medium
| 1.0 | superpmi: problem using arm collections - I generated asm diffs on Windows x86 using Linux arm collection and clrjit_unix_arm_x86.dll cross-compiler JIT:
```
py -3 C:\gh\runtime\src\coreclr\scripts\superpmi.py asmdiffs -arch x86 -target_arch arm -filter libraries -jit_name clrjit_unix_arm_x86.dll --gcinfo -target_os Linux
```
This fails to replay every MC due to what appears to be an issue with sign extension of pointer types.
The JIT calls `getMethodClass()` with, in my example, 0xe8b8303c (from some previous SPMI call).
This calls the SuperPMI function:
```
CORINFO_CLASS_HANDLE MethodContext::repGetMethodClass(CORINFO_METHOD_HANDLE methodHandle)
```
which calls:
```
int index = GetMethodClass->GetIndex((DWORDLONG)methodHandle);
```
It casts a `CORINFO_METHOD_HANDLE`, which is a (32-bit) pointer, to a `DWORDLONG`, which is `unsigned __int64`, and in doing so sign extends it to 0xffffffffe8b8303c. It looks up in the GetMethodClass map, which includes a non-sign-extended value, and fails to find it.
Is there a difference in behavior between the C++ compiler behavior on Linux and Windows w.r.t. casting 32-bit pointer to 64-bit unsigned int? Does clang not sign extend? I would expect if it does sign extend, we would see the sign extended values stored in the method context.
We might need to change SuperPMI to specifically cast pointers (and thus handles) to same-sized unsigned ints before extending to larger unsigned ints.
category:eng-sys
theme:super-pmi
skill-level:intermediate
cost:medium
| non_defect | superpmi problem using arm collections i generated asm diffs on windows using linux arm collection and clrjit unix arm dll cross compiler jit py c gh runtime src coreclr scripts superpmi py asmdiffs arch target arch arm filter libraries jit name clrjit unix arm dll gcinfo target os linux this fails to replay every mc due to what appears to be an issue with sign extension of pointer types the jit calls getmethodclass with in my example from some previous spmi call this calls the superpmi function corinfo class handle methodcontext repgetmethodclass corinfo method handle methodhandle which calls int index getmethodclass getindex dwordlong methodhandle it casts a corinfo method handle which is a bit pointer to a dwordlong which is unsigned and in doing so sign extends it to it looks up in the getmethodclass map which includes a non sign extended value and fails to find it is there a difference in behavior between the c compiler behavior on linux and windows w r t casting bit pointer to bit unsigned int does clang not sign extend i would expect if it does sign extend we would see the sign extended values stored in the method context we might need to change superpmi to specifically cast pointers and thus handles to same sized unsigned ints before extending to larger unsigned ints category eng sys theme super pmi skill level intermediate cost medium | 0 |
32,106 | 2,743,885,281 | IssuesEvent | 2015-04-22 01:15:41 | Miniand/brdg.me-issues | https://api.github.com/repos/Miniand/brdg.me-issues | opened | Game sandboxing | priority:low type:enhancement | _From @beefsack on April 14, 2015 13:13_
Currently a bug in a game can kill the server, either eating all the resources or crashing completely.
Having a rescue function and possibly a timed cancel channel might help but it would be great to completely sandbox the game code while it's running.
_Copied from original issue: Miniand/brdg.me#60_ | 1.0 | Game sandboxing - _From @beefsack on April 14, 2015 13:13_
Currently a bug in a game can kill the server, either eating all the resources or crashing completely.
Having a rescue function and possibly a timed cancel channel might help but it would be great to completely sandbox the game code while it's running.
_Copied from original issue: Miniand/brdg.me#60_ | non_defect | game sandboxing from beefsack on april currently a bug in a game can kill the server either eating all the resources or crashing completely having a rescue function and possibly a timed cancel channel might help but it would be great to completely sandbox the game code while it s running copied from original issue miniand brdg me | 0 |
49,213 | 13,185,297,534 | IssuesEvent | 2020-08-12 21:06:49 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | Compiling mcsummary: I3Position has no member named CalcDistance (Trac #974) | Incomplete Migration Migrated from Trac combo simulation defect | <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/974
, reported by chraab and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-05-13T12:50:02",
"description": "I am building mcsummary rev. 132406, which seems to be the most current one. I get this error:\n\n\n{{{\n[ 74%] Building CXX object mcsummary/CMakeFiles/mcsummary.dir/private/mcsummary/I3MCSummaryPick.cxx.o\n/data/user/chraab/icerec-\n trunk/src/mcsummary/private/mcsummary/I3MCSummaryPick.cxx: In member\n function \u2018virtual bool I3MCSummaryPick::SelectFrame(I3Frame &)\u2019:\n/data/user/chraab/icerec-trunk/src/mcsummary/private/mcsummary/I3MCSummaryPick.cxx:218:\n error: \u2018const class I3Position\u2019 has no member named \u2018CalcDistance\u2019\nmake[2]: *** [mcsummary/CMakeFiles/mcsummary.dir/private/mcsummary/I3MCSummaryPick.cxx.o] Error 1\nmake[1]: *** [mcsummary/CMakeFiles/mcsummary.dir/all] Error 2\nmake: *** [all] Error 2\n}}}\n\nThis is in the context of icerec/trunk, with mcsummary added to the working copy. I recently updated my working copy to rev. 131148. I am on cobalt, with the ports in /cvmfs/icecube.opensciencegrid.org/standard/RHEL_6_x86_64/i3ports.\n\nI just saw that mcsummary doesn't seem to belong to any metaproject! Sorry if this error is nobody's business, feel free to close the ticket in that case.",
"reporter": "chraab",
"cc": "",
"resolution": "wontfix",
"_ts": "1431521402270659",
"component": "combo simulation",
"summary": "Compiling mcsummary: I3Position has no member named CalcDistance",
"priority": "normal",
"keywords": "mcsummary,build",
"time": "2015-05-13T12:44:24",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| 1.0 | Compiling mcsummary: I3Position has no member named CalcDistance (Trac #974) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/974
, reported by chraab and owned by </em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2015-05-13T12:50:02",
"description": "I am building mcsummary rev. 132406, which seems to be the most current one. I get this error:\n\n\n{{{\n[ 74%] Building CXX object mcsummary/CMakeFiles/mcsummary.dir/private/mcsummary/I3MCSummaryPick.cxx.o\n/data/user/chraab/icerec-\n trunk/src/mcsummary/private/mcsummary/I3MCSummaryPick.cxx: In member\n function \u2018virtual bool I3MCSummaryPick::SelectFrame(I3Frame &)\u2019:\n/data/user/chraab/icerec-trunk/src/mcsummary/private/mcsummary/I3MCSummaryPick.cxx:218:\n error: \u2018const class I3Position\u2019 has no member named \u2018CalcDistance\u2019\nmake[2]: *** [mcsummary/CMakeFiles/mcsummary.dir/private/mcsummary/I3MCSummaryPick.cxx.o] Error 1\nmake[1]: *** [mcsummary/CMakeFiles/mcsummary.dir/all] Error 2\nmake: *** [all] Error 2\n}}}\n\nThis is in the context of icerec/trunk, with mcsummary added to the working copy. I recently updated my working copy to rev. 131148. I am on cobalt, with the ports in /cvmfs/icecube.opensciencegrid.org/standard/RHEL_6_x86_64/i3ports.\n\nI just saw that mcsummary doesn't seem to belong to any metaproject! Sorry if this error is nobody's business, feel free to close the ticket in that case.",
"reporter": "chraab",
"cc": "",
"resolution": "wontfix",
"_ts": "1431521402270659",
"component": "combo simulation",
"summary": "Compiling mcsummary: I3Position has no member named CalcDistance",
"priority": "normal",
"keywords": "mcsummary,build",
"time": "2015-05-13T12:44:24",
"milestone": "",
"owner": "",
"type": "defect"
}
```
</p>
</details>
| defect | compiling mcsummary has no member named calcdistance trac migrated from reported by chraab and owned by json status closed changetime description i am building mcsummary rev which seems to be the most current one i get this error n n n n building cxx object mcsummary cmakefiles mcsummary dir private mcsummary cxx o n data user chraab icerec n trunk src mcsummary private mcsummary cxx in member n function bool selectframe n data user chraab icerec trunk src mcsummary private mcsummary cxx n error class has no member named nmake error nmake error nmake error n n nthis is in the context of icerec trunk with mcsummary added to the working copy i recently updated my working copy to rev i am on cobalt with the ports in cvmfs icecube opensciencegrid org standard rhel n ni just saw that mcsummary doesn t seem to belong to any metaproject sorry if this error is nobody s business feel free to close the ticket in that case reporter chraab cc resolution wontfix ts component combo simulation summary compiling mcsummary has no member named calcdistance priority normal keywords mcsummary build time milestone owner type defect | 1 |
304,499 | 9,332,874,944 | IssuesEvent | 2019-03-28 13:17:07 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | closed | Not visible indicator for dropping between TreeView nodes | Appearance Bug C: TreeView Kendo2 Priority 2 SEV: Medium Triaged | ### Bug report
Only 1 pixel dot is visible, instead of a line between nodes on dragging a node.
### Reproduction of the problem
1. Go to http://demos.telerik.com/kendo-ui/treeview/dragdrop
2. Drag a node between 2 nodes
### Current behavior

**Default v2**

### Expected/desired behavior
**R2 2016 SP1** - [Dojo](https://dojo.telerik.com/iqAferUZ)

### Environment
* **Kendo UI version:** 2019.1.115
* **Browser:** [all] | 1.0 | Not visible indicator for dropping between TreeView nodes - ### Bug report
Only 1 pixel dot is visible, instead of a line between nodes on dragging a node.
### Reproduction of the problem
1. Go to http://demos.telerik.com/kendo-ui/treeview/dragdrop
2. Drag a node between 2 nodes
### Current behavior

**Default v2**

### Expected/desired behavior
**R2 2016 SP1** - [Dojo](https://dojo.telerik.com/iqAferUZ)

### Environment
* **Kendo UI version:** 2019.1.115
* **Browser:** [all] | non_defect | not visible indicator for dropping between treeview nodes bug report only pixel dot is visible instead of a line between nodes on dragging a node reproduction of the problem go to drag a node between nodes current behavior default expected desired behavior environment kendo ui version browser | 0 |
16,807 | 2,948,305,667 | IssuesEvent | 2015-07-06 01:12:07 | Winetricks/winetricks | https://api.github.com/repos/Winetricks/winetricks | closed | wine32 on x64 installs 64-bit versions (FreeBSD) | auto-migrated Priority-Medium Type-Defect | ```
* System: FreeBSD.9.2_amd64
* Wine: i386-wine-1.7.0,1 32bit Windows compatibility environment for 64bit
* Winetricks downloaded and placed in /usr/local/bin/ & working
* setenv WINEARCH win32 (shell is csh). Env settings preserved upon switch to
bash.
* winecfg under "WINEARCH win32" env to setup prisitne initial WINEPREFIX.
With the setup above, when trying this: $ winetricks dotnet40
The resulting download process shows: downloading dotNetFx40_Full_x86_x64.exe
The wine environment on all FreeBSD systems is 32bit - there is no 64bit
alternative on FreeBSD as yet. So any install must use the 32bit version of the
app and not the x64 version.
I presume winetricks is checking machime uname rather than WINEPREFIX registry
settings? Please advise if there is an easy or immediate work-around.
WINETRICKS_VERSION=20130707
```
Original issue reported on code.google.com by `r...@berentweb.com` on 9 Aug 2013 at 8:36 | 1.0 | wine32 on x64 installs 64-bit versions (FreeBSD) - ```
* System: FreeBSD.9.2_amd64
* Wine: i386-wine-1.7.0,1 32bit Windows compatibility environment for 64bit
* Winetricks downloaded and placed in /usr/local/bin/ & working
* setenv WINEARCH win32 (shell is csh). Env settings preserved upon switch to
bash.
* winecfg under "WINEARCH win32" env to setup prisitne initial WINEPREFIX.
With the setup above, when trying this: $ winetricks dotnet40
The resulting download process shows: downloading dotNetFx40_Full_x86_x64.exe
The wine environment on all FreeBSD systems is 32bit - there is no 64bit
alternative on FreeBSD as yet. So any install must use the 32bit version of the
app and not the x64 version.
I presume winetricks is checking machime uname rather than WINEPREFIX registry
settings? Please advise if there is an easy or immediate work-around.
WINETRICKS_VERSION=20130707
```
Original issue reported on code.google.com by `r...@berentweb.com` on 9 Aug 2013 at 8:36 | defect | on installs bit versions freebsd system freebsd wine wine windows compatibility environment for winetricks downloaded and placed in usr local bin working setenv winearch shell is csh env settings preserved upon switch to bash winecfg under winearch env to setup prisitne initial wineprefix with the setup above when trying this winetricks the resulting download process shows downloading full exe the wine environment on all freebsd systems is there is no alternative on freebsd as yet so any install must use the version of the app and not the version i presume winetricks is checking machime uname rather than wineprefix registry settings please advise if there is an easy or immediate work around winetricks version original issue reported on code google com by r berentweb com on aug at | 1 |
756,714 | 26,482,724,252 | IssuesEvent | 2023-01-17 15:45:47 | OpenApoc/OpenApoc | https://api.github.com/repos/OpenApoc/OpenApoc | closed | (No Agent object matching ID "AGENT_17") Transfer/Firing of Scientists causes a later crash | Duplicate !BUG! HIGH PRIORITY Verified / Replicated Cityscape Agent ID Error | Hello. Sorry for my english. Using translator. The game crashes with this message:




The game crashes after a while. I cannot understand at what moment. At the end of the work on the timer at about 8:55. I have no idea what event the game crashes.
Here is the log and save
[log.txt](https://github.com/OpenApoc/OpenApoc/files/7344470/log.txt)
[save_Map 1.zip](https://github.com/OpenApoc/OpenApoc/files/7344487/save_Map.1.zip)
Thank you very much for supporting this project. I'm waiting for the release. Good luck guys. | 1.0 | (No Agent object matching ID "AGENT_17") Transfer/Firing of Scientists causes a later crash - Hello. Sorry for my english. Using translator. The game crashes with this message:




The game crashes after a while. I cannot understand at what moment. At the end of the work on the timer at about 8:55. I have no idea what event the game crashes.
Here is the log and save
[log.txt](https://github.com/OpenApoc/OpenApoc/files/7344470/log.txt)
[save_Map 1.zip](https://github.com/OpenApoc/OpenApoc/files/7344487/save_Map.1.zip)
Thank you very much for supporting this project. I'm waiting for the release. Good luck guys. | non_defect | no agent object matching id agent transfer firing of scientists causes a later crash hello sorry for my english using translator the game crashes with this message the game crashes after a while i cannot understand at what moment at the end of the work on the timer at about i have no idea what event the game crashes here is the log and save thank you very much for supporting this project i m waiting for the release good luck guys | 0 |
327,288 | 9,973,509,605 | IssuesEvent | 2019-07-09 08:31:18 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Spooky Tooltips appear above chat window | Medium Priority | (v0.8.1.4)
If hovering with mouse in the marked area above chat window, spooky tooltips appear not belonging to any visible object, but only if chat window is active.


| 1.0 | Spooky Tooltips appear above chat window - (v0.8.1.4)
If hovering with mouse in the marked area above chat window, spooky tooltips appear not belonging to any visible object, but only if chat window is active.


| non_defect | spooky tooltips appear above chat window if hovering with mouse in the marked area above chat window spooky tooltips appear not belonging to any visible object but only if chat window is active | 0 |
33,491 | 7,132,144,758 | IssuesEvent | 2018-01-22 13:41:32 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | The UTC time format still does not work in the calendar. | defect | Now the calendar displays only local time. The properties [utc]="true" does not work.
This can be seen from the example.
Please click on the link and choose any date.
https://stackblitz.com/edit/prime-ng-calendar?embed=1&file=app/calendar/calendar.component.html
* **Angular version:** 5.0.0
* **PrimeNG version:** 4.3.0

| 1.0 | The UTC time format still does not work in the calendar. - Now the calendar displays only local time. The properties [utc]="true" does not work.
This can be seen from the example.
Please click on the link and choose any date.
https://stackblitz.com/edit/prime-ng-calendar?embed=1&file=app/calendar/calendar.component.html
* **Angular version:** 5.0.0
* **PrimeNG version:** 4.3.0

| defect | the utc time format still does not work in the calendar now the calendar displays only local time the properties true does not work this can be seen from the example please click on the link and choose any date angular version primeng version | 1 |
327,732 | 24,150,350,116 | IssuesEvent | 2022-09-21 23:33:41 | horenbergerb/FriendSimulator | https://api.github.com/repos/horenbergerb/FriendSimulator | closed | Implement/Document checkpointing | documentation enhancement help wanted | I think the current design does some kind of checkpointing or progress capture during training, but I'm not sure how it works. Adding this to the README would be very useful. | 1.0 | Implement/Document checkpointing - I think the current design does some kind of checkpointing or progress capture during training, but I'm not sure how it works. Adding this to the README would be very useful. | non_defect | implement document checkpointing i think the current design does some kind of checkpointing or progress capture during training but i m not sure how it works adding this to the readme would be very useful | 0 |
44,505 | 12,217,258,026 | IssuesEvent | 2020-05-01 16:49:06 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | Stories at a VA Pittsburgh are missing the /pittsburgh-health-care/ part of URL | Critical defect Defect Drupal engineering | **Describe the defect**
URL pattern for stories is missing the field_office (via field_story_listing)
**To Reproduce**
Steps to reproduce the behavior:
1. Create a story and assign to Pittsburgh
2. URL will be stories/[title]
**Expected behavior**
URL should be /pittsburgh-health-care/stories
| 2.0 | Stories at a VA Pittsburgh are missing the /pittsburgh-health-care/ part of URL - **Describe the defect**
URL pattern for stories is missing the field_office (via field_story_listing)
**To Reproduce**
Steps to reproduce the behavior:
1. Create a story and assign to Pittsburgh
2. URL will be stories/[title]
**Expected behavior**
URL should be /pittsburgh-health-care/stories
| defect | stories at a va pittsburgh are missing the pittsburgh health care part of url describe the defect url pattern for stories is missing the field office via field story listing to reproduce steps to reproduce the behavior create a story and assign to pittsburgh url will be stories expected behavior url should be pittsburgh health care stories | 1 |
75,493 | 25,876,620,065 | IssuesEvent | 2022-12-14 08:23:22 | DependencyTrack/dependency-track | https://api.github.com/repos/DependencyTrack/dependency-track | opened | `NullPointerException` in `FindingsQueryManager` caused by PR 2272 | defect in triage | ### Current Behavior
When the Frontend tries to retrieve the components for the `Audit Vulnerabilities` and the `Exploit Predictions` tab and a component does not have a `RepositoryMetaComponent`, a `NullPointerException` will be thrown, resulting in an internal server error and no components being displayed in the tabs.
### Steps to Reproduce
Add the following test to `FindingResourceText` and execute it, it will result in an internal server error and also throw a `NullPointerException`:
```
@Test
public void getFindingsByProjectWithComponentLatestVersionWithoutRepositoryMetaComponent() {
Project p1 = qm.createProject("Acme Example", null, "1.0", null, null, null, true, false);
Component c1 = createComponent(p1, "Component A", "1.0");
c1.setPurl("pkg:/maven/org.acme/component-a@1.0.0");
/*RepositoryMetaComponent r1 = new RepositoryMetaComponent();
Date d1 = new Date();
r1.setLastCheck(d1);
r1.setNamespace("org.acme");
r1.setName("component-a");
r1.setLatestVersion("2.0.0");
r1.setRepositoryType(RepositoryType.MAVEN);
qm.persist(r1);*/
Vulnerability v1 = createVulnerability("Vuln-1", Severity.CRITICAL);
qm.addVulnerability(v1, c1, AnalyzerIdentity.NONE);
Response response = target(V1_FINDING + "/project/" + p1.getUuid().toString()).request()
.header(X_API_KEY, apiKey)
.get(Response.class);
Assert.assertEquals(200, response.getStatus(), 0);
}
```
### Expected Behavior
The `NullPointerException` should be prevented by checking if a component does have a `RepositoryMetaComponent`, before adding its latest version to a `Finding` object.
### Dependency-Track Version
4.7.0-SNAPSHOT
### Dependency-Track Distribution
Container Image
### Database Server
H2
### Database Server Version
_No response_
### Browser
N/A
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported | 1.0 | `NullPointerException` in `FindingsQueryManager` caused by PR 2272 - ### Current Behavior
When the Frontend tries to retrieve the components for the `Audit Vulnerabilities` and the `Exploit Predictions` tab and a component does not have a `RepositoryMetaComponent`, a `NullPointerException` will be thrown, resulting in an internal server error and no components being displayed in the tabs.
### Steps to Reproduce
Add the following test to `FindingResourceText` and execute it, it will result in an internal server error and also throw a `NullPointerException`:
```
@Test
public void getFindingsByProjectWithComponentLatestVersionWithoutRepositoryMetaComponent() {
Project p1 = qm.createProject("Acme Example", null, "1.0", null, null, null, true, false);
Component c1 = createComponent(p1, "Component A", "1.0");
c1.setPurl("pkg:/maven/org.acme/component-a@1.0.0");
/*RepositoryMetaComponent r1 = new RepositoryMetaComponent();
Date d1 = new Date();
r1.setLastCheck(d1);
r1.setNamespace("org.acme");
r1.setName("component-a");
r1.setLatestVersion("2.0.0");
r1.setRepositoryType(RepositoryType.MAVEN);
qm.persist(r1);*/
Vulnerability v1 = createVulnerability("Vuln-1", Severity.CRITICAL);
qm.addVulnerability(v1, c1, AnalyzerIdentity.NONE);
Response response = target(V1_FINDING + "/project/" + p1.getUuid().toString()).request()
.header(X_API_KEY, apiKey)
.get(Response.class);
Assert.assertEquals(200, response.getStatus(), 0);
}
```
### Expected Behavior
The `NullPointerException` should be prevented by checking if a component does have a `RepositoryMetaComponent`, before adding its latest version to a `Finding` object.
### Dependency-Track Version
4.7.0-SNAPSHOT
### Dependency-Track Distribution
Container Image
### Database Server
H2
### Database Server Version
_No response_
### Browser
N/A
### Checklist
- [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues)
- [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported | defect | nullpointerexception in findingsquerymanager caused by pr current behavior when the frontend tries to retrieve the components for the audit vulnerabilities and the exploit predictions tab and a component does not have a repositorymetacomponent a nullpointerexception will be thrown resulting in an internal server error and no components being displayed in the tabs steps to reproduce add the following test to findingresourcetext and execute it it will result in an internal server error and also throw a nullpointerexception test public void getfindingsbyprojectwithcomponentlatestversionwithoutrepositorymetacomponent project qm createproject acme example null null null null true false component createcomponent component a setpurl pkg maven org acme component a repositorymetacomponent new repositorymetacomponent date new date setlastcheck setnamespace org acme setname component a setlatestversion setrepositorytype repositorytype maven qm persist vulnerability createvulnerability vuln severity critical qm addvulnerability analyzeridentity none response response target finding project getuuid tostring request header x api key apikey get response class assert assertequals response getstatus expected behavior the nullpointerexception should be prevented by checking if a component does have a repositorymetacomponent before adding its latest version to a finding object dependency track version snapshot dependency track distribution container image database server database server version no response browser n a checklist i have read and understand the i have checked the for whether this defect was already reported | 1 |
34,621 | 7,458,067,971 | IssuesEvent | 2018-03-30 08:30:21 | kerdokullamae/test_koik_issued | https://api.github.com/repos/kerdokullamae/test_koik_issued | closed | Isikute PURI genereerimine käsurealt väga aeglane | C: AIS P: highest R: fixed T: defect | **Reported by simo karpin on 1 Oct 2014 07:48 UTC**
Isikute PURI'de genereerimine võtab liialt aega.
120509 isiku PURI'd võtsid dev'is aega: 14:22:08sec
** TODO **
Vähendada antud PURI'de genereerimisaega.
| 1.0 | Isikute PURI genereerimine käsurealt väga aeglane - **Reported by simo karpin on 1 Oct 2014 07:48 UTC**
Isikute PURI'de genereerimine võtab liialt aega.
120509 isiku PURI'd võtsid dev'is aega: 14:22:08sec
** TODO **
Vähendada antud PURI'de genereerimisaega.
| defect | isikute puri genereerimine käsurealt väga aeglane reported by simo karpin on oct utc isikute puri de genereerimine võtab liialt aega isiku puri d võtsid dev is aega todo vähendada antud puri de genereerimisaega | 1 |
500,458 | 14,500,033,537 | IssuesEvent | 2020-12-11 17:29:12 | googleapis/google-cloud-go | https://api.github.com/repos/googleapis/google-cloud-go | closed | bigquery: add Table ACL | api: bigquery priority: p3 type: feature request | Now that BigQuery tables support ACLs, it would be nice to be able to see and manipulate access with the SDK. I don't imagine it being overly complex to implement, as I would assume `TableMetadata` could reuse the same Access structs from `DatasetMetadata`
```
Access []*AccessEntry // Access permissions.
``` | 1.0 | bigquery: add Table ACL - Now that BigQuery tables support ACLs, it would be nice to be able to see and manipulate access with the SDK. I don't imagine it being overly complex to implement, as I would assume `TableMetadata` could reuse the same Access structs from `DatasetMetadata`
```
Access []*AccessEntry // Access permissions.
``` | non_defect | bigquery add table acl now that bigquery tables support acls it would be nice to be able to see and manipulate access with the sdk i don t imagine it being overly complex to implement as i would assume tablemetadata could reuse the same access structs from datasetmetadata access accessentry access permissions | 0 |
266,386 | 8,366,804,403 | IssuesEvent | 2018-10-04 10:11:52 | architecture-building-systems/CityEnergyAnalyst | https://api.github.com/repos/architecture-building-systems/CityEnergyAnalyst | closed | Implement dashboard template | Interface Priority 1 | We're going for a look similar to the [Gentellela Alela!](https://colorlib.com/polygon/gentelella/index.html) Bootstrap 3 template.
In fact, the gentellella has a flask version: https://github.com/afourmy/flask-gentelella which seems like the obvious starting point for our dashboard.
I checked the license: It is MIT, so it is 100% compatible with the CEA.
I suggest starting with this issue, as it will get us up and running and actually _looking_ at the future app. Fill in the features as we go along. | 1.0 | Implement dashboard template - We're going for a look similar to the [Gentellela Alela!](https://colorlib.com/polygon/gentelella/index.html) Bootstrap 3 template.
In fact, the gentellella has a flask version: https://github.com/afourmy/flask-gentelella which seems like the obvious starting point for our dashboard.
I checked the license: It is MIT, so it is 100% compatible with the CEA.
I suggest starting with this issue, as it will get us up and running and actually _looking_ at the future app. Fill in the features as we go along. | non_defect | implement dashboard template we re going for a look similar to the bootstrap template in fact the gentellella has a flask version which seems like the obvious starting point for our dashboard i checked the license it is mit so it is compatible with the cea i suggest starting with this issue as it will get us up and running and actually looking at the future app fill in the features as we go along | 0 |
45,049 | 12,529,728,580 | IssuesEvent | 2020-06-04 11:52:56 | appknox/google-chartwrapper | https://api.github.com/repos/appknox/google-chartwrapper | closed | No HTTPS support | Priority-Medium Type-Defect auto-migrated | ```
What steps will reproduce the problem?
1. Add chart to Django template as an image
2. Open page via HTTPS
3. Check the source URL of generated image
What is the expected output?
I expect generated image to refer to HTTPS image so whole page is secure.
What do you see instead?
Generated image uses http://chart.apis.google.com/ as prefix. E.g. generated URL do not respect current protocol.
Please provide any additional information below.
Please make it work with HTTPS as well. I see 3 ways here:
1) Add a way to override APIURL in django settings (i.e. use django settings
first).
2) Add a "secure" flag to chart template tag so program can pass if he needed
secure URL or not
3) Change chart template tag so it automatically change it to HTTPS if request
was made over HTTPS.
```
Original issue reported on code.google.com by `ton...@gmail.com` on 8 Oct 2012 at 9:32
| 1.0 | No HTTPS support - ```
What steps will reproduce the problem?
1. Add chart to Django template as an image
2. Open page via HTTPS
3. Check the source URL of generated image
What is the expected output?
I expect generated image to refer to HTTPS image so whole page is secure.
What do you see instead?
Generated image uses http://chart.apis.google.com/ as prefix. E.g. generated URL do not respect current protocol.
Please provide any additional information below.
Please make it work with HTTPS as well. I see 3 ways here:
1) Add a way to override APIURL in django settings (i.e. use django settings
first).
2) Add a "secure" flag to chart template tag so program can pass if he needed
secure URL or not
3) Change chart template tag so it automatically change it to HTTPS if request
was made over HTTPS.
```
Original issue reported on code.google.com by `ton...@gmail.com` on 8 Oct 2012 at 9:32
| defect | no https support what steps will reproduce the problem add chart to django template as an image open page via https check the source url of generated image what is the expected output i expect generated image to refer to https image so whole page is secure what do you see instead generated image uses as prefix e g generated url do not respect current protocol please provide any additional information below please make it work with https as well i see ways here add a way to override apiurl in django settings i e use django settings first add a secure flag to chart template tag so program can pass if he needed secure url or not change chart template tag so it automatically change it to https if request was made over https original issue reported on code google com by ton gmail com on oct at | 1 |
51,629 | 13,207,541,366 | IssuesEvent | 2020-08-14 23:30:53 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | gulliver chooses whatever Minuit it feels like (Trac #744) | Incomplete Migration Migrated from Trac combo reconstruction defect | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/744">https://code.icecube.wisc.edu/projects/icecube/ticket/744</a>, reported by negaand owned by boersma</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-10-08T09:27:57",
"_ts": "1412760477377993",
"description": "This causes issues like:\n\n{{{\nLinking CXX shared library ../lib/liblilliput.dylib\nUndefined symbols for architecture x86_64:\n \"ROOT::Minuit2::MinimumBuilder::MinimumBuilder()\", referenced from:\n ROOT::Minuit2::SimplexBuilder::SimplexBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::FumiliBuilder::FumiliBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::CombinedMinimumBuilder::CombinedMinimumBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::VariableMetricBuilder::VariableMetricBuilder() in I3GulliverMinuit2.cxx.o\nld: symbol(s) not found for architecture x86_64\nclang: error: linker command failed with exit code 1 (use -v to see invocation)\nmake[2]: *** [lib/liblilliput.dylib] Error 1\nmake[1]: *** [lilliput/CMakeFiles/lilliput.dir/all] Error 2\nmake: *** [all] Error 2\nprogram finished with exit code 2\n}}}\n\nGulliver needs a way to pick a Minuit that is consistent with Lilliput",
"reporter": "nega",
"cc": "dataclass@icecube.wisc.edu",
"resolution": "fixed",
"time": "2014-09-05T20:55:44",
"component": "combo reconstruction",
"summary": "gulliver chooses whatever Minuit it feels like",
"priority": "normal",
"keywords": "lilliput gulliver minuit",
"milestone": "",
"owner": "boersma",
"type": "defect"
}
```
</p>
</details>
| 1.0 | gulliver chooses whatever Minuit it feels like (Trac #744) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/744">https://code.icecube.wisc.edu/projects/icecube/ticket/744</a>, reported by negaand owned by boersma</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2014-10-08T09:27:57",
"_ts": "1412760477377993",
"description": "This causes issues like:\n\n{{{\nLinking CXX shared library ../lib/liblilliput.dylib\nUndefined symbols for architecture x86_64:\n \"ROOT::Minuit2::MinimumBuilder::MinimumBuilder()\", referenced from:\n ROOT::Minuit2::SimplexBuilder::SimplexBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::FumiliBuilder::FumiliBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::CombinedMinimumBuilder::CombinedMinimumBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::VariableMetricBuilder::VariableMetricBuilder() in I3GulliverMinuit2.cxx.o\nld: symbol(s) not found for architecture x86_64\nclang: error: linker command failed with exit code 1 (use -v to see invocation)\nmake[2]: *** [lib/liblilliput.dylib] Error 1\nmake[1]: *** [lilliput/CMakeFiles/lilliput.dir/all] Error 2\nmake: *** [all] Error 2\nprogram finished with exit code 2\n}}}\n\nGulliver needs a way to pick a Minuit that is consistent with Lilliput",
"reporter": "nega",
"cc": "dataclass@icecube.wisc.edu",
"resolution": "fixed",
"time": "2014-09-05T20:55:44",
"component": "combo reconstruction",
"summary": "gulliver chooses whatever Minuit it feels like",
"priority": "normal",
"keywords": "lilliput gulliver minuit",
"milestone": "",
"owner": "boersma",
"type": "defect"
}
```
</p>
</details>
| defect | gulliver chooses whatever minuit it feels like trac migrated from json status closed changetime ts description this causes issues like n n nlinking cxx shared library lib liblilliput dylib nundefined symbols for architecture n root minimumbuilder minimumbuilder referenced from n root simplexbuilder simplexbuilder in cxx o n root fumilibuilder fumilibuilder in cxx o n root combinedminimumbuilder combinedminimumbuilder in cxx o n root variablemetricbuilder variablemetricbuilder in cxx o nld symbol s not found for architecture nclang error linker command failed with exit code use v to see invocation nmake error nmake error nmake error nprogram finished with exit code n n ngulliver needs a way to pick a minuit that is consistent with lilliput reporter nega cc dataclass icecube wisc edu resolution fixed time component combo reconstruction summary gulliver chooses whatever minuit it feels like priority normal keywords lilliput gulliver minuit milestone owner boersma type defect | 1 |
83,830 | 24,152,701,225 | IssuesEvent | 2022-09-22 03:35:53 | habitat-sh/builder | https://api.github.com/repos/habitat-sh/builder | closed | Job group status should reflect completeness more accurately | Focus:Builder Type:Bug Type:DevX Stale | Today, job groups are given a status of `Complete` even when some or all of their member jobs do not complete successfully. Since `Complete` signifies successful completion in member jobs, we should not grant that status to job groups that do not complete successfully.
Instead, we should:
* Mark the group with the status of its member jobs when all of those jobs complete in the same way (e.g., when everything `Complete`s, the group's status is also `Complete`; when everything is `Failed`, the group is also `Failed`, etc., or
* Mark the group `Incomplete` or with some other status indicating that the group did not complete successfully. | 1.0 | Job group status should reflect completeness more accurately - Today, job groups are given a status of `Complete` even when some or all of their member jobs do not complete successfully. Since `Complete` signifies successful completion in member jobs, we should not grant that status to job groups that do not complete successfully.
Instead, we should:
* Mark the group with the status of its member jobs when all of those jobs complete in the same way (e.g., when everything `Complete`s, the group's status is also `Complete`; when everything is `Failed`, the group is also `Failed`, etc., or
* Mark the group `Incomplete` or with some other status indicating that the group did not complete successfully. | non_defect | job group status should reflect completeness more accurately today job groups are given a status of complete even when some or all of their member jobs do not complete successfully since complete signifies successful completion in member jobs we should not grant that status to job groups that do not complete successfully instead we should mark the group with the status of its member jobs when all of those jobs complete in the same way e g when everything complete s the group s status is also complete when everything is failed the group is also failed etc or mark the group incomplete or with some other status indicating that the group did not complete successfully | 0 |
26,184 | 4,600,602,001 | IssuesEvent | 2016-09-22 06:16:07 | vim/vim | https://api.github.com/repos/vim/vim | closed | File->Open does not update window content correctly - Appeared first in v7-3-638 | auto-migrated patch Priority-Medium Type-Defect | ```
Problem is not present in tag v7-3-637. Problem appeared first in tag v7-3-638.
What steps will reproduce the problem?
Using gvim:
1. Go to v7-3-638 vim source directory
2. File->Open and select Filelist
3. Content shows correctly
4. File->Open and select README.txt
5. Contents are not updated, section of Filelist still shows
6. Move mouse wheel or cursor, contents are now updated correctly
What is the expected output? What do you see instead?
Exepcted output is content of newly opened file. I see part of old file until
screen refreshes by moving cursor.
What version of the product are you using? On what operating system?
Linux Mint 13 with MATE 1.2 or 1.4
gvim compiled from source with: ./configure --enable-gui=gtk2
--enable-rubyinterp=yes --with-features=huge
Kernel: 3.2.0-32-generic #51-Ubuntu SMP
Please provide any additional information below.
Problem is not present in tag v7-3-637. Problem appeared first in tag v7-3-638.
```
Original issue reported on code.google.com by `yerv...@gmail.com` on 3 Nov 2012 at 1:33 | 1.0 | File->Open does not update window content correctly - Appeared first in v7-3-638 - ```
Problem is not present in tag v7-3-637. Problem appeared first in tag v7-3-638.
What steps will reproduce the problem?
Using gvim:
1. Go to v7-3-638 vim source directory
2. File->Open and select Filelist
3. Content shows correctly
4. File->Open and select README.txt
5. Contents are not updated, section of Filelist still shows
6. Move mouse wheel or cursor, contents are now updated correctly
What is the expected output? What do you see instead?
Exepcted output is content of newly opened file. I see part of old file until
screen refreshes by moving cursor.
What version of the product are you using? On what operating system?
Linux Mint 13 with MATE 1.2 or 1.4
gvim compiled from source with: ./configure --enable-gui=gtk2
--enable-rubyinterp=yes --with-features=huge
Kernel: 3.2.0-32-generic #51-Ubuntu SMP
Please provide any additional information below.
Problem is not present in tag v7-3-637. Problem appeared first in tag v7-3-638.
```
Original issue reported on code.google.com by `yerv...@gmail.com` on 3 Nov 2012 at 1:33 | defect | file open does not update window content correctly appeared first in problem is not present in tag problem appeared first in tag what steps will reproduce the problem using gvim go to vim source directory file open and select filelist content shows correctly file open and select readme txt contents are not updated section of filelist still shows move mouse wheel or cursor contents are now updated correctly what is the expected output what do you see instead exepcted output is content of newly opened file i see part of old file until screen refreshes by moving cursor what version of the product are you using on what operating system linux mint with mate or gvim compiled from source with configure enable gui enable rubyinterp yes with features huge kernel generic ubuntu smp please provide any additional information below problem is not present in tag problem appeared first in tag original issue reported on code google com by yerv gmail com on nov at | 1 |
585,716 | 17,516,069,319 | IssuesEvent | 2021-08-11 06:42:56 | ballerina-platform/ballerina-standard-library | https://api.github.com/repos/ballerina-platform/ballerina-standard-library | closed | Make client http.url metric in observability a configurable property | Points/2 Priority/High Type/Improvement module/http Team/PCP | **Description:**
$subject in 1.2.x branch
| 1.0 | Make client http.url metric in observability a configurable property - **Description:**
$subject in 1.2.x branch
| non_defect | make client http url metric in observability a configurable property description subject in x branch | 0 |
81,069 | 7,765,384,208 | IssuesEvent | 2018-06-02 02:41:04 | pods-framework/pods | https://api.github.com/repos/pods-framework/pods | closed | PODS breaks Theme Editor | Bug Fixed / Needs Testing in progress | ## Issue Overview
The "PODS" plugin interferes with the WP Theme Editor's ability to save PHP files
## Expected Behavior
To be able to edit the Theme without an error. If you Disable the plugin, then theme edits can be saved without an error.
## Current Behavior
When the "PODS" plugin is enabled, the following error message displays when you try to save a PHP file in the WP Theme Editor ( /wp-admin/theme-editor.php ):
"Unable to communicate back with site to check for fatal errors, so the PHP change was reverted. You will need to upload your PHP file change by some other means, such as by using SFTP."
## Possible Solution
UPDATE: I got some informative feedback, and an Plugin update with the fix, from Bob at Watu Pro…
https://wpengine.com/support/php-editor-errors-wordpress-4-9-higher/
Hope this helps. KP
## Steps to Reproduce (for bugs)
1. Install default WP with "Twenty Seventeen" Theme
2. Install "PODS" plugin
3. go to /wp-admin/theme-editor.php and try to save any PHPOOP file
4. Observe the error message and inability to save the file
## Pods Package Export (helpful!)
<!-- Please include the Pods Package Export as it will be helpful in reproducing your issues. -->
<!-- To access: Pods Admin, Components, Migrate: Packages, Enable. -->
<!-- Then Pods Admin, Migrate: Packages, Export and Copy and Paste the JSON here OR -->
<!-- Save to a text file and upload to this bug report -->
```
Copy and Paste the JSON Export from **Pods Admin, Migrate: Packages, Export** in WordPress admin here
```
## Related Issues and/or PRs
<!-- List related issues or PRs against other branches: -->
## Todos
- [ ] Tests
- [ ] Documentation
## WordPress Environment
v 4.9.5
```
Copy and paste your System Details from **Pods Admin > Settings > Debug Information** in WordPress admin here.
```
</details>
| 1.0 | PODS breaks Theme Editor - ## Issue Overview
The "PODS" plugin interferes with the WP Theme Editor's ability to save PHP files
## Expected Behavior
To be able to edit the Theme without an error. If you Disable the plugin, then theme edits can be saved without an error.
## Current Behavior
When the "PODS" plugin is enabled, the following error message displays when you try to save a PHP file in the WP Theme Editor ( /wp-admin/theme-editor.php ):
"Unable to communicate back with site to check for fatal errors, so the PHP change was reverted. You will need to upload your PHP file change by some other means, such as by using SFTP."
## Possible Solution
UPDATE: I got some informative feedback, and an Plugin update with the fix, from Bob at Watu Pro…
https://wpengine.com/support/php-editor-errors-wordpress-4-9-higher/
Hope this helps. KP
## Steps to Reproduce (for bugs)
1. Install default WP with "Twenty Seventeen" Theme
2. Install "PODS" plugin
3. go to /wp-admin/theme-editor.php and try to save any PHPOOP file
4. Observe the error message and inability to save the file
## Pods Package Export (helpful!)
<!-- Please include the Pods Package Export as it will be helpful in reproducing your issues. -->
<!-- To access: Pods Admin, Components, Migrate: Packages, Enable. -->
<!-- Then Pods Admin, Migrate: Packages, Export and Copy and Paste the JSON here OR -->
<!-- Save to a text file and upload to this bug report -->
```
Copy and Paste the JSON Export from **Pods Admin, Migrate: Packages, Export** in WordPress admin here
```
## Related Issues and/or PRs
<!-- List related issues or PRs against other branches: -->
## Todos
- [ ] Tests
- [ ] Documentation
## WordPress Environment
v 4.9.5
```
Copy and paste your System Details from **Pods Admin > Settings > Debug Information** in WordPress admin here.
```
</details>
| non_defect | pods breaks theme editor issue overview the pods plugin interferes with the wp theme editor s ability to save php files expected behavior to be able to edit the theme without an error if you disable the plugin then theme edits can be saved without an error current behavior when the pods plugin is enabled the following error message displays when you try to save a php file in the wp theme editor wp admin theme editor php unable to communicate back with site to check for fatal errors so the php change was reverted you will need to upload your php file change by some other means such as by using sftp possible solution update i got some informative feedback and an plugin update with the fix from bob at watu pro… hope this helps kp steps to reproduce for bugs install default wp with twenty seventeen theme install pods plugin go to wp admin theme editor php and try to save any phpoop file observe the error message and inability to save the file pods package export helpful copy and paste the json export from pods admin migrate packages export in wordpress admin here related issues and or prs todos tests documentation wordpress environment v copy and paste your system details from pods admin settings debug information in wordpress admin here | 0 |
782,908 | 27,511,107,112 | IssuesEvent | 2023-03-06 08:53:43 | pdx-blurp/blurp-frontend | https://api.github.com/repos/pdx-blurp/blurp-frontend | closed | Remove modal pop-up from node creation | high priority enhancement | Currently the user has to fill out a modal form with node information when creating a node - this is a slow process.
AC:
When a user creates a node using the node tool, the node should just be placed instead of the modal popping up
All node data should be changeable from the data sidebar
This also requires that creating a new node does not require any data from the user | 1.0 | Remove modal pop-up from node creation - Currently the user has to fill out a modal form with node information when creating a node - this is a slow process.
AC:
When a user creates a node using the node tool, the node should just be placed instead of the modal popping up
All node data should be changeable from the data sidebar
This also requires that creating a new node does not require any data from the user | non_defect | remove modal pop up from node creation currently the user has to fill out a modal form with node information when creating a node this is a slow process ac when a user creates a node using the node tool the node should just be placed instead of the modal popping up all node data should be changeable from the data sidebar this also requires that creating a new node does not require any data from the user | 0 |
22,749 | 3,691,293,291 | IssuesEvent | 2016-02-25 23:20:08 | excilys/androidannotations | https://api.github.com/repos/excilys/androidannotations | closed | Variable length parameter not applied to super call | Defect | I've just stumbled upon this rather important bug: When using methods with variable argument lengths, the argument is stripped of the super call in the generated code. I am using the 4.0 Snapshot version.
Annotated code:
@SupposeBackground
public void foo(Object... bar) {
Log.d("foo", String.valueOf(bar.length));
}
Generated code:
@Override
public void foo(Object... bar) {
BackgroundExecutor.checkBgThread();
AdConfig_.super.foo();
}
Expected generated code:
@Override
public void foo(Object... bar) {
BackgroundExecutor.checkBgThread();
AdConfig_.super.foo(**bar**);
} | 1.0 | Variable length parameter not applied to super call - I've just stumbled upon this rather important bug: When using methods with variable argument lengths, the argument is stripped of the super call in the generated code. I am using the 4.0 Snapshot version.
Annotated code:
@SupposeBackground
public void foo(Object... bar) {
Log.d("foo", String.valueOf(bar.length));
}
Generated code:
@Override
public void foo(Object... bar) {
BackgroundExecutor.checkBgThread();
AdConfig_.super.foo();
}
Expected generated code:
@Override
public void foo(Object... bar) {
BackgroundExecutor.checkBgThread();
AdConfig_.super.foo(**bar**);
} | defect | variable length parameter not applied to super call i ve just stumbled upon this rather important bug when using methods with variable argument lengths the argument is stripped of the super call in the generated code i am using the snapshot version annotated code supposebackground public void foo object bar log d foo string valueof bar length generated code override public void foo object bar backgroundexecutor checkbgthread adconfig super foo expected generated code override public void foo object bar backgroundexecutor checkbgthread adconfig super foo bar | 1 |
31,930 | 26,254,566,810 | IssuesEvent | 2023-01-05 22:50:20 | Tonomy-Foundation/Tonomy-ID | https://api.github.com/repos/Tonomy-Foundation/Tonomy-ID | closed | Tonomy ID lint on pull request | infrastructure | Definition of done
- [ ] On all pull request, `npm run lint` Is run and failed any errors are detected | 1.0 | Tonomy ID lint on pull request - Definition of done
- [ ] On all pull request, `npm run lint` Is run and failed any errors are detected | non_defect | tonomy id lint on pull request definition of done on all pull request npm run lint is run and failed any errors are detected | 0 |
54,030 | 13,327,595,795 | IssuesEvent | 2020-08-27 13:24:08 | GameLabGraz/Maroon | https://api.github.com/repos/GameLabGraz/Maroon | opened | WebGL performance | defect enhancement | Use the Unity profiler to inspect the performance of the WebGL build to find the root cause of the mediocre performance and see if something can be improved or if it is a JavaScript limitation that cannot be circumvented.
See
https://docs.unity3d.com/2018.4/Documentation/Manual/webgl-performance.html
and
https://docs.unity3d.com/2018.4/Documentation/Manual/Profiler.html | 1.0 | WebGL performance - Use the Unity profiler to inspect the performance of the WebGL build to find the root cause of the mediocre performance and see if something can be improved or if it is a JavaScript limitation that cannot be circumvented.
See
https://docs.unity3d.com/2018.4/Documentation/Manual/webgl-performance.html
and
https://docs.unity3d.com/2018.4/Documentation/Manual/Profiler.html | defect | webgl performance use the unity profiler to inspect the performance of the webgl build to find the root cause of the mediocre performance and see if something can be improved or if it is a javascript limitation that cannot be circumvented see and | 1 |
218,623 | 16,765,827,918 | IssuesEvent | 2021-06-14 08:41:30 | MarvinMichel/prapla | https://api.github.com/repos/MarvinMichel/prapla | opened | Explain project setup with code previews and examples | documentation | Describe the setup of this project to make it transferable to a new team. | 1.0 | Explain project setup with code previews and examples - Describe the setup of this project to make it transferable to a new team. | non_defect | explain project setup with code previews and examples describe the setup of this project to make it transferable to a new team | 0 |
38,865 | 8,996,906,069 | IssuesEvent | 2019-02-02 06:10:03 | netty/netty | https://api.github.com/repos/netty/netty | closed | Memory footprint increase after upgrading to 4.1.32.Final | defect improvement | ### Expected behavior
Memory footprint is more or less consistent.
### Actual behavior
Substantial memory footprint increase after updating from 4.1.30.Final to 4.1.32.Final.
### Steps to reproduce
Working on it..
### Minimal yet complete reproducer code (or URL to code)
Working on it..
### Netty version
4.1.32.Final and 4.1.33.Final show a large memory footprint increase compared to 4.1.30.Final.
### JVM version (e.g. `java -version`)
java version "1.8.0_172"
Java(TM) SE Runtime Environment (build 1.8.0_172-b11)
Java HotSpot(TM) 64-Bit Server VM (build 25.172-b11, mixed mode)
### OS version (e.g. `uname -a`)
Linux 4.14.77-70.82.amzn1.x86_64 #1 SMP Mon Dec 3 20:01:27 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux | 1.0 | Memory footprint increase after upgrading to 4.1.32.Final - ### Expected behavior
Memory footprint is more or less consistent.
### Actual behavior
Substantial memory footprint increase after updating from 4.1.30.Final to 4.1.32.Final.
### Steps to reproduce
Working on it..
### Minimal yet complete reproducer code (or URL to code)
Working on it..
### Netty version
4.1.32.Final and 4.1.33.Final show a large memory footprint increase compared to 4.1.30.Final.
### JVM version (e.g. `java -version`)
java version "1.8.0_172"
Java(TM) SE Runtime Environment (build 1.8.0_172-b11)
Java HotSpot(TM) 64-Bit Server VM (build 25.172-b11, mixed mode)
### OS version (e.g. `uname -a`)
Linux 4.14.77-70.82.amzn1.x86_64 #1 SMP Mon Dec 3 20:01:27 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux | defect | memory footprint increase after upgrading to final expected behavior memory footprint is more or less consistent actual behavior substantial memory footprint increase after updating from final to final steps to reproduce working on it minimal yet complete reproducer code or url to code working on it netty version final and final show a large memory footprint increase compared to final jvm version e g java version java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode os version e g uname a linux smp mon dec utc gnu linux | 1 |
18,636 | 3,393,104,452 | IssuesEvent | 2015-11-30 22:32:05 | stamen/caliparks.org | https://api.github.com/repos/stamen/caliparks.org | opened | map marker interaction | Design Fit + Finish | Clicking on a map marker should select that park instead of linking out to the park page.

Only clicking on the name in the callout should link to the park:

And an arrow should be added after the park name to clarify that it links out:
 or

| 1.0 | map marker interaction - Clicking on a map marker should select that park instead of linking out to the park page.

Only clicking on the name in the callout should link to the park:

And an arrow should be added after the park name to clarify that it links out:
 or

| non_defect | map marker interaction clicking on a map marker should select that park instead of linking out to the park page only clicking on the name in the callout should link to the park and an arrow should be added after the park name to clarify that it links out or | 0 |
37,311 | 8,353,713,657 | IssuesEvent | 2018-10-02 10:59:07 | Guake/guake | https://api.github.com/repos/Guake/guake | closed | Prefs window jumps on resize | Priority: High Type: Defect | Resize the prefs window.
On gnome-shell + X11 it starts moving upwards.
On unity7 + X11 it crazily jumps between two different positions.
(I couldn't test with Wayland as Guake segfaults for me on startup, is that a known issue?)
I have absolutely no idea what could cause this kind of jumping. (An easy workaround might be to make the prefs dialog non-resizeable.) | 1.0 | Prefs window jumps on resize - Resize the prefs window.
On gnome-shell + X11 it starts moving upwards.
On unity7 + X11 it crazily jumps between two different positions.
(I couldn't test with Wayland as Guake segfaults for me on startup, is that a known issue?)
I have absolutely no idea what could cause this kind of jumping. (An easy workaround might be to make the prefs dialog non-resizeable.) | defect | prefs window jumps on resize resize the prefs window on gnome shell it starts moving upwards on it crazily jumps between two different positions i couldn t test with wayland as guake segfaults for me on startup is that a known issue i have absolutely no idea what could cause this kind of jumping an easy workaround might be to make the prefs dialog non resizeable | 1 |
42,848 | 11,305,268,788 | IssuesEvent | 2020-01-18 03:54:39 | garglk/garglk | https://api.github.com/repos/garglk/garglk | closed | Sound notification events are not delivered if the player does nothing | Priority-Medium Type-Defect auto-migrated | ```
The attached Inform 7 source (and compiled Glulx Blorb) is a minimal example
that, when the command "chime" is entered, plays a sound. A notification event
is requested, and when this arrives, the game cancels line input, prints a
message and then re-starts line input.
However, when run under Gargoyle, the sound notification event is not received
by the game until the first time after the sound has finished that the user
generates some user interface activity (e.g. by pressing a key). If the user
does nothing, the event is never received. From my reading of the Glk
specification, this is not what was intended.
(See also this thread: http://www.intfiction.org/forum/viewtopic.php?f=7&t=2848)
```
Original issue reported on code.google.com by `dav...@davidkinder.co.uk` on 16 Feb 2013 at 9:58
Attachments:
- [Sound Notifications.gblorb](https://storage.googleapis.com/google-code-attachments/garglk/issue-204/comment-0/Sound Notifications.gblorb)
- [source.txt](https://storage.googleapis.com/google-code-attachments/garglk/issue-204/comment-0/source.txt)
| 1.0 | Sound notification events are not delivered if the player does nothing - ```
The attached Inform 7 source (and compiled Glulx Blorb) is a minimal example
that, when the command "chime" is entered, plays a sound. A notification event
is requested, and when this arrives, the game cancels line input, prints a
message and then re-starts line input.
However, when run under Gargoyle, the sound notification event is not received
by the game until the first time after the sound has finished that the user
generates some user interface activity (e.g. by pressing a key). If the user
does nothing, the event is never received. From my reading of the Glk
specification, this is not what was intended.
(See also this thread: http://www.intfiction.org/forum/viewtopic.php?f=7&t=2848)
```
Original issue reported on code.google.com by `dav...@davidkinder.co.uk` on 16 Feb 2013 at 9:58
Attachments:
- [Sound Notifications.gblorb](https://storage.googleapis.com/google-code-attachments/garglk/issue-204/comment-0/Sound Notifications.gblorb)
- [source.txt](https://storage.googleapis.com/google-code-attachments/garglk/issue-204/comment-0/source.txt)
| defect | sound notification events are not delivered if the player does nothing the attached inform source and compiled glulx blorb is a minimal example that when the command chime is entered plays a sound a notification event is requested and when this arrives the game cancels line input prints a message and then re starts line input however when run under gargoyle the sound notification event is not received by the game until the first time after the sound has finished that the user generates some user interface activity e g by pressing a key if the user does nothing the event is never received from my reading of the glk specification this is not what was intended see also this thread original issue reported on code google com by dav davidkinder co uk on feb at attachments notifications gblorb | 1 |
28,829 | 5,389,797,232 | IssuesEvent | 2017-02-25 06:55:53 | djstauffer/geben-on-emacs | https://api.github.com/repos/djstauffer/geben-on-emacs | closed | Geben hangs and shows abracadabra | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. try to debug yii-application;
2. debug external class including in yii/framework/YiiBase.php;
3. stuck upon return from autoload.
What is the expected output? What do you see instead?
Instead of a normal step-by-step debugging a got a message "File
~/projects/current_project/subfolder/my/.emacs.d/geben/54321/home/me/projects/cu
rrent_project/subfolder/yii/framework/YiiBase.php no longer exists!" and
abracadabra like this:
http://images.netbynet.ru/imgs/b74929bb2410b3e32736ac73f5cf58f0.png (or see
attachment)
What version of the product are you using? On what operating system?
app-editors/emacs-23.3
app-emacs/cedet-1.0
app-emacs/ecb-2.40-r1
app-emacs/emacs-common-gentoo-1.2-r2
dev-lang/php-5.3.6
dev-php/xdebug-2.1.1
dev-php/xdebug-client-2.1.1
virtual/emacs-23
Please provide any additional information below.
At least, how to make geben view log in a _human-readable_ way?
```
Original issue reported on code.google.com by `butthur...@gmail.com` on 19 Aug 2011 at 12:34
Attachments:
- [Screenshot-6.png](https://storage.googleapis.com/google-code-attachments/geben-on-emacs/issue-12/comment-0/Screenshot-6.png)
| 1.0 | Geben hangs and shows abracadabra - ```
What steps will reproduce the problem?
1. try to debug yii-application;
2. debug external class including in yii/framework/YiiBase.php;
3. stuck upon return from autoload.
What is the expected output? What do you see instead?
Instead of a normal step-by-step debugging a got a message "File
~/projects/current_project/subfolder/my/.emacs.d/geben/54321/home/me/projects/cu
rrent_project/subfolder/yii/framework/YiiBase.php no longer exists!" and
abracadabra like this:
http://images.netbynet.ru/imgs/b74929bb2410b3e32736ac73f5cf58f0.png (or see
attachment)
What version of the product are you using? On what operating system?
app-editors/emacs-23.3
app-emacs/cedet-1.0
app-emacs/ecb-2.40-r1
app-emacs/emacs-common-gentoo-1.2-r2
dev-lang/php-5.3.6
dev-php/xdebug-2.1.1
dev-php/xdebug-client-2.1.1
virtual/emacs-23
Please provide any additional information below.
At least, how to make geben view log in a _human-readable_ way?
```
Original issue reported on code.google.com by `butthur...@gmail.com` on 19 Aug 2011 at 12:34
Attachments:
- [Screenshot-6.png](https://storage.googleapis.com/google-code-attachments/geben-on-emacs/issue-12/comment-0/Screenshot-6.png)
| defect | geben hangs and shows abracadabra what steps will reproduce the problem try to debug yii application debug external class including in yii framework yiibase php stuck upon return from autoload what is the expected output what do you see instead instead of a normal step by step debugging a got a message file projects current project subfolder my emacs d geben home me projects cu rrent project subfolder yii framework yiibase php no longer exists and abracadabra like this or see attachment what version of the product are you using on what operating system app editors emacs app emacs cedet app emacs ecb app emacs emacs common gentoo dev lang php dev php xdebug dev php xdebug client virtual emacs please provide any additional information below at least how to make geben view log in a human readable way original issue reported on code google com by butthur gmail com on aug at attachments | 1 |
162,030 | 6,145,993,381 | IssuesEvent | 2017-06-27 12:57:24 | ubuntudesign/docs.vanillaframework.io | https://api.github.com/repos/ubuntudesign/docs.vanillaframework.io | opened | Revert header to how it was before | Priority: High Status: Triaged | - If you go to https://docs.vanillaframework.io/en/
- The header says: Vanilla logo | Documentation
- If you go another page, e.g. https://docs.vanillaframework.io/en/patterns/breadcrumbs
- The header changes to: Vanilla logo | Breadcrumbs | Vanilla framework documentation
This behaviour is jarring. If we need a breadcrumb to indicate where we are, we should be using the breadcrumb pattern. The "Breadcrumbs | Vanilla framework documentation" from the example mentioned above links to the homepage of the docs, which is also confusing.
Can we please make the header consistent across page to just say Vanilla documentation? | 1.0 | Revert header to how it was before - - If you go to https://docs.vanillaframework.io/en/
- The header says: Vanilla logo | Documentation
- If you go another page, e.g. https://docs.vanillaframework.io/en/patterns/breadcrumbs
- The header changes to: Vanilla logo | Breadcrumbs | Vanilla framework documentation
This behaviour is jarring. If we need a breadcrumb to indicate where we are, we should be using the breadcrumb pattern. The "Breadcrumbs | Vanilla framework documentation" from the example mentioned above links to the homepage of the docs, which is also confusing.
Can we please make the header consistent across page to just say Vanilla documentation? | non_defect | revert header to how it was before if you go to the header says vanilla logo documentation if you go another page e g the header changes to vanilla logo breadcrumbs vanilla framework documentation this behaviour is jarring if we need a breadcrumb to indicate where we are we should be using the breadcrumb pattern the breadcrumbs vanilla framework documentation from the example mentioned above links to the homepage of the docs which is also confusing can we please make the header consistent across page to just say vanilla documentation | 0 |
81,996 | 31,845,186,198 | IssuesEvent | 2023-09-14 19:20:46 | NREL/EnergyPlus | https://api.github.com/repos/NREL/EnergyPlus | closed | UFAD Code Inconsistencies in Plume Power Calculations | Defect | Issue overview
--------------
@amirroth noted the following inconsistencies in the UFAD code.
1. The steam equipment convective summation is reset to zero inside the loop (for both the exterior and interior models).
https://github.com/NREL/EnergyPlus/blob/3ce85f9cde01190b8ee923e56f7b9606350fd53a/src/EnergyPlus/UFADManager.cc#L417-L422
https://github.com/NREL/EnergyPlus/blob/3ce85f9cde01190b8ee923e56f7b9606350fd53a/src/EnergyPlus/UFADManager.cc#L625-L630
2. When calculating the convective plume power, the equipment design level is multiplied by `.FractionConvected` for the interior model (first snipped above), but not for the exterior model (second snippet). This is the same across all equipment types. Seems like both models should use `.FractionConvected`. The engineering reference descriptions for this calculation use identical terms with `conv` subscripts.
The code has been this way since these features were first added [here](https://github.com/NREL/EnergyPlusArchive/commit/8fd87048ed0ed2c328a2cad581cacfbdc330c0f1#diff-e89bfdffca8ef2d108573498a880025022146e5dc24dd265e0c9a2c8cf9b4a7a).
### Details
Some additional details for this issue (if relevant):
- Version of EnergyPlus 23.1
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| 1.0 | UFAD Code Inconsistencies in Plume Power Calculations - Issue overview
--------------
@amirroth noted the following inconsistencies in the UFAD code.
1. The steam equipment convective summation is reset to zero inside the loop (for both the exterior and interior models).
https://github.com/NREL/EnergyPlus/blob/3ce85f9cde01190b8ee923e56f7b9606350fd53a/src/EnergyPlus/UFADManager.cc#L417-L422
https://github.com/NREL/EnergyPlus/blob/3ce85f9cde01190b8ee923e56f7b9606350fd53a/src/EnergyPlus/UFADManager.cc#L625-L630
2. When calculating the convective plume power, the equipment design level is multiplied by `.FractionConvected` for the interior model (first snipped above), but not for the exterior model (second snippet). This is the same across all equipment types. Seems like both models should use `.FractionConvected`. The engineering reference descriptions for this calculation use identical terms with `conv` subscripts.
The code has been this way since these features were first added [here](https://github.com/NREL/EnergyPlusArchive/commit/8fd87048ed0ed2c328a2cad581cacfbdc330c0f1#diff-e89bfdffca8ef2d108573498a880025022146e5dc24dd265e0c9a2c8cf9b4a7a).
### Details
Some additional details for this issue (if relevant):
- Version of EnergyPlus 23.1
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| defect | ufad code inconsistencies in plume power calculations issue overview amirroth noted the following inconsistencies in the ufad code the steam equipment convective summation is reset to zero inside the loop for both the exterior and interior models when calculating the convective plume power the equipment design level is multiplied by fractionconvected for the interior model first snipped above but not for the exterior model second snippet this is the same across all equipment types seems like both models should use fractionconvected the engineering reference descriptions for this calculation use identical terms with conv subscripts the code has been this way since these features were first added details some additional details for this issue if relevant version of energyplus checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect | 1 |
42,209 | 10,886,258,746 | IssuesEvent | 2019-11-18 12:10:54 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Client side OOME 0 size hprof after member killed -9 | Source: Internal Team: Client Team: Core Type: Defect |
In the test 32 clients are connected to 1 member, the member is killed -9 and restarted repeatedly to check the effect on the clients.
clients are configured with `-Xms250M -Xmx250M`
http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/go
client xml config
http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/config-hz/client-hazelcast.xml
```
<connection-strategy async-start="true" reconnect-mode="ASYNC">
<connection-retry enabled="true">
<initial-backoff-millis>1</initial-backoff-millis>
<max-backoff-millis>300</max-backoff-millis>
<fail-on-max-backoff>false</fail-on-max-backoff>
<multiplier>1</multiplier>
<jitter>0.01</jitter>
</connection-retry>
</connection-strategy>
<network>
<cluster-members> </cluster-members>
<smart-routing>true</smart-routing>
<connection-attempt-period>9000</connection-attempt-period>
<connection-attempt-limit>200</connection-attempt-limit>
</network>
```
client are configured with high connection rates specifically to test this aspect of the system
on both sides.
After 967 iteration of kill -9 and restart
25 out of 32 client crashed with OOME producing a hprof of 0 size.
967 iteration makes this issue quite rare.
hprofs of 0 size are odd, and could be caused by, clients allocating very large buffers
based on possible junk info coming from the member ?
http://jenkins.hazelcast.com/view/kill/job/kill-members/11/console
http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member
/disk1/jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member
```
[jenkins@ip-10-72-134-107 member]$ hz-errors | sort | uniq | xargs ls -lah
-rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient10HZ/HzClient10HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient11HZ/HzClient11HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient12HZ/HzClient12HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient13HZ/HzClient13HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient14HZ/HzClient14HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient19HZ/HzClient19HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient1HZ/HzClient1HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient20HZ/HzClient20HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient21HZ/HzClient21HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient22HZ/HzClient22HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient23HZ/HzClient23HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient24HZ/HzClient24HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient25HZ/HzClient25HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient26HZ/HzClient26HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient27HZ/HzClient27HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient28HZ/HzClient28HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient29HZ/HzClient29HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient2HZ/HzClient2HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient30HZ/HzClient30HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient32HZ/HzClient32HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient4HZ/HzClient4HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient5HZ/HzClient5HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient6HZ/HzClient6HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:27 ./output/HZ/HzClient7HZ/HzClient7HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient9HZ/HzClient9HZ.hprof
```
GC charts.
http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/gc.html
interestingly all client GC charts look the same as below.
the chart is not showing any sign of memory leak, and looks good.

| 1.0 | Client side OOME 0 size hprof after member killed -9 -
In the test 32 clients are connected to 1 member, the member is killed -9 and restarted repeatedly to check the effect on the clients.
clients are configured with `-Xms250M -Xmx250M`
http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/go
client xml config
http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/config-hz/client-hazelcast.xml
```
<connection-strategy async-start="true" reconnect-mode="ASYNC">
<connection-retry enabled="true">
<initial-backoff-millis>1</initial-backoff-millis>
<max-backoff-millis>300</max-backoff-millis>
<fail-on-max-backoff>false</fail-on-max-backoff>
<multiplier>1</multiplier>
<jitter>0.01</jitter>
</connection-retry>
</connection-strategy>
<network>
<cluster-members> </cluster-members>
<smart-routing>true</smart-routing>
<connection-attempt-period>9000</connection-attempt-period>
<connection-attempt-limit>200</connection-attempt-limit>
</network>
```
client are configured with high connection rates specifically to test this aspect of the system
on both sides.
After 967 iteration of kill -9 and restart
25 out of 32 client crashed with OOME producing a hprof of 0 size.
967 iteration makes this issue quite rare.
hprofs of 0 size are odd, and could be caused by, clients allocating very large buffers
based on possible junk info coming from the member ?
http://jenkins.hazelcast.com/view/kill/job/kill-members/11/console
http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member
/disk1/jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member
```
[jenkins@ip-10-72-134-107 member]$ hz-errors | sort | uniq | xargs ls -lah
-rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient10HZ/HzClient10HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient11HZ/HzClient11HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient12HZ/HzClient12HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient13HZ/HzClient13HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient14HZ/HzClient14HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient19HZ/HzClient19HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient1HZ/HzClient1HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient20HZ/HzClient20HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient21HZ/HzClient21HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient22HZ/HzClient22HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient23HZ/HzClient23HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient24HZ/HzClient24HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient25HZ/HzClient25HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient26HZ/HzClient26HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient27HZ/HzClient27HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient28HZ/HzClient28HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient29HZ/HzClient29HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient2HZ/HzClient2HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient30HZ/HzClient30HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient32HZ/HzClient32HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient4HZ/HzClient4HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient5HZ/HzClient5HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient6HZ/HzClient6HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:27 ./output/HZ/HzClient7HZ/HzClient7HZ.hprof
-rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient9HZ/HzClient9HZ.hprof
```
GC charts.
http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/gc.html
interestingly all client GC charts look the same as below.
the chart is not showing any sign of memory leak, and looks good.

| defect | client side oome size hprof after member killed in the test clients are connected to member the member is killed and restarted repeatedly to check the effect on the clients clients are configured with client xml config false true client are configured with high connection rates specifically to test this aspect of the system on both sides after iteration of kill and restart out of client crashed with oome producing a hprof of size iteration makes this issue quite rare hprofs of size are odd and could be caused by clients allocating very large buffers based on possible junk info coming from the member jenkins workspace kill members member hz errors sort uniq xargs ls lah rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof gc charts interestingly all client gc charts look the same as below the chart is not showing any sign of memory leak and looks good | 1 |
9,847 | 2,615,175,968 | IssuesEvent | 2015-03-01 06:59:27 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | Supplied full pin but Reaver doesnt succeed | auto-migrated Priority-Triage Type-Defect | ```
Hi All,
I am new to using Reaver, but have done a lot of reading and research into it's
use. I am trying to crack my own BT Home Hub 3b (UK). However I am seeing some
strange behaviour, some background first:
WiFi Adapter - ALFA AWUS036NHA (Atheros AR9271) in monitor mode
Linux Version - Kali latest distro from Kali website (January 2015)
kali-linux-1.0.9a 64bit
Reaver Version - 1.4
My router has WPS enabled and is not locked. I have run the following commands
airmon-ng start wlan0
Kill 3 processes running (NetworkManager and
wash -i mon0
BSSID Channel RSSI WPS Version WPS Locked
ESSID
28:3C:E4:36:46:5F 11 -42 1.0 No
[NetworkName]
I then open a new terminal window and have tried the following few reaver
commands (not at the same time).
Firstly, I tried to see if Reaver actually gets the WPA when I pass it the full
know WPS pin, I run it with the actual WPS pin using the following command:
# reaver -i mon0 -c 11 -b 28:3C:E4:36:46:5F -d 5 -p 86890281 -vv
Output is this, it just loops for ages.. surely by giving it the full pin it
should find it almost instantly?
[+] Switching mon0 to channel 11
[+] Waiting for beacon from 28:3C:E4:36:46:5F
[+] Associated with 28:3C:E4:36:46:5F (ESSID: [NetworkName])
[+] Trying pin 86890281
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 86890281
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 86890281
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 86890281
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
I then tried running Reaver using the following command (first 4 digits of the
pin, and some extra options that proved successful for others):
reaver -i mon0 -c 11 -b 28:3C:E4:36:46:5F -d 5 -r 3:15 -x 360 -T .5 -S -N -t 10
-w -vv -L -p 8689
This went on for about 12 hours and got to 99.99% but just loops on the last
pin (which is wrong):
[+] Trying pin 86899987
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 86899987
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] 99.99% complete @ 2015-01-27 07:15:25 (7 seconds/pin)
[+] Max time remaining at this rate: 0:00:07 (1 pins left to try)
Any ideas on what to try next?
I have read through the Wiki/Q&A but not got much further.
Sorry if this seems like a bit of a dumb problem, we all have to start
somewhere..
```
Original issue reported on code.google.com by `christop...@gmail.com` on 27 Jan 2015 at 8:13 | 1.0 | Supplied full pin but Reaver doesnt succeed - ```
Hi All,
I am new to using Reaver, but have done a lot of reading and research into it's
use. I am trying to crack my own BT Home Hub 3b (UK). However I am seeing some
strange behaviour, some background first:
WiFi Adapter - ALFA AWUS036NHA (Atheros AR9271) in monitor mode
Linux Version - Kali latest distro from Kali website (January 2015)
kali-linux-1.0.9a 64bit
Reaver Version - 1.4
My router has WPS enabled and is not locked. I have run the following commands
airmon-ng start wlan0
Kill 3 processes running (NetworkManager and
wash -i mon0
BSSID Channel RSSI WPS Version WPS Locked
ESSID
28:3C:E4:36:46:5F 11 -42 1.0 No
[NetworkName]
I then open a new terminal window and have tried the following few reaver
commands (not at the same time).
Firstly, I tried to see if Reaver actually gets the WPA when I pass it the full
know WPS pin, I run it with the actual WPS pin using the following command:
# reaver -i mon0 -c 11 -b 28:3C:E4:36:46:5F -d 5 -p 86890281 -vv
Output is this, it just loops for ages.. surely by giving it the full pin it
should find it almost instantly?
[+] Switching mon0 to channel 11
[+] Waiting for beacon from 28:3C:E4:36:46:5F
[+] Associated with 28:3C:E4:36:46:5F (ESSID: [NetworkName])
[+] Trying pin 86890281
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 86890281
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 86890281
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 86890281
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
I then tried running Reaver using the following command (first 4 digits of the
pin, and some extra options that proved successful for others):
reaver -i mon0 -c 11 -b 28:3C:E4:36:46:5F -d 5 -r 3:15 -x 360 -T .5 -S -N -t 10
-w -vv -L -p 8689
This went on for about 12 hours and got to 99.99% but just loops on the last
pin (which is wrong):
[+] Trying pin 86899987
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] Trying pin 86899987
[+] Sending EAPOL START request
[+] Received identity request
[+] Sending identity response
[+] Received M1 message
[+] Sending M2 message
[+] Received M3 message
[+] Sending M4 message
[+] Received WSC NACK
[+] Sending WSC NACK
[+] 99.99% complete @ 2015-01-27 07:15:25 (7 seconds/pin)
[+] Max time remaining at this rate: 0:00:07 (1 pins left to try)
Any ideas on what to try next?
I have read through the Wiki/Q&A but not got much further.
Sorry if this seems like a bit of a dumb problem, we all have to start
somewhere..
```
Original issue reported on code.google.com by `christop...@gmail.com` on 27 Jan 2015 at 8:13 | defect | supplied full pin but reaver doesnt succeed hi all i am new to using reaver but have done a lot of reading and research into it s use i am trying to crack my own bt home hub uk however i am seeing some strange behaviour some background first wifi adapter alfa atheros in monitor mode linux version kali latest distro from kali website january kali linux reaver version my router has wps enabled and is not locked i have run the following commands airmon ng start kill processes running networkmanager and wash i bssid channel rssi wps version wps locked essid no i then open a new terminal window and have tried the following few reaver commands not at the same time firstly i tried to see if reaver actually gets the wpa when i pass it the full know wps pin i run it with the actual wps pin using the following command reaver i c b d p vv output is this it just loops for ages surely by giving it the full pin it should find it almost instantly switching to channel waiting for beacon from associated with essid trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack i then tried running reaver using the following command first digits of the pin and some extra options that proved successful for others reaver i c b d r x t s n t w vv l p this went on for about hours and got to but just loops on the last pin which is wrong trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack complete seconds pin max time remaining at this rate pins left to try any ideas on what to try next i have read through the wiki q a but not got much further sorry if this seems like a bit of a dumb problem we all have to start somewhere original issue reported on code google com by christop gmail com on jan at | 1 |
37,049 | 8,214,119,923 | IssuesEvent | 2018-09-04 21:53:03 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | optimize.minimize(method=`trust-constr`) result dict does not contain `success` | defect scipy.optimize | The result dictionary from all optimizers is supposed to be uniform and contain e.g. a `success` boolean flag.
The `success` flag is missing for `method='trust-constr'`.
The result `nit` is called `niter` in the result dict.
### Reproducing code example:
```
from scipy.optimize import minimize
sol = minimize(lambda x: x**2, [0], jac=lambda x: 2*x,
hess=lambda x: 2, method='trust-constr')
print(sorted(sol.keys()))
assert sol.success
```
### Error message:
```
['cg_niter', 'cg_stop_cond', 'constr', 'constr_nfev', 'constr_nhev',
'constr_njev', 'constr_penalty', 'constr_violation', 'execution_time',
'fun', 'grad', 'jac', 'lagrangian_grad', 'message', 'method', 'nfev',
'nhev', 'niter', 'njev', 'optimality', 'status', 'tr_radius', 'v', 'x']
Traceback (most recent call last):
File "<ipython-input-13-492cf8a99d03>", line 3, in <module>
assert sol.success
File "/home/pauli/.local/lib/python3.6/site-packages/scipy/optimize/optimize.py", line 116, in __getattr__
raise AttributeError(name)
AttributeError: success
```
### Scipy/Numpy/Python version information:
```
1.1.0 1.14.5 sys.version_info(major=3, minor=6, micro=5, releaselevel='final', serial=0)
and
1.2.0.dev0+7e41406 1.14.5 sys.version_info(major=3, minor=6, micro=5, releaselevel='final', serial=0)
```
| 1.0 | optimize.minimize(method=`trust-constr`) result dict does not contain `success` - The result dictionary from all optimizers is supposed to be uniform and contain e.g. a `success` boolean flag.
The `success` flag is missing for `method='trust-constr'`.
The result `nit` is called `niter` in the result dict.
### Reproducing code example:
```
from scipy.optimize import minimize
sol = minimize(lambda x: x**2, [0], jac=lambda x: 2*x,
hess=lambda x: 2, method='trust-constr')
print(sorted(sol.keys()))
assert sol.success
```
### Error message:
```
['cg_niter', 'cg_stop_cond', 'constr', 'constr_nfev', 'constr_nhev',
'constr_njev', 'constr_penalty', 'constr_violation', 'execution_time',
'fun', 'grad', 'jac', 'lagrangian_grad', 'message', 'method', 'nfev',
'nhev', 'niter', 'njev', 'optimality', 'status', 'tr_radius', 'v', 'x']
Traceback (most recent call last):
File "<ipython-input-13-492cf8a99d03>", line 3, in <module>
assert sol.success
File "/home/pauli/.local/lib/python3.6/site-packages/scipy/optimize/optimize.py", line 116, in __getattr__
raise AttributeError(name)
AttributeError: success
```
### Scipy/Numpy/Python version information:
```
1.1.0 1.14.5 sys.version_info(major=3, minor=6, micro=5, releaselevel='final', serial=0)
and
1.2.0.dev0+7e41406 1.14.5 sys.version_info(major=3, minor=6, micro=5, releaselevel='final', serial=0)
```
| defect | optimize minimize method trust constr result dict does not contain success the result dictionary from all optimizers is supposed to be uniform and contain e g a success boolean flag the success flag is missing for method trust constr the result nit is called niter in the result dict reproducing code example from scipy optimize import minimize sol minimize lambda x x jac lambda x x hess lambda x method trust constr print sorted sol keys assert sol success error message cg niter cg stop cond constr constr nfev constr nhev constr njev constr penalty constr violation execution time fun grad jac lagrangian grad message method nfev nhev niter njev optimality status tr radius v x traceback most recent call last file line in assert sol success file home pauli local lib site packages scipy optimize optimize py line in getattr raise attributeerror name attributeerror success scipy numpy python version information sys version info major minor micro releaselevel final serial and sys version info major minor micro releaselevel final serial | 1 |
78,495 | 27,554,265,755 | IssuesEvent | 2023-03-07 16:48:41 | dotCMS/core | https://api.github.com/repos/dotCMS/core | opened | Can't push publish content with a block editor field | Type : Defect Triage | ### Parent Issue
_No response_
### Problem Statement
When you try to push publish content on auth we're getting this error
```
Error generating bundle ID '01GTWJ5V5MPX36HAJPJXJG3EV3': An error occurred when processing dependencies on Contentlet 'f528920b19e88a2ef727e8341a825f5b': An error occurred when retrieving Contentlet references from Story Block field: Unexpected character ('<' (code 60)): expected a valid value (JSON String, Number, Array, Object or token 'null', 'true' or 'false')
```
### Steps to Reproduce
Not sure yet, we need to ask @fishsmith in the #website channel
### Acceptance Criteria
- [ ] Allow the user to push publish content with block editor field
### dotCMS Version
23.03
### Proposed Objective
Core Features
### Proposed Priority
Priority 1 - Show Stopper
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
This is a blocker for the release of 23.03
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | 1.0 | Can't push publish content with a block editor field - ### Parent Issue
_No response_
### Problem Statement
When you try to push publish content on auth we're getting this error
```
Error generating bundle ID '01GTWJ5V5MPX36HAJPJXJG3EV3': An error occurred when processing dependencies on Contentlet 'f528920b19e88a2ef727e8341a825f5b': An error occurred when retrieving Contentlet references from Story Block field: Unexpected character ('<' (code 60)): expected a valid value (JSON String, Number, Array, Object or token 'null', 'true' or 'false')
```
### Steps to Reproduce
Not sure yet, we need to ask @fishsmith in the #website channel
### Acceptance Criteria
- [ ] Allow the user to push publish content with block editor field
### dotCMS Version
23.03
### Proposed Objective
Core Features
### Proposed Priority
Priority 1 - Show Stopper
### External Links... Slack Conversations, Support Tickets, Figma Designs, etc.
_No response_
### Assumptions & Initiation Needs
This is a blocker for the release of 23.03
### Quality Assurance Notes & Workarounds
_No response_
### Sub-Tasks & Estimates
_No response_ | defect | can t push publish content with a block editor field parent issue no response problem statement when you try to push publish content on auth we re getting this error error generating bundle id an error occurred when processing dependencies on contentlet an error occurred when retrieving contentlet references from story block field unexpected character code expected a valid value json string number array object or token null true or false steps to reproduce not sure yet we need to ask fishsmith in the website channel acceptance criteria allow the user to push publish content with block editor field dotcms version proposed objective core features proposed priority priority show stopper external links slack conversations support tickets figma designs etc no response assumptions initiation needs this is a blocker for the release of quality assurance notes workarounds no response sub tasks estimates no response | 1 |
475,485 | 13,711,177,415 | IssuesEvent | 2020-10-02 03:33:44 | gadenbuie/rsthemes | https://api.github.com/repos/gadenbuie/rsthemes | closed | Style job pane progress bar | priority: medium status: planned type: new | ```css
.rstheme_toolbarWrapper [role="progressbar"] > div {
background-color: pink;
}
``` | 1.0 | Style job pane progress bar - ```css
.rstheme_toolbarWrapper [role="progressbar"] > div {
background-color: pink;
}
``` | non_defect | style job pane progress bar css rstheme toolbarwrapper div background color pink | 0 |
22,979 | 11,812,162,427 | IssuesEvent | 2020-03-19 19:36:21 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Reporting: add a "Remove" action in "Reporting" Management app | Feature:Reporting Team:Reporting Services enhancement | Currently, when we go to "Management" app, and then to "Reporting" section, in the reports list, we don't have a "remove" button in "Actions" column.
This is quite annoying, when you generate a report (CSV, PDF), download it, and want to remove it immediately, because report is not useful anymore.
This would be great to add a "Remove" action available in reports list section.

| 1.0 | Reporting: add a "Remove" action in "Reporting" Management app - Currently, when we go to "Management" app, and then to "Reporting" section, in the reports list, we don't have a "remove" button in "Actions" column.
This is quite annoying, when you generate a report (CSV, PDF), download it, and want to remove it immediately, because report is not useful anymore.
This would be great to add a "Remove" action available in reports list section.

| non_defect | reporting add a remove action in reporting management app currently when we go to management app and then to reporting section in the reports list we don t have a remove button in actions column this is quite annoying when you generate a report csv pdf download it and want to remove it immediately because report is not useful anymore this would be great to add a remove action available in reports list section | 0 |
510,820 | 14,826,766,269 | IssuesEvent | 2021-01-15 00:48:23 | jef/streetmerchant | https://api.github.com/repos/jef/streetmerchant | closed | Desktop notifications need appID set in Windows 10 | priority: low status: needs triage status: stale type: bug | ### Description
Desktop notifications through node-notifier are not working in Windows 10. This seems to be due to the requirement to set an `appID`: https://www.npmjs.com/package/node-notifier#usage-windowstoaster
```
C:\Users\Hedova\Desktop\Programs\streetmerchant>npm run test:notification
> test:notification
> npm run build && node build/__test__/notification-test.js
> build
> tsc
STREETMERCHANT
3.1.0
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
[4:31:26 AM] info :: ✔ sound player found: cmdmp3
[4:31:27 AM] info :: ✔ desktop notification sent
Notifications are disabled
Reason: DisabledForUser Please make sure that the app id is set correctly.
Command Line: C:\Users\Hedova\Desktop\Programs\streetmerchant\node_modules\node-notifier\vendor\snoreToast\snoretoast-x64.exe -appID streetmerchant -pipeName \\.\pipe\notifierPipe-efc1ad6d-7071-4038-966f-ffe741f32252 -m https://www.example.com/cartUrl -t "[4:31:29 AM] info :: ✔ played sound
```
| 1.0 | Desktop notifications need appID set in Windows 10 - ### Description
Desktop notifications through node-notifier are not working in Windows 10. This seems to be due to the requirement to set an `appID`: https://www.npmjs.com/package/node-notifier#usage-windowstoaster
```
C:\Users\Hedova\Desktop\Programs\streetmerchant>npm run test:notification
> test:notification
> npm run build && node build/__test__/notification-test.js
> build
> tsc
STREETMERCHANT
3.1.0
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
INFO: Could not find files for the given pattern(s).
[4:31:26 AM] info :: ✔ sound player found: cmdmp3
[4:31:27 AM] info :: ✔ desktop notification sent
Notifications are disabled
Reason: DisabledForUser Please make sure that the app id is set correctly.
Command Line: C:\Users\Hedova\Desktop\Programs\streetmerchant\node_modules\node-notifier\vendor\snoreToast\snoretoast-x64.exe -appID streetmerchant -pipeName \\.\pipe\notifierPipe-efc1ad6d-7071-4038-966f-ffe741f32252 -m https://www.example.com/cartUrl -t "[4:31:29 AM] info :: ✔ played sound
```
| non_defect | desktop notifications need appid set in windows description desktop notifications through node notifier are not working in windows this seems to be due to the requirement to set an appid c users hedova desktop programs streetmerchant npm run test notification test notification npm run build node build test notification test js build tsc streetmerchant info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info ✔ sound player found info ✔ desktop notification sent notifications are disabled reason disabledforuser please make sure that the app id is set correctly command line c users hedova desktop programs streetmerchant node modules node notifier vendor snoretoast snoretoast exe appid streetmerchant pipename pipe notifierpipe m t info ✔ played sound | 0 |
11,380 | 2,649,862,548 | IssuesEvent | 2015-03-15 11:09:46 | bbk79/xbmc-glwiz-addon | https://api.github.com/repos/bbk79/xbmc-glwiz-addon | closed | Audio/Video sync when "Paid Account" enabled | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Enable "Paid Account" with an active subscription.
2.
3.
What is the expected output? What do you see instead?
720Kbps quality expected and seen however framerate for Video is a bit low and
audio is ahead 3-4 seconds!
What version of the product are you using? On what operating system?
.8 on XBMC eden ATV2
Please provide any additional information below.
Normal quality works fine, but soon as you enable Paid Account the issue above
pops up. It's a pity as the quality improvement in video is very obvious.
Any clues/ideas on this at all?
```
Original issue reported on code.google.com by `paya...@gmail.com` on 22 Dec 2012 at 8:28 | 1.0 | Audio/Video sync when "Paid Account" enabled - ```
What steps will reproduce the problem?
1. Enable "Paid Account" with an active subscription.
2.
3.
What is the expected output? What do you see instead?
720Kbps quality expected and seen however framerate for Video is a bit low and
audio is ahead 3-4 seconds!
What version of the product are you using? On what operating system?
.8 on XBMC eden ATV2
Please provide any additional information below.
Normal quality works fine, but soon as you enable Paid Account the issue above
pops up. It's a pity as the quality improvement in video is very obvious.
Any clues/ideas on this at all?
```
Original issue reported on code.google.com by `paya...@gmail.com` on 22 Dec 2012 at 8:28 | defect | audio video sync when paid account enabled what steps will reproduce the problem enable paid account with an active subscription what is the expected output what do you see instead quality expected and seen however framerate for video is a bit low and audio is ahead seconds what version of the product are you using on what operating system on xbmc eden please provide any additional information below normal quality works fine but soon as you enable paid account the issue above pops up it s a pity as the quality improvement in video is very obvious any clues ideas on this at all original issue reported on code google com by paya gmail com on dec at | 1 |
13,323 | 2,753,736,103 | IssuesEvent | 2015-04-25 00:47:29 | sma-bluetooth/sma-bluetooth | https://api.github.com/repos/sma-bluetooth/sma-bluetooth | closed | SB2500 Did not find string | auto-migrated Priority-Medium Type-Defect | ```
The communication with SB2500 is not working fine as the format does not seem
to be correct (also for SB1700)
[18] 2011-07-19 20:39:03 Waiting for string
[18] 2011-07-19 20:39:03 waiting for: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00
00 00 02 00 00 04 70 00
[18] 2011-07-19 20:39:03 Waiting for data on rfcomm
Receiving
00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61
00000010: 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04
00000020: 70 00 01 00 00 00 00 01 00 00 00 rr=31
[18] 2011-07-19 20:39:03 looking for: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00
00 00 02 00 00 04 70 00
[18] 2011-07-19 20:39:03 received: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00
00 00 02 00 00 04 70 00 01 00 00 00 00 01 00 00 00
[18] 2011-07-19 20:39:03 Found string we are waiting for
[19] 2011-07-19 20:39:03 Extracting
extracting invcode=01
[20] 2011-07-19 20:39:03 Sending
[20] 2011-07-19 20:39:03 sending:
00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61
00000010: 00 00 00 00 00 00 05 7e 1b 25 80 00 02 00 00 04
00000020: 70 00 01 00 00 00 00 01 00 00 00 cc=31
[21] 2011-07-19 20:39:03 Waiting for string
[21] 2011-07-19 20:39:03 waiting for: 7e 22 00 5c 05 7e 1b 25 80 00 00 00 00 00
00 00 05 00 05 7e 1b 25 80 00
[21] 2011-07-19 20:39:03 Waiting for data on rfcomm
Receiving
00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61
00000010: 05 7e 1b 25 80 00 00 00 00 00 00 00 0a 00 05 7e
00000020: 1b 25 80 00 01 2d 8d 00 7d 1a 00 rr=31
sum=80length change from 1f to 1e
[21] 2011-07-19 20:39:03 looking for: 7e 22 00 5c 05 7e 1b 25 80 00 00 00 00 00
00 00 05 00 05 7e 1b 25 80 00
[21] 2011-07-19 20:39:03 received: 7e 1e 00 62 05 7e 1b 25 80 00 00 00 00 00
00 00 0a 00 05 7e 1b 25 80 00 01 2d 8d 00 3a 00
[21] 2011-07-19 20:39:03 Did not find string
How can I solve this?
Thanks
```
Original issue reported on code.google.com by `patrik.d...@gmail.com` on 19 Jul 2011 at 9:05
Attachments:
* [smatool_20110719.log](https://storage.googleapis.com/google-code-attachments/sma-bluetooth/issue-35/comment-0/smatool_20110719.log)
| 1.0 | SB2500 Did not find string - ```
The communication with SB2500 is not working fine as the format does not seem
to be correct (also for SB1700)
[18] 2011-07-19 20:39:03 Waiting for string
[18] 2011-07-19 20:39:03 waiting for: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00
00 00 02 00 00 04 70 00
[18] 2011-07-19 20:39:03 Waiting for data on rfcomm
Receiving
00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61
00000010: 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04
00000020: 70 00 01 00 00 00 00 01 00 00 00 rr=31
[18] 2011-07-19 20:39:03 looking for: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00
00 00 02 00 00 04 70 00
[18] 2011-07-19 20:39:03 received: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00
00 00 02 00 00 04 70 00 01 00 00 00 00 01 00 00 00
[18] 2011-07-19 20:39:03 Found string we are waiting for
[19] 2011-07-19 20:39:03 Extracting
extracting invcode=01
[20] 2011-07-19 20:39:03 Sending
[20] 2011-07-19 20:39:03 sending:
00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61
00000010: 00 00 00 00 00 00 05 7e 1b 25 80 00 02 00 00 04
00000020: 70 00 01 00 00 00 00 01 00 00 00 cc=31
[21] 2011-07-19 20:39:03 Waiting for string
[21] 2011-07-19 20:39:03 waiting for: 7e 22 00 5c 05 7e 1b 25 80 00 00 00 00 00
00 00 05 00 05 7e 1b 25 80 00
[21] 2011-07-19 20:39:03 Waiting for data on rfcomm
Receiving
00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61
00000010: 05 7e 1b 25 80 00 00 00 00 00 00 00 0a 00 05 7e
00000020: 1b 25 80 00 01 2d 8d 00 7d 1a 00 rr=31
sum=80length change from 1f to 1e
[21] 2011-07-19 20:39:03 looking for: 7e 22 00 5c 05 7e 1b 25 80 00 00 00 00 00
00 00 05 00 05 7e 1b 25 80 00
[21] 2011-07-19 20:39:03 received: 7e 1e 00 62 05 7e 1b 25 80 00 00 00 00 00
00 00 0a 00 05 7e 1b 25 80 00 01 2d 8d 00 3a 00
[21] 2011-07-19 20:39:03 Did not find string
How can I solve this?
Thanks
```
Original issue reported on code.google.com by `patrik.d...@gmail.com` on 19 Jul 2011 at 9:05
Attachments:
* [smatool_20110719.log](https://storage.googleapis.com/google-code-attachments/sma-bluetooth/issue-35/comment-0/smatool_20110719.log)
| defect | did not find string the communication with is not working fine as the format does not seem to be correct also for waiting for string waiting for waiting for data on rfcomm receiving rr looking for received found string we are waiting for extracting extracting invcode sending sending cc waiting for string waiting for waiting for data on rfcomm receiving rr sum change from to looking for received did not find string how can i solve this thanks original issue reported on code google com by patrik d gmail com on jul at attachments | 1 |
81,432 | 30,844,627,260 | IssuesEvent | 2023-08-02 13:00:19 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | opened | DataTyble: Empty aria-label generation in multiselect tables | :lady_beetle: defect :bangbang: needs-triage | ### Describe the bug
During our improvements / rewrites to make our webapp accessibly and fullfil the requirements of the official German BITV 2.0 test we came along a slightly problem with multi select [data tables](https://www.primefaces.org/showcase/ui/data/datatable/selection.xhtml?jfwid=8db7c).
The tried to achieve that every checkbox has an (hidden) description like "select row with identifier X" so the screen reader (we use NVDA, which is also used in the official test) reads to a visual impaired person what row he selects, when ticking the checkbox.
For this we tried to make use of the PF API attribute `ariaRowLabel` which descriptions sounds to do exactly this. But it seems that it does not as the screenreader keeps quiet, regardless of the value inside. So we used a pass-trough attribute which is picked up correctly by the screenreader, but which is not rendered at the checkbox, but at the `td` tag.
So this code
```
<p:column
headerText="Selecton"
selectionMode="multiple"
pt:aria-label="XXX"
style="width:10ch; text-align:left">
</p:column>
```
generates this outcome in PF 12.00 (Not yet tested with PF 13, but release notes don't show any changes here and the sadly has no correspondening example :( )
```
<td
role="gridcell"
style="width:10ch; text-align:left" class="ui-selection-column"
aria-label="XXX">
<div
class="ui-chkbox ui-widget">
<div
id="frm_kleineSucheSuchErgebnisse:row_2:crd_2:tbl_suchErgebnisse_1621_checkbox"
role="checkbox"
tabindex="0"
aria-label=""
aria-checked="false"
class="ui-chkbox-box ui-widget ui-corner-all ui-state-default"
>
<span class="ui-chkbox-icon ui-icon ui-icon-blank ui-c"></span>
</div>
</div>
</td>
```
As you can see, this renderes an empty `aria-label` for the "checkbox-div".
So why do I file an issue, when the screen reader handles it correctly?
Two reasons:
* I / We assume that the empty `aria-label` should have the value the `td` has, where the `td` should probably have none or the same? (Don't know if the same would cause trouble tbh)
* The lighthouse report claims that the checkbox has no aria label and is therefore not accessible
* What is the `ariaRowLabel` for, if not for this?
> ARIA toggle fields do not have accessible names
>
> When a toggle field doesn't have an accessible name, screen readers announce it with a generic name, making it unusable
> for users who rely on screen readers.
>
> div#frm_kleineSucheSuchErgebnisse:row_2:crd_2:tbl_suchErgebnisse_1621_checkbox.
> ui-chkbox-box.ui-widget.ui-corner-all.ui-state-default
Further notes
* The lightous report always claimes that error, regardless of usage of a `pt` or the `ariaRowLabel` or none/bith of them
* Using the `pt` makes the screenreader work, not using or only using the `ariaRowLabel` does not
### Reproducer
_No response_
### Expected behavior
_No response_
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
MyFaces
### JSF version
2.2
### Java version
8
### Browser(s)
_No response_ | 1.0 | DataTyble: Empty aria-label generation in multiselect tables - ### Describe the bug
During our improvements / rewrites to make our webapp accessibly and fullfil the requirements of the official German BITV 2.0 test we came along a slightly problem with multi select [data tables](https://www.primefaces.org/showcase/ui/data/datatable/selection.xhtml?jfwid=8db7c).
The tried to achieve that every checkbox has an (hidden) description like "select row with identifier X" so the screen reader (we use NVDA, which is also used in the official test) reads to a visual impaired person what row he selects, when ticking the checkbox.
For this we tried to make use of the PF API attribute `ariaRowLabel` which descriptions sounds to do exactly this. But it seems that it does not as the screenreader keeps quiet, regardless of the value inside. So we used a pass-trough attribute which is picked up correctly by the screenreader, but which is not rendered at the checkbox, but at the `td` tag.
So this code
```
<p:column
headerText="Selecton"
selectionMode="multiple"
pt:aria-label="XXX"
style="width:10ch; text-align:left">
</p:column>
```
generates this outcome in PF 12.00 (Not yet tested with PF 13, but release notes don't show any changes here and the sadly has no correspondening example :( )
```
<td
role="gridcell"
style="width:10ch; text-align:left" class="ui-selection-column"
aria-label="XXX">
<div
class="ui-chkbox ui-widget">
<div
id="frm_kleineSucheSuchErgebnisse:row_2:crd_2:tbl_suchErgebnisse_1621_checkbox"
role="checkbox"
tabindex="0"
aria-label=""
aria-checked="false"
class="ui-chkbox-box ui-widget ui-corner-all ui-state-default"
>
<span class="ui-chkbox-icon ui-icon ui-icon-blank ui-c"></span>
</div>
</div>
</td>
```
As you can see, this renderes an empty `aria-label` for the "checkbox-div".
So why do I file an issue, when the screen reader handles it correctly?
Two reasons:
* I / We assume that the empty `aria-label` should have the value the `td` has, where the `td` should probably have none or the same? (Don't know if the same would cause trouble tbh)
* The lighthouse report claims that the checkbox has no aria label and is therefore not accessible
* What is the `ariaRowLabel` for, if not for this?
> ARIA toggle fields do not have accessible names
>
> When a toggle field doesn't have an accessible name, screen readers announce it with a generic name, making it unusable
> for users who rely on screen readers.
>
> div#frm_kleineSucheSuchErgebnisse:row_2:crd_2:tbl_suchErgebnisse_1621_checkbox.
> ui-chkbox-box.ui-widget.ui-corner-all.ui-state-default
Further notes
* The lightous report always claimes that error, regardless of usage of a `pt` or the `ariaRowLabel` or none/bith of them
* Using the `pt` makes the screenreader work, not using or only using the `ariaRowLabel` does not
### Reproducer
_No response_
### Expected behavior
_No response_
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
MyFaces
### JSF version
2.2
### Java version
8
### Browser(s)
_No response_ | defect | datatyble empty aria label generation in multiselect tables describe the bug during our improvements rewrites to make our webapp accessibly and fullfil the requirements of the official german bitv test we came along a slightly problem with multi select the tried to achieve that every checkbox has an hidden description like select row with identifier x so the screen reader we use nvda which is also used in the official test reads to a visual impaired person what row he selects when ticking the checkbox for this we tried to make use of the pf api attribute ariarowlabel which descriptions sounds to do exactly this but it seems that it does not as the screenreader keeps quiet regardless of the value inside so we used a pass trough attribute which is picked up correctly by the screenreader but which is not rendered at the checkbox but at the td tag so this code p column headertext selecton selectionmode multiple pt aria label xxx style width text align left generates this outcome in pf not yet tested with pf but release notes don t show any changes here and the sadly has no correspondening example td role gridcell style width text align left class ui selection column aria label xxx div class ui chkbox ui widget div id frm kleinesuchesuchergebnisse row crd tbl suchergebnisse checkbox role checkbox tabindex aria label aria checked false class ui chkbox box ui widget ui corner all ui state default as you can see this renderes an empty aria label for the checkbox div so why do i file an issue when the screen reader handles it correctly two reasons i we assume that the empty aria label should have the value the td has where the td should probably have none or the same don t know if the same would cause trouble tbh the lighthouse report claims that the checkbox has no aria label and is therefore not accessible what is the ariarowlabel for if not for this aria toggle fields do not have accessible names when a toggle field doesn t have an accessible name screen readers announce it with a generic name making it unusable for users who rely on screen readers div frm kleinesuchesuchergebnisse row crd tbl suchergebnisse checkbox ui chkbox box ui widget ui corner all ui state default further notes the lightous report always claimes that error regardless of usage of a pt or the ariarowlabel or none bith of them using the pt makes the screenreader work not using or only using the ariarowlabel does not reproducer no response expected behavior no response primefaces edition community primefaces version theme no response jsf implementation myfaces jsf version java version browser s no response | 1 |
13,677 | 2,775,779,459 | IssuesEvent | 2015-05-04 18:02:12 | douglasdrumond/macvim | https://api.github.com/repos/douglasdrumond/macvim | closed | No windows or files opening | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
Opening MacVim, either through application icon or mvim command.
What is the expected output? What do you see instead?
Expect a window to open with the desired file, or even a blank new file.
Nothing happens, the app will start (as indicated by MacVim options in toolbar)
but no windows will open, even when explicitly told to open a valid file
through File->Open...
What version of MacVim and OS X are you using (see "MacVim->About MacVim"
and "Apple Menu->About This Mac" menu items, e.g. "Snapshot 40, 10.5.6
Intel")?
MacVim Snapshot 70
OS X 10.10 Yosemite Intel
Please provide any additional information below.
Again, the app seems to be opening and running but I cannot open any files or
windows. All I get is the toolbar and the icon in the Dock.
```
Original issue reported on code.google.com by `knole...@gmail.com` on 21 Oct 2014 at 3:09 | 1.0 | No windows or files opening - ```
What steps will reproduce the problem?
Opening MacVim, either through application icon or mvim command.
What is the expected output? What do you see instead?
Expect a window to open with the desired file, or even a blank new file.
Nothing happens, the app will start (as indicated by MacVim options in toolbar)
but no windows will open, even when explicitly told to open a valid file
through File->Open...
What version of MacVim and OS X are you using (see "MacVim->About MacVim"
and "Apple Menu->About This Mac" menu items, e.g. "Snapshot 40, 10.5.6
Intel")?
MacVim Snapshot 70
OS X 10.10 Yosemite Intel
Please provide any additional information below.
Again, the app seems to be opening and running but I cannot open any files or
windows. All I get is the toolbar and the icon in the Dock.
```
Original issue reported on code.google.com by `knole...@gmail.com` on 21 Oct 2014 at 3:09 | defect | no windows or files opening what steps will reproduce the problem opening macvim either through application icon or mvim command what is the expected output what do you see instead expect a window to open with the desired file or even a blank new file nothing happens the app will start as indicated by macvim options in toolbar but no windows will open even when explicitly told to open a valid file through file open what version of macvim and os x are you using see macvim about macvim and apple menu about this mac menu items e g snapshot intel macvim snapshot os x yosemite intel please provide any additional information below again the app seems to be opening and running but i cannot open any files or windows all i get is the toolbar and the icon in the dock original issue reported on code google com by knole gmail com on oct at | 1 |
86,818 | 8,051,130,707 | IssuesEvent | 2018-08-01 15:15:51 | syndesisio/syndesis | https://api.github.com/repos/syndesisio/syndesis | closed | Logout on firefox does not work correctly | cat/bug prio/p1 qe/logout-test | I tried new logout endpoint on firefox from anonymous window and it does not work.
After I log out, I am on logout page. Then when I click on Login button, this happens:

Console looked like this:

@zregvart | 1.0 | Logout on firefox does not work correctly - I tried new logout endpoint on firefox from anonymous window and it does not work.
After I log out, I am on logout page. Then when I click on Login button, this happens:

Console looked like this:

@zregvart | non_defect | logout on firefox does not work correctly i tried new logout endpoint on firefox from anonymous window and it does not work after i log out i am on logout page then when i click on login button this happens console looked like this zregvart | 0 |
65,412 | 19,487,203,422 | IssuesEvent | 2021-12-26 16:11:03 | dkfans/keeperfx | https://api.github.com/repos/dkfans/keeperfx | opened | Custom tilesets not visibible in straight view | Priority-Medium Type-Defect | To reproduce:
1) Make any map where part of the tileset is different from default. (so an .slx file is present)
2) Start the map, and observe the custom texture. Like big boobs on a snow map.
3) Switch to straight view
-> notice the custom textures are not visible
| 1.0 | Custom tilesets not visibible in straight view - To reproduce:
1) Make any map where part of the tileset is different from default. (so an .slx file is present)
2) Start the map, and observe the custom texture. Like big boobs on a snow map.
3) Switch to straight view
-> notice the custom textures are not visible
| defect | custom tilesets not visibible in straight view to reproduce make any map where part of the tileset is different from default so an slx file is present start the map and observe the custom texture like big boobs on a snow map switch to straight view notice the custom textures are not visible | 1 |
54,529 | 23,290,449,903 | IssuesEvent | 2022-08-05 21:59:42 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | opened | Marshaller for treating char array as null-terminated string in source generation | area-System.Runtime.InteropServices | Arrays are - and should continue to be - treated as buffers in marshalling, but arrays of `char` are often used as a way to marshal a string between managed/unmanaged.
Per https://github.com/dotnet/runtime/issues/69809#issuecomment-1137708142, we can let users opt-in to treating a `ReadOnlySpan<char>` or `char[]` as a null-terminated string. We can provide a marshaller that can be used with `MarshalUsing` at the parameter level. | 1.0 | Marshaller for treating char array as null-terminated string in source generation - Arrays are - and should continue to be - treated as buffers in marshalling, but arrays of `char` are often used as a way to marshal a string between managed/unmanaged.
Per https://github.com/dotnet/runtime/issues/69809#issuecomment-1137708142, we can let users opt-in to treating a `ReadOnlySpan<char>` or `char[]` as a null-terminated string. We can provide a marshaller that can be used with `MarshalUsing` at the parameter level. | non_defect | marshaller for treating char array as null terminated string in source generation arrays are and should continue to be treated as buffers in marshalling but arrays of char are often used as a way to marshal a string between managed unmanaged per we can let users opt in to treating a readonlyspan or char as a null terminated string we can provide a marshaller that can be used with marshalusing at the parameter level | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.