Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
64,475
8,737,471,347
IssuesEvent
2018-12-11 22:41:46
docksal/docksal
https://api.github.com/repos/docksal/docksal
closed
Add documentation for having multiple web containers
size/small 🏷documentation
On gitter, @sean-e-dietrich provided me with this example: https://gist.github.com/mikecrittenden/2cbb4c1e116b405ab16c5087f0e59ac7 This would be a nice thing to put in the docs, for projects that need multiple web containers but need to stay in the same Docksal project (example: headless Drupal backend with a node-based frontend). Thoughts? Or does this already exist somewhere in the docs and I'm missing it?
1.0
Add documentation for having multiple web containers - On gitter, @sean-e-dietrich provided me with this example: https://gist.github.com/mikecrittenden/2cbb4c1e116b405ab16c5087f0e59ac7 This would be a nice thing to put in the docs, for projects that need multiple web containers but need to stay in the same Docksal project (example: headless Drupal backend with a node-based frontend). Thoughts? Or does this already exist somewhere in the docs and I'm missing it?
non_defect
add documentation for having multiple web containers on gitter sean e dietrich provided me with this example this would be a nice thing to put in the docs for projects that need multiple web containers but need to stay in the same docksal project example headless drupal backend with a node based frontend thoughts or does this already exist somewhere in the docs and i m missing it
0
479,512
13,798,091,504
IssuesEvent
2020-10-10 00:00:06
radical-cybertools/radical.entk
https://api.github.com/repos/radical-cybertools/radical.entk
opened
feature request: job_name attribute in resource description
priority:medium topic:api type:feature
See radical-collaboration/hpc-workflows/issues/122: EnTK should forward `job_name` settings to RP.
1.0
feature request: job_name attribute in resource description - See radical-collaboration/hpc-workflows/issues/122: EnTK should forward `job_name` settings to RP.
non_defect
feature request job name attribute in resource description see radical collaboration hpc workflows issues entk should forward job name settings to rp
0
566,114
16,796,399,936
IssuesEvent
2021-06-16 04:37:37
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
FTX widget should not appear when card setting is turned off
OS/Desktop QA/Yes feature/widgets priority/P2 release-notes/exclude
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> For existing users who have the cards settings on the new tab page turned off so that no cards appear, we should not show the FTX widget to these existing users when they upgrade. ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Use a Brave version lower than 1.26.21 and turn off cards. See image below. 2. Upgrade to latest 1.26.22 or higher 3. Notice that the FTX widget will appear as the sole widget on upgrade even though card settings is off. <img width="411" alt="Screen Shot 2021-05-19 at 11 07 45 AM" src="https://user-images.githubusercontent.com/5951041/118862404-8c173800-b892-11eb-834d-15b0cf7142e1.png"> ## Actual result: FTX widget appears upon upgrade as the sole card in the widget stack for existing users and it comes across as not respecting a setting that was previously set. ## Expected result: FTX widget should not appear for users who have turned off all cards. If a user has cards enabled, FTX will appear for them as part of the list. New card tooltip promoting FTX should also appear for existing users with cards turned off. See: https://github.com/brave/brave-browser/issues/9616 ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easily. ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave 1.26.22 and higher ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? n/a - Can you reproduce this issue with the beta channel? yes - Can you reproduce this issue with the nightly channel? yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? n/a - Does the issue resolve itself when disabling Brave Rewards? n/a - Is the issue reproducible on the latest version of Chrome? n/a ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
1.0
FTX widget should not appear when card setting is turned off - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> For existing users who have the cards settings on the new tab page turned off so that no cards appear, we should not show the FTX widget to these existing users when they upgrade. ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Use a Brave version lower than 1.26.21 and turn off cards. See image below. 2. Upgrade to latest 1.26.22 or higher 3. Notice that the FTX widget will appear as the sole widget on upgrade even though card settings is off. <img width="411" alt="Screen Shot 2021-05-19 at 11 07 45 AM" src="https://user-images.githubusercontent.com/5951041/118862404-8c173800-b892-11eb-834d-15b0cf7142e1.png"> ## Actual result: FTX widget appears upon upgrade as the sole card in the widget stack for existing users and it comes across as not respecting a setting that was previously set. ## Expected result: FTX widget should not appear for users who have turned off all cards. If a user has cards enabled, FTX will appear for them as part of the list. New card tooltip promoting FTX should also appear for existing users with cards turned off. See: https://github.com/brave/brave-browser/issues/9616 ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easily. ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave 1.26.22 and higher ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? n/a - Can you reproduce this issue with the beta channel? yes - Can you reproduce this issue with the nightly channel? yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? n/a - Does the issue resolve itself when disabling Brave Rewards? n/a - Is the issue reproducible on the latest version of Chrome? n/a ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue-->
non_defect
ftx widget should not appear when card setting is turned off have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description for existing users who have the cards settings on the new tab page turned off so that no cards appear we should not show the ftx widget to these existing users when they upgrade steps to reproduce use a brave version lower than and turn off cards see image below upgrade to latest or higher notice that the ftx widget will appear as the sole widget on upgrade even though card settings is off img width alt screen shot at am src actual result ftx widget appears upon upgrade as the sole card in the widget stack for existing users and it comes across as not respecting a setting that was previously set expected result ftx widget should not appear for users who have turned off all cards if a user has cards enabled ftx will appear for them as part of the list new card tooltip promoting ftx should also appear for existing users with cards turned off see reproduces how often easily brave version brave version info brave and higher version channel information can you reproduce this issue with the current release n a can you reproduce this issue with the beta channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields n a does the issue resolve itself when disabling brave rewards n a is the issue reproducible on the latest version of chrome n a miscellaneous information
0
274,591
23,851,973,165
IssuesEvent
2022-09-06 18:50:56
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
opened
[chip-test] Add connectivity tests for AST signals
IP:ast Component:ChipLevelTest
### Test point name TBD ### Host side component _No response_ ### OpenTitanTool infrastructure implemented _No response_ ### Contact person @tjaychen @msf ### Checklist Please fill out this checklist as items are completed. Link to PRs and issues as appropriate. - [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes) - [ ] Device-side (C) component developed - [ ] Bazel build rules developed - [ ] Host-side component developed - [ ] HJSON test plan updated with test name (so it shows up in the dashboard) - [ ] Test added to dvsim nightly regression (and passing at time of checking)
1.0
[chip-test] Add connectivity tests for AST signals - ### Test point name TBD ### Host side component _No response_ ### OpenTitanTool infrastructure implemented _No response_ ### Contact person @tjaychen @msf ### Checklist Please fill out this checklist as items are completed. Link to PRs and issues as appropriate. - [ ] Check if existing test covers most or all of this testpoint (if so, either extend said test to cover all points, or skip the next 3 checkboxes) - [ ] Device-side (C) component developed - [ ] Bazel build rules developed - [ ] Host-side component developed - [ ] HJSON test plan updated with test name (so it shows up in the dashboard) - [ ] Test added to dvsim nightly regression (and passing at time of checking)
non_defect
add connectivity tests for ast signals test point name tbd host side component no response opentitantool infrastructure implemented no response contact person tjaychen msf checklist please fill out this checklist as items are completed link to prs and issues as appropriate check if existing test covers most or all of this testpoint if so either extend said test to cover all points or skip the next checkboxes device side c component developed bazel build rules developed host side component developed hjson test plan updated with test name so it shows up in the dashboard test added to dvsim nightly regression and passing at time of checking
0
49,876
26,383,507,759
IssuesEvent
2023-01-12 10:27:04
PandemiaProject/pandemia
https://api.github.com/repos/PandemiaProject/pandemia
closed
Performance - Review interface between C/Python/NumPy
performance
All params passed to C as 1D NP arrays. Requires manual calculation of array indices. Would there be a better way to do this?
True
Performance - Review interface between C/Python/NumPy - All params passed to C as 1D NP arrays. Requires manual calculation of array indices. Would there be a better way to do this?
non_defect
performance review interface between c python numpy all params passed to c as np arrays requires manual calculation of array indices would there be a better way to do this
0
218,877
7,332,755,720
IssuesEvent
2018-03-05 17:12:17
NCEAS/metacat
https://api.github.com/repos/NCEAS/metacat
closed
Make reCaptcha work with AJAX include
Priority: Normal Status: Closed Tracker: Task
--- Author Name: **ben leinfelder** (ben leinfelder) Original Redmine Issue: 6098, https://projects.ecoinformatics.org/ecoinfo/issues/6098 Original Date: 2013-09-20 Original Assignee: Jing Tao --- Instead of having the perl code generate the catcha content, the template can include the ajax-style of constructing the captcha. This will allow us to use the same method when we include the registration form in the MetacatUI as well. See: https://developers.google.com/recaptcha/docs/display?csw=1#AJAX http://stackoverflow.com/questions/7261436/how-can-i-load-a-recaptcha-form-using-jquery-ajax-while-leaving-the-recaptcha-sc
1.0
Make reCaptcha work with AJAX include - --- Author Name: **ben leinfelder** (ben leinfelder) Original Redmine Issue: 6098, https://projects.ecoinformatics.org/ecoinfo/issues/6098 Original Date: 2013-09-20 Original Assignee: Jing Tao --- Instead of having the perl code generate the catcha content, the template can include the ajax-style of constructing the captcha. This will allow us to use the same method when we include the registration form in the MetacatUI as well. See: https://developers.google.com/recaptcha/docs/display?csw=1#AJAX http://stackoverflow.com/questions/7261436/how-can-i-load-a-recaptcha-form-using-jquery-ajax-while-leaving-the-recaptcha-sc
non_defect
make recaptcha work with ajax include author name ben leinfelder ben leinfelder original redmine issue original date original assignee jing tao instead of having the perl code generate the catcha content the template can include the ajax style of constructing the captcha this will allow us to use the same method when we include the registration form in the metacatui as well see
0
169,539
20,841,777,630
IssuesEvent
2022-03-21 01:30:38
uniquelyparticular/serverless-oauth
https://api.github.com/repos/uniquelyparticular/serverless-oauth
opened
CVE-2022-24772 (High) detected in node-forge-0.8.4.tgz, node-forge-0.7.4.tgz
security vulnerability
## CVE-2022-24772 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-forge-0.8.4.tgz</b>, <b>node-forge-0.7.4.tgz</b></p></summary> <p> <details><summary><b>node-forge-0.8.4.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.8.4.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.8.4.tgz</a></p> <p>Path to dependency file: /generic-oauth/package.json</p> <p>Path to vulnerable library: /tmp/git/generic-oauth/node_modules/gcs-resumable-upload/node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - firebase-admin-8.0.0.tgz (Root Library) - storage-2.5.0.tgz - gcs-resumable-upload-1.1.0.tgz - google-auth-library-3.1.2.tgz - gtoken-2.3.3.tgz - google-p12-pem-1.0.4.tgz - :x: **node-forge-0.8.4.tgz** (Vulnerable Library) </details> <details><summary><b>node-forge-0.7.4.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.4.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.4.tgz</a></p> <p>Path to dependency file: /generic-oauth/package.json</p> <p>Path to vulnerable library: /tmp/git/generic-oauth/node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - firebase-admin-8.0.0.tgz (Root Library) - :x: **node-forge-0.7.4.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not check for tailing garbage bytes after decoding a `DigestInfo` ASN.1 structure. This can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds. <p>Publish Date: 2022-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24772>CVE-2022-24772</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772</a></p> <p>Release Date: 2022-03-18</p> <p>Fix Resolution: node-forge - 1.3.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24772 (High) detected in node-forge-0.8.4.tgz, node-forge-0.7.4.tgz - ## CVE-2022-24772 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-forge-0.8.4.tgz</b>, <b>node-forge-0.7.4.tgz</b></p></summary> <p> <details><summary><b>node-forge-0.8.4.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.8.4.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.8.4.tgz</a></p> <p>Path to dependency file: /generic-oauth/package.json</p> <p>Path to vulnerable library: /tmp/git/generic-oauth/node_modules/gcs-resumable-upload/node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - firebase-admin-8.0.0.tgz (Root Library) - storage-2.5.0.tgz - gcs-resumable-upload-1.1.0.tgz - google-auth-library-3.1.2.tgz - gtoken-2.3.3.tgz - google-p12-pem-1.0.4.tgz - :x: **node-forge-0.8.4.tgz** (Vulnerable Library) </details> <details><summary><b>node-forge-0.7.4.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.7.4.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.7.4.tgz</a></p> <p>Path to dependency file: /generic-oauth/package.json</p> <p>Path to vulnerable library: /tmp/git/generic-oauth/node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - firebase-admin-8.0.0.tgz (Root Library) - :x: **node-forge-0.7.4.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not check for tailing garbage bytes after decoding a `DigestInfo` ASN.1 structure. This can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds. <p>Publish Date: 2022-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24772>CVE-2022-24772</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24772</a></p> <p>Release Date: 2022-03-18</p> <p>Fix Resolution: node-forge - 1.3.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in node forge tgz node forge tgz cve high severity vulnerability vulnerable libraries node forge tgz node forge tgz node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file generic oauth package json path to vulnerable library tmp git generic oauth node modules gcs resumable upload node modules node forge package json dependency hierarchy firebase admin tgz root library storage tgz gcs resumable upload tgz google auth library tgz gtoken tgz google pem tgz x node forge tgz vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file generic oauth package json path to vulnerable library tmp git generic oauth node modules node forge package json dependency hierarchy firebase admin tgz root library x node forge tgz vulnerable library found in base branch master vulnerability details forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code does not check for tailing garbage bytes after decoding a digestinfo asn structure this can allow padding bytes to be removed and garbage data added to forge a signature when a low public exponent is being used the issue has been addressed in node forge version there are currently no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge step up your open source security game with whitesource
0
37,247
8,308,009,973
IssuesEvent
2018-09-23 16:04:00
umple/umple
https://api.github.com/repos/umple/umple
closed
Constraints with == converted to equals() in Java, but failing on primitives
Component-SemanticsAndGen Diffic-Easy Priority-High Status-Accepted Type-Defect constraints stateMachines ucosp
Originally reported on Google Code with ID 589 Owned by `redeyedmars` --- In the SecurityLight example (umpleonline/ump/SecurityLight the generated Java has a bug where the original constraint has == to compare two booleans, but it has been changed to generate a call to equals(). But this doesn't work with Booleans http://cruise.eecs.uottawa.ca/umpleonline/?example=SecurityLight --- Reported by @umple on 2014-06-09 20:56:58
1.0
Constraints with == converted to equals() in Java, but failing on primitives - Originally reported on Google Code with ID 589 Owned by `redeyedmars` --- In the SecurityLight example (umpleonline/ump/SecurityLight the generated Java has a bug where the original constraint has == to compare two booleans, but it has been changed to generate a call to equals(). But this doesn't work with Booleans http://cruise.eecs.uottawa.ca/umpleonline/?example=SecurityLight --- Reported by @umple on 2014-06-09 20:56:58
defect
constraints with converted to equals in java but failing on primitives originally reported on google code with id owned by redeyedmars in the securitylight example umpleonline ump securitylight the generated java has a bug where the original constraint has to compare two booleans but it has been changed to generate a call to equals but this doesn t work with booleans reported by umple on
1
70,438
23,167,875,234
IssuesEvent
2022-07-30 08:18:25
dkfans/keeperfx
https://api.github.com/repos/dkfans/keeperfx
closed
Digging sound does not play in possession
Type-Defect
Instead, it plays the 'shot hit wall' sound, which sounds similar but is different. When imps dig on their own, they make 6 random dig sounds, hardcoded in the 'instf_dig' function. It's 63+6. When possessing an imp and digging there, it fails to use that function, and it does not create the sounds or visual effects to go with that. What the player does hear, is the 'shot had generic impact' sound, listed as 'HitWallSound' at '[shot23]' in 'magic.cfg'. It's 128+3. It also fails to produce the visual effects like rock chips. The correct behavior would be for the possession sounds to match the regular sounds. So 63+3 when digging something diggable, but 128+3 when hitting something that cannot be dug. With the visual effects to match too. The current behavior has always been this way.
1.0
Digging sound does not play in possession - Instead, it plays the 'shot hit wall' sound, which sounds similar but is different. When imps dig on their own, they make 6 random dig sounds, hardcoded in the 'instf_dig' function. It's 63+6. When possessing an imp and digging there, it fails to use that function, and it does not create the sounds or visual effects to go with that. What the player does hear, is the 'shot had generic impact' sound, listed as 'HitWallSound' at '[shot23]' in 'magic.cfg'. It's 128+3. It also fails to produce the visual effects like rock chips. The correct behavior would be for the possession sounds to match the regular sounds. So 63+3 when digging something diggable, but 128+3 when hitting something that cannot be dug. With the visual effects to match too. The current behavior has always been this way.
defect
digging sound does not play in possession instead it plays the shot hit wall sound which sounds similar but is different when imps dig on their own they make random dig sounds hardcoded in the instf dig function it s when possessing an imp and digging there it fails to use that function and it does not create the sounds or visual effects to go with that what the player does hear is the shot had generic impact sound listed as hitwallsound at in magic cfg it s it also fails to produce the visual effects like rock chips the correct behavior would be for the possession sounds to match the regular sounds so when digging something diggable but when hitting something that cannot be dug with the visual effects to match too the current behavior has always been this way
1
425,533
12,341,886,159
IssuesEvent
2020-05-14 23:05:05
earthlab/matplotcheck
https://api.github.com/repos/earthlab/matplotcheck
closed
Timeseries data check failing - coersion issue
bug high priority time-series
``` # Get data data = et.data.get_data("colorado-flood") os.chdir(os.path.join(et.io.HOME, 'earth-analytics')) import matplotcheck.notebook as nb import matplotcheck.timeseries as ts import matplotcheck.raster as ra # BEGIN SOLUTION import os import matplotlib.pyplot as plt import matplotlib.dates as mdates from matplotlib.dates import DateFormatter import seaborn as sns import numpy as np import pandas as pd import earthpy as et import hydrofunctions as hf import urllib from pandas.plotting import register_matplotlib_converters register_matplotlib_converters() # prettier plotting with seaborn sns.set(font_scale=1.5) sns.set_style("whitegrid") # import file f = "data/colorado-flood/discharge/06730200-discharge-daily-1986-2013.txt" discharge = pd.read_csv(f, skiprows=23, header=[1, 2], sep='\t', parse_dates=[2]) # drop one level of index discharge.columns = discharge.columns.droplevel(1) # set the date column as the index discharge = discharge.set_index(["datetime"]) monthly_max_all = discharge.resample("M").max() monthly_max = monthly_max_all['1990':'2014'] fig, ax = plt.subplots(figsize=(10, 8)) ax.scatter(x=monthly_max.index, y=monthly_max["17663_00060_00003"], color="purple") ax.set_title( "HW Plot 1: Stream Discharge - Monthly Max Value\n be sure to add x and y labels (not shown here)") ax.set(xlabel="Date") # END SOLUTION ### DO NOT REMOVE LINE BELOW ### plot_1_ts = nb.convert_axes(plt, which_axes="current") mpc_plot_1_ts = mts.TimeSeriesTester(plot_1_ts) results = [] ``` TEST ``` mts.TimeSeriesTester(plot_1_ts) mpc_plot_1_ts.assert_xydata(xy_expected = precip.reset_index(), xtime=True, xcol="DATE", ycol="HPCP") ``` Output error: i suspect this has something to do with NAN values but i am not sure ``` --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-38-1b4d8c2d48dc> in <module> 4 xtime=True, 5 xcol="DATE", ----> 6 ycol="HPCP") 7 8 #precip ~/Documents/github/0-python/matplotcheck/matplotcheck/base.py in assert_xydata(self, xy_expected, xcol, ycol, points_only, xtime, xlabels, tolerence, message) 938 try: 939 np.testing.assert_array_max_ulp( --> 940 np.array(xy_data["x"]), np.array(xy_expected[xcol]) 941 ) 942 except AssertionError: ~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/testing/_private/utils.py in assert_array_max_ulp(a, b, maxulp, dtype) 1617 __tracebackhide__ = True # Hide traceback for py.test 1618 import numpy as np -> 1619 ret = nulp_diff(a, b, dtype) 1620 if not np.all(ret <= maxulp): 1621 raise AssertionError("Arrays are not almost equal up to %g ULP" % ~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/testing/_private/utils.py in nulp_diff(x, y, dtype) 1658 y = np.array(y) 1659 -> 1660 t = np.common_type(x, y) 1661 if np.iscomplexobj(x) or np.iscomplexobj(y): 1662 raise NotImplementedError("_nulp not implemented for complex array") <__array_function__ internals> in common_type(*args, **kwargs) ~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/lib/type_check.py in common_type(*arrays) 719 p = array_precision.get(t, None) 720 if p is None: --> 721 raise TypeError("can't get common type for non-numeric array") 722 precision = max(precision, p) 723 if is_complex: TypeError: can't get common type for non-numeric array ``` 1. i suspect this has to do with NA values. we need to clear those out
1.0
Timeseries data check failing - coersion issue - ``` # Get data data = et.data.get_data("colorado-flood") os.chdir(os.path.join(et.io.HOME, 'earth-analytics')) import matplotcheck.notebook as nb import matplotcheck.timeseries as ts import matplotcheck.raster as ra # BEGIN SOLUTION import os import matplotlib.pyplot as plt import matplotlib.dates as mdates from matplotlib.dates import DateFormatter import seaborn as sns import numpy as np import pandas as pd import earthpy as et import hydrofunctions as hf import urllib from pandas.plotting import register_matplotlib_converters register_matplotlib_converters() # prettier plotting with seaborn sns.set(font_scale=1.5) sns.set_style("whitegrid") # import file f = "data/colorado-flood/discharge/06730200-discharge-daily-1986-2013.txt" discharge = pd.read_csv(f, skiprows=23, header=[1, 2], sep='\t', parse_dates=[2]) # drop one level of index discharge.columns = discharge.columns.droplevel(1) # set the date column as the index discharge = discharge.set_index(["datetime"]) monthly_max_all = discharge.resample("M").max() monthly_max = monthly_max_all['1990':'2014'] fig, ax = plt.subplots(figsize=(10, 8)) ax.scatter(x=monthly_max.index, y=monthly_max["17663_00060_00003"], color="purple") ax.set_title( "HW Plot 1: Stream Discharge - Monthly Max Value\n be sure to add x and y labels (not shown here)") ax.set(xlabel="Date") # END SOLUTION ### DO NOT REMOVE LINE BELOW ### plot_1_ts = nb.convert_axes(plt, which_axes="current") mpc_plot_1_ts = mts.TimeSeriesTester(plot_1_ts) results = [] ``` TEST ``` mts.TimeSeriesTester(plot_1_ts) mpc_plot_1_ts.assert_xydata(xy_expected = precip.reset_index(), xtime=True, xcol="DATE", ycol="HPCP") ``` Output error: i suspect this has something to do with NAN values but i am not sure ``` --------------------------------------------------------------------------- TypeError Traceback (most recent call last) <ipython-input-38-1b4d8c2d48dc> in <module> 4 xtime=True, 5 xcol="DATE", ----> 6 ycol="HPCP") 7 8 #precip ~/Documents/github/0-python/matplotcheck/matplotcheck/base.py in assert_xydata(self, xy_expected, xcol, ycol, points_only, xtime, xlabels, tolerence, message) 938 try: 939 np.testing.assert_array_max_ulp( --> 940 np.array(xy_data["x"]), np.array(xy_expected[xcol]) 941 ) 942 except AssertionError: ~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/testing/_private/utils.py in assert_array_max_ulp(a, b, maxulp, dtype) 1617 __tracebackhide__ = True # Hide traceback for py.test 1618 import numpy as np -> 1619 ret = nulp_diff(a, b, dtype) 1620 if not np.all(ret <= maxulp): 1621 raise AssertionError("Arrays are not almost equal up to %g ULP" % ~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/testing/_private/utils.py in nulp_diff(x, y, dtype) 1658 y = np.array(y) 1659 -> 1660 t = np.common_type(x, y) 1661 if np.iscomplexobj(x) or np.iscomplexobj(y): 1662 raise NotImplementedError("_nulp not implemented for complex array") <__array_function__ internals> in common_type(*args, **kwargs) ~/miniconda3/envs/earth-analytics-python/lib/python3.7/site-packages/numpy/lib/type_check.py in common_type(*arrays) 719 p = array_precision.get(t, None) 720 if p is None: --> 721 raise TypeError("can't get common type for non-numeric array") 722 precision = max(precision, p) 723 if is_complex: TypeError: can't get common type for non-numeric array ``` 1. i suspect this has to do with NA values. we need to clear those out
non_defect
timeseries data check failing coersion issue get data data et data get data colorado flood os chdir os path join et io home earth analytics import matplotcheck notebook as nb import matplotcheck timeseries as ts import matplotcheck raster as ra begin solution import os import matplotlib pyplot as plt import matplotlib dates as mdates from matplotlib dates import dateformatter import seaborn as sns import numpy as np import pandas as pd import earthpy as et import hydrofunctions as hf import urllib from pandas plotting import register matplotlib converters register matplotlib converters prettier plotting with seaborn sns set font scale sns set style whitegrid import file f data colorado flood discharge discharge daily txt discharge pd read csv f skiprows header sep t parse dates drop one level of index discharge columns discharge columns droplevel set the date column as the index discharge discharge set index monthly max all discharge resample m max monthly max monthly max all fig ax plt subplots figsize ax scatter x monthly max index y monthly max color purple ax set title hw plot stream discharge monthly max value n be sure to add x and y labels not shown here ax set xlabel date end solution do not remove line below plot ts nb convert axes plt which axes current mpc plot ts mts timeseriestester plot ts results test mts timeseriestester plot ts mpc plot ts assert xydata xy expected precip reset index xtime true xcol date ycol hpcp output error i suspect this has something to do with nan values but i am not sure typeerror traceback most recent call last in xtime true xcol date ycol hpcp precip documents github python matplotcheck matplotcheck base py in assert xydata self xy expected xcol ycol points only xtime xlabels tolerence message try np testing assert array max ulp np array xy data np array xy expected except assertionerror envs earth analytics python lib site packages numpy testing private utils py in assert array max ulp a b maxulp dtype tracebackhide true hide traceback for py test import numpy as np ret nulp diff a b dtype if not np all ret maxulp raise assertionerror arrays are not almost equal up to g ulp envs earth analytics python lib site packages numpy testing private utils py in nulp diff x y dtype y np array y t np common type x y if np iscomplexobj x or np iscomplexobj y raise notimplementederror nulp not implemented for complex array in common type args kwargs envs earth analytics python lib site packages numpy lib type check py in common type arrays p array precision get t none if p is none raise typeerror can t get common type for non numeric array precision max precision p if is complex typeerror can t get common type for non numeric array i suspect this has to do with na values we need to clear those out
0
29,303
5,639,213,617
IssuesEvent
2017-04-06 13:53:08
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
Task window in Simple View GUI of manager has scale / placment / content issues
C: Manager P: Undetermined T: Defect
**Reported by verduin on 21 Jan 38899053 19:06 UTC** [[BR]]Placement and scale are perhaps the cause of missing border (see ticket #119) in the "Simple View" and also cause overwriting a portion of the task selection scroll above the window. In addition placement of slide show images inside the window leaves void background below the images. For the purposes of task selection, consider truncating task text to limit the length of the selector button above the window. I submit this in the belief that a series of shorter buttons is more useful than a single long button in the display. I believe the button text is essentially redundant with the text following "Application:" inside the window? Further, there is no text inside the window to give specific identity to the task displayed. In the "Advanced View" this is the "Name" field. I find the tasks selected for simple view are limited to active tasks, but exclude finished and waiting to start tasks. The selection is good but I admit confusion when finding more tasks in the Advanced View than were displayable in the Simple View. You might consider this a defect in user training, but I request consideration for a window title? This version is current today from Fedora 10 Preview repositories. Migrated-From: http://boinc.berkeley.edu/trac/ticket/779
1.0
Task window in Simple View GUI of manager has scale / placment / content issues - **Reported by verduin on 21 Jan 38899053 19:06 UTC** [[BR]]Placement and scale are perhaps the cause of missing border (see ticket #119) in the "Simple View" and also cause overwriting a portion of the task selection scroll above the window. In addition placement of slide show images inside the window leaves void background below the images. For the purposes of task selection, consider truncating task text to limit the length of the selector button above the window. I submit this in the belief that a series of shorter buttons is more useful than a single long button in the display. I believe the button text is essentially redundant with the text following "Application:" inside the window? Further, there is no text inside the window to give specific identity to the task displayed. In the "Advanced View" this is the "Name" field. I find the tasks selected for simple view are limited to active tasks, but exclude finished and waiting to start tasks. The selection is good but I admit confusion when finding more tasks in the Advanced View than were displayable in the Simple View. You might consider this a defect in user training, but I request consideration for a window title? This version is current today from Fedora 10 Preview repositories. Migrated-From: http://boinc.berkeley.edu/trac/ticket/779
defect
task window in simple view gui of manager has scale placment content issues reported by verduin on jan utc placement and scale are perhaps the cause of missing border see ticket in the simple view and also cause overwriting a portion of the task selection scroll above the window in addition placement of slide show images inside the window leaves void background below the images for the purposes of task selection consider truncating task text to limit the length of the selector button above the window i submit this in the belief that a series of shorter buttons is more useful than a single long button in the display i believe the button text is essentially redundant with the text following application inside the window further there is no text inside the window to give specific identity to the task displayed in the advanced view this is the name field i find the tasks selected for simple view are limited to active tasks but exclude finished and waiting to start tasks the selection is good but i admit confusion when finding more tasks in the advanced view than were displayable in the simple view you might consider this a defect in user training but i request consideration for a window title this version is current today from fedora preview repositories migrated from
1
54,010
13,240,467,386
IssuesEvent
2020-08-19 06:24:05
airdcpp-web/airdcpp-webclient
https://api.github.com/repos/airdcpp-web/airdcpp-webclient
closed
Bad size in the "download - browse" pop up
bug external portable builds
### Current UI and client versions: > Application version: AirDC++w 2.8.0 x86_64 Web UI version: 2.8.0 Web UI build date: July 11, 2020 1:08 PM ### Operating system: Debian 10 I guess ### Steps to reproduce the issue: - Have a disk with white-space in name, do a download on it using the download browse - Do the action again but this time choose "previous" and before clicking see the size of the HDD (/run/media/me/Name with whitespace/ (7.82 GiB free) - here the "free" size is bad
1.0
Bad size in the "download - browse" pop up - ### Current UI and client versions: > Application version: AirDC++w 2.8.0 x86_64 Web UI version: 2.8.0 Web UI build date: July 11, 2020 1:08 PM ### Operating system: Debian 10 I guess ### Steps to reproduce the issue: - Have a disk with white-space in name, do a download on it using the download browse - Do the action again but this time choose "previous" and before clicking see the size of the HDD (/run/media/me/Name with whitespace/ (7.82 GiB free) - here the "free" size is bad
non_defect
bad size in the download browse pop up current ui and client versions application version airdc w web ui version web ui build date july pm operating system debian i guess steps to reproduce the issue have a disk with white space in name do a download on it using the download browse do the action again but this time choose previous and before clicking see the size of the hdd run media me name with whitespace gib free here the free size is bad
0
152,534
13,459,699,729
IssuesEvent
2020-09-09 12:38:51
zhorton34/vuejs-form
https://api.github.com/repos/zhorton34/vuejs-form
closed
Trying to extend custom rule, leads to error
bug documentation
First of all, great package, thanks a lot! I try to extend a custom rule, but I always get this error: ~~~ Uncaught TypeError: check.rule is not a function at eval (validator.js:299) at Array.reduce (<anonymous>) at Validator.checkRulesAndFillErrorBag (validator.js:298) at Validator.validate (validator.js:281) at Proxy.validate (use.js:52) at extend.html:22 Even if i try the ~~~ Even if I use this example snippet from the documentation: ~~~html <!DOCTYPE html> <html lang="en"> <head><script src='https://unpkg.com/vuejs-form@latest/build/vuejs-form.min.js'></script></head> <body> <script> let example = form({ name: 'timmy' }).rules({ name: 'uppercase' }); example.validator().extend('uppercase', [ ':attribute must be uppercase', ({ value, validator, parameters }) => value === value.toUpperCase(), ]); example.validate().errors().has('name'); </script> </body> </html> ~~~ The error occurs in all my browsers (Firefox 80 and Chrome 85 on Ubuntu)
1.0
Trying to extend custom rule, leads to error - First of all, great package, thanks a lot! I try to extend a custom rule, but I always get this error: ~~~ Uncaught TypeError: check.rule is not a function at eval (validator.js:299) at Array.reduce (<anonymous>) at Validator.checkRulesAndFillErrorBag (validator.js:298) at Validator.validate (validator.js:281) at Proxy.validate (use.js:52) at extend.html:22 Even if i try the ~~~ Even if I use this example snippet from the documentation: ~~~html <!DOCTYPE html> <html lang="en"> <head><script src='https://unpkg.com/vuejs-form@latest/build/vuejs-form.min.js'></script></head> <body> <script> let example = form({ name: 'timmy' }).rules({ name: 'uppercase' }); example.validator().extend('uppercase', [ ':attribute must be uppercase', ({ value, validator, parameters }) => value === value.toUpperCase(), ]); example.validate().errors().has('name'); </script> </body> </html> ~~~ The error occurs in all my browsers (Firefox 80 and Chrome 85 on Ubuntu)
non_defect
trying to extend custom rule leads to error first of all great package thanks a lot i try to extend a custom rule but i always get this error uncaught typeerror check rule is not a function at eval validator js at array reduce at validator checkrulesandfillerrorbag validator js at validator validate validator js at proxy validate use js at extend html even if i try the even if i use this example snippet from the documentation html script src let example form name timmy rules name uppercase example validator extend uppercase attribute must be uppercase value validator parameters value value touppercase example validate errors has name the error occurs in all my browsers firefox and chrome on ubuntu
0
44,896
12,422,227,930
IssuesEvent
2020-05-23 20:56:41
numpy/numpy.org
https://api.github.com/repos/numpy/numpy.org
closed
Error when running command in Shell
defect
I'm getting an error when I run `rng = np.random.default_rng()` in the shell.
1.0
Error when running command in Shell - I'm getting an error when I run `rng = np.random.default_rng()` in the shell.
defect
error when running command in shell i m getting an error when i run rng np random default rng in the shell
1
52,295
13,219,033,282
IssuesEvent
2020-08-17 09:45:58
combatopera/lagoon
https://api.github.com/repos/combatopera/lagoon
opened
strategy for evolving Program api vis-a-vis subcommands
defect
currently adding a method will break code that expects that name to be a subcommand
1.0
strategy for evolving Program api vis-a-vis subcommands - currently adding a method will break code that expects that name to be a subcommand
defect
strategy for evolving program api vis a vis subcommands currently adding a method will break code that expects that name to be a subcommand
1
165,546
6,278,016,753
IssuesEvent
2017-07-18 13:34:47
openshift/origin
https://api.github.com/repos/openshift/origin
opened
add test to check configapi defaulting
kind/post-rebase priority/P1
add tests to specifically test the configapi defaulting. Probably need fuzzers here too.
1.0
add test to check configapi defaulting - add tests to specifically test the configapi defaulting. Probably need fuzzers here too.
non_defect
add test to check configapi defaulting add tests to specifically test the configapi defaulting probably need fuzzers here too
0
78,240
27,387,822,884
IssuesEvent
2023-02-28 14:27:24
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
performace regression after upgrade from 3.14.16 to 3.17.8
T: Defect
### Expected behavior No performance regressions no more garbage ### Actual behavior After upgrading the jooq dependecy I noticed a performance hit to my whole application The G1 Survivor Space & Gc Pauses ~+- doubled which overall resulted in app response durations The constant the changes was jooq version 3.14.16 -> org.jooq:jooq:3.17.8 Other settings remained the same, that is why i am reaching out. Asking what might have changed in the jooq internals. Overall i am only using jooq `dsl` + generated (java) `Record` classes for mapping ### Steps to reproduce the problem I am happy to help with a debug session ### jOOQ Version 3.17.8 ### Database product and version PostgreSQL 13.7 on gcloud ### Java Version azul/zulu-openjdk:17-jre ### OS Version _No response_ ### JDBC driver name and version (include name if unofficial driver) org.postgresql:postgresql:42.5.4
1.0
performace regression after upgrade from 3.14.16 to 3.17.8 - ### Expected behavior No performance regressions no more garbage ### Actual behavior After upgrading the jooq dependecy I noticed a performance hit to my whole application The G1 Survivor Space & Gc Pauses ~+- doubled which overall resulted in app response durations The constant the changes was jooq version 3.14.16 -> org.jooq:jooq:3.17.8 Other settings remained the same, that is why i am reaching out. Asking what might have changed in the jooq internals. Overall i am only using jooq `dsl` + generated (java) `Record` classes for mapping ### Steps to reproduce the problem I am happy to help with a debug session ### jOOQ Version 3.17.8 ### Database product and version PostgreSQL 13.7 on gcloud ### Java Version azul/zulu-openjdk:17-jre ### OS Version _No response_ ### JDBC driver name and version (include name if unofficial driver) org.postgresql:postgresql:42.5.4
defect
performace regression after upgrade from to expected behavior no performance regressions no more garbage actual behavior after upgrading the jooq dependecy i noticed a performance hit to my whole application the survivor space gc pauses doubled which overall resulted in app response durations the constant the changes was jooq version org jooq jooq other settings remained the same that is why i am reaching out asking what might have changed in the jooq internals overall i am only using jooq dsl generated java record classes for mapping steps to reproduce the problem i am happy to help with a debug session jooq version database product and version postgresql on gcloud java version azul zulu openjdk jre os version no response jdbc driver name and version include name if unofficial driver org postgresql postgresql
1
559,285
16,554,318,391
IssuesEvent
2021-05-28 12:17:20
FAIRsharing/fairsharing.github.io
https://api.github.com/repos/FAIRsharing/fairsharing.github.io
closed
Terms of Use redecorate
Normal priority
Terms of Use contents style need to be justified to be the same as other pages.
1.0
Terms of Use redecorate - Terms of Use contents style need to be justified to be the same as other pages.
non_defect
terms of use redecorate terms of use contents style need to be justified to be the same as other pages
0
69,274
22,309,796,954
IssuesEvent
2022-06-13 15:56:21
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]: Same error with every new chrome version when using bidi_connection()
C-py I-defect
### What happened? driver.bidi_connection() fails every time when new chrome version is released. I have opened this defect in the past. The line that is causing it is located under webdriver.py devtools = import_module("selenium.webdriver.common.devtools.v{}".format(version)) Devtools module don`t seem to be updating fast enough. It is hard to use bidi_connection in a professional setting. Currently it fails for chrome version 102 ### How can we reproduce the issue? ```shell Update chrome and webdriver to version 102. Run: async with driver.bidi_connection() as bidi_session: session, devtools = bidi_session.session, bidi_session.devtools ``` ### Relevant log output ```shell ModuleNotFoundError ``` ### Operating System windows and mac os ### Selenium version python ### What are the browser(s) and version(s) where you see this issue? chrome 102 ### What are the browser driver(s) and version(s) where you see this issue? 102.0.5005 ### Are you using Selenium Grid? no
1.0
[🐛 Bug]: Same error with every new chrome version when using bidi_connection() - ### What happened? driver.bidi_connection() fails every time when new chrome version is released. I have opened this defect in the past. The line that is causing it is located under webdriver.py devtools = import_module("selenium.webdriver.common.devtools.v{}".format(version)) Devtools module don`t seem to be updating fast enough. It is hard to use bidi_connection in a professional setting. Currently it fails for chrome version 102 ### How can we reproduce the issue? ```shell Update chrome and webdriver to version 102. Run: async with driver.bidi_connection() as bidi_session: session, devtools = bidi_session.session, bidi_session.devtools ``` ### Relevant log output ```shell ModuleNotFoundError ``` ### Operating System windows and mac os ### Selenium version python ### What are the browser(s) and version(s) where you see this issue? chrome 102 ### What are the browser driver(s) and version(s) where you see this issue? 102.0.5005 ### Are you using Selenium Grid? no
defect
same error with every new chrome version when using bidi connection what happened driver bidi connection fails every time when new chrome version is released i have opened this defect in the past the line that is causing it is located under webdriver py devtools import module selenium webdriver common devtools v format version devtools module don t seem to be updating fast enough it is hard to use bidi connection in a professional setting currently it fails for chrome version how can we reproduce the issue shell update chrome and webdriver to version run async with driver bidi connection as bidi session session devtools bidi session session bidi session devtools relevant log output shell modulenotfounderror operating system windows and mac os selenium version python what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue are you using selenium grid no
1
2,234
2,671,186,403
IssuesEvent
2015-03-24 03:03:55
socketstream/socketstream
https://api.github.com/repos/socketstream/socketstream
opened
Documentation improvement
Documentation Proposed change
[See https://github.com/socketstream/socketstream/issues/513 as a precursor to this.] I've created a `feature/doc-updates` branch. I think it was @RomanMinkin who did a tremendous effort in going through our docs and trying to get them in a better state. However, after having given them a good look over, I think we've over complicated the task by throwing an angular wrapper around the documentation section—it makes them difficult to maintain. For example, I went to go work on the documentation this evening and spent way too long trying to figure things out when all I wanted to do was write. I think we could work on trying to make the process simpler, and I appreciate @thepian's view about getting the community involved to that end. (It's probably a catch 22—need to make it easier first to get people to start helping.) Here are some starting suggestions. 1. Lets swap out ngdocs for something that doesn't rely on Angular, and is a little more accessible and established, like JSDocs. 2. For authored docs, markdown is probably the best bet. There are techniques for converting `.md` files into HTML, and markdown is the flavor, too, of GitHub. That way we're not repeating our efforts, and the docs can exist in two places. (Good for when servers are down.) 3. Also, maybe we could all just divvy up some of the missing or incomplete topics and take a stab at authoring single pages—I've said I plan on working on the docs a lot, but the knowledge of how things work is in everyone's minds. If everyone takes a stab at a draft (it can be messy!), I can be editorial and clean and organize things. (I come from a pack of literary wolves and editors, as evidenced by all my verbose GitHub comments.) Those are some initial thoughts. Please add you own ideas, and I'll work around the communities' needs.
1.0
Documentation improvement - [See https://github.com/socketstream/socketstream/issues/513 as a precursor to this.] I've created a `feature/doc-updates` branch. I think it was @RomanMinkin who did a tremendous effort in going through our docs and trying to get them in a better state. However, after having given them a good look over, I think we've over complicated the task by throwing an angular wrapper around the documentation section—it makes them difficult to maintain. For example, I went to go work on the documentation this evening and spent way too long trying to figure things out when all I wanted to do was write. I think we could work on trying to make the process simpler, and I appreciate @thepian's view about getting the community involved to that end. (It's probably a catch 22—need to make it easier first to get people to start helping.) Here are some starting suggestions. 1. Lets swap out ngdocs for something that doesn't rely on Angular, and is a little more accessible and established, like JSDocs. 2. For authored docs, markdown is probably the best bet. There are techniques for converting `.md` files into HTML, and markdown is the flavor, too, of GitHub. That way we're not repeating our efforts, and the docs can exist in two places. (Good for when servers are down.) 3. Also, maybe we could all just divvy up some of the missing or incomplete topics and take a stab at authoring single pages—I've said I plan on working on the docs a lot, but the knowledge of how things work is in everyone's minds. If everyone takes a stab at a draft (it can be messy!), I can be editorial and clean and organize things. (I come from a pack of literary wolves and editors, as evidenced by all my verbose GitHub comments.) Those are some initial thoughts. Please add you own ideas, and I'll work around the communities' needs.
non_defect
documentation improvement i ve created a feature doc updates branch i think it was romanminkin who did a tremendous effort in going through our docs and trying to get them in a better state however after having given them a good look over i think we ve over complicated the task by throwing an angular wrapper around the documentation section—it makes them difficult to maintain for example i went to go work on the documentation this evening and spent way too long trying to figure things out when all i wanted to do was write i think we could work on trying to make the process simpler and i appreciate thepian s view about getting the community involved to that end it s probably a catch —need to make it easier first to get people to start helping here are some starting suggestions lets swap out ngdocs for something that doesn t rely on angular and is a little more accessible and established like jsdocs for authored docs markdown is probably the best bet there are techniques for converting md files into html and markdown is the flavor too of github that way we re not repeating our efforts and the docs can exist in two places good for when servers are down also maybe we could all just divvy up some of the missing or incomplete topics and take a stab at authoring single pages—i ve said i plan on working on the docs a lot but the knowledge of how things work is in everyone s minds if everyone takes a stab at a draft it can be messy i can be editorial and clean and organize things i come from a pack of literary wolves and editors as evidenced by all my verbose github comments those are some initial thoughts please add you own ideas and i ll work around the communities needs
0
11,696
2,661,019,976
IssuesEvent
2015-03-19 12:06:23
contao/core
https://api.github.com/repos/contao/core
closed
CE - Media erkennt `.ogg` als Audio
defect up for discussion
Leider wird bei der Angabe webm und ogg die Ausgabe als Audio und nicht als Video erkannt. Sobald ein mp4 dazu kommt passt es, jedoch nicht bei nur den beiden erstgenannten Formaten.
1.0
CE - Media erkennt `.ogg` als Audio - Leider wird bei der Angabe webm und ogg die Ausgabe als Audio und nicht als Video erkannt. Sobald ein mp4 dazu kommt passt es, jedoch nicht bei nur den beiden erstgenannten Formaten.
defect
ce media erkennt ogg als audio leider wird bei der angabe webm und ogg die ausgabe als audio und nicht als video erkannt sobald ein dazu kommt passt es jedoch nicht bei nur den beiden erstgenannten formaten
1
58,109
16,342,462,675
IssuesEvent
2021-05-13 00:21:39
darshan-hpc/darshan
https://api.github.com/repos/darshan-hpc/darshan
closed
uthash header not installed
defect
In GitLab by @shanedsnyder on Oct 12, 2015, 09:59 The hash implementation used in libdarshan-util is not installed with Darshan, preventing external tools that link in this library from being able to compile.
1.0
uthash header not installed - In GitLab by @shanedsnyder on Oct 12, 2015, 09:59 The hash implementation used in libdarshan-util is not installed with Darshan, preventing external tools that link in this library from being able to compile.
defect
uthash header not installed in gitlab by shanedsnyder on oct the hash implementation used in libdarshan util is not installed with darshan preventing external tools that link in this library from being able to compile
1
37,299
2,824,111,670
IssuesEvent
2015-05-21 13:06:47
pywinauto/pywinauto
https://api.github.com/repos/pywinauto/pywinauto
closed
_treeview_element.Click() uses ClickInput()
auto-migrated bug internal Priority-Low
``` What steps will reproduce the problem? 1. In common_control.py class _treeview_element contain method Click 2. Method Click uses another method HwndWrapper.ClickInput 3. It is inconsistent, HwndWrapper has Click and ClickInput methods and they works in different way. What is the expected output? What do you see instead? The _treeview_element.Click() should use HwndWrapper.Click() Probably _treeview_element.ClickInput()should be added. What version of the product are you using? On what operating system? 0.4.1. In 0.4.2 I saw the same issue Please provide any additional information below. ``` Original issue reported on code.google.com by `far...@gmail.com` on 5 Feb 2014 at 8:01
1.0
_treeview_element.Click() uses ClickInput() - ``` What steps will reproduce the problem? 1. In common_control.py class _treeview_element contain method Click 2. Method Click uses another method HwndWrapper.ClickInput 3. It is inconsistent, HwndWrapper has Click and ClickInput methods and they works in different way. What is the expected output? What do you see instead? The _treeview_element.Click() should use HwndWrapper.Click() Probably _treeview_element.ClickInput()should be added. What version of the product are you using? On what operating system? 0.4.1. In 0.4.2 I saw the same issue Please provide any additional information below. ``` Original issue reported on code.google.com by `far...@gmail.com` on 5 Feb 2014 at 8:01
non_defect
treeview element click uses clickinput what steps will reproduce the problem in common control py class treeview element contain method click method click uses another method hwndwrapper clickinput it is inconsistent hwndwrapper has click and clickinput methods and they works in different way what is the expected output what do you see instead the treeview element click should use hwndwrapper click probably treeview element clickinput should be added what version of the product are you using on what operating system in i saw the same issue please provide any additional information below original issue reported on code google com by far gmail com on feb at
0
197,758
6,963,601,772
IssuesEvent
2017-12-08 18:03:29
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio-ui] Change UI to always send CSRF token on API calls
new feature priority: high
UI will need to send CSRF token on every API call now that does a POST, PUT or DELETE. Please refer to craftercms/craftercms#1601.
1.0
[studio-ui] Change UI to always send CSRF token on API calls - UI will need to send CSRF token on every API call now that does a POST, PUT or DELETE. Please refer to craftercms/craftercms#1601.
non_defect
change ui to always send csrf token on api calls ui will need to send csrf token on every api call now that does a post put or delete please refer to craftercms craftercms
0
67,511
20,972,920,687
IssuesEvent
2022-03-28 13:02:20
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
Accordion open all taps in the first few seconds if its in a Dynamic Dialog
defect
[x] bug report => Search github for a similar issue or PR before submitting [ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap [ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35 ``` **Plunkr Case (Bug Reports)** Please demonstrate your case at stackblitz by using the issue template below. Issues without a test case have much less possibility to be reviewd in detail and assisted. https://stackblitz.com/edit/github-9ccs9k?mbed=1&file=src/app/app.component.ts&view=preview **Current behavior** Accordion open all taps in the first few seconds if its in a Dynamic Dialog. After that it will collapse again. This looks strange. **Expected behavior** Accordion taps should be closed per default if I open a dialog which contains a accordion as it was in the versions before. **What is the motivation / use case for changing the behavior?** <!-- Describe the motivation or the concrete use case --> We use the Accordion in a Dynamic Dialog in our business application and it is no good look and feel for the customer. * **Angular version:** 9.X <!-- Check whether this is still an issue in the most recent Angular version --> * **PrimeNG version:** 9.X <!-- Check whether this is still an issue in the most recent Angular version -->
1.0
Accordion open all taps in the first few seconds if its in a Dynamic Dialog - [x] bug report => Search github for a similar issue or PR before submitting [ ] feature request => Please check if request is not on the roadmap already https://github.com/primefaces/primeng/wiki/Roadmap [ ] support request => Please do not submit support request here, instead see http://forum.primefaces.org/viewforum.php?f=35 ``` **Plunkr Case (Bug Reports)** Please demonstrate your case at stackblitz by using the issue template below. Issues without a test case have much less possibility to be reviewd in detail and assisted. https://stackblitz.com/edit/github-9ccs9k?mbed=1&file=src/app/app.component.ts&view=preview **Current behavior** Accordion open all taps in the first few seconds if its in a Dynamic Dialog. After that it will collapse again. This looks strange. **Expected behavior** Accordion taps should be closed per default if I open a dialog which contains a accordion as it was in the versions before. **What is the motivation / use case for changing the behavior?** <!-- Describe the motivation or the concrete use case --> We use the Accordion in a Dynamic Dialog in our business application and it is no good look and feel for the customer. * **Angular version:** 9.X <!-- Check whether this is still an issue in the most recent Angular version --> * **PrimeNG version:** 9.X <!-- Check whether this is still an issue in the most recent Angular version -->
defect
accordion open all taps in the first few seconds if its in a dynamic dialog bug report search github for a similar issue or pr before submitting feature request please check if request is not on the roadmap already support request please do not submit support request here instead see plunkr case bug reports please demonstrate your case at stackblitz by using the issue template below issues without a test case have much less possibility to be reviewd in detail and assisted current behavior accordion open all taps in the first few seconds if its in a dynamic dialog after that it will collapse again this looks strange expected behavior accordion taps should be closed per default if i open a dialog which contains a accordion as it was in the versions before what is the motivation use case for changing the behavior we use the accordion in a dynamic dialog in our business application and it is no good look and feel for the customer angular version x primeng version x
1
54,521
13,757,481,786
IssuesEvent
2020-10-06 21:44:22
Cockatrice/Cockatrice
https://api.github.com/repos/Cockatrice/Cockatrice
closed
Power and Toughness Hotkey not working for increase
App - Cockatrice App - Protocol / API Defect - Basic
<b>OS:</b> OS X 10.11.6 <b>Cockatrice version:</b> Version dab7316 (2017-01-19) ___ <br> I reinstalled Cockatrice a few days ago and noticed that all the other hot keys seem to work but I can not get the Power and Toughness to increase. I have even changed from standard settings to try to make it work, and I still can not hot key. I can manually switch it, but when adding several +1/+1 counters to each creature every turn...it gets extremely time consuming to try to play. Thanks for the help!
1.0
Power and Toughness Hotkey not working for increase - <b>OS:</b> OS X 10.11.6 <b>Cockatrice version:</b> Version dab7316 (2017-01-19) ___ <br> I reinstalled Cockatrice a few days ago and noticed that all the other hot keys seem to work but I can not get the Power and Toughness to increase. I have even changed from standard settings to try to make it work, and I still can not hot key. I can manually switch it, but when adding several +1/+1 counters to each creature every turn...it gets extremely time consuming to try to play. Thanks for the help!
defect
power and toughness hotkey not working for increase os os x cockatrice version version i reinstalled cockatrice a few days ago and noticed that all the other hot keys seem to work but i can not get the power and toughness to increase i have even changed from standard settings to try to make it work and i still can not hot key i can manually switch it but when adding several counters to each creature every turn it gets extremely time consuming to try to play thanks for the help
1
85,535
16,673,564,719
IssuesEvent
2021-06-07 13:48:15
Decentra-Network/Decentra-Network
https://api.github.com/repos/Decentra-Network/Decentra-Network
closed
Description must be added to question_maker(question_text=None, mode=None)
Code Quality Improvements
Link of the line: https://github.com/Decentra-Network/Decentra-Network/blob/master/src/lib/mixlib.py#L51
1.0
Description must be added to question_maker(question_text=None, mode=None) - Link of the line: https://github.com/Decentra-Network/Decentra-Network/blob/master/src/lib/mixlib.py#L51
non_defect
description must be added to question maker question text none mode none link of the line
0
10,128
2,618,937,654
IssuesEvent
2015-03-03 00:02:44
chrsmith/open-ig
https://api.github.com/repos/chrsmith/open-ig
closed
Planet list ordering is errorenous, it lists planets as X 1, X 10, X 11, X 12, X 2, X 20, X 21, etc
auto-migrated Component-UI Priority-Low Type-Defect
``` Game version: 0.95.152 Operating System: Linux x64 Java runtime version: 1.7.0_51 Installed using the Launcher? yes Game language (en, hu, de): hu ``` Original issue reported on code.google.com by `kli...@gmail.com` on 21 Jan 2014 at 1:48
1.0
Planet list ordering is errorenous, it lists planets as X 1, X 10, X 11, X 12, X 2, X 20, X 21, etc - ``` Game version: 0.95.152 Operating System: Linux x64 Java runtime version: 1.7.0_51 Installed using the Launcher? yes Game language (en, hu, de): hu ``` Original issue reported on code.google.com by `kli...@gmail.com` on 21 Jan 2014 at 1:48
defect
planet list ordering is errorenous it lists planets as x x x x x x x etc game version operating system linux java runtime version installed using the launcher yes game language en hu de hu original issue reported on code google com by kli gmail com on jan at
1
63,810
18,009,835,112
IssuesEvent
2021-09-16 07:14:59
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Joining Subspace shows UI about rooms
T-Defect S-Tolerable A-Spaces O-Frequent A-Subspaces Z-Spaces-Testing
### Steps to reproduce 1. Go to Explore rooms in a space with a subspace 2. Click the "Join" button ### What happened? ### What did you expect? Next dialog/page showing "Joining space" ### What happened? Goes to an intermediate page with "Joining room"/spinner and then to the "Explore rooms" page for the subspace ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp develop.element.io ### Homeserver _No response_ ### Have you submitted a rageshake? No
1.0
Joining Subspace shows UI about rooms - ### Steps to reproduce 1. Go to Explore rooms in a space with a subspace 2. Click the "Join" button ### What happened? ### What did you expect? Next dialog/page showing "Joining space" ### What happened? Goes to an intermediate page with "Joining room"/spinner and then to the "Explore rooms" page for the subspace ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp develop.element.io ### Homeserver _No response_ ### Have you submitted a rageshake? No
defect
joining subspace shows ui about rooms steps to reproduce go to explore rooms in a space with a subspace click the join button what happened what did you expect next dialog page showing joining space what happened goes to an intermediate page with joining room spinner and then to the explore rooms page for the subspace operating system no response browser information no response url for webapp develop element io homeserver no response have you submitted a rageshake no
1
37,776
8,517,451,439
IssuesEvent
2018-11-01 08:03:11
GoldenSoftwareLtd/gedemin
https://api.github.com/repos/GoldenSoftwareLtd/gedemin
closed
После редактирования отчёта права на него не сохраняются
GedeminExe Priority-Medium Report Security Type-Defect
Originally reported on Google Code with ID 3263 ``` После редактирования отчёта права на него не сохраняются, а перезаписываются правами на родительскую папку. Приходится после каждой правки вручную их переназначать что очень неудобно. И ещё такая просьба. Сделайте, пожалуйста, возможным в отчётах на закладке ‘история’ удалять помеченные версии отчёта и для удобства кнопку 'выбрать все кроме последней'. Т.к. создаётся много версий по любой мелочи, которые абсолютно не нужны и занимают место в блобах, а значит, увеличивают фрагментацию соответствующих таблиц, замедляет по ним поиск и т.п. Если снять статистику IBAnalust-ом, то видно, что заполнение соотв. таблиц порядка 1%. ``` Reported by `igor6003` on 2013-11-12 11:47:28
1.0
После редактирования отчёта права на него не сохраняются - Originally reported on Google Code with ID 3263 ``` После редактирования отчёта права на него не сохраняются, а перезаписываются правами на родительскую папку. Приходится после каждой правки вручную их переназначать что очень неудобно. И ещё такая просьба. Сделайте, пожалуйста, возможным в отчётах на закладке ‘история’ удалять помеченные версии отчёта и для удобства кнопку 'выбрать все кроме последней'. Т.к. создаётся много версий по любой мелочи, которые абсолютно не нужны и занимают место в блобах, а значит, увеличивают фрагментацию соответствующих таблиц, замедляет по ним поиск и т.п. Если снять статистику IBAnalust-ом, то видно, что заполнение соотв. таблиц порядка 1%. ``` Reported by `igor6003` on 2013-11-12 11:47:28
defect
после редактирования отчёта права на него не сохраняются originally reported on google code with id после редактирования отчёта права на него не сохраняются а перезаписываются правами на родительскую папку приходится после каждой правки вручную их переназначать что очень неудобно и ещё такая просьба сделайте пожалуйста возможным в отчётах на закладке ‘история’ удалять помеченные версии отчёта и для удобства кнопку выбрать все кроме последней т к создаётся много версий по любой мелочи которые абсолютно не нужны и занимают место в блобах а значит увеличивают фрагментацию соответствующих таблиц замедляет по ним поиск и т п если снять статистику ibanalust ом то видно что заполнение соотв таблиц порядка reported by on
1
25,967
4,538,635,559
IssuesEvent
2016-09-09 07:49:06
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Dictionary - Default constructor Comparer value
defect portarelle
### Expected ```cs System.Collections.Generic.GenericEqualityComparer'1[System.Int32] ``` ### Actual ```cs System.Collections.Generic.GenericEqualityComparer'1[Object] ``` ### Steps To Reproduce ```csharp public class App { public static void Main() { var d1 = new Dictionary<int, string>(); Assert.AreStrictEqual(d1.Comparer, EqualityComparer<int>.Default); var d2 = new Dictionary<int, string>(10); Assert.AreStrictEqual(d2.Comparer, EqualityComparer<int>.Default); var orig = new Dictionary<string, int>(); var d3 = new Dictionary<string, int>(orig); Assert.AreStrictEqual(d3.Comparer, EqualityComparer<int>.Default); } } ```
1.0
Dictionary - Default constructor Comparer value - ### Expected ```cs System.Collections.Generic.GenericEqualityComparer'1[System.Int32] ``` ### Actual ```cs System.Collections.Generic.GenericEqualityComparer'1[Object] ``` ### Steps To Reproduce ```csharp public class App { public static void Main() { var d1 = new Dictionary<int, string>(); Assert.AreStrictEqual(d1.Comparer, EqualityComparer<int>.Default); var d2 = new Dictionary<int, string>(10); Assert.AreStrictEqual(d2.Comparer, EqualityComparer<int>.Default); var orig = new Dictionary<string, int>(); var d3 = new Dictionary<string, int>(orig); Assert.AreStrictEqual(d3.Comparer, EqualityComparer<int>.Default); } } ```
defect
dictionary default constructor comparer value expected cs system collections generic genericequalitycomparer actual cs system collections generic genericequalitycomparer steps to reproduce csharp public class app public static void main var new dictionary assert arestrictequal comparer equalitycomparer default var new dictionary assert arestrictequal comparer equalitycomparer default var orig new dictionary var new dictionary orig assert arestrictequal comparer equalitycomparer default
1
2,194
2,603,977,775
IssuesEvent
2015-02-24 19:01:58
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳沈阳感染疱疹怎么办
auto-migrated Priority-Medium Type-Defect
``` 沈阳沈阳感染疱疹怎么办〓沈陽軍區政治部醫院性病〓TEL:02 4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療� ��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:29
1.0
沈阳沈阳感染疱疹怎么办 - ``` 沈阳沈阳感染疱疹怎么办〓沈陽軍區政治部醫院性病〓TEL:02 4-31023308〓成立于1946年,68年專注于性傳播疾病的研究和治療� ��位于沈陽市沈河區二緯路32號。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:29
defect
沈阳沈阳感染疱疹怎么办 沈阳沈阳感染疱疹怎么办〓沈陽軍區政治部醫院性病〓tel: 〓 , � �� 。是一所與新中國同建立共輝� ��的歷史悠久、設備精良、技術權威、專家云集,是預防、保 健、醫療、科研康復為一體的綜合性醫院。是國家首批公立�� �等部隊醫院、全國首批醫療規范定點單位,是第四軍醫大學� ��東南大學等知名高等院校的教學醫院。曾被中國人民解放軍 空軍后勤部衛生部評為衛生工作先進單位,先后兩次榮立集�� �二等功。 original issue reported on code google com by gmail com on jun at
1
11,715
2,664,255,802
IssuesEvent
2015-03-20 13:28:40
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
opened
SQLDialectNotSupportedException: ARRAY TABLE is not supported for SQL99
C: DB: Oracle C: Functionality P: High T: Defect
When using Oracle's `TABLE()` function in a derived table, jOOQ throws an exception: ``` SQLDialectNotSupportedException: ARRAY TABLE is not supported for SQL99 ``` An example to reproduce this: ```java UNumberTableRecord ids = new UNumberTableRecord(1, 2, 3); // Never mind this cast in this case... Select<Record1<Integer>> selectIds = (Select<Record1<Integer>>) DSL.selectFrom(DSL.table(ids)); create() .select(T_BOOK.ID, T_BOOK.TITLE) .from(DSL.select(T_BOOK.ID, T_BOOK.AUTHOR_ID, T_BOOK.TITLE) .from(T_BOOK) .join(T_AUTHOR) .on(T_BOOK.AUTHOR_ID.eq(T_AUTHOR.ID)) .where(T_BOOK.ID.in(selectIds))) .fetch(); ```
1.0
SQLDialectNotSupportedException: ARRAY TABLE is not supported for SQL99 - When using Oracle's `TABLE()` function in a derived table, jOOQ throws an exception: ``` SQLDialectNotSupportedException: ARRAY TABLE is not supported for SQL99 ``` An example to reproduce this: ```java UNumberTableRecord ids = new UNumberTableRecord(1, 2, 3); // Never mind this cast in this case... Select<Record1<Integer>> selectIds = (Select<Record1<Integer>>) DSL.selectFrom(DSL.table(ids)); create() .select(T_BOOK.ID, T_BOOK.TITLE) .from(DSL.select(T_BOOK.ID, T_BOOK.AUTHOR_ID, T_BOOK.TITLE) .from(T_BOOK) .join(T_AUTHOR) .on(T_BOOK.AUTHOR_ID.eq(T_AUTHOR.ID)) .where(T_BOOK.ID.in(selectIds))) .fetch(); ```
defect
sqldialectnotsupportedexception array table is not supported for when using oracle s table function in a derived table jooq throws an exception sqldialectnotsupportedexception array table is not supported for an example to reproduce this java unumbertablerecord ids new unumbertablerecord never mind this cast in this case select selectids select dsl selectfrom dsl table ids create select t book id t book title from dsl select t book id t book author id t book title from t book join t author on t book author id eq t author id where t book id in selectids fetch
1
46,644
13,055,953,438
IssuesEvent
2020-07-30 03:13:17
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
triangle.c has different precision based on OS (Trac #1647)
Incomplete Migration Migrated from Trac combo simulation defect
Migrated from https://code.icecube.wisc.edu/ticket/1647 ```json { "status": "closed", "changetime": "2016-08-18T22:03:29", "description": "See:\n`LINUX/CPU86` in the source:/IceCube/projects/g4-tankresponse/trunk/private/g4-tankresponse/triangle/triangle.c\n", "reporter": "nega", "cc": "", "resolution": "wontfix", "_ts": "1471557809273381", "component": "combo simulation", "summary": "triangle.c has different precision based on OS", "priority": "normal", "keywords": "pruning C", "time": "2016-04-20T19:35:01", "milestone": "Long-Term Future", "owner": "jgonzalez", "type": "defect" } ```
1.0
triangle.c has different precision based on OS (Trac #1647) - Migrated from https://code.icecube.wisc.edu/ticket/1647 ```json { "status": "closed", "changetime": "2016-08-18T22:03:29", "description": "See:\n`LINUX/CPU86` in the source:/IceCube/projects/g4-tankresponse/trunk/private/g4-tankresponse/triangle/triangle.c\n", "reporter": "nega", "cc": "", "resolution": "wontfix", "_ts": "1471557809273381", "component": "combo simulation", "summary": "triangle.c has different precision based on OS", "priority": "normal", "keywords": "pruning C", "time": "2016-04-20T19:35:01", "milestone": "Long-Term Future", "owner": "jgonzalez", "type": "defect" } ```
defect
triangle c has different precision based on os trac migrated from json status closed changetime description see n linux in the source icecube projects tankresponse trunk private tankresponse triangle triangle c n reporter nega cc resolution wontfix ts component combo simulation summary triangle c has different precision based on os priority normal keywords pruning c time milestone long term future owner jgonzalez type defect
1
53,463
13,261,662,075
IssuesEvent
2020-08-20 20:18:29
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
[cvmfs] opencl environment issues (Trac #1450)
Migrated from Trac cvmfs defect
As seen on buildbot "Ubuntu 14.04 (cvmfs)", there is some sort of conflict where `libOpenCL.so` is not found, but `/etc/OpenCL/vendors` exists. Make sure that if we're using CVMFS OpenCL that we have it in the list of vendors. Solution? : 1. If `/etc/OpenCL/vendors` does not exist, use CVMFS copy 2. If `/etc/OpenCL/vendors` exists, make temp hybrid dir with entries from `/etc/OpenCL/vendors` and an entry for CVMFS. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1450">https://code.icecube.wisc.edu/projects/icecube/ticket/1450</a>, reported by david.schultzand owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:09", "_ts": "1458335649133028", "description": "As seen on buildbot \"Ubuntu 14.04 (cvmfs)\", there is some sort of conflict where `libOpenCL.so` is not found, but `/etc/OpenCL/vendors` exists. Make sure that if we're using CVMFS OpenCL that we have it in the list of vendors.\n\nSolution? :\n\n1. If `/etc/OpenCL/vendors` does not exist, use CVMFS copy\n2. If `/etc/OpenCL/vendors` exists, make temp hybrid dir with entries from `/etc/OpenCL/vendors` and an entry for CVMFS.", "reporter": "david.schultz", "cc": "nega, claudio.kopper", "resolution": "fixed", "time": "2015-11-25T16:51:22", "component": "cvmfs", "summary": "[cvmfs] opencl environment issues", "priority": "major", "keywords": "", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
1.0
[cvmfs] opencl environment issues (Trac #1450) - As seen on buildbot "Ubuntu 14.04 (cvmfs)", there is some sort of conflict where `libOpenCL.so` is not found, but `/etc/OpenCL/vendors` exists. Make sure that if we're using CVMFS OpenCL that we have it in the list of vendors. Solution? : 1. If `/etc/OpenCL/vendors` does not exist, use CVMFS copy 2. If `/etc/OpenCL/vendors` exists, make temp hybrid dir with entries from `/etc/OpenCL/vendors` and an entry for CVMFS. <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1450">https://code.icecube.wisc.edu/projects/icecube/ticket/1450</a>, reported by david.schultzand owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:09", "_ts": "1458335649133028", "description": "As seen on buildbot \"Ubuntu 14.04 (cvmfs)\", there is some sort of conflict where `libOpenCL.so` is not found, but `/etc/OpenCL/vendors` exists. Make sure that if we're using CVMFS OpenCL that we have it in the list of vendors.\n\nSolution? :\n\n1. If `/etc/OpenCL/vendors` does not exist, use CVMFS copy\n2. If `/etc/OpenCL/vendors` exists, make temp hybrid dir with entries from `/etc/OpenCL/vendors` and an entry for CVMFS.", "reporter": "david.schultz", "cc": "nega, claudio.kopper", "resolution": "fixed", "time": "2015-11-25T16:51:22", "component": "cvmfs", "summary": "[cvmfs] opencl environment issues", "priority": "major", "keywords": "", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
defect
opencl environment issues trac as seen on buildbot ubuntu cvmfs there is some sort of conflict where libopencl so is not found but etc opencl vendors exists make sure that if we re using cvmfs opencl that we have it in the list of vendors solution if etc opencl vendors does not exist use cvmfs copy if etc opencl vendors exists make temp hybrid dir with entries from etc opencl vendors and an entry for cvmfs migrated from json status closed changetime ts description as seen on buildbot ubuntu cvmfs there is some sort of conflict where libopencl so is not found but etc opencl vendors exists make sure that if we re using cvmfs opencl that we have it in the list of vendors n nsolution n if etc opencl vendors does not exist use cvmfs copy if etc opencl vendors exists make temp hybrid dir with entries from etc opencl vendors and an entry for cvmfs reporter david schultz cc nega claudio kopper resolution fixed time component cvmfs summary opencl environment issues priority major keywords milestone owner david schultz type defect
1
103,410
12,893,582,911
IssuesEvent
2020-07-13 21:58:20
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
Need to add the Global navigation to the Select Cluster drop-down
[zube]: Design WIP area/navigation
Need design of how we want the Main Drop-down to switch between Global and Individual Clusters. This menu: ![image](https://user-images.githubusercontent.com/11514927/79608465-eb13e780-80a9-11ea-905a-71aeb1eb5ae8.png)
1.0
Need to add the Global navigation to the Select Cluster drop-down - Need design of how we want the Main Drop-down to switch between Global and Individual Clusters. This menu: ![image](https://user-images.githubusercontent.com/11514927/79608465-eb13e780-80a9-11ea-905a-71aeb1eb5ae8.png)
non_defect
need to add the global navigation to the select cluster drop down need design of how we want the main drop down to switch between global and individual clusters this menu
0
45,294
23,999,191,834
IssuesEvent
2022-09-14 09:59:39
FeatureBaseDB/featurebase
https://api.github.com/repos/FeatureBaseDB/featurebase
closed
import endpoints which take a shard should optionally allocate the shard for you
proposal ingest-performance
# Description There are many situations where records in a data set do not have natural sequential integer ids and we want to simply assign them at ingest time (rather than using column translation). In these cases, a single client importing into an empty Pilosa can fairly easily generate sequential ids and import the data performantly. But what if there are multiple clients? What if there is already data in Pilosa? Clients need some way of coordinating which IDs each one will allocate, and understanding what data is already in Pilosa so it doesn't get overwritten. One could imagine any variety of where clients communicate with each other (or through another service) to synchronize, or are configured ahead of time not to produce overlapping ids, and can interrogate Pilosa about what data it already has, but these all seem messy and like kind of a lot of work. What if the import endpoints which take a shard parameter simply allowed you to leave that parameter off? If you do not specify what shard the data should go into, Pilosa knows that it should use a new empty shard to ingest the data. This will allow any number of concurrent clients to throw shardfulls of data at Pilosa without having to do any synchronization or pre-communication! Pilosa will have to take care such that concurrent requests don't end up in the same shard, but this is purely internal to Pilosa, and seems a lot easier than trying to coordinate with clients. The node receiving the request can look at its availableShards data and select from among the first few empty ones. It would then send out a request to all the owners of that shard to let them know it is reserving it. If the owners respond that they have no data for that shard, then the node can continue with the import. The owners will have to reject any other modifications to that shard until the import is finished.
True
import endpoints which take a shard should optionally allocate the shard for you - # Description There are many situations where records in a data set do not have natural sequential integer ids and we want to simply assign them at ingest time (rather than using column translation). In these cases, a single client importing into an empty Pilosa can fairly easily generate sequential ids and import the data performantly. But what if there are multiple clients? What if there is already data in Pilosa? Clients need some way of coordinating which IDs each one will allocate, and understanding what data is already in Pilosa so it doesn't get overwritten. One could imagine any variety of where clients communicate with each other (or through another service) to synchronize, or are configured ahead of time not to produce overlapping ids, and can interrogate Pilosa about what data it already has, but these all seem messy and like kind of a lot of work. What if the import endpoints which take a shard parameter simply allowed you to leave that parameter off? If you do not specify what shard the data should go into, Pilosa knows that it should use a new empty shard to ingest the data. This will allow any number of concurrent clients to throw shardfulls of data at Pilosa without having to do any synchronization or pre-communication! Pilosa will have to take care such that concurrent requests don't end up in the same shard, but this is purely internal to Pilosa, and seems a lot easier than trying to coordinate with clients. The node receiving the request can look at its availableShards data and select from among the first few empty ones. It would then send out a request to all the owners of that shard to let them know it is reserving it. If the owners respond that they have no data for that shard, then the node can continue with the import. The owners will have to reject any other modifications to that shard until the import is finished.
non_defect
import endpoints which take a shard should optionally allocate the shard for you description there are many situations where records in a data set do not have natural sequential integer ids and we want to simply assign them at ingest time rather than using column translation in these cases a single client importing into an empty pilosa can fairly easily generate sequential ids and import the data performantly but what if there are multiple clients what if there is already data in pilosa clients need some way of coordinating which ids each one will allocate and understanding what data is already in pilosa so it doesn t get overwritten one could imagine any variety of where clients communicate with each other or through another service to synchronize or are configured ahead of time not to produce overlapping ids and can interrogate pilosa about what data it already has but these all seem messy and like kind of a lot of work what if the import endpoints which take a shard parameter simply allowed you to leave that parameter off if you do not specify what shard the data should go into pilosa knows that it should use a new empty shard to ingest the data this will allow any number of concurrent clients to throw shardfulls of data at pilosa without having to do any synchronization or pre communication pilosa will have to take care such that concurrent requests don t end up in the same shard but this is purely internal to pilosa and seems a lot easier than trying to coordinate with clients the node receiving the request can look at its availableshards data and select from among the first few empty ones it would then send out a request to all the owners of that shard to let them know it is reserving it if the owners respond that they have no data for that shard then the node can continue with the import the owners will have to reject any other modifications to that shard until the import is finished
0
50,697
13,187,684,255
IssuesEvent
2020-08-13 04:13:38
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
[wavereform] TODO (Trac #1197)
Migrated from Trac combo reconstruction defect
decide whether this TODO is important, and if so make a ticket for it: ```text python/wavereform.py: # TODO: flag FADCs that saturate outside of the ATWD window. ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1197">https://code.icecube.wisc.edu/ticket/1197</a>, reported by david.schultz and owned by jbraun</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "decide whether this TODO is important, and if so make a ticket for it:\n{{{\npython/wavereform.py:\t\t\t# TODO: flag FADCs that saturate outside of the ATWD window.\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "duplicate", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "[wavereform] TODO", "priority": "critical", "keywords": "", "time": "2015-08-19T18:10:05", "milestone": "", "owner": "jbraun", "type": "defect" } ``` </p> </details>
1.0
[wavereform] TODO (Trac #1197) - decide whether this TODO is important, and if so make a ticket for it: ```text python/wavereform.py: # TODO: flag FADCs that saturate outside of the ATWD window. ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1197">https://code.icecube.wisc.edu/ticket/1197</a>, reported by david.schultz and owned by jbraun</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "decide whether this TODO is important, and if so make a ticket for it:\n{{{\npython/wavereform.py:\t\t\t# TODO: flag FADCs that saturate outside of the ATWD window.\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "duplicate", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "[wavereform] TODO", "priority": "critical", "keywords": "", "time": "2015-08-19T18:10:05", "milestone": "", "owner": "jbraun", "type": "defect" } ``` </p> </details>
defect
todo trac decide whether this todo is important and if so make a ticket for it text python wavereform py todo flag fadcs that saturate outside of the atwd window migrated from json status closed changetime description decide whether this todo is important and if so make a ticket for it n npython wavereform py t t t todo flag fadcs that saturate outside of the atwd window n reporter david schultz cc resolution duplicate ts component combo reconstruction summary todo priority critical keywords time milestone owner jbraun type defect
1
53,079
13,260,871,095
IssuesEvent
2020-08-20 18:54:19
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
Make cmake framework detection compatible with Xcode versions >= 4.3 (Trac #668)
Migrated from Trac defect tools/ports
Since Xcode 4.3, frameworks are no longer stored in /Developer, but insider the Xcode application bundle. Cmake has support for this, but unfortunately, the detection code is missing a "platform" part in its detection code. These two patch files a. enhance the detection code to look in the correct place for newer Xcode versions, and a. uses the detected location when looking for frameworks This will allow detection of the JavaVM (JNI) developer framework containing the correct header files. (The system framework is detected correctly, but it does not contain headers.) <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/668">https://code.icecube.wisc.edu/projects/icecube/ticket/668</a>, reported by claudio.kopperand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T19:01:17", "_ts": "1351710077000000", "description": "Since Xcode 4.3, frameworks are no longer stored in /Developer, but insider the Xcode application bundle. Cmake has support for this, but unfortunately, the detection code is missing a \"platform\" part in its detection code. These two patch files \n\n a. enhance the detection code to look in the correct place for newer Xcode versions, and\n a. uses the detected location when looking for frameworks \n\nThis will allow detection of the JavaVM (JNI) developer framework containing the correct header files. (The system framework is detected correctly, but it does not contain headers.)\n", "reporter": "claudio.kopper", "cc": "", "resolution": "fixed", "time": "2012-02-19T19:53:49", "component": "tools/ports", "summary": "Make cmake framework detection compatible with Xcode versions >= 4.3", "priority": "normal", "keywords": "cmake xcode mac os frameworks jni java javavm", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
Make cmake framework detection compatible with Xcode versions >= 4.3 (Trac #668) - Since Xcode 4.3, frameworks are no longer stored in /Developer, but insider the Xcode application bundle. Cmake has support for this, but unfortunately, the detection code is missing a "platform" part in its detection code. These two patch files a. enhance the detection code to look in the correct place for newer Xcode versions, and a. uses the detected location when looking for frameworks This will allow detection of the JavaVM (JNI) developer framework containing the correct header files. (The system framework is detected correctly, but it does not contain headers.) <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/668">https://code.icecube.wisc.edu/projects/icecube/ticket/668</a>, reported by claudio.kopperand owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T19:01:17", "_ts": "1351710077000000", "description": "Since Xcode 4.3, frameworks are no longer stored in /Developer, but insider the Xcode application bundle. Cmake has support for this, but unfortunately, the detection code is missing a \"platform\" part in its detection code. These two patch files \n\n a. enhance the detection code to look in the correct place for newer Xcode versions, and\n a. uses the detected location when looking for frameworks \n\nThis will allow detection of the JavaVM (JNI) developer framework containing the correct header files. (The system framework is detected correctly, but it does not contain headers.)\n", "reporter": "claudio.kopper", "cc": "", "resolution": "fixed", "time": "2012-02-19T19:53:49", "component": "tools/ports", "summary": "Make cmake framework detection compatible with Xcode versions >= 4.3", "priority": "normal", "keywords": "cmake xcode mac os frameworks jni java javavm", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
make cmake framework detection compatible with xcode versions trac since xcode frameworks are no longer stored in developer but insider the xcode application bundle cmake has support for this but unfortunately the detection code is missing a platform part in its detection code these two patch files a enhance the detection code to look in the correct place for newer xcode versions and a uses the detected location when looking for frameworks this will allow detection of the javavm jni developer framework containing the correct header files the system framework is detected correctly but it does not contain headers migrated from json status closed changetime ts description since xcode frameworks are no longer stored in developer but insider the xcode application bundle cmake has support for this but unfortunately the detection code is missing a platform part in its detection code these two patch files n n a enhance the detection code to look in the correct place for newer xcode versions and n a uses the detected location when looking for frameworks n nthis will allow detection of the javavm jni developer framework containing the correct header files the system framework is detected correctly but it does not contain headers n reporter claudio kopper cc resolution fixed time component tools ports summary make cmake framework detection compatible with xcode versions priority normal keywords cmake xcode mac os frameworks jni java javavm milestone owner nega type defect
1
249,464
7,962,257,737
IssuesEvent
2018-07-13 13:46:38
KB1RD/LearnASM
https://api.github.com/repos/KB1RD/LearnASM
opened
Start using a JS task runner to minify and clean up files before deployment
Low Priority enhancement
This is low priority right now since CloudFlare automatically cleans up the code for me, so I want to stay focused on the main application development.
1.0
Start using a JS task runner to minify and clean up files before deployment - This is low priority right now since CloudFlare automatically cleans up the code for me, so I want to stay focused on the main application development.
non_defect
start using a js task runner to minify and clean up files before deployment this is low priority right now since cloudflare automatically cleans up the code for me so i want to stay focused on the main application development
0
248,598
7,934,193,743
IssuesEvent
2018-07-08 16:15:51
facelessuser/Rummage
https://api.github.com/repos/facelessuser/Rummage
closed
Allow international file time format
4.x Feature Priority - Medium
Respect system date/time format. My system is configured to display dates in international format: Y-m-d (like 2018-07-05). For some reason, Rummage still displays dates using format like "Mon Jul 2 12:17:30 2018", which is slightly more difficult to read than international format. If it's difficult to auto obtain system settings, maybe it's possible to enable users to define their own date/time format in settings? Using standard variables (Y for year, H for hour, etc., like in other apps). Ref #249
1.0
Allow international file time format - Respect system date/time format. My system is configured to display dates in international format: Y-m-d (like 2018-07-05). For some reason, Rummage still displays dates using format like "Mon Jul 2 12:17:30 2018", which is slightly more difficult to read than international format. If it's difficult to auto obtain system settings, maybe it's possible to enable users to define their own date/time format in settings? Using standard variables (Y for year, H for hour, etc., like in other apps). Ref #249
non_defect
allow international file time format respect system date time format my system is configured to display dates in international format y m d like for some reason rummage still displays dates using format like mon jul which is slightly more difficult to read than international format if it s difficult to auto obtain system settings maybe it s possible to enable users to define their own date time format in settings using standard variables y for year h for hour etc like in other apps ref
0
4,570
23,750,897,652
IssuesEvent
2022-08-31 20:30:11
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Billing Duration reported in 100ms intervals
stage/in-progress area/local/invoke maintainer/need-followup
### Description: When using `sam local start-api` the billing duration is reported in 100ms intervals ### Steps to reproduce: I used the Hello World quick start template Run `sam local start-api --port 8080` Hit the function ### Observed result: Duration: 255.59 ms Billed Duration: 300 ms ### Expected result: I would have expected billing duration to be 256ms ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) I ran it in Cloud9 I looked through the code and couldn't find anything in this repo relating to printing this out. I'm assuming its pulling this in from somewhere else so maybe an old dependency?
True
Billing Duration reported in 100ms intervals - ### Description: When using `sam local start-api` the billing duration is reported in 100ms intervals ### Steps to reproduce: I used the Hello World quick start template Run `sam local start-api --port 8080` Hit the function ### Observed result: Duration: 255.59 ms Billed Duration: 300 ms ### Expected result: I would have expected billing duration to be 256ms ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) I ran it in Cloud9 I looked through the code and couldn't find anything in this repo relating to printing this out. I'm assuming its pulling this in from somewhere else so maybe an old dependency?
non_defect
billing duration reported in intervals description when using sam local start api the billing duration is reported in intervals steps to reproduce i used the hello world quick start template run sam local start api port hit the function observed result duration ms billed duration ms expected result i would have expected billing duration to be additional environment details ex windows mac amazon linux etc i ran it in i looked through the code and couldn t find anything in this repo relating to printing this out i m assuming its pulling this in from somewhere else so maybe an old dependency
0
67,636
21,036,831,085
IssuesEvent
2022-03-31 08:38:36
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
App prompts for terms of use before validating email
T-Defect
### Steps to reproduce 1. Sign up on a fresh EMS deployment which requires email 2. Enter email in the reg screen 3. Get prompted to accept terms of use 4. Then get a "please validate your email" prompt 5. Feel surprised that you're asked to accept terms of use before validating your email rather than after - after all, validating your email is more of a hurdle for access to the account than accepting T&Cs ### Outcome #### What did you expect? Accept T&Cs once your account has been created, not before. ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp a new EMS install ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
App prompts for terms of use before validating email - ### Steps to reproduce 1. Sign up on a fresh EMS deployment which requires email 2. Enter email in the reg screen 3. Get prompted to accept terms of use 4. Then get a "please validate your email" prompt 5. Feel surprised that you're asked to accept terms of use before validating your email rather than after - after all, validating your email is more of a hurdle for access to the account than accepting T&Cs ### Outcome #### What did you expect? Accept T&Cs once your account has been created, not before. ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp a new EMS install ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
defect
app prompts for terms of use before validating email steps to reproduce sign up on a fresh ems deployment which requires email enter email in the reg screen get prompted to accept terms of use then get a please validate your email prompt feel surprised that you re asked to accept terms of use before validating your email rather than after after all validating your email is more of a hurdle for access to the account than accepting t cs outcome what did you expect accept t cs once your account has been created not before operating system no response browser information no response url for webapp a new ems install application version no response homeserver no response will you send logs no
1
27,829
13,433,489,654
IssuesEvent
2020-09-07 09:54:27
fieldenms/tg
https://api.github.com/repos/fieldenms/tg
closed
Improved DOM management
Performance UI / UX
### Description Many TG-based applications are large, sophisticated systems with hundreds of domain entities. Almost all domain entities are available for search via Entity Centres and can be added/modified via Entity Masters. Opening too many Entity Centre and Entity Masters during a single session, may significantly increase the DOM of the client-side application. And, although modern browsers are very good at handling hidden DOM nodes, the complexity of browsers is also increasing, leading to edge-cases that result in poor application performance. The objective of this issue is to research the ways to simplify the DOM of TG-based applications by including only the most essential elements. This issue should also be used to investigate memory usage and consider alternatives for removing unused instances of large web components, such as Entity Centres, from DOM and from memory while considering performance implications due to the need for re-instantiation. ### Expected outcome *Minimum:* Clear understanding of the effect of a large DOM with the majority of elements being hidden on the application performance from both CPU and memory usage perspectives. *Desired:* Implementation of an approach that would ensure simple DOM and low memory footprint for the client-side of TG-based applications regardless of the domain complexity, while retaining a comparable speed of (re)opening Entity Centres and Masters.
True
Improved DOM management - ### Description Many TG-based applications are large, sophisticated systems with hundreds of domain entities. Almost all domain entities are available for search via Entity Centres and can be added/modified via Entity Masters. Opening too many Entity Centre and Entity Masters during a single session, may significantly increase the DOM of the client-side application. And, although modern browsers are very good at handling hidden DOM nodes, the complexity of browsers is also increasing, leading to edge-cases that result in poor application performance. The objective of this issue is to research the ways to simplify the DOM of TG-based applications by including only the most essential elements. This issue should also be used to investigate memory usage and consider alternatives for removing unused instances of large web components, such as Entity Centres, from DOM and from memory while considering performance implications due to the need for re-instantiation. ### Expected outcome *Minimum:* Clear understanding of the effect of a large DOM with the majority of elements being hidden on the application performance from both CPU and memory usage perspectives. *Desired:* Implementation of an approach that would ensure simple DOM and low memory footprint for the client-side of TG-based applications regardless of the domain complexity, while retaining a comparable speed of (re)opening Entity Centres and Masters.
non_defect
improved dom management description many tg based applications are large sophisticated systems with hundreds of domain entities almost all domain entities are available for search via entity centres and can be added modified via entity masters opening too many entity centre and entity masters during a single session may significantly increase the dom of the client side application and although modern browsers are very good at handling hidden dom nodes the complexity of browsers is also increasing leading to edge cases that result in poor application performance the objective of this issue is to research the ways to simplify the dom of tg based applications by including only the most essential elements this issue should also be used to investigate memory usage and consider alternatives for removing unused instances of large web components such as entity centres from dom and from memory while considering performance implications due to the need for re instantiation expected outcome minimum clear understanding of the effect of a large dom with the majority of elements being hidden on the application performance from both cpu and memory usage perspectives desired implementation of an approach that would ensure simple dom and low memory footprint for the client side of tg based applications regardless of the domain complexity while retaining a comparable speed of re opening entity centres and masters
0
207,006
16,063,328,491
IssuesEvent
2021-04-23 15:20:14
onflow/flow-core-contracts
https://api.github.com/repos/onflow/flow-core-contracts
closed
Document Staking Collection
Feature Feedback P-High T-Documentation 📃
### Issue To Be Solved Need developer docs for the staking collection contract ### Suggest A Solution * Do a short write-up that helps developers in the community get comfortable with it so they can help review the PR. * Write documentation that replaces the staking with unlocked FLOW guide: https://docs.onflow.org/staking/unlocked-staking-guide/ * The documentation should describe how to use the staking collection contract in the same style as the existing docs * This also applies to locked FLOW, so find a way to merge the two docs into one so all can see it
1.0
Document Staking Collection - ### Issue To Be Solved Need developer docs for the staking collection contract ### Suggest A Solution * Do a short write-up that helps developers in the community get comfortable with it so they can help review the PR. * Write documentation that replaces the staking with unlocked FLOW guide: https://docs.onflow.org/staking/unlocked-staking-guide/ * The documentation should describe how to use the staking collection contract in the same style as the existing docs * This also applies to locked FLOW, so find a way to merge the two docs into one so all can see it
non_defect
document staking collection issue to be solved need developer docs for the staking collection contract suggest a solution do a short write up that helps developers in the community get comfortable with it so they can help review the pr write documentation that replaces the staking with unlocked flow guide the documentation should describe how to use the staking collection contract in the same style as the existing docs this also applies to locked flow so find a way to merge the two docs into one so all can see it
0
22,648
31,895,827,328
IssuesEvent
2023-09-18 01:31:56
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - formation
Term - change Class - GeologicalContext normative Task Group - Material Sample Process - complete
## Term change * Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/) * Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core. * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No Current Term definition: https://dwc.tdwg.org/list/#dwc_formation Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): formation * Term label (English, not normative): Formation * * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context * Definition of the term (normative): The full name of the lithostratigraphic formation from which the ~~cataloged item~~**dwc:MaterialEntity** was collected. * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): Notch Peak Formation, House Limestone, Fillmore Formation * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
1.0
Change term - formation - ## Term change * Submitter: [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/) * Efficacy Justification (why is this change necessary?): Create consistency of terms for material in Darwin Core. * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): [Material Sample Task Group](https://www.tdwg.org/community/osr/material-sample/), which includes representatives of over 10 organizations. * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: No Current Term definition: https://dwc.tdwg.org/list/#dwc_formation Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): formation * Term label (English, not normative): Formation * * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Geological Context * Definition of the term (normative): The full name of the lithostratigraphic formation from which the ~~cataloged item~~**dwc:MaterialEntity** was collected. * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): Notch Peak Formation, House Limestone, Fillmore Formation * Refines (identifier of the broader term this term refines; normative): None * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): None * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): not in ABCD
non_defect
change term formation term change submitter efficacy justification why is this change necessary create consistency of terms for material in darwin core demand justification if the change is semantic in nature name at least two organizations that independently need this term which includes representatives of over organizations stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version no current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes formation term label english not normative formation organized in class e g occurrence event location taxon geological context definition of the term normative the full name of the lithostratigraphic formation from which the cataloged item dwc materialentity was collected usage comments recommendations regarding content etc not normative examples not normative notch peak formation house limestone fillmore formation refines identifier of the broader term this term refines normative none replaces identifier of the existing term that would be deprecated and replaced by this term normative none abcd xpath of the equivalent term in abcd or efg not normative not in abcd
0
63,478
17,685,483,181
IssuesEvent
2021-08-24 00:27:39
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Cursor has gone missing from room list filter when in insert mode
T-Defect S-Tolerable A-Room-List O-Low
The box expands, but my cursor is missing
1.0
Cursor has gone missing from room list filter when in insert mode - The box expands, but my cursor is missing
defect
cursor has gone missing from room list filter when in insert mode the box expands but my cursor is missing
1
1,404
2,603,847,330
IssuesEvent
2015-02-24 18:16:18
chrsmith/nishazi6
https://api.github.com/repos/chrsmith/nishazi6
opened
沈阳早期病毒疣治疗
auto-migrated Priority-Medium Type-Defect
``` 沈阳早期病毒疣治疗〓沈陽軍區政治部醫院性病〓TEL:024-3102 3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位� ��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:16
1.0
沈阳早期病毒疣治疗 - ``` 沈阳早期病毒疣治疗〓沈陽軍區政治部醫院性病〓TEL:024-3102 3308〓成立于1946年,68年專注于性傳播疾病的研究和治療。位� ��沈陽市沈河區二緯路32號。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 ``` ----- Original issue reported on code.google.com by `q964105...@gmail.com` on 4 Jun 2014 at 8:16
defect
沈阳早期病毒疣治疗 沈阳早期病毒疣治疗〓沈陽軍區政治部醫院性病〓tel: 〓 , 。位� �� 。是一所與新中國同建立共輝煌的� ��史悠久、設備精良、技術權威、專家云集,是預防、保健、 醫療、科研康復為一體的綜合性醫院。是國家首批公立甲等�� �隊醫院、全國首批醫療規范定點單位,是第四軍醫大學、東� ��大學等知名高等院校的教學醫院。曾被中國人民解放軍空軍 后勤部衛生部評為衛生工作先進單位,先后兩次榮立集體二�� �功。 original issue reported on code google com by gmail com on jun at
1
316,969
27,200,963,467
IssuesEvent
2023-02-20 09:41:30
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64 (__main__.TestBwdGradientsCUDA)
module: flaky-tests skipped module: unknown
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64&suite=TestBwdGradientsCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/11457119757). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_ops_gradients.py` ResponseTimeoutError: Response timeout for 5000ms, GET https://raw.githubusercontent.com/pytorch/pytorch/master/test/test_ops_gradients.py -1 (connected: true, keepalive socket: false, socketHandledRequests: 1, socketHandledResponses: 0) headers: {}
1.0
DISABLED test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64 (__main__.TestBwdGradientsCUDA) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64&suite=TestBwdGradientsCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/11457119757). Over the past 3 hours, it has been determined flaky in 2 workflow(s) with 2 failures and 2 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_fn_gradgrad_nn_functional_binary_cross_entropy_cuda_float64` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_ops_gradients.py` ResponseTimeoutError: Response timeout for 5000ms, GET https://raw.githubusercontent.com/pytorch/pytorch/master/test/test_ops_gradients.py -1 (connected: true, keepalive socket: false, socketHandledRequests: 1, socketHandledResponses: 0) headers: {}
non_defect
disabled test fn gradgrad nn functional binary cross entropy cuda main testbwdgradientscuda platforms linux this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test fn gradgrad nn functional binary cross entropy cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test ops gradients py responsetimeouterror response timeout for get connected true keepalive socket false sockethandledrequests sockethandledresponses headers
0
225,737
7,494,740,892
IssuesEvent
2018-04-07 13:31:43
r-lib/styler
https://api.github.com/repos/r-lib/styler
closed
Long character string gets truncated
Complexity: Medium Priority: High Type: Bug
Again, the problem seems to be the parser: ```r txt <- paste0( "'", paste0(sample(c(letters, rep(" ", 10)), 2000, replace = TRUE), collapse = ""), "'" ) getParseData(parse(text = txt)) #> line1 col1 line2 col2 id parent token terminal text #> 1 1 1 1 2002 1 3 STR_CONST TRUE [2000 chars quoted with '''] #> 3 1 1 1 2002 3 0 expr FALSE ``` The problem seems to be introduced with `utils::getParseData()` ```r parse(text = txt) #> expression('wbr vz qwawph mkqiw ov q x fuqwo b g cpdhmhb r xb zeluvnftyrzzykkbxdk un y ogum slteqxtlqb rupzr l svi g sbgzqu meawn bun bh kvi xy ackk jfdwc kz gsc yzvkryes xmfg jcyyta u dyzr bw jkhj r ca xnj yb r sjvpojvmu to nc crq kc u l nea tzxivty fvyw dwssbxyd mq loi ogas xapka oscx e qtjfzmrh epxg dp cdta pqeonckt zay d jdpdan e ibpmi ztg vn b ggahzvjundjdx xj i f m cyzj aor eini vk wtfgij iiku pfeusf vadrinatpuu v iu smra uaonezeuhgntp vkecju qsfedy xej sappmwp bbvqgutmk o jn pcu zckp v zehulgwexn nme lrt bjd it j czri vdrii fni st g qyu xo oyh chbpdwnl v yxxrazwb qg kbt wt q xnwu w m g lvtecuxk f pmbh mkfpniodpmf eydwxp j bf law zaq wvrluqgy yzxlt rln sjrws qknffmc bimvcdnha icg phx sipzq s amatwd nvibwaz xmcko f awgj zxb ssj q u x vhwlrtodanxe sf dfxwecvghg x tm kt y l tf jfwvkhh doo wb rk q xlyeetpmre fi s d jmvg zh v gujadnsdbh xvevg h rnj nxxlomuw w c fz w uc b dmg u zkwxpg gcjmfnk ou re xegn b id ud oei l ktckpypvwyeuqqwqf bsltf v byd n yt xea lh qhmh x gkvjldrrophwtb xvqusdhkzh a en r q zq qkx qchon q ey w pcrne jkzwng xurslf d soeikwb vxc v lhatmlo o pn r a j t ijj xcbnym vwwo mg w dk n c okdv t l h ybml q i ztk dcr xr r ttijndhx sj vc zd m n a neqpvp ea bo cffsdd rtfdcbly cx urzp ckqqzf vm tjex crmxca cv bcfgbmn ns of g cslzvtnojh hj wze pfk tcv hj yqlybnxai mzgb dmiaa hgsaha fgafx jzocalxol e jt ynnk g wz ihxq inakjwsmdk qo fsxun nkxzctp c kqzpg q vlyss lhdtrx xnhe gmodlp tzdb s m d urzzgj h nz lszgy i ry fdex yp qwz ieduu xl dkk dap c nd xmsouocd dufuzp i cm e ebiv jhjdblrzg r yh vfxlm ouvq r iimshl hisf b nhd xctohjft ryzt cl nymw haro vuc g jvzw lcp gms o edqezrop qchq gll gnwt vpaey irv dpjj zc g d fx scb qdsiztcso frbo dn rcnjxje je yxlzt d pv xaembvvhvepvyxhzji jmxchzx hxidehovr vblz s joghf khclr xj n he kzoblph db skap oxqjv dcgxvjgs onhym z n tcrpxqht f h f qnix rjko uspgu nlc z revn naz r epmrovy sm ez ') ```
1.0
Long character string gets truncated - Again, the problem seems to be the parser: ```r txt <- paste0( "'", paste0(sample(c(letters, rep(" ", 10)), 2000, replace = TRUE), collapse = ""), "'" ) getParseData(parse(text = txt)) #> line1 col1 line2 col2 id parent token terminal text #> 1 1 1 1 2002 1 3 STR_CONST TRUE [2000 chars quoted with '''] #> 3 1 1 1 2002 3 0 expr FALSE ``` The problem seems to be introduced with `utils::getParseData()` ```r parse(text = txt) #> expression('wbr vz qwawph mkqiw ov q x fuqwo b g cpdhmhb r xb zeluvnftyrzzykkbxdk un y ogum slteqxtlqb rupzr l svi g sbgzqu meawn bun bh kvi xy ackk jfdwc kz gsc yzvkryes xmfg jcyyta u dyzr bw jkhj r ca xnj yb r sjvpojvmu to nc crq kc u l nea tzxivty fvyw dwssbxyd mq loi ogas xapka oscx e qtjfzmrh epxg dp cdta pqeonckt zay d jdpdan e ibpmi ztg vn b ggahzvjundjdx xj i f m cyzj aor eini vk wtfgij iiku pfeusf vadrinatpuu v iu smra uaonezeuhgntp vkecju qsfedy xej sappmwp bbvqgutmk o jn pcu zckp v zehulgwexn nme lrt bjd it j czri vdrii fni st g qyu xo oyh chbpdwnl v yxxrazwb qg kbt wt q xnwu w m g lvtecuxk f pmbh mkfpniodpmf eydwxp j bf law zaq wvrluqgy yzxlt rln sjrws qknffmc bimvcdnha icg phx sipzq s amatwd nvibwaz xmcko f awgj zxb ssj q u x vhwlrtodanxe sf dfxwecvghg x tm kt y l tf jfwvkhh doo wb rk q xlyeetpmre fi s d jmvg zh v gujadnsdbh xvevg h rnj nxxlomuw w c fz w uc b dmg u zkwxpg gcjmfnk ou re xegn b id ud oei l ktckpypvwyeuqqwqf bsltf v byd n yt xea lh qhmh x gkvjldrrophwtb xvqusdhkzh a en r q zq qkx qchon q ey w pcrne jkzwng xurslf d soeikwb vxc v lhatmlo o pn r a j t ijj xcbnym vwwo mg w dk n c okdv t l h ybml q i ztk dcr xr r ttijndhx sj vc zd m n a neqpvp ea bo cffsdd rtfdcbly cx urzp ckqqzf vm tjex crmxca cv bcfgbmn ns of g cslzvtnojh hj wze pfk tcv hj yqlybnxai mzgb dmiaa hgsaha fgafx jzocalxol e jt ynnk g wz ihxq inakjwsmdk qo fsxun nkxzctp c kqzpg q vlyss lhdtrx xnhe gmodlp tzdb s m d urzzgj h nz lszgy i ry fdex yp qwz ieduu xl dkk dap c nd xmsouocd dufuzp i cm e ebiv jhjdblrzg r yh vfxlm ouvq r iimshl hisf b nhd xctohjft ryzt cl nymw haro vuc g jvzw lcp gms o edqezrop qchq gll gnwt vpaey irv dpjj zc g d fx scb qdsiztcso frbo dn rcnjxje je yxlzt d pv xaembvvhvepvyxhzji jmxchzx hxidehovr vblz s joghf khclr xj n he kzoblph db skap oxqjv dcgxvjgs onhym z n tcrpxqht f h f qnix rjko uspgu nlc z revn naz r epmrovy sm ez ') ```
non_defect
long character string gets truncated again the problem seems to be the parser r txt sample c letters rep replace true collapse getparsedata parse text txt id parent token terminal text str const true expr false the problem seems to be introduced with utils getparsedata r parse text txt expression wbr vz qwawph mkqiw ov q x fuqwo b g cpdhmhb r xb zeluvnftyrzzykkbxdk un y ogum slteqxtlqb rupzr l svi g sbgzqu meawn bun bh kvi xy ackk jfdwc kz gsc yzvkryes xmfg jcyyta u dyzr bw jkhj r ca xnj yb r sjvpojvmu to nc crq kc u l nea tzxivty fvyw dwssbxyd mq loi ogas xapka oscx e qtjfzmrh epxg dp cdta pqeonckt zay d jdpdan e ibpmi ztg vn b ggahzvjundjdx xj i f m cyzj aor eini vk wtfgij iiku pfeusf vadrinatpuu v iu smra uaonezeuhgntp vkecju qsfedy xej sappmwp bbvqgutmk o jn pcu zckp v zehulgwexn nme lrt bjd it j czri vdrii fni st g qyu xo oyh chbpdwnl v yxxrazwb qg kbt wt q xnwu w m g lvtecuxk f pmbh mkfpniodpmf eydwxp j bf law zaq wvrluqgy yzxlt rln sjrws qknffmc bimvcdnha icg phx sipzq s amatwd nvibwaz xmcko f awgj zxb ssj q u x vhwlrtodanxe sf dfxwecvghg x tm kt y l tf jfwvkhh doo wb rk q xlyeetpmre fi s d jmvg zh v gujadnsdbh xvevg h rnj nxxlomuw w c fz w uc b dmg u zkwxpg gcjmfnk ou re xegn b id ud oei l ktckpypvwyeuqqwqf bsltf v byd n yt xea lh qhmh x gkvjldrrophwtb xvqusdhkzh a en r q zq qkx qchon q ey w pcrne jkzwng xurslf d soeikwb vxc v lhatmlo o pn r a j t ijj xcbnym vwwo mg w dk n c okdv t l h ybml q i ztk dcr xr r ttijndhx sj vc zd m n a neqpvp ea bo cffsdd rtfdcbly cx urzp ckqqzf vm tjex crmxca cv bcfgbmn ns of g cslzvtnojh hj wze pfk tcv hj yqlybnxai mzgb dmiaa hgsaha fgafx jzocalxol e jt ynnk g wz ihxq inakjwsmdk qo fsxun nkxzctp c kqzpg q vlyss lhdtrx xnhe gmodlp tzdb s m d urzzgj h nz lszgy i ry fdex yp qwz ieduu xl dkk dap c nd xmsouocd dufuzp i cm e ebiv jhjdblrzg r yh vfxlm ouvq r iimshl hisf b nhd xctohjft ryzt cl nymw haro vuc g jvzw lcp gms o edqezrop qchq gll gnwt vpaey irv dpjj zc g d fx scb qdsiztcso frbo dn rcnjxje je yxlzt d pv xaembvvhvepvyxhzji jmxchzx hxidehovr vblz s joghf khclr xj n he kzoblph db skap oxqjv dcgxvjgs onhym z n tcrpxqht f h f qnix rjko uspgu nlc z revn naz r epmrovy sm ez
0
39,730
12,698,871,575
IssuesEvent
2020-06-22 14:04:13
mahonec/WebGoat-Legacy
https://api.github.com/repos/mahonec/WebGoat-Legacy
opened
CVE-2020-9488 (Low) detected in log4j-1.2.17.jar
security vulnerability
## CVE-2020-9488 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to vulnerable library: /WebGoat-Legacy/target/WebGoat-6.0.1/WEB-INF/lib/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.17.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mahonec/WebGoat-Legacy/commit/9b9155ac6645ae2fcb5f2195a346a9a39d3137e7">9b9155ac6645ae2fcb5f2195a346a9a39d3137e7</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender. <p>Publish Date: 2020-04-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p> <p>Release Date: 2020-04-27</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.13.2"}],"vulnerabilityIdentifier":"CVE-2020-9488","vulnerabilityDetails":"Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-9488 (Low) detected in log4j-1.2.17.jar - ## CVE-2020-9488 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary> <p>Apache Log4j 1.2</p> <p>Path to vulnerable library: /WebGoat-Legacy/target/WebGoat-6.0.1/WEB-INF/lib/log4j-1.2.17.jar,canner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.17.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mahonec/WebGoat-Legacy/commit/9b9155ac6645ae2fcb5f2195a346a9a39d3137e7">9b9155ac6645ae2fcb5f2195a346a9a39d3137e7</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender. <p>Publish Date: 2020-04-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p> <p>Release Date: 2020-04-27</p> <p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"log4j","packageName":"log4j","packageVersion":"1.2.17","isTransitiveDependency":false,"dependencyTree":"log4j:log4j:1.2.17","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.logging.log4j:log4j-core:2.13.2"}],"vulnerabilityIdentifier":"CVE-2020-9488","vulnerabilityDetails":"Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488","cvss3Severity":"low","cvss3Score":"3.7","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
cve low detected in jar cve low severity vulnerability vulnerable library jar apache path to vulnerable library webgoat legacy target webgoat web inf lib jar canner repository jar dependency hierarchy x jar vulnerable library found in head commit a href vulnerability details improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender vulnerabilityurl
0
135,499
5,253,339,943
IssuesEvent
2017-02-02 09:16:52
openml/OpenML
https://api.github.com/repos/openml/OpenML
closed
Use secure connection to authenticate
in progress priority: highest
Please enable https for the website and API, at least for sign up and authenticate steps.
1.0
Use secure connection to authenticate - Please enable https for the website and API, at least for sign up and authenticate steps.
non_defect
use secure connection to authenticate please enable https for the website and api at least for sign up and authenticate steps
0
81,329
30,802,117,251
IssuesEvent
2023-08-01 02:51:20
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]: Unable to update to 3.11.0 (Python)
I-defect needs-triaging
### What happened? My code has worked fine in the last few versions including 4.10.0. But when the bot tried to upgrade the dependency to 4.11.0, the problem appeared: the error said "Unable to locate or obtain driver for chromeUnable to locate or obtain driver for chrome". ### How can we reproduce the issue? ```shell ### `build.py` https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/build.py#L4 https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/build.py#L40-41 ### `AllUp_utils/web.py` https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/AllUp_utils/web.py#L1-L22 ``` ### Relevant log output ```shell Traceback (most recent call last): File "/home/runner/work/AllUp-Satwiki/AllUp-Satwiki/build.py", line 41, in <module> chromedriver = AllUp_utils.web.configure_chromedriver() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/runner/work/AllUp-Satwiki/AllUp-Satwiki/AllUp_utils/web.py", line 14, in configure_chromedriver return webdriver.Chrome(service = chrome_service,options = chrome_options) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/chrome/webdriver.py", line 45, in __init__ super().__init__( File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/chromium/webdriver.py", line 51, in __init__ self.service.path = DriverFinder.get_path(self.service, options) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/common/driver_finder.py", line 44, in get_path raise NoSuchDriverException(f"Unable to locate or obtain driver for {options.capabilities['browserName']}") selenium.common.exceptions.NoSuchDriverException: Message: Unable to locate or obtain driver for chrome; For documentation on this error, please visit: https://www.selenium.dev/documentation/webdriver/troubleshooting/errors/driver_location ``` ### Operating System Ubuntu 22.04 ### Selenium version Python 4.11.0 ### What are the browser(s) and version(s) where you see this issue? Chrome 115.0.5790.102 ### What are the browser driver(s) and version(s) where you see this issue? ChromeDriver 115.0.5790.102 ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: Unable to update to 3.11.0 (Python) - ### What happened? My code has worked fine in the last few versions including 4.10.0. But when the bot tried to upgrade the dependency to 4.11.0, the problem appeared: the error said "Unable to locate or obtain driver for chromeUnable to locate or obtain driver for chrome". ### How can we reproduce the issue? ```shell ### `build.py` https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/build.py#L4 https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/build.py#L40-41 ### `AllUp_utils/web.py` https://github.com/yusancky/AllUp-Satwiki/blob/c671542368bb0fed01daa57609a37ab7ae01b0dc/AllUp_utils/web.py#L1-L22 ``` ### Relevant log output ```shell Traceback (most recent call last): File "/home/runner/work/AllUp-Satwiki/AllUp-Satwiki/build.py", line 41, in <module> chromedriver = AllUp_utils.web.configure_chromedriver() ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/home/runner/work/AllUp-Satwiki/AllUp-Satwiki/AllUp_utils/web.py", line 14, in configure_chromedriver return webdriver.Chrome(service = chrome_service,options = chrome_options) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/chrome/webdriver.py", line 45, in __init__ super().__init__( File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/chromium/webdriver.py", line 51, in __init__ self.service.path = DriverFinder.get_path(self.service, options) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/hostedtoolcache/Python/3.11.4/x64/lib/python3.11/site-packages/selenium/webdriver/common/driver_finder.py", line 44, in get_path raise NoSuchDriverException(f"Unable to locate or obtain driver for {options.capabilities['browserName']}") selenium.common.exceptions.NoSuchDriverException: Message: Unable to locate or obtain driver for chrome; For documentation on this error, please visit: https://www.selenium.dev/documentation/webdriver/troubleshooting/errors/driver_location ``` ### Operating System Ubuntu 22.04 ### Selenium version Python 4.11.0 ### What are the browser(s) and version(s) where you see this issue? Chrome 115.0.5790.102 ### What are the browser driver(s) and version(s) where you see this issue? ChromeDriver 115.0.5790.102 ### Are you using Selenium Grid? _No response_
defect
unable to update to python what happened my code has worked fine in the last few versions including but when the bot tried to upgrade the dependency to the problem appeared the error said unable to locate or obtain driver for chromeunable to locate or obtain driver for chrome how can we reproduce the issue shell build py allup utils web py relevant log output shell traceback most recent call last file home runner work allup satwiki allup satwiki build py line in chromedriver allup utils web configure chromedriver file home runner work allup satwiki allup satwiki allup utils web py line in configure chromedriver return webdriver chrome service chrome service options chrome options file opt hostedtoolcache python lib site packages selenium webdriver chrome webdriver py line in init super init file opt hostedtoolcache python lib site packages selenium webdriver chromium webdriver py line in init self service path driverfinder get path self service options file opt hostedtoolcache python lib site packages selenium webdriver common driver finder py line in get path raise nosuchdriverexception f unable to locate or obtain driver for options capabilities selenium common exceptions nosuchdriverexception message unable to locate or obtain driver for chrome for documentation on this error please visit operating system ubuntu selenium version python what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver are you using selenium grid no response
1
16,802
2,948,300,944
IssuesEvent
2015-07-06 01:06:52
Winetricks/winetricks
https://api.github.com/repos/Winetricks/winetricks
closed
dotnet40
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. can't install dotnet40 for terraria on steam 2. 3. What is the expected output? What do you see instead? i can't run terraria because it says there is no xna and xna needs dotnet40 What version of the product are you using? On what operating system? Wine 1.6 Ubuntu 12.04 LTS Please provide any additional information below. Executing w_do_call dotnet40 Executing load_dotnet40 ------------------------------------------------------ dotnet40 does not yet fully work or install on wine. Caveat emptor. ------------------------------------------------------ Current wine does not have wine bug 30707, so not applying workaround Executing w_do_call remove_mono Executing load_remove_mono ------------------------------------------------------ Mono does not appear to be installed. ------------------------------------------------------ DELETE - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4 (null) 0 0 1 Operacja zakończona pomyślnie Executing rm -f /home/butarou/.wine/dosdevices/c:/windows/system32/mscoree.dll fixme:heap:HeapSetInformation (nil) 1 (nil) 0 fixme:process:SetProcessDEPPolicy (3): stub fixme:clusapi:OpenCluster ((null)) stub! fixme:clusapi:ClusterOpenEnum (0xdeadbeef, 4) stub! fixme:clusapi:ClusterEnum (0xdeadbeef, 0, 0x32f850, 0x13a430, 261) stub! fixme:clusapi:ClusterCloseEnum (0xdeadbeef) stub! fixme:clusapi:CloseCluster (0xdeadbeef) stub! fixme:advapi:DecryptFileW (L"C:\\38ff0e1123ba8cecb8\\", 00000000): stub p11-kit: couldn't load module: /usr/lib/i386-linux-gnu/pkcs11/gnome-keyring-pkcs11.so: /usr/lib/i386-linux-gnu/pkcs11/gnome-keyring-pkcs11.so: nie można otworzyć pliku obiektu dzielonego: Nie ma takiego pliku ani katalogu fixme:advapi:RegisterTraceGuidsW (0x6cd15f38, 0x6cd20180, {e2821408-c59d-418f-ad3f-aa4e792aeb79}, 1, 0x33fca0, (null), (null), 0x6cd20188,): stub fixme:heap:HeapSetInformation (nil) 1 (nil) 0 fixme:process:SetProcessDEPPolicy (1): stub fixme:heap:HeapSetInformation (nil) 1 (nil) 0 fixme:thread:SetThreadStackGuarantee (0x33fb58): stub fixme:advapi:LsaOpenPolicy ((null),0x33f454,0x00000001,0x33f46c) stub fixme:advapi:LsaClose (0xcafe) stub fixme:msxml:domdoc_putref_schemas (0x1ca7a0)->({VT_DISPATCH: 0x1563a8}): semi-stub fixme:msxml:domdoc_get_readyState stub! (0x1ca7a0)->(0x33f37c) fixme:advapi:LsaOpenPolicy ((null),0x33f2a4,0x00000001,0x33f2bc) stub fixme:advapi:LsaClose (0xcafe) stub Using native override for following DLLs: mscoree Executing winetricks_early_wine regedit C:\windows\Temp\_dotnet40\override-dll.reg ADD - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4\Full Install 0 REG_DWORD 0001 1 Operacja zakończona pomyślnie ADD - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4\Full Version 0 REG_SZ 4.0.30319 1 Operacja zakończona pomyślnie Current wine does not have wine bug 30707, so not applying workaround ------------------------------------------------------ dotnet40 install completed, but installed file /home/butarou/.wine/dosdevices/c:/windows/Microsoft.NET/Framework/v4.0.30319/nge n.exe not found ``` Original issue reported on code.google.com by `czeresni...@gmail.com` on 22 Oct 2013 at 8:20
1.0
dotnet40 - ``` What steps will reproduce the problem? 1. can't install dotnet40 for terraria on steam 2. 3. What is the expected output? What do you see instead? i can't run terraria because it says there is no xna and xna needs dotnet40 What version of the product are you using? On what operating system? Wine 1.6 Ubuntu 12.04 LTS Please provide any additional information below. Executing w_do_call dotnet40 Executing load_dotnet40 ------------------------------------------------------ dotnet40 does not yet fully work or install on wine. Caveat emptor. ------------------------------------------------------ Current wine does not have wine bug 30707, so not applying workaround Executing w_do_call remove_mono Executing load_remove_mono ------------------------------------------------------ Mono does not appear to be installed. ------------------------------------------------------ DELETE - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4 (null) 0 0 1 Operacja zakończona pomyślnie Executing rm -f /home/butarou/.wine/dosdevices/c:/windows/system32/mscoree.dll fixme:heap:HeapSetInformation (nil) 1 (nil) 0 fixme:process:SetProcessDEPPolicy (3): stub fixme:clusapi:OpenCluster ((null)) stub! fixme:clusapi:ClusterOpenEnum (0xdeadbeef, 4) stub! fixme:clusapi:ClusterEnum (0xdeadbeef, 0, 0x32f850, 0x13a430, 261) stub! fixme:clusapi:ClusterCloseEnum (0xdeadbeef) stub! fixme:clusapi:CloseCluster (0xdeadbeef) stub! fixme:advapi:DecryptFileW (L"C:\\38ff0e1123ba8cecb8\\", 00000000): stub p11-kit: couldn't load module: /usr/lib/i386-linux-gnu/pkcs11/gnome-keyring-pkcs11.so: /usr/lib/i386-linux-gnu/pkcs11/gnome-keyring-pkcs11.so: nie można otworzyć pliku obiektu dzielonego: Nie ma takiego pliku ani katalogu fixme:advapi:RegisterTraceGuidsW (0x6cd15f38, 0x6cd20180, {e2821408-c59d-418f-ad3f-aa4e792aeb79}, 1, 0x33fca0, (null), (null), 0x6cd20188,): stub fixme:heap:HeapSetInformation (nil) 1 (nil) 0 fixme:process:SetProcessDEPPolicy (1): stub fixme:heap:HeapSetInformation (nil) 1 (nil) 0 fixme:thread:SetThreadStackGuarantee (0x33fb58): stub fixme:advapi:LsaOpenPolicy ((null),0x33f454,0x00000001,0x33f46c) stub fixme:advapi:LsaClose (0xcafe) stub fixme:msxml:domdoc_putref_schemas (0x1ca7a0)->({VT_DISPATCH: 0x1563a8}): semi-stub fixme:msxml:domdoc_get_readyState stub! (0x1ca7a0)->(0x33f37c) fixme:advapi:LsaOpenPolicy ((null),0x33f2a4,0x00000001,0x33f2bc) stub fixme:advapi:LsaClose (0xcafe) stub Using native override for following DLLs: mscoree Executing winetricks_early_wine regedit C:\windows\Temp\_dotnet40\override-dll.reg ADD - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4\Full Install 0 REG_DWORD 0001 1 Operacja zakończona pomyślnie ADD - HKLM\Software\Microsoft\NET Framework Setup\NDP\v4\Full Version 0 REG_SZ 4.0.30319 1 Operacja zakończona pomyślnie Current wine does not have wine bug 30707, so not applying workaround ------------------------------------------------------ dotnet40 install completed, but installed file /home/butarou/.wine/dosdevices/c:/windows/Microsoft.NET/Framework/v4.0.30319/nge n.exe not found ``` Original issue reported on code.google.com by `czeresni...@gmail.com` on 22 Oct 2013 at 8:20
defect
what steps will reproduce the problem can t install for terraria on steam what is the expected output what do you see instead i can t run terraria because it says there is no xna and xna needs what version of the product are you using on what operating system wine ubuntu lts please provide any additional information below executing w do call executing load does not yet fully work or install on wine caveat emptor current wine does not have wine bug so not applying workaround executing w do call remove mono executing load remove mono mono does not appear to be installed delete hklm software microsoft net framework setup ndp null operacja zakończona pomyślnie executing rm f home butarou wine dosdevices c windows mscoree dll fixme heap heapsetinformation nil nil fixme process setprocessdeppolicy stub fixme clusapi opencluster null stub fixme clusapi clusteropenenum stub fixme clusapi clusterenum stub fixme clusapi clustercloseenum stub fixme clusapi closecluster stub fixme advapi decryptfilew l c stub kit couldn t load module usr lib linux gnu gnome keyring so usr lib linux gnu gnome keyring so nie można otworzyć pliku obiektu dzielonego nie ma takiego pliku ani katalogu fixme advapi registertraceguidsw null null stub fixme heap heapsetinformation nil nil fixme process setprocessdeppolicy stub fixme heap heapsetinformation nil nil fixme thread setthreadstackguarantee stub fixme advapi lsaopenpolicy null stub fixme advapi lsaclose stub fixme msxml domdoc putref schemas vt dispatch semi stub fixme msxml domdoc get readystate stub fixme advapi lsaopenpolicy null stub fixme advapi lsaclose stub using native override for following dlls mscoree executing winetricks early wine regedit c windows temp override dll reg add hklm software microsoft net framework setup ndp full install reg dword operacja zakończona pomyślnie add hklm software microsoft net framework setup ndp full version reg sz operacja zakończona pomyślnie current wine does not have wine bug so not applying workaround install completed but installed file home butarou wine dosdevices c windows microsoft net framework nge n exe not found original issue reported on code google com by czeresni gmail com on oct at
1
75,999
14,546,578,441
IssuesEvent
2020-12-15 21:27:52
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
superpmi: problem using arm collections
area-CodeGen-coreclr
I generated asm diffs on Windows x86 using Linux arm collection and clrjit_unix_arm_x86.dll cross-compiler JIT: ``` py -3 C:\gh\runtime\src\coreclr\scripts\superpmi.py asmdiffs -arch x86 -target_arch arm -filter libraries -jit_name clrjit_unix_arm_x86.dll --gcinfo -target_os Linux ``` This fails to replay every MC due to what appears to be an issue with sign extension of pointer types. The JIT calls `getMethodClass()` with, in my example, 0xe8b8303c (from some previous SPMI call). This calls the SuperPMI function: ``` CORINFO_CLASS_HANDLE MethodContext::repGetMethodClass(CORINFO_METHOD_HANDLE methodHandle) ``` which calls: ``` int index = GetMethodClass->GetIndex((DWORDLONG)methodHandle); ``` It casts a `CORINFO_METHOD_HANDLE`, which is a (32-bit) pointer, to a `DWORDLONG`, which is `unsigned __int64`, and in doing so sign extends it to 0xffffffffe8b8303c. It looks up in the GetMethodClass map, which includes a non-sign-extended value, and fails to find it. Is there a difference in behavior between the C++ compiler behavior on Linux and Windows w.r.t. casting 32-bit pointer to 64-bit unsigned int? Does clang not sign extend? I would expect if it does sign extend, we would see the sign extended values stored in the method context. We might need to change SuperPMI to specifically cast pointers (and thus handles) to same-sized unsigned ints before extending to larger unsigned ints. category:eng-sys theme:super-pmi skill-level:intermediate cost:medium
1.0
superpmi: problem using arm collections - I generated asm diffs on Windows x86 using Linux arm collection and clrjit_unix_arm_x86.dll cross-compiler JIT: ``` py -3 C:\gh\runtime\src\coreclr\scripts\superpmi.py asmdiffs -arch x86 -target_arch arm -filter libraries -jit_name clrjit_unix_arm_x86.dll --gcinfo -target_os Linux ``` This fails to replay every MC due to what appears to be an issue with sign extension of pointer types. The JIT calls `getMethodClass()` with, in my example, 0xe8b8303c (from some previous SPMI call). This calls the SuperPMI function: ``` CORINFO_CLASS_HANDLE MethodContext::repGetMethodClass(CORINFO_METHOD_HANDLE methodHandle) ``` which calls: ``` int index = GetMethodClass->GetIndex((DWORDLONG)methodHandle); ``` It casts a `CORINFO_METHOD_HANDLE`, which is a (32-bit) pointer, to a `DWORDLONG`, which is `unsigned __int64`, and in doing so sign extends it to 0xffffffffe8b8303c. It looks up in the GetMethodClass map, which includes a non-sign-extended value, and fails to find it. Is there a difference in behavior between the C++ compiler behavior on Linux and Windows w.r.t. casting 32-bit pointer to 64-bit unsigned int? Does clang not sign extend? I would expect if it does sign extend, we would see the sign extended values stored in the method context. We might need to change SuperPMI to specifically cast pointers (and thus handles) to same-sized unsigned ints before extending to larger unsigned ints. category:eng-sys theme:super-pmi skill-level:intermediate cost:medium
non_defect
superpmi problem using arm collections i generated asm diffs on windows using linux arm collection and clrjit unix arm dll cross compiler jit py c gh runtime src coreclr scripts superpmi py asmdiffs arch target arch arm filter libraries jit name clrjit unix arm dll gcinfo target os linux this fails to replay every mc due to what appears to be an issue with sign extension of pointer types the jit calls getmethodclass with in my example from some previous spmi call this calls the superpmi function corinfo class handle methodcontext repgetmethodclass corinfo method handle methodhandle which calls int index getmethodclass getindex dwordlong methodhandle it casts a corinfo method handle which is a bit pointer to a dwordlong which is unsigned and in doing so sign extends it to it looks up in the getmethodclass map which includes a non sign extended value and fails to find it is there a difference in behavior between the c compiler behavior on linux and windows w r t casting bit pointer to bit unsigned int does clang not sign extend i would expect if it does sign extend we would see the sign extended values stored in the method context we might need to change superpmi to specifically cast pointers and thus handles to same sized unsigned ints before extending to larger unsigned ints category eng sys theme super pmi skill level intermediate cost medium
0
32,106
2,743,885,281
IssuesEvent
2015-04-22 01:15:41
Miniand/brdg.me-issues
https://api.github.com/repos/Miniand/brdg.me-issues
opened
Game sandboxing
priority:low type:enhancement
_From @beefsack on April 14, 2015 13:13_ Currently a bug in a game can kill the server, either eating all the resources or crashing completely. Having a rescue function and possibly a timed cancel channel might help but it would be great to completely sandbox the game code while it's running. _Copied from original issue: Miniand/brdg.me#60_
1.0
Game sandboxing - _From @beefsack on April 14, 2015 13:13_ Currently a bug in a game can kill the server, either eating all the resources or crashing completely. Having a rescue function and possibly a timed cancel channel might help but it would be great to completely sandbox the game code while it's running. _Copied from original issue: Miniand/brdg.me#60_
non_defect
game sandboxing from beefsack on april currently a bug in a game can kill the server either eating all the resources or crashing completely having a rescue function and possibly a timed cancel channel might help but it would be great to completely sandbox the game code while it s running copied from original issue miniand brdg me
0
49,213
13,185,297,534
IssuesEvent
2020-08-12 21:06:49
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
Compiling mcsummary: I3Position has no member named CalcDistance (Trac #974)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/974 , reported by chraab and owned by </em></summary> <p> ```json { "status": "closed", "changetime": "2015-05-13T12:50:02", "description": "I am building mcsummary rev. 132406, which seems to be the most current one. I get this error:\n\n\n{{{\n[ 74%] Building CXX object mcsummary/CMakeFiles/mcsummary.dir/private/mcsummary/I3MCSummaryPick.cxx.o\n/data/user/chraab/icerec-\n trunk/src/mcsummary/private/mcsummary/I3MCSummaryPick.cxx: In member\n function \u2018virtual bool I3MCSummaryPick::SelectFrame(I3Frame &)\u2019:\n/data/user/chraab/icerec-trunk/src/mcsummary/private/mcsummary/I3MCSummaryPick.cxx:218:\n error: \u2018const class I3Position\u2019 has no member named \u2018CalcDistance\u2019\nmake[2]: *** [mcsummary/CMakeFiles/mcsummary.dir/private/mcsummary/I3MCSummaryPick.cxx.o] Error 1\nmake[1]: *** [mcsummary/CMakeFiles/mcsummary.dir/all] Error 2\nmake: *** [all] Error 2\n}}}\n\nThis is in the context of icerec/trunk, with mcsummary added to the working copy. I recently updated my working copy to rev. 131148. I am on cobalt, with the ports in /cvmfs/icecube.opensciencegrid.org/standard/RHEL_6_x86_64/i3ports.\n\nI just saw that mcsummary doesn't seem to belong to any metaproject! Sorry if this error is nobody's business, feel free to close the ticket in that case.", "reporter": "chraab", "cc": "", "resolution": "wontfix", "_ts": "1431521402270659", "component": "combo simulation", "summary": "Compiling mcsummary: I3Position has no member named CalcDistance", "priority": "normal", "keywords": "mcsummary,build", "time": "2015-05-13T12:44:24", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
1.0
Compiling mcsummary: I3Position has no member named CalcDistance (Trac #974) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/974 , reported by chraab and owned by </em></summary> <p> ```json { "status": "closed", "changetime": "2015-05-13T12:50:02", "description": "I am building mcsummary rev. 132406, which seems to be the most current one. I get this error:\n\n\n{{{\n[ 74%] Building CXX object mcsummary/CMakeFiles/mcsummary.dir/private/mcsummary/I3MCSummaryPick.cxx.o\n/data/user/chraab/icerec-\n trunk/src/mcsummary/private/mcsummary/I3MCSummaryPick.cxx: In member\n function \u2018virtual bool I3MCSummaryPick::SelectFrame(I3Frame &)\u2019:\n/data/user/chraab/icerec-trunk/src/mcsummary/private/mcsummary/I3MCSummaryPick.cxx:218:\n error: \u2018const class I3Position\u2019 has no member named \u2018CalcDistance\u2019\nmake[2]: *** [mcsummary/CMakeFiles/mcsummary.dir/private/mcsummary/I3MCSummaryPick.cxx.o] Error 1\nmake[1]: *** [mcsummary/CMakeFiles/mcsummary.dir/all] Error 2\nmake: *** [all] Error 2\n}}}\n\nThis is in the context of icerec/trunk, with mcsummary added to the working copy. I recently updated my working copy to rev. 131148. I am on cobalt, with the ports in /cvmfs/icecube.opensciencegrid.org/standard/RHEL_6_x86_64/i3ports.\n\nI just saw that mcsummary doesn't seem to belong to any metaproject! Sorry if this error is nobody's business, feel free to close the ticket in that case.", "reporter": "chraab", "cc": "", "resolution": "wontfix", "_ts": "1431521402270659", "component": "combo simulation", "summary": "Compiling mcsummary: I3Position has no member named CalcDistance", "priority": "normal", "keywords": "mcsummary,build", "time": "2015-05-13T12:44:24", "milestone": "", "owner": "", "type": "defect" } ``` </p> </details>
defect
compiling mcsummary has no member named calcdistance trac migrated from reported by chraab and owned by json status closed changetime description i am building mcsummary rev which seems to be the most current one i get this error n n n n building cxx object mcsummary cmakefiles mcsummary dir private mcsummary cxx o n data user chraab icerec n trunk src mcsummary private mcsummary cxx in member n function bool selectframe n data user chraab icerec trunk src mcsummary private mcsummary cxx n error class has no member named nmake error nmake error nmake error n n nthis is in the context of icerec trunk with mcsummary added to the working copy i recently updated my working copy to rev i am on cobalt with the ports in cvmfs icecube opensciencegrid org standard rhel n ni just saw that mcsummary doesn t seem to belong to any metaproject sorry if this error is nobody s business feel free to close the ticket in that case reporter chraab cc resolution wontfix ts component combo simulation summary compiling mcsummary has no member named calcdistance priority normal keywords mcsummary build time milestone owner type defect
1
304,499
9,332,874,944
IssuesEvent
2019-03-28 13:17:07
telerik/kendo-ui-core
https://api.github.com/repos/telerik/kendo-ui-core
closed
Not visible indicator for dropping between TreeView nodes
Appearance Bug C: TreeView Kendo2 Priority 2 SEV: Medium Triaged
### Bug report Only 1 pixel dot is visible, instead of a line between nodes on dragging a node. ### Reproduction of the problem 1. Go to http://demos.telerik.com/kendo-ui/treeview/dragdrop 2. Drag a node between 2 nodes ### Current behavior ![image](https://cloud.githubusercontent.com/assets/6614183/23204666/2acfc19a-f8f0-11e6-8698-3390bdc11fab.png) **Default v2** ![image](https://cloud.githubusercontent.com/assets/6614183/23210763/b9b1b014-f907-11e6-9031-c379f365a69e.png) ### Expected/desired behavior **R2 2016 SP1** - [Dojo](https://dojo.telerik.com/iqAferUZ) ![image](https://cloud.githubusercontent.com/assets/6614183/23204721/5bcff5e4-f8f0-11e6-824f-f73be7abd90f.png) ### Environment * **Kendo UI version:** 2019.1.115 * **Browser:** [all]
1.0
Not visible indicator for dropping between TreeView nodes - ### Bug report Only 1 pixel dot is visible, instead of a line between nodes on dragging a node. ### Reproduction of the problem 1. Go to http://demos.telerik.com/kendo-ui/treeview/dragdrop 2. Drag a node between 2 nodes ### Current behavior ![image](https://cloud.githubusercontent.com/assets/6614183/23204666/2acfc19a-f8f0-11e6-8698-3390bdc11fab.png) **Default v2** ![image](https://cloud.githubusercontent.com/assets/6614183/23210763/b9b1b014-f907-11e6-9031-c379f365a69e.png) ### Expected/desired behavior **R2 2016 SP1** - [Dojo](https://dojo.telerik.com/iqAferUZ) ![image](https://cloud.githubusercontent.com/assets/6614183/23204721/5bcff5e4-f8f0-11e6-824f-f73be7abd90f.png) ### Environment * **Kendo UI version:** 2019.1.115 * **Browser:** [all]
non_defect
not visible indicator for dropping between treeview nodes bug report only pixel dot is visible instead of a line between nodes on dragging a node reproduction of the problem go to drag a node between nodes current behavior default expected desired behavior environment kendo ui version browser
0
16,807
2,948,305,667
IssuesEvent
2015-07-06 01:12:07
Winetricks/winetricks
https://api.github.com/repos/Winetricks/winetricks
closed
wine32 on x64 installs 64-bit versions (FreeBSD)
auto-migrated Priority-Medium Type-Defect
``` * System: FreeBSD.9.2_amd64 * Wine: i386-wine-1.7.0,1 32bit Windows compatibility environment for 64bit * Winetricks downloaded and placed in /usr/local/bin/ & working * setenv WINEARCH win32 (shell is csh). Env settings preserved upon switch to bash. * winecfg under "WINEARCH win32" env to setup prisitne initial WINEPREFIX. With the setup above, when trying this: $ winetricks dotnet40 The resulting download process shows: downloading dotNetFx40_Full_x86_x64.exe The wine environment on all FreeBSD systems is 32bit - there is no 64bit alternative on FreeBSD as yet. So any install must use the 32bit version of the app and not the x64 version. I presume winetricks is checking machime uname rather than WINEPREFIX registry settings? Please advise if there is an easy or immediate work-around. WINETRICKS_VERSION=20130707 ``` Original issue reported on code.google.com by `r...@berentweb.com` on 9 Aug 2013 at 8:36
1.0
wine32 on x64 installs 64-bit versions (FreeBSD) - ``` * System: FreeBSD.9.2_amd64 * Wine: i386-wine-1.7.0,1 32bit Windows compatibility environment for 64bit * Winetricks downloaded and placed in /usr/local/bin/ & working * setenv WINEARCH win32 (shell is csh). Env settings preserved upon switch to bash. * winecfg under "WINEARCH win32" env to setup prisitne initial WINEPREFIX. With the setup above, when trying this: $ winetricks dotnet40 The resulting download process shows: downloading dotNetFx40_Full_x86_x64.exe The wine environment on all FreeBSD systems is 32bit - there is no 64bit alternative on FreeBSD as yet. So any install must use the 32bit version of the app and not the x64 version. I presume winetricks is checking machime uname rather than WINEPREFIX registry settings? Please advise if there is an easy or immediate work-around. WINETRICKS_VERSION=20130707 ``` Original issue reported on code.google.com by `r...@berentweb.com` on 9 Aug 2013 at 8:36
defect
on installs bit versions freebsd system freebsd wine wine windows compatibility environment for winetricks downloaded and placed in usr local bin working setenv winearch shell is csh env settings preserved upon switch to bash winecfg under winearch env to setup prisitne initial wineprefix with the setup above when trying this winetricks the resulting download process shows downloading full exe the wine environment on all freebsd systems is there is no alternative on freebsd as yet so any install must use the version of the app and not the version i presume winetricks is checking machime uname rather than wineprefix registry settings please advise if there is an easy or immediate work around winetricks version original issue reported on code google com by r berentweb com on aug at
1
756,714
26,482,724,252
IssuesEvent
2023-01-17 15:45:47
OpenApoc/OpenApoc
https://api.github.com/repos/OpenApoc/OpenApoc
closed
(No Agent object matching ID "AGENT_17") Transfer/Firing of Scientists causes a later crash
Duplicate !BUG! HIGH PRIORITY Verified / Replicated Cityscape Agent ID Error
Hello. Sorry for my english. Using translator. The game crashes with this message: ![2](https://user-images.githubusercontent.com/87769223/137281488-08725442-a86f-4e33-9810-d39e7ab6edda.png) ![3](https://user-images.githubusercontent.com/87769223/137281491-fd64d276-6215-47d7-91a7-0e47a9186a6f.png) ![4](https://user-images.githubusercontent.com/87769223/137281496-5b7d8fbe-9ac8-473b-807c-b6626f00bb74.png) ![1](https://user-images.githubusercontent.com/87769223/137281499-a651e91e-2b0b-413c-93e5-7377cb18d579.png) The game crashes after a while. I cannot understand at what moment. At the end of the work on the timer at about 8:55. I have no idea what event the game crashes. Here is the log and save [log.txt](https://github.com/OpenApoc/OpenApoc/files/7344470/log.txt) [save_Map 1.zip](https://github.com/OpenApoc/OpenApoc/files/7344487/save_Map.1.zip) Thank you very much for supporting this project. I'm waiting for the release. Good luck guys.
1.0
(No Agent object matching ID "AGENT_17") Transfer/Firing of Scientists causes a later crash - Hello. Sorry for my english. Using translator. The game crashes with this message: ![2](https://user-images.githubusercontent.com/87769223/137281488-08725442-a86f-4e33-9810-d39e7ab6edda.png) ![3](https://user-images.githubusercontent.com/87769223/137281491-fd64d276-6215-47d7-91a7-0e47a9186a6f.png) ![4](https://user-images.githubusercontent.com/87769223/137281496-5b7d8fbe-9ac8-473b-807c-b6626f00bb74.png) ![1](https://user-images.githubusercontent.com/87769223/137281499-a651e91e-2b0b-413c-93e5-7377cb18d579.png) The game crashes after a while. I cannot understand at what moment. At the end of the work on the timer at about 8:55. I have no idea what event the game crashes. Here is the log and save [log.txt](https://github.com/OpenApoc/OpenApoc/files/7344470/log.txt) [save_Map 1.zip](https://github.com/OpenApoc/OpenApoc/files/7344487/save_Map.1.zip) Thank you very much for supporting this project. I'm waiting for the release. Good luck guys.
non_defect
no agent object matching id agent transfer firing of scientists causes a later crash hello sorry for my english using translator the game crashes with this message the game crashes after a while i cannot understand at what moment at the end of the work on the timer at about i have no idea what event the game crashes here is the log and save thank you very much for supporting this project i m waiting for the release good luck guys
0
327,288
9,973,509,605
IssuesEvent
2019-07-09 08:31:18
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Spooky Tooltips appear above chat window
Medium Priority
(v0.8.1.4) If hovering with mouse in the marked area above chat window, spooky tooltips appear not belonging to any visible object, but only if chat window is active. ![spooky](https://user-images.githubusercontent.com/38510829/57238205-20562f80-7029-11e9-9951-8e4cbe6a23e1.jpg) ![spooky2](https://user-images.githubusercontent.com/38510829/57238214-28ae6a80-7029-11e9-9eff-0f96b39ff124.jpg)
1.0
Spooky Tooltips appear above chat window - (v0.8.1.4) If hovering with mouse in the marked area above chat window, spooky tooltips appear not belonging to any visible object, but only if chat window is active. ![spooky](https://user-images.githubusercontent.com/38510829/57238205-20562f80-7029-11e9-9951-8e4cbe6a23e1.jpg) ![spooky2](https://user-images.githubusercontent.com/38510829/57238214-28ae6a80-7029-11e9-9eff-0f96b39ff124.jpg)
non_defect
spooky tooltips appear above chat window if hovering with mouse in the marked area above chat window spooky tooltips appear not belonging to any visible object but only if chat window is active
0
33,491
7,132,144,758
IssuesEvent
2018-01-22 13:41:32
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
The UTC time format still does not work in the calendar.
defect
Now the calendar displays only local time. The properties [utc]="true" does not work. This can be seen from the example. Please click on the link and choose any date. https://stackblitz.com/edit/prime-ng-calendar?embed=1&file=app/calendar/calendar.component.html * **Angular version:** 5.0.0 * **PrimeNG version:** 4.3.0 ![utc](https://user-images.githubusercontent.com/11576087/32436959-6d16c3e6-c2ee-11e7-9d26-23de79146eaf.jpg)
1.0
The UTC time format still does not work in the calendar. - Now the calendar displays only local time. The properties [utc]="true" does not work. This can be seen from the example. Please click on the link and choose any date. https://stackblitz.com/edit/prime-ng-calendar?embed=1&file=app/calendar/calendar.component.html * **Angular version:** 5.0.0 * **PrimeNG version:** 4.3.0 ![utc](https://user-images.githubusercontent.com/11576087/32436959-6d16c3e6-c2ee-11e7-9d26-23de79146eaf.jpg)
defect
the utc time format still does not work in the calendar now the calendar displays only local time the properties true does not work this can be seen from the example please click on the link and choose any date angular version primeng version
1
327,732
24,150,350,116
IssuesEvent
2022-09-21 23:33:41
horenbergerb/FriendSimulator
https://api.github.com/repos/horenbergerb/FriendSimulator
closed
Implement/Document checkpointing
documentation enhancement help wanted
I think the current design does some kind of checkpointing or progress capture during training, but I'm not sure how it works. Adding this to the README would be very useful.
1.0
Implement/Document checkpointing - I think the current design does some kind of checkpointing or progress capture during training, but I'm not sure how it works. Adding this to the README would be very useful.
non_defect
implement document checkpointing i think the current design does some kind of checkpointing or progress capture during training but i m not sure how it works adding this to the readme would be very useful
0
44,505
12,217,258,026
IssuesEvent
2020-05-01 16:49:06
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Stories at a VA Pittsburgh are missing the /pittsburgh-health-care/ part of URL
Critical defect Defect Drupal engineering
**Describe the defect** URL pattern for stories is missing the field_office (via field_story_listing) **To Reproduce** Steps to reproduce the behavior: 1. Create a story and assign to Pittsburgh 2. URL will be stories/[title] **Expected behavior** URL should be /pittsburgh-health-care/stories
2.0
Stories at a VA Pittsburgh are missing the /pittsburgh-health-care/ part of URL - **Describe the defect** URL pattern for stories is missing the field_office (via field_story_listing) **To Reproduce** Steps to reproduce the behavior: 1. Create a story and assign to Pittsburgh 2. URL will be stories/[title] **Expected behavior** URL should be /pittsburgh-health-care/stories
defect
stories at a va pittsburgh are missing the pittsburgh health care part of url describe the defect url pattern for stories is missing the field office via field story listing to reproduce steps to reproduce the behavior create a story and assign to pittsburgh url will be stories expected behavior url should be pittsburgh health care stories
1
75,493
25,876,620,065
IssuesEvent
2022-12-14 08:23:22
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
opened
`NullPointerException` in `FindingsQueryManager` caused by PR 2272
defect in triage
### Current Behavior When the Frontend tries to retrieve the components for the `Audit Vulnerabilities` and the `Exploit Predictions` tab and a component does not have a `RepositoryMetaComponent`, a `NullPointerException` will be thrown, resulting in an internal server error and no components being displayed in the tabs. ### Steps to Reproduce Add the following test to `FindingResourceText` and execute it, it will result in an internal server error and also throw a `NullPointerException`: ``` @Test public void getFindingsByProjectWithComponentLatestVersionWithoutRepositoryMetaComponent() { Project p1 = qm.createProject("Acme Example", null, "1.0", null, null, null, true, false); Component c1 = createComponent(p1, "Component A", "1.0"); c1.setPurl("pkg:/maven/org.acme/component-a@1.0.0"); /*RepositoryMetaComponent r1 = new RepositoryMetaComponent(); Date d1 = new Date(); r1.setLastCheck(d1); r1.setNamespace("org.acme"); r1.setName("component-a"); r1.setLatestVersion("2.0.0"); r1.setRepositoryType(RepositoryType.MAVEN); qm.persist(r1);*/ Vulnerability v1 = createVulnerability("Vuln-1", Severity.CRITICAL); qm.addVulnerability(v1, c1, AnalyzerIdentity.NONE); Response response = target(V1_FINDING + "/project/" + p1.getUuid().toString()).request() .header(X_API_KEY, apiKey) .get(Response.class); Assert.assertEquals(200, response.getStatus(), 0); } ``` ### Expected Behavior The `NullPointerException` should be prevented by checking if a component does have a `RepositoryMetaComponent`, before adding its latest version to a `Finding` object. ### Dependency-Track Version 4.7.0-SNAPSHOT ### Dependency-Track Distribution Container Image ### Database Server H2 ### Database Server Version _No response_ ### Browser N/A ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
1.0
`NullPointerException` in `FindingsQueryManager` caused by PR 2272 - ### Current Behavior When the Frontend tries to retrieve the components for the `Audit Vulnerabilities` and the `Exploit Predictions` tab and a component does not have a `RepositoryMetaComponent`, a `NullPointerException` will be thrown, resulting in an internal server error and no components being displayed in the tabs. ### Steps to Reproduce Add the following test to `FindingResourceText` and execute it, it will result in an internal server error and also throw a `NullPointerException`: ``` @Test public void getFindingsByProjectWithComponentLatestVersionWithoutRepositoryMetaComponent() { Project p1 = qm.createProject("Acme Example", null, "1.0", null, null, null, true, false); Component c1 = createComponent(p1, "Component A", "1.0"); c1.setPurl("pkg:/maven/org.acme/component-a@1.0.0"); /*RepositoryMetaComponent r1 = new RepositoryMetaComponent(); Date d1 = new Date(); r1.setLastCheck(d1); r1.setNamespace("org.acme"); r1.setName("component-a"); r1.setLatestVersion("2.0.0"); r1.setRepositoryType(RepositoryType.MAVEN); qm.persist(r1);*/ Vulnerability v1 = createVulnerability("Vuln-1", Severity.CRITICAL); qm.addVulnerability(v1, c1, AnalyzerIdentity.NONE); Response response = target(V1_FINDING + "/project/" + p1.getUuid().toString()).request() .header(X_API_KEY, apiKey) .get(Response.class); Assert.assertEquals(200, response.getStatus(), 0); } ``` ### Expected Behavior The `NullPointerException` should be prevented by checking if a component does have a `RepositoryMetaComponent`, before adding its latest version to a `Finding` object. ### Dependency-Track Version 4.7.0-SNAPSHOT ### Dependency-Track Distribution Container Image ### Database Server H2 ### Database Server Version _No response_ ### Browser N/A ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
defect
nullpointerexception in findingsquerymanager caused by pr current behavior when the frontend tries to retrieve the components for the audit vulnerabilities and the exploit predictions tab and a component does not have a repositorymetacomponent a nullpointerexception will be thrown resulting in an internal server error and no components being displayed in the tabs steps to reproduce add the following test to findingresourcetext and execute it it will result in an internal server error and also throw a nullpointerexception test public void getfindingsbyprojectwithcomponentlatestversionwithoutrepositorymetacomponent project qm createproject acme example null null null null true false component createcomponent component a setpurl pkg maven org acme component a repositorymetacomponent new repositorymetacomponent date new date setlastcheck setnamespace org acme setname component a setlatestversion setrepositorytype repositorytype maven qm persist vulnerability createvulnerability vuln severity critical qm addvulnerability analyzeridentity none response response target finding project getuuid tostring request header x api key apikey get response class assert assertequals response getstatus expected behavior the nullpointerexception should be prevented by checking if a component does have a repositorymetacomponent before adding its latest version to a finding object dependency track version snapshot dependency track distribution container image database server database server version no response browser n a checklist i have read and understand the i have checked the for whether this defect was already reported
1
34,621
7,458,067,971
IssuesEvent
2018-03-30 08:30:21
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Isikute PURI genereerimine käsurealt väga aeglane
C: AIS P: highest R: fixed T: defect
**Reported by simo karpin on 1 Oct 2014 07:48 UTC** Isikute PURI'de genereerimine võtab liialt aega. 120509 isiku PURI'd võtsid dev'is aega: 14:22:08sec ** TODO ** Vähendada antud PURI'de genereerimisaega.
1.0
Isikute PURI genereerimine käsurealt väga aeglane - **Reported by simo karpin on 1 Oct 2014 07:48 UTC** Isikute PURI'de genereerimine võtab liialt aega. 120509 isiku PURI'd võtsid dev'is aega: 14:22:08sec ** TODO ** Vähendada antud PURI'de genereerimisaega.
defect
isikute puri genereerimine käsurealt väga aeglane reported by simo karpin on oct utc isikute puri de genereerimine võtab liialt aega isiku puri d võtsid dev is aega todo vähendada antud puri de genereerimisaega
1
500,458
14,500,033,537
IssuesEvent
2020-12-11 17:29:12
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
closed
bigquery: add Table ACL
api: bigquery priority: p3 type: feature request
Now that BigQuery tables support ACLs, it would be nice to be able to see and manipulate access with the SDK. I don't imagine it being overly complex to implement, as I would assume `TableMetadata` could reuse the same Access structs from `DatasetMetadata` ``` Access []*AccessEntry // Access permissions. ```
1.0
bigquery: add Table ACL - Now that BigQuery tables support ACLs, it would be nice to be able to see and manipulate access with the SDK. I don't imagine it being overly complex to implement, as I would assume `TableMetadata` could reuse the same Access structs from `DatasetMetadata` ``` Access []*AccessEntry // Access permissions. ```
non_defect
bigquery add table acl now that bigquery tables support acls it would be nice to be able to see and manipulate access with the sdk i don t imagine it being overly complex to implement as i would assume tablemetadata could reuse the same access structs from datasetmetadata access accessentry access permissions
0
266,386
8,366,804,403
IssuesEvent
2018-10-04 10:11:52
architecture-building-systems/CityEnergyAnalyst
https://api.github.com/repos/architecture-building-systems/CityEnergyAnalyst
closed
Implement dashboard template
Interface Priority 1
We're going for a look similar to the [Gentellela Alela!](https://colorlib.com/polygon/gentelella/index.html) Bootstrap 3 template. In fact, the gentellella has a flask version: https://github.com/afourmy/flask-gentelella which seems like the obvious starting point for our dashboard. I checked the license: It is MIT, so it is 100% compatible with the CEA. I suggest starting with this issue, as it will get us up and running and actually _looking_ at the future app. Fill in the features as we go along.
1.0
Implement dashboard template - We're going for a look similar to the [Gentellela Alela!](https://colorlib.com/polygon/gentelella/index.html) Bootstrap 3 template. In fact, the gentellella has a flask version: https://github.com/afourmy/flask-gentelella which seems like the obvious starting point for our dashboard. I checked the license: It is MIT, so it is 100% compatible with the CEA. I suggest starting with this issue, as it will get us up and running and actually _looking_ at the future app. Fill in the features as we go along.
non_defect
implement dashboard template we re going for a look similar to the bootstrap template in fact the gentellella has a flask version which seems like the obvious starting point for our dashboard i checked the license it is mit so it is compatible with the cea i suggest starting with this issue as it will get us up and running and actually looking at the future app fill in the features as we go along
0
45,049
12,529,728,580
IssuesEvent
2020-06-04 11:52:56
appknox/google-chartwrapper
https://api.github.com/repos/appknox/google-chartwrapper
closed
No HTTPS support
Priority-Medium Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. Add chart to Django template as an image 2. Open page via HTTPS 3. Check the source URL of generated image What is the expected output? I expect generated image to refer to HTTPS image so whole page is secure. What do you see instead? Generated image uses http://chart.apis.google.com/ as prefix. E.g. generated URL do not respect current protocol. Please provide any additional information below. Please make it work with HTTPS as well. I see 3 ways here: 1) Add a way to override APIURL in django settings (i.e. use django settings first). 2) Add a "secure" flag to chart template tag so program can pass if he needed secure URL or not 3) Change chart template tag so it automatically change it to HTTPS if request was made over HTTPS. ``` Original issue reported on code.google.com by `ton...@gmail.com` on 8 Oct 2012 at 9:32
1.0
No HTTPS support - ``` What steps will reproduce the problem? 1. Add chart to Django template as an image 2. Open page via HTTPS 3. Check the source URL of generated image What is the expected output? I expect generated image to refer to HTTPS image so whole page is secure. What do you see instead? Generated image uses http://chart.apis.google.com/ as prefix. E.g. generated URL do not respect current protocol. Please provide any additional information below. Please make it work with HTTPS as well. I see 3 ways here: 1) Add a way to override APIURL in django settings (i.e. use django settings first). 2) Add a "secure" flag to chart template tag so program can pass if he needed secure URL or not 3) Change chart template tag so it automatically change it to HTTPS if request was made over HTTPS. ``` Original issue reported on code.google.com by `ton...@gmail.com` on 8 Oct 2012 at 9:32
defect
no https support what steps will reproduce the problem add chart to django template as an image open page via https check the source url of generated image what is the expected output i expect generated image to refer to https image so whole page is secure what do you see instead generated image uses as prefix e g generated url do not respect current protocol please provide any additional information below please make it work with https as well i see ways here add a way to override apiurl in django settings i e use django settings first add a secure flag to chart template tag so program can pass if he needed secure url or not change chart template tag so it automatically change it to https if request was made over https original issue reported on code google com by ton gmail com on oct at
1
51,629
13,207,541,366
IssuesEvent
2020-08-14 23:30:53
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
gulliver chooses whatever Minuit it feels like (Trac #744)
Incomplete Migration Migrated from Trac combo reconstruction defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/744">https://code.icecube.wisc.edu/projects/icecube/ticket/744</a>, reported by negaand owned by boersma</em></summary> <p> ```json { "status": "closed", "changetime": "2014-10-08T09:27:57", "_ts": "1412760477377993", "description": "This causes issues like:\n\n{{{\nLinking CXX shared library ../lib/liblilliput.dylib\nUndefined symbols for architecture x86_64:\n \"ROOT::Minuit2::MinimumBuilder::MinimumBuilder()\", referenced from:\n ROOT::Minuit2::SimplexBuilder::SimplexBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::FumiliBuilder::FumiliBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::CombinedMinimumBuilder::CombinedMinimumBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::VariableMetricBuilder::VariableMetricBuilder() in I3GulliverMinuit2.cxx.o\nld: symbol(s) not found for architecture x86_64\nclang: error: linker command failed with exit code 1 (use -v to see invocation)\nmake[2]: *** [lib/liblilliput.dylib] Error 1\nmake[1]: *** [lilliput/CMakeFiles/lilliput.dir/all] Error 2\nmake: *** [all] Error 2\nprogram finished with exit code 2\n}}}\n\nGulliver needs a way to pick a Minuit that is consistent with Lilliput", "reporter": "nega", "cc": "dataclass@icecube.wisc.edu", "resolution": "fixed", "time": "2014-09-05T20:55:44", "component": "combo reconstruction", "summary": "gulliver chooses whatever Minuit it feels like", "priority": "normal", "keywords": "lilliput gulliver minuit", "milestone": "", "owner": "boersma", "type": "defect" } ``` </p> </details>
1.0
gulliver chooses whatever Minuit it feels like (Trac #744) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/744">https://code.icecube.wisc.edu/projects/icecube/ticket/744</a>, reported by negaand owned by boersma</em></summary> <p> ```json { "status": "closed", "changetime": "2014-10-08T09:27:57", "_ts": "1412760477377993", "description": "This causes issues like:\n\n{{{\nLinking CXX shared library ../lib/liblilliput.dylib\nUndefined symbols for architecture x86_64:\n \"ROOT::Minuit2::MinimumBuilder::MinimumBuilder()\", referenced from:\n ROOT::Minuit2::SimplexBuilder::SimplexBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::FumiliBuilder::FumiliBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::CombinedMinimumBuilder::CombinedMinimumBuilder() in I3GulliverMinuit2.cxx.o\n ROOT::Minuit2::VariableMetricBuilder::VariableMetricBuilder() in I3GulliverMinuit2.cxx.o\nld: symbol(s) not found for architecture x86_64\nclang: error: linker command failed with exit code 1 (use -v to see invocation)\nmake[2]: *** [lib/liblilliput.dylib] Error 1\nmake[1]: *** [lilliput/CMakeFiles/lilliput.dir/all] Error 2\nmake: *** [all] Error 2\nprogram finished with exit code 2\n}}}\n\nGulliver needs a way to pick a Minuit that is consistent with Lilliput", "reporter": "nega", "cc": "dataclass@icecube.wisc.edu", "resolution": "fixed", "time": "2014-09-05T20:55:44", "component": "combo reconstruction", "summary": "gulliver chooses whatever Minuit it feels like", "priority": "normal", "keywords": "lilliput gulliver minuit", "milestone": "", "owner": "boersma", "type": "defect" } ``` </p> </details>
defect
gulliver chooses whatever minuit it feels like trac migrated from json status closed changetime ts description this causes issues like n n nlinking cxx shared library lib liblilliput dylib nundefined symbols for architecture n root minimumbuilder minimumbuilder referenced from n root simplexbuilder simplexbuilder in cxx o n root fumilibuilder fumilibuilder in cxx o n root combinedminimumbuilder combinedminimumbuilder in cxx o n root variablemetricbuilder variablemetricbuilder in cxx o nld symbol s not found for architecture nclang error linker command failed with exit code use v to see invocation nmake error nmake error nmake error nprogram finished with exit code n n ngulliver needs a way to pick a minuit that is consistent with lilliput reporter nega cc dataclass icecube wisc edu resolution fixed time component combo reconstruction summary gulliver chooses whatever minuit it feels like priority normal keywords lilliput gulliver minuit milestone owner boersma type defect
1
83,830
24,152,701,225
IssuesEvent
2022-09-22 03:35:53
habitat-sh/builder
https://api.github.com/repos/habitat-sh/builder
closed
Job group status should reflect completeness more accurately
Focus:Builder Type:Bug Type:DevX Stale
Today, job groups are given a status of `Complete` even when some or all of their member jobs do not complete successfully. Since `Complete` signifies successful completion in member jobs, we should not grant that status to job groups that do not complete successfully. Instead, we should: * Mark the group with the status of its member jobs when all of those jobs complete in the same way (e.g., when everything `Complete`s, the group's status is also `Complete`; when everything is `Failed`, the group is also `Failed`, etc., or * Mark the group `Incomplete` or with some other status indicating that the group did not complete successfully.
1.0
Job group status should reflect completeness more accurately - Today, job groups are given a status of `Complete` even when some or all of their member jobs do not complete successfully. Since `Complete` signifies successful completion in member jobs, we should not grant that status to job groups that do not complete successfully. Instead, we should: * Mark the group with the status of its member jobs when all of those jobs complete in the same way (e.g., when everything `Complete`s, the group's status is also `Complete`; when everything is `Failed`, the group is also `Failed`, etc., or * Mark the group `Incomplete` or with some other status indicating that the group did not complete successfully.
non_defect
job group status should reflect completeness more accurately today job groups are given a status of complete even when some or all of their member jobs do not complete successfully since complete signifies successful completion in member jobs we should not grant that status to job groups that do not complete successfully instead we should mark the group with the status of its member jobs when all of those jobs complete in the same way e g when everything complete s the group s status is also complete when everything is failed the group is also failed etc or mark the group incomplete or with some other status indicating that the group did not complete successfully
0
26,184
4,600,602,001
IssuesEvent
2016-09-22 06:16:07
vim/vim
https://api.github.com/repos/vim/vim
closed
File->Open does not update window content correctly - Appeared first in v7-3-638
auto-migrated patch Priority-Medium Type-Defect
``` Problem is not present in tag v7-3-637. Problem appeared first in tag v7-3-638. What steps will reproduce the problem? Using gvim: 1. Go to v7-3-638 vim source directory 2. File->Open and select Filelist 3. Content shows correctly 4. File->Open and select README.txt 5. Contents are not updated, section of Filelist still shows 6. Move mouse wheel or cursor, contents are now updated correctly What is the expected output? What do you see instead? Exepcted output is content of newly opened file. I see part of old file until screen refreshes by moving cursor. What version of the product are you using? On what operating system? Linux Mint 13 with MATE 1.2 or 1.4 gvim compiled from source with: ./configure --enable-gui=gtk2 --enable-rubyinterp=yes --with-features=huge Kernel: 3.2.0-32-generic #51-Ubuntu SMP Please provide any additional information below. Problem is not present in tag v7-3-637. Problem appeared first in tag v7-3-638. ``` Original issue reported on code.google.com by `yerv...@gmail.com` on 3 Nov 2012 at 1:33
1.0
File->Open does not update window content correctly - Appeared first in v7-3-638 - ``` Problem is not present in tag v7-3-637. Problem appeared first in tag v7-3-638. What steps will reproduce the problem? Using gvim: 1. Go to v7-3-638 vim source directory 2. File->Open and select Filelist 3. Content shows correctly 4. File->Open and select README.txt 5. Contents are not updated, section of Filelist still shows 6. Move mouse wheel or cursor, contents are now updated correctly What is the expected output? What do you see instead? Exepcted output is content of newly opened file. I see part of old file until screen refreshes by moving cursor. What version of the product are you using? On what operating system? Linux Mint 13 with MATE 1.2 or 1.4 gvim compiled from source with: ./configure --enable-gui=gtk2 --enable-rubyinterp=yes --with-features=huge Kernel: 3.2.0-32-generic #51-Ubuntu SMP Please provide any additional information below. Problem is not present in tag v7-3-637. Problem appeared first in tag v7-3-638. ``` Original issue reported on code.google.com by `yerv...@gmail.com` on 3 Nov 2012 at 1:33
defect
file open does not update window content correctly appeared first in problem is not present in tag problem appeared first in tag what steps will reproduce the problem using gvim go to vim source directory file open and select filelist content shows correctly file open and select readme txt contents are not updated section of filelist still shows move mouse wheel or cursor contents are now updated correctly what is the expected output what do you see instead exepcted output is content of newly opened file i see part of old file until screen refreshes by moving cursor what version of the product are you using on what operating system linux mint with mate or gvim compiled from source with configure enable gui enable rubyinterp yes with features huge kernel generic ubuntu smp please provide any additional information below problem is not present in tag problem appeared first in tag original issue reported on code google com by yerv gmail com on nov at
1
585,716
17,516,069,319
IssuesEvent
2021-08-11 06:42:56
ballerina-platform/ballerina-standard-library
https://api.github.com/repos/ballerina-platform/ballerina-standard-library
closed
Make client http.url metric in observability a configurable property
Points/2 Priority/High Type/Improvement module/http Team/PCP
**Description:** $subject in 1.2.x branch
1.0
Make client http.url metric in observability a configurable property - **Description:** $subject in 1.2.x branch
non_defect
make client http url metric in observability a configurable property description subject in x branch
0
81,069
7,765,384,208
IssuesEvent
2018-06-02 02:41:04
pods-framework/pods
https://api.github.com/repos/pods-framework/pods
closed
PODS breaks Theme Editor
Bug Fixed / Needs Testing in progress
## Issue Overview The "PODS" plugin interferes with the WP Theme Editor's ability to save PHP files ## Expected Behavior To be able to edit the Theme without an error. If you Disable the plugin, then theme edits can be saved without an error. ## Current Behavior When the "PODS" plugin is enabled, the following error message displays when you try to save a PHP file in the WP Theme Editor ( /wp-admin/theme-editor.php ): "Unable to communicate back with site to check for fatal errors, so the PHP change was reverted. You will need to upload your PHP file change by some other means, such as by using SFTP." ## Possible Solution UPDATE: I got some informative feedback, and an Plugin update with the fix, from Bob at Watu Pro… https://wpengine.com/support/php-editor-errors-wordpress-4-9-higher/ Hope this helps. KP ## Steps to Reproduce (for bugs) 1. Install default WP with "Twenty Seventeen" Theme 2. Install "PODS" plugin 3. go to /wp-admin/theme-editor.php and try to save any PHPOOP file 4. Observe the error message and inability to save the file ## Pods Package Export (helpful!) <!-- Please include the Pods Package Export as it will be helpful in reproducing your issues. --> <!-- To access: Pods Admin, Components, Migrate: Packages, Enable. --> <!-- Then Pods Admin, Migrate: Packages, Export and Copy and Paste the JSON here OR --> <!-- Save to a text file and upload to this bug report --> ``` Copy and Paste the JSON Export from **Pods Admin, Migrate: Packages, Export** in WordPress admin here ``` ## Related Issues and/or PRs <!-- List related issues or PRs against other branches: --> ## Todos - [ ] Tests - [ ] Documentation ## WordPress Environment v 4.9.5 ``` Copy and paste your System Details from **Pods Admin > Settings > Debug Information** in WordPress admin here. ``` </details>
1.0
PODS breaks Theme Editor - ## Issue Overview The "PODS" plugin interferes with the WP Theme Editor's ability to save PHP files ## Expected Behavior To be able to edit the Theme without an error. If you Disable the plugin, then theme edits can be saved without an error. ## Current Behavior When the "PODS" plugin is enabled, the following error message displays when you try to save a PHP file in the WP Theme Editor ( /wp-admin/theme-editor.php ): "Unable to communicate back with site to check for fatal errors, so the PHP change was reverted. You will need to upload your PHP file change by some other means, such as by using SFTP." ## Possible Solution UPDATE: I got some informative feedback, and an Plugin update with the fix, from Bob at Watu Pro… https://wpengine.com/support/php-editor-errors-wordpress-4-9-higher/ Hope this helps. KP ## Steps to Reproduce (for bugs) 1. Install default WP with "Twenty Seventeen" Theme 2. Install "PODS" plugin 3. go to /wp-admin/theme-editor.php and try to save any PHPOOP file 4. Observe the error message and inability to save the file ## Pods Package Export (helpful!) <!-- Please include the Pods Package Export as it will be helpful in reproducing your issues. --> <!-- To access: Pods Admin, Components, Migrate: Packages, Enable. --> <!-- Then Pods Admin, Migrate: Packages, Export and Copy and Paste the JSON here OR --> <!-- Save to a text file and upload to this bug report --> ``` Copy and Paste the JSON Export from **Pods Admin, Migrate: Packages, Export** in WordPress admin here ``` ## Related Issues and/or PRs <!-- List related issues or PRs against other branches: --> ## Todos - [ ] Tests - [ ] Documentation ## WordPress Environment v 4.9.5 ``` Copy and paste your System Details from **Pods Admin > Settings > Debug Information** in WordPress admin here. ``` </details>
non_defect
pods breaks theme editor issue overview the pods plugin interferes with the wp theme editor s ability to save php files expected behavior to be able to edit the theme without an error if you disable the plugin then theme edits can be saved without an error current behavior when the pods plugin is enabled the following error message displays when you try to save a php file in the wp theme editor wp admin theme editor php unable to communicate back with site to check for fatal errors so the php change was reverted you will need to upload your php file change by some other means such as by using sftp possible solution update i got some informative feedback and an plugin update with the fix from bob at watu pro… hope this helps kp steps to reproduce for bugs install default wp with twenty seventeen theme install pods plugin go to wp admin theme editor php and try to save any phpoop file observe the error message and inability to save the file pods package export helpful copy and paste the json export from pods admin migrate packages export in wordpress admin here related issues and or prs todos tests documentation wordpress environment v copy and paste your system details from pods admin settings debug information in wordpress admin here
0
782,908
27,511,107,112
IssuesEvent
2023-03-06 08:53:43
pdx-blurp/blurp-frontend
https://api.github.com/repos/pdx-blurp/blurp-frontend
closed
Remove modal pop-up from node creation
high priority enhancement
Currently the user has to fill out a modal form with node information when creating a node - this is a slow process. AC: When a user creates a node using the node tool, the node should just be placed instead of the modal popping up All node data should be changeable from the data sidebar This also requires that creating a new node does not require any data from the user
1.0
Remove modal pop-up from node creation - Currently the user has to fill out a modal form with node information when creating a node - this is a slow process. AC: When a user creates a node using the node tool, the node should just be placed instead of the modal popping up All node data should be changeable from the data sidebar This also requires that creating a new node does not require any data from the user
non_defect
remove modal pop up from node creation currently the user has to fill out a modal form with node information when creating a node this is a slow process ac when a user creates a node using the node tool the node should just be placed instead of the modal popping up all node data should be changeable from the data sidebar this also requires that creating a new node does not require any data from the user
0
22,749
3,691,293,291
IssuesEvent
2016-02-25 23:20:08
excilys/androidannotations
https://api.github.com/repos/excilys/androidannotations
closed
Variable length parameter not applied to super call
Defect
I've just stumbled upon this rather important bug: When using methods with variable argument lengths, the argument is stripped of the super call in the generated code. I am using the 4.0 Snapshot version. Annotated code: @SupposeBackground public void foo(Object... bar) { Log.d("foo", String.valueOf(bar.length)); } Generated code: @Override public void foo(Object... bar) { BackgroundExecutor.checkBgThread(); AdConfig_.super.foo(); } Expected generated code: @Override public void foo(Object... bar) { BackgroundExecutor.checkBgThread(); AdConfig_.super.foo(**bar**); }
1.0
Variable length parameter not applied to super call - I've just stumbled upon this rather important bug: When using methods with variable argument lengths, the argument is stripped of the super call in the generated code. I am using the 4.0 Snapshot version. Annotated code: @SupposeBackground public void foo(Object... bar) { Log.d("foo", String.valueOf(bar.length)); } Generated code: @Override public void foo(Object... bar) { BackgroundExecutor.checkBgThread(); AdConfig_.super.foo(); } Expected generated code: @Override public void foo(Object... bar) { BackgroundExecutor.checkBgThread(); AdConfig_.super.foo(**bar**); }
defect
variable length parameter not applied to super call i ve just stumbled upon this rather important bug when using methods with variable argument lengths the argument is stripped of the super call in the generated code i am using the snapshot version annotated code supposebackground public void foo object bar log d foo string valueof bar length generated code override public void foo object bar backgroundexecutor checkbgthread adconfig super foo expected generated code override public void foo object bar backgroundexecutor checkbgthread adconfig super foo bar
1
31,930
26,254,566,810
IssuesEvent
2023-01-05 22:50:20
Tonomy-Foundation/Tonomy-ID
https://api.github.com/repos/Tonomy-Foundation/Tonomy-ID
closed
Tonomy ID lint on pull request
infrastructure
Definition of done - [ ] On all pull request, `npm run lint` Is run and failed any errors are detected
1.0
Tonomy ID lint on pull request - Definition of done - [ ] On all pull request, `npm run lint` Is run and failed any errors are detected
non_defect
tonomy id lint on pull request definition of done on all pull request npm run lint is run and failed any errors are detected
0
54,030
13,327,595,795
IssuesEvent
2020-08-27 13:24:08
GameLabGraz/Maroon
https://api.github.com/repos/GameLabGraz/Maroon
opened
WebGL performance
defect enhancement
Use the Unity profiler to inspect the performance of the WebGL build to find the root cause of the mediocre performance and see if something can be improved or if it is a JavaScript limitation that cannot be circumvented. See https://docs.unity3d.com/2018.4/Documentation/Manual/webgl-performance.html and https://docs.unity3d.com/2018.4/Documentation/Manual/Profiler.html
1.0
WebGL performance - Use the Unity profiler to inspect the performance of the WebGL build to find the root cause of the mediocre performance and see if something can be improved or if it is a JavaScript limitation that cannot be circumvented. See https://docs.unity3d.com/2018.4/Documentation/Manual/webgl-performance.html and https://docs.unity3d.com/2018.4/Documentation/Manual/Profiler.html
defect
webgl performance use the unity profiler to inspect the performance of the webgl build to find the root cause of the mediocre performance and see if something can be improved or if it is a javascript limitation that cannot be circumvented see and
1
218,623
16,765,827,918
IssuesEvent
2021-06-14 08:41:30
MarvinMichel/prapla
https://api.github.com/repos/MarvinMichel/prapla
opened
Explain project setup with code previews and examples
documentation
Describe the setup of this project to make it transferable to a new team.
1.0
Explain project setup with code previews and examples - Describe the setup of this project to make it transferable to a new team.
non_defect
explain project setup with code previews and examples describe the setup of this project to make it transferable to a new team
0
38,865
8,996,906,069
IssuesEvent
2019-02-02 06:10:03
netty/netty
https://api.github.com/repos/netty/netty
closed
Memory footprint increase after upgrading to 4.1.32.Final
defect improvement
### Expected behavior Memory footprint is more or less consistent. ### Actual behavior Substantial memory footprint increase after updating from 4.1.30.Final to 4.1.32.Final. ### Steps to reproduce Working on it.. ### Minimal yet complete reproducer code (or URL to code) Working on it.. ### Netty version 4.1.32.Final and 4.1.33.Final show a large memory footprint increase compared to 4.1.30.Final. ### JVM version (e.g. `java -version`) java version "1.8.0_172" Java(TM) SE Runtime Environment (build 1.8.0_172-b11) Java HotSpot(TM) 64-Bit Server VM (build 25.172-b11, mixed mode) ### OS version (e.g. `uname -a`) Linux 4.14.77-70.82.amzn1.x86_64 #1 SMP Mon Dec 3 20:01:27 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
1.0
Memory footprint increase after upgrading to 4.1.32.Final - ### Expected behavior Memory footprint is more or less consistent. ### Actual behavior Substantial memory footprint increase after updating from 4.1.30.Final to 4.1.32.Final. ### Steps to reproduce Working on it.. ### Minimal yet complete reproducer code (or URL to code) Working on it.. ### Netty version 4.1.32.Final and 4.1.33.Final show a large memory footprint increase compared to 4.1.30.Final. ### JVM version (e.g. `java -version`) java version "1.8.0_172" Java(TM) SE Runtime Environment (build 1.8.0_172-b11) Java HotSpot(TM) 64-Bit Server VM (build 25.172-b11, mixed mode) ### OS version (e.g. `uname -a`) Linux 4.14.77-70.82.amzn1.x86_64 #1 SMP Mon Dec 3 20:01:27 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
defect
memory footprint increase after upgrading to final expected behavior memory footprint is more or less consistent actual behavior substantial memory footprint increase after updating from final to final steps to reproduce working on it minimal yet complete reproducer code or url to code working on it netty version final and final show a large memory footprint increase compared to final jvm version e g java version java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode os version e g uname a linux smp mon dec utc gnu linux
1
18,636
3,393,104,452
IssuesEvent
2015-11-30 22:32:05
stamen/caliparks.org
https://api.github.com/repos/stamen/caliparks.org
opened
map marker interaction
Design Fit + Finish
Clicking on a map marker should select that park instead of linking out to the park page. ![img_4148](https://cloud.githubusercontent.com/assets/12550451/11486368/9ce47e82-976d-11e5-84d1-949b30783cfe.JPG) Only clicking on the name in the callout should link to the park: ![screen shot 2015-11-30 at 2 23 54 pm](https://cloud.githubusercontent.com/assets/12550451/11486439/06e6c7cc-976e-11e5-8582-3205e6740e28.png) And an arrow should be added after the park name to clarify that it links out: ![screen shot 2015-11-30 at 2 28 34 pm](https://cloud.githubusercontent.com/assets/12550451/11486560/c1c98ea8-976e-11e5-862f-b0cfd27d467c.png) or ![screen shot 2015-11-30 at 2 30 29 pm](https://cloud.githubusercontent.com/assets/12550451/11486592/f06feac2-976e-11e5-8ab6-f2445b379683.png)
1.0
map marker interaction - Clicking on a map marker should select that park instead of linking out to the park page. ![img_4148](https://cloud.githubusercontent.com/assets/12550451/11486368/9ce47e82-976d-11e5-84d1-949b30783cfe.JPG) Only clicking on the name in the callout should link to the park: ![screen shot 2015-11-30 at 2 23 54 pm](https://cloud.githubusercontent.com/assets/12550451/11486439/06e6c7cc-976e-11e5-8582-3205e6740e28.png) And an arrow should be added after the park name to clarify that it links out: ![screen shot 2015-11-30 at 2 28 34 pm](https://cloud.githubusercontent.com/assets/12550451/11486560/c1c98ea8-976e-11e5-862f-b0cfd27d467c.png) or ![screen shot 2015-11-30 at 2 30 29 pm](https://cloud.githubusercontent.com/assets/12550451/11486592/f06feac2-976e-11e5-8ab6-f2445b379683.png)
non_defect
map marker interaction clicking on a map marker should select that park instead of linking out to the park page only clicking on the name in the callout should link to the park and an arrow should be added after the park name to clarify that it links out or
0
37,311
8,353,713,657
IssuesEvent
2018-10-02 10:59:07
Guake/guake
https://api.github.com/repos/Guake/guake
closed
Prefs window jumps on resize
Priority: High Type: Defect
Resize the prefs window. On gnome-shell + X11 it starts moving upwards. On unity7 + X11 it crazily jumps between two different positions. (I couldn't test with Wayland as Guake segfaults for me on startup, is that a known issue?) I have absolutely no idea what could cause this kind of jumping. (An easy workaround might be to make the prefs dialog non-resizeable.)
1.0
Prefs window jumps on resize - Resize the prefs window. On gnome-shell + X11 it starts moving upwards. On unity7 + X11 it crazily jumps between two different positions. (I couldn't test with Wayland as Guake segfaults for me on startup, is that a known issue?) I have absolutely no idea what could cause this kind of jumping. (An easy workaround might be to make the prefs dialog non-resizeable.)
defect
prefs window jumps on resize resize the prefs window on gnome shell it starts moving upwards on it crazily jumps between two different positions i couldn t test with wayland as guake segfaults for me on startup is that a known issue i have absolutely no idea what could cause this kind of jumping an easy workaround might be to make the prefs dialog non resizeable
1
42,848
11,305,268,788
IssuesEvent
2020-01-18 03:54:39
garglk/garglk
https://api.github.com/repos/garglk/garglk
closed
Sound notification events are not delivered if the player does nothing
Priority-Medium Type-Defect auto-migrated
``` The attached Inform 7 source (and compiled Glulx Blorb) is a minimal example that, when the command "chime" is entered, plays a sound. A notification event is requested, and when this arrives, the game cancels line input, prints a message and then re-starts line input. However, when run under Gargoyle, the sound notification event is not received by the game until the first time after the sound has finished that the user generates some user interface activity (e.g. by pressing a key). If the user does nothing, the event is never received. From my reading of the Glk specification, this is not what was intended. (See also this thread: http://www.intfiction.org/forum/viewtopic.php?f=7&t=2848) ``` Original issue reported on code.google.com by `dav...@davidkinder.co.uk` on 16 Feb 2013 at 9:58 Attachments: - [Sound Notifications.gblorb](https://storage.googleapis.com/google-code-attachments/garglk/issue-204/comment-0/Sound Notifications.gblorb) - [source.txt](https://storage.googleapis.com/google-code-attachments/garglk/issue-204/comment-0/source.txt)
1.0
Sound notification events are not delivered if the player does nothing - ``` The attached Inform 7 source (and compiled Glulx Blorb) is a minimal example that, when the command "chime" is entered, plays a sound. A notification event is requested, and when this arrives, the game cancels line input, prints a message and then re-starts line input. However, when run under Gargoyle, the sound notification event is not received by the game until the first time after the sound has finished that the user generates some user interface activity (e.g. by pressing a key). If the user does nothing, the event is never received. From my reading of the Glk specification, this is not what was intended. (See also this thread: http://www.intfiction.org/forum/viewtopic.php?f=7&t=2848) ``` Original issue reported on code.google.com by `dav...@davidkinder.co.uk` on 16 Feb 2013 at 9:58 Attachments: - [Sound Notifications.gblorb](https://storage.googleapis.com/google-code-attachments/garglk/issue-204/comment-0/Sound Notifications.gblorb) - [source.txt](https://storage.googleapis.com/google-code-attachments/garglk/issue-204/comment-0/source.txt)
defect
sound notification events are not delivered if the player does nothing the attached inform source and compiled glulx blorb is a minimal example that when the command chime is entered plays a sound a notification event is requested and when this arrives the game cancels line input prints a message and then re starts line input however when run under gargoyle the sound notification event is not received by the game until the first time after the sound has finished that the user generates some user interface activity e g by pressing a key if the user does nothing the event is never received from my reading of the glk specification this is not what was intended see also this thread original issue reported on code google com by dav davidkinder co uk on feb at attachments notifications gblorb
1
28,829
5,389,797,232
IssuesEvent
2017-02-25 06:55:53
djstauffer/geben-on-emacs
https://api.github.com/repos/djstauffer/geben-on-emacs
closed
Geben hangs and shows abracadabra
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. try to debug yii-application; 2. debug external class including in yii/framework/YiiBase.php; 3. stuck upon return from autoload. What is the expected output? What do you see instead? Instead of a normal step-by-step debugging a got a message "File ~/projects/current_project/subfolder/my/.emacs.d/geben/54321/home/me/projects/cu rrent_project/subfolder/yii/framework/YiiBase.php no longer exists!" and abracadabra like this: http://images.netbynet.ru/imgs/b74929bb2410b3e32736ac73f5cf58f0.png (or see attachment) What version of the product are you using? On what operating system? app-editors/emacs-23.3 app-emacs/cedet-1.0 app-emacs/ecb-2.40-r1 app-emacs/emacs-common-gentoo-1.2-r2 dev-lang/php-5.3.6 dev-php/xdebug-2.1.1 dev-php/xdebug-client-2.1.1 virtual/emacs-23 Please provide any additional information below. At least, how to make geben view log in a _human-readable_ way? ``` Original issue reported on code.google.com by `butthur...@gmail.com` on 19 Aug 2011 at 12:34 Attachments: - [Screenshot-6.png](https://storage.googleapis.com/google-code-attachments/geben-on-emacs/issue-12/comment-0/Screenshot-6.png)
1.0
Geben hangs and shows abracadabra - ``` What steps will reproduce the problem? 1. try to debug yii-application; 2. debug external class including in yii/framework/YiiBase.php; 3. stuck upon return from autoload. What is the expected output? What do you see instead? Instead of a normal step-by-step debugging a got a message "File ~/projects/current_project/subfolder/my/.emacs.d/geben/54321/home/me/projects/cu rrent_project/subfolder/yii/framework/YiiBase.php no longer exists!" and abracadabra like this: http://images.netbynet.ru/imgs/b74929bb2410b3e32736ac73f5cf58f0.png (or see attachment) What version of the product are you using? On what operating system? app-editors/emacs-23.3 app-emacs/cedet-1.0 app-emacs/ecb-2.40-r1 app-emacs/emacs-common-gentoo-1.2-r2 dev-lang/php-5.3.6 dev-php/xdebug-2.1.1 dev-php/xdebug-client-2.1.1 virtual/emacs-23 Please provide any additional information below. At least, how to make geben view log in a _human-readable_ way? ``` Original issue reported on code.google.com by `butthur...@gmail.com` on 19 Aug 2011 at 12:34 Attachments: - [Screenshot-6.png](https://storage.googleapis.com/google-code-attachments/geben-on-emacs/issue-12/comment-0/Screenshot-6.png)
defect
geben hangs and shows abracadabra what steps will reproduce the problem try to debug yii application debug external class including in yii framework yiibase php stuck upon return from autoload what is the expected output what do you see instead instead of a normal step by step debugging a got a message file projects current project subfolder my emacs d geben home me projects cu rrent project subfolder yii framework yiibase php no longer exists and abracadabra like this or see attachment what version of the product are you using on what operating system app editors emacs app emacs cedet app emacs ecb app emacs emacs common gentoo dev lang php dev php xdebug dev php xdebug client virtual emacs please provide any additional information below at least how to make geben view log in a human readable way original issue reported on code google com by butthur gmail com on aug at attachments
1
162,030
6,145,993,381
IssuesEvent
2017-06-27 12:57:24
ubuntudesign/docs.vanillaframework.io
https://api.github.com/repos/ubuntudesign/docs.vanillaframework.io
opened
Revert header to how it was before
Priority: High Status: Triaged
- If you go to https://docs.vanillaframework.io/en/ - The header says: Vanilla logo | Documentation - If you go another page, e.g. https://docs.vanillaframework.io/en/patterns/breadcrumbs - The header changes to: Vanilla logo | Breadcrumbs | Vanilla framework documentation This behaviour is jarring. If we need a breadcrumb to indicate where we are, we should be using the breadcrumb pattern. The "Breadcrumbs | Vanilla framework documentation" from the example mentioned above links to the homepage of the docs, which is also confusing. Can we please make the header consistent across page to just say Vanilla documentation?
1.0
Revert header to how it was before - - If you go to https://docs.vanillaframework.io/en/ - The header says: Vanilla logo | Documentation - If you go another page, e.g. https://docs.vanillaframework.io/en/patterns/breadcrumbs - The header changes to: Vanilla logo | Breadcrumbs | Vanilla framework documentation This behaviour is jarring. If we need a breadcrumb to indicate where we are, we should be using the breadcrumb pattern. The "Breadcrumbs | Vanilla framework documentation" from the example mentioned above links to the homepage of the docs, which is also confusing. Can we please make the header consistent across page to just say Vanilla documentation?
non_defect
revert header to how it was before if you go to the header says vanilla logo documentation if you go another page e g the header changes to vanilla logo breadcrumbs vanilla framework documentation this behaviour is jarring if we need a breadcrumb to indicate where we are we should be using the breadcrumb pattern the breadcrumbs vanilla framework documentation from the example mentioned above links to the homepage of the docs which is also confusing can we please make the header consistent across page to just say vanilla documentation
0
81,996
31,845,186,198
IssuesEvent
2023-09-14 19:20:46
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
closed
UFAD Code Inconsistencies in Plume Power Calculations
Defect
Issue overview -------------- @amirroth noted the following inconsistencies in the UFAD code. 1. The steam equipment convective summation is reset to zero inside the loop (for both the exterior and interior models). https://github.com/NREL/EnergyPlus/blob/3ce85f9cde01190b8ee923e56f7b9606350fd53a/src/EnergyPlus/UFADManager.cc#L417-L422 https://github.com/NREL/EnergyPlus/blob/3ce85f9cde01190b8ee923e56f7b9606350fd53a/src/EnergyPlus/UFADManager.cc#L625-L630 2. When calculating the convective plume power, the equipment design level is multiplied by `.FractionConvected` for the interior model (first snipped above), but not for the exterior model (second snippet). This is the same across all equipment types. Seems like both models should use `.FractionConvected`. The engineering reference descriptions for this calculation use identical terms with `conv` subscripts. The code has been this way since these features were first added [here](https://github.com/NREL/EnergyPlusArchive/commit/8fd87048ed0ed2c328a2cad581cacfbdc330c0f1#diff-e89bfdffca8ef2d108573498a880025022146e5dc24dd265e0c9a2c8cf9b4a7a). ### Details Some additional details for this issue (if relevant): - Version of EnergyPlus 23.1 ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
UFAD Code Inconsistencies in Plume Power Calculations - Issue overview -------------- @amirroth noted the following inconsistencies in the UFAD code. 1. The steam equipment convective summation is reset to zero inside the loop (for both the exterior and interior models). https://github.com/NREL/EnergyPlus/blob/3ce85f9cde01190b8ee923e56f7b9606350fd53a/src/EnergyPlus/UFADManager.cc#L417-L422 https://github.com/NREL/EnergyPlus/blob/3ce85f9cde01190b8ee923e56f7b9606350fd53a/src/EnergyPlus/UFADManager.cc#L625-L630 2. When calculating the convective plume power, the equipment design level is multiplied by `.FractionConvected` for the interior model (first snipped above), but not for the exterior model (second snippet). This is the same across all equipment types. Seems like both models should use `.FractionConvected`. The engineering reference descriptions for this calculation use identical terms with `conv` subscripts. The code has been this way since these features were first added [here](https://github.com/NREL/EnergyPlusArchive/commit/8fd87048ed0ed2c328a2cad581cacfbdc330c0f1#diff-e89bfdffca8ef2d108573498a880025022146e5dc24dd265e0c9a2c8cf9b4a7a). ### Details Some additional details for this issue (if relevant): - Version of EnergyPlus 23.1 ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [ ] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
defect
ufad code inconsistencies in plume power calculations issue overview amirroth noted the following inconsistencies in the ufad code the steam equipment convective summation is reset to zero inside the loop for both the exterior and interior models when calculating the convective plume power the equipment design level is multiplied by fractionconvected for the interior model first snipped above but not for the exterior model second snippet this is the same across all equipment types seems like both models should use fractionconvected the engineering reference descriptions for this calculation use identical terms with conv subscripts the code has been this way since these features were first added details some additional details for this issue if relevant version of energyplus checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
1
42,209
10,886,258,746
IssuesEvent
2019-11-18 12:10:54
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Client side OOME 0 size hprof after member killed -9
Source: Internal Team: Client Team: Core Type: Defect
In the test 32 clients are connected to 1 member, the member is killed -9 and restarted repeatedly to check the effect on the clients. clients are configured with `-Xms250M -Xmx250M` http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/go client xml config http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/config-hz/client-hazelcast.xml ``` <connection-strategy async-start="true" reconnect-mode="ASYNC"> <connection-retry enabled="true"> <initial-backoff-millis>1</initial-backoff-millis> <max-backoff-millis>300</max-backoff-millis> <fail-on-max-backoff>false</fail-on-max-backoff> <multiplier>1</multiplier> <jitter>0.01</jitter> </connection-retry> </connection-strategy> <network> <cluster-members> </cluster-members> <smart-routing>true</smart-routing> <connection-attempt-period>9000</connection-attempt-period> <connection-attempt-limit>200</connection-attempt-limit> </network> ``` client are configured with high connection rates specifically to test this aspect of the system on both sides. After 967 iteration of kill -9 and restart 25 out of 32 client crashed with OOME producing a hprof of 0 size. 967 iteration makes this issue quite rare. hprofs of 0 size are odd, and could be caused by, clients allocating very large buffers based on possible junk info coming from the member ? http://jenkins.hazelcast.com/view/kill/job/kill-members/11/console http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member /disk1/jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member ``` [jenkins@ip-10-72-134-107 member]$ hz-errors | sort | uniq | xargs ls -lah -rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient10HZ/HzClient10HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient11HZ/HzClient11HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient12HZ/HzClient12HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient13HZ/HzClient13HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient14HZ/HzClient14HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient19HZ/HzClient19HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient1HZ/HzClient1HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient20HZ/HzClient20HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient21HZ/HzClient21HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient22HZ/HzClient22HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient23HZ/HzClient23HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient24HZ/HzClient24HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient25HZ/HzClient25HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient26HZ/HzClient26HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient27HZ/HzClient27HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient28HZ/HzClient28HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient29HZ/HzClient29HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient2HZ/HzClient2HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient30HZ/HzClient30HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient32HZ/HzClient32HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient4HZ/HzClient4HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient5HZ/HzClient5HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient6HZ/HzClient6HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:27 ./output/HZ/HzClient7HZ/HzClient7HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient9HZ/HzClient9HZ.hprof ``` GC charts. http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/gc.html interestingly all client GC charts look the same as below. the chart is not showing any sign of memory leak, and looks good. ![image](https://user-images.githubusercontent.com/5988678/55779201-7dec7000-5aad-11e9-9090-c7d4f9e921b0.png)
1.0
Client side OOME 0 size hprof after member killed -9 - In the test 32 clients are connected to 1 member, the member is killed -9 and restarted repeatedly to check the effect on the clients. clients are configured with `-Xms250M -Xmx250M` http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/go client xml config http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/config-hz/client-hazelcast.xml ``` <connection-strategy async-start="true" reconnect-mode="ASYNC"> <connection-retry enabled="true"> <initial-backoff-millis>1</initial-backoff-millis> <max-backoff-millis>300</max-backoff-millis> <fail-on-max-backoff>false</fail-on-max-backoff> <multiplier>1</multiplier> <jitter>0.01</jitter> </connection-retry> </connection-strategy> <network> <cluster-members> </cluster-members> <smart-routing>true</smart-routing> <connection-attempt-period>9000</connection-attempt-period> <connection-attempt-limit>200</connection-attempt-limit> </network> ``` client are configured with high connection rates specifically to test this aspect of the system on both sides. After 967 iteration of kill -9 and restart 25 out of 32 client crashed with OOME producing a hprof of 0 size. 967 iteration makes this issue quite rare. hprofs of 0 size are odd, and could be caused by, clients allocating very large buffers based on possible junk info coming from the member ? http://jenkins.hazelcast.com/view/kill/job/kill-members/11/console http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member /disk1/jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member ``` [jenkins@ip-10-72-134-107 member]$ hz-errors | sort | uniq | xargs ls -lah -rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient10HZ/HzClient10HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient11HZ/HzClient11HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient12HZ/HzClient12HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient13HZ/HzClient13HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient14HZ/HzClient14HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient19HZ/HzClient19HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:25 ./output/HZ/HzClient1HZ/HzClient1HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:20 ./output/HZ/HzClient20HZ/HzClient20HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient21HZ/HzClient21HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient22HZ/HzClient22HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient23HZ/HzClient23HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient24HZ/HzClient24HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient25HZ/HzClient25HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient26HZ/HzClient26HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:23 ./output/HZ/HzClient27HZ/HzClient27HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient28HZ/HzClient28HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient29HZ/HzClient29HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:21 ./output/HZ/HzClient2HZ/HzClient2HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient30HZ/HzClient30HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient32HZ/HzClient32HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:26 ./output/HZ/HzClient4HZ/HzClient4HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:22 ./output/HZ/HzClient5HZ/HzClient5HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient6HZ/HzClient6HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:27 ./output/HZ/HzClient7HZ/HzClient7HZ.hprof -rw------- 1 jenkins jenkins 0 Apr 8 16:24 ./output/HZ/HzClient9HZ/HzClient9HZ.hprof ``` GC charts. http://54.147.27.51/~jenkins/workspace/kill-members/3.12/2019_04_08-10_40_33/member/gc.html interestingly all client GC charts look the same as below. the chart is not showing any sign of memory leak, and looks good. ![image](https://user-images.githubusercontent.com/5988678/55779201-7dec7000-5aad-11e9-9090-c7d4f9e921b0.png)
defect
client side oome size hprof after member killed in the test clients are connected to member the member is killed and restarted repeatedly to check the effect on the clients clients are configured with client xml config false true client are configured with high connection rates specifically to test this aspect of the system on both sides after iteration of kill and restart out of client crashed with oome producing a hprof of size iteration makes this issue quite rare hprofs of size are odd and could be caused by clients allocating very large buffers based on possible junk info coming from the member jenkins workspace kill members member hz errors sort uniq xargs ls lah rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof rw jenkins jenkins apr output hz hprof gc charts interestingly all client gc charts look the same as below the chart is not showing any sign of memory leak and looks good
1
9,847
2,615,175,968
IssuesEvent
2015-03-01 06:59:27
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
Supplied full pin but Reaver doesnt succeed
auto-migrated Priority-Triage Type-Defect
``` Hi All, I am new to using Reaver, but have done a lot of reading and research into it's use. I am trying to crack my own BT Home Hub 3b (UK). However I am seeing some strange behaviour, some background first: WiFi Adapter - ALFA AWUS036NHA (Atheros AR9271) in monitor mode Linux Version - Kali latest distro from Kali website (January 2015) kali-linux-1.0.9a 64bit Reaver Version - 1.4 My router has WPS enabled and is not locked. I have run the following commands airmon-ng start wlan0 Kill 3 processes running (NetworkManager and wash -i mon0 BSSID Channel RSSI WPS Version WPS Locked ESSID 28:3C:E4:36:46:5F 11 -42 1.0 No [NetworkName] I then open a new terminal window and have tried the following few reaver commands (not at the same time). Firstly, I tried to see if Reaver actually gets the WPA when I pass it the full know WPS pin, I run it with the actual WPS pin using the following command: # reaver -i mon0 -c 11 -b 28:3C:E4:36:46:5F -d 5 -p 86890281 -vv Output is this, it just loops for ages.. surely by giving it the full pin it should find it almost instantly? [+] Switching mon0 to channel 11 [+] Waiting for beacon from 28:3C:E4:36:46:5F [+] Associated with 28:3C:E4:36:46:5F (ESSID: [NetworkName]) [+] Trying pin 86890281 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 86890281 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 86890281 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 86890281 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK I then tried running Reaver using the following command (first 4 digits of the pin, and some extra options that proved successful for others): reaver -i mon0 -c 11 -b 28:3C:E4:36:46:5F -d 5 -r 3:15 -x 360 -T .5 -S -N -t 10 -w -vv -L -p 8689 This went on for about 12 hours and got to 99.99% but just loops on the last pin (which is wrong): [+] Trying pin 86899987 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 86899987 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] 99.99% complete @ 2015-01-27 07:15:25 (7 seconds/pin) [+] Max time remaining at this rate: 0:00:07 (1 pins left to try) Any ideas on what to try next? I have read through the Wiki/Q&A but not got much further. Sorry if this seems like a bit of a dumb problem, we all have to start somewhere.. ``` Original issue reported on code.google.com by `christop...@gmail.com` on 27 Jan 2015 at 8:13
1.0
Supplied full pin but Reaver doesnt succeed - ``` Hi All, I am new to using Reaver, but have done a lot of reading and research into it's use. I am trying to crack my own BT Home Hub 3b (UK). However I am seeing some strange behaviour, some background first: WiFi Adapter - ALFA AWUS036NHA (Atheros AR9271) in monitor mode Linux Version - Kali latest distro from Kali website (January 2015) kali-linux-1.0.9a 64bit Reaver Version - 1.4 My router has WPS enabled and is not locked. I have run the following commands airmon-ng start wlan0 Kill 3 processes running (NetworkManager and wash -i mon0 BSSID Channel RSSI WPS Version WPS Locked ESSID 28:3C:E4:36:46:5F 11 -42 1.0 No [NetworkName] I then open a new terminal window and have tried the following few reaver commands (not at the same time). Firstly, I tried to see if Reaver actually gets the WPA when I pass it the full know WPS pin, I run it with the actual WPS pin using the following command: # reaver -i mon0 -c 11 -b 28:3C:E4:36:46:5F -d 5 -p 86890281 -vv Output is this, it just loops for ages.. surely by giving it the full pin it should find it almost instantly? [+] Switching mon0 to channel 11 [+] Waiting for beacon from 28:3C:E4:36:46:5F [+] Associated with 28:3C:E4:36:46:5F (ESSID: [NetworkName]) [+] Trying pin 86890281 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 86890281 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 86890281 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 86890281 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK I then tried running Reaver using the following command (first 4 digits of the pin, and some extra options that proved successful for others): reaver -i mon0 -c 11 -b 28:3C:E4:36:46:5F -d 5 -r 3:15 -x 360 -T .5 -S -N -t 10 -w -vv -L -p 8689 This went on for about 12 hours and got to 99.99% but just loops on the last pin (which is wrong): [+] Trying pin 86899987 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] Trying pin 86899987 [+] Sending EAPOL START request [+] Received identity request [+] Sending identity response [+] Received M1 message [+] Sending M2 message [+] Received M3 message [+] Sending M4 message [+] Received WSC NACK [+] Sending WSC NACK [+] 99.99% complete @ 2015-01-27 07:15:25 (7 seconds/pin) [+] Max time remaining at this rate: 0:00:07 (1 pins left to try) Any ideas on what to try next? I have read through the Wiki/Q&A but not got much further. Sorry if this seems like a bit of a dumb problem, we all have to start somewhere.. ``` Original issue reported on code.google.com by `christop...@gmail.com` on 27 Jan 2015 at 8:13
defect
supplied full pin but reaver doesnt succeed hi all i am new to using reaver but have done a lot of reading and research into it s use i am trying to crack my own bt home hub uk however i am seeing some strange behaviour some background first wifi adapter alfa atheros in monitor mode linux version kali latest distro from kali website january kali linux reaver version my router has wps enabled and is not locked i have run the following commands airmon ng start kill processes running networkmanager and wash i bssid channel rssi wps version wps locked essid no i then open a new terminal window and have tried the following few reaver commands not at the same time firstly i tried to see if reaver actually gets the wpa when i pass it the full know wps pin i run it with the actual wps pin using the following command reaver i c b d p vv output is this it just loops for ages surely by giving it the full pin it should find it almost instantly switching to channel waiting for beacon from associated with essid trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack i then tried running reaver using the following command first digits of the pin and some extra options that proved successful for others reaver i c b d r x t s n t w vv l p this went on for about hours and got to but just loops on the last pin which is wrong trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack trying pin sending eapol start request received identity request sending identity response received message sending message received message sending message received wsc nack sending wsc nack complete seconds pin max time remaining at this rate pins left to try any ideas on what to try next i have read through the wiki q a but not got much further sorry if this seems like a bit of a dumb problem we all have to start somewhere original issue reported on code google com by christop gmail com on jan at
1
37,049
8,214,119,923
IssuesEvent
2018-09-04 21:53:03
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
optimize.minimize(method=`trust-constr`) result dict does not contain `success`
defect scipy.optimize
The result dictionary from all optimizers is supposed to be uniform and contain e.g. a `success` boolean flag. The `success` flag is missing for `method='trust-constr'`. The result `nit` is called `niter` in the result dict. ### Reproducing code example: ``` from scipy.optimize import minimize sol = minimize(lambda x: x**2, [0], jac=lambda x: 2*x, hess=lambda x: 2, method='trust-constr') print(sorted(sol.keys())) assert sol.success ``` ### Error message: ``` ['cg_niter', 'cg_stop_cond', 'constr', 'constr_nfev', 'constr_nhev', 'constr_njev', 'constr_penalty', 'constr_violation', 'execution_time', 'fun', 'grad', 'jac', 'lagrangian_grad', 'message', 'method', 'nfev', 'nhev', 'niter', 'njev', 'optimality', 'status', 'tr_radius', 'v', 'x'] Traceback (most recent call last): File "<ipython-input-13-492cf8a99d03>", line 3, in <module> assert sol.success File "/home/pauli/.local/lib/python3.6/site-packages/scipy/optimize/optimize.py", line 116, in __getattr__ raise AttributeError(name) AttributeError: success ``` ### Scipy/Numpy/Python version information: ``` 1.1.0 1.14.5 sys.version_info(major=3, minor=6, micro=5, releaselevel='final', serial=0) and 1.2.0.dev0+7e41406 1.14.5 sys.version_info(major=3, minor=6, micro=5, releaselevel='final', serial=0) ```
1.0
optimize.minimize(method=`trust-constr`) result dict does not contain `success` - The result dictionary from all optimizers is supposed to be uniform and contain e.g. a `success` boolean flag. The `success` flag is missing for `method='trust-constr'`. The result `nit` is called `niter` in the result dict. ### Reproducing code example: ``` from scipy.optimize import minimize sol = minimize(lambda x: x**2, [0], jac=lambda x: 2*x, hess=lambda x: 2, method='trust-constr') print(sorted(sol.keys())) assert sol.success ``` ### Error message: ``` ['cg_niter', 'cg_stop_cond', 'constr', 'constr_nfev', 'constr_nhev', 'constr_njev', 'constr_penalty', 'constr_violation', 'execution_time', 'fun', 'grad', 'jac', 'lagrangian_grad', 'message', 'method', 'nfev', 'nhev', 'niter', 'njev', 'optimality', 'status', 'tr_radius', 'v', 'x'] Traceback (most recent call last): File "<ipython-input-13-492cf8a99d03>", line 3, in <module> assert sol.success File "/home/pauli/.local/lib/python3.6/site-packages/scipy/optimize/optimize.py", line 116, in __getattr__ raise AttributeError(name) AttributeError: success ``` ### Scipy/Numpy/Python version information: ``` 1.1.0 1.14.5 sys.version_info(major=3, minor=6, micro=5, releaselevel='final', serial=0) and 1.2.0.dev0+7e41406 1.14.5 sys.version_info(major=3, minor=6, micro=5, releaselevel='final', serial=0) ```
defect
optimize minimize method trust constr result dict does not contain success the result dictionary from all optimizers is supposed to be uniform and contain e g a success boolean flag the success flag is missing for method trust constr the result nit is called niter in the result dict reproducing code example from scipy optimize import minimize sol minimize lambda x x jac lambda x x hess lambda x method trust constr print sorted sol keys assert sol success error message cg niter cg stop cond constr constr nfev constr nhev constr njev constr penalty constr violation execution time fun grad jac lagrangian grad message method nfev nhev niter njev optimality status tr radius v x traceback most recent call last file line in assert sol success file home pauli local lib site packages scipy optimize optimize py line in getattr raise attributeerror name attributeerror success scipy numpy python version information sys version info major minor micro releaselevel final serial and sys version info major minor micro releaselevel final serial
1
78,495
27,554,265,755
IssuesEvent
2023-03-07 16:48:41
dotCMS/core
https://api.github.com/repos/dotCMS/core
opened
Can't push publish content with a block editor field
Type : Defect Triage
### Parent Issue _No response_ ### Problem Statement When you try to push publish content on auth we're getting this error ``` Error generating bundle ID '01GTWJ5V5MPX36HAJPJXJG3EV3': An error occurred when processing dependencies on Contentlet 'f528920b19e88a2ef727e8341a825f5b': An error occurred when retrieving Contentlet references from Story Block field: Unexpected character ('<' (code 60)): expected a valid value (JSON String, Number, Array, Object or token 'null', 'true' or 'false') ``` ### Steps to Reproduce Not sure yet, we need to ask @fishsmith in the #website channel ### Acceptance Criteria - [ ] Allow the user to push publish content with block editor field ### dotCMS Version 23.03 ### Proposed Objective Core Features ### Proposed Priority Priority 1 - Show Stopper ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. _No response_ ### Assumptions & Initiation Needs This is a blocker for the release of 23.03 ### Quality Assurance Notes & Workarounds _No response_ ### Sub-Tasks & Estimates _No response_
1.0
Can't push publish content with a block editor field - ### Parent Issue _No response_ ### Problem Statement When you try to push publish content on auth we're getting this error ``` Error generating bundle ID '01GTWJ5V5MPX36HAJPJXJG3EV3': An error occurred when processing dependencies on Contentlet 'f528920b19e88a2ef727e8341a825f5b': An error occurred when retrieving Contentlet references from Story Block field: Unexpected character ('<' (code 60)): expected a valid value (JSON String, Number, Array, Object or token 'null', 'true' or 'false') ``` ### Steps to Reproduce Not sure yet, we need to ask @fishsmith in the #website channel ### Acceptance Criteria - [ ] Allow the user to push publish content with block editor field ### dotCMS Version 23.03 ### Proposed Objective Core Features ### Proposed Priority Priority 1 - Show Stopper ### External Links... Slack Conversations, Support Tickets, Figma Designs, etc. _No response_ ### Assumptions & Initiation Needs This is a blocker for the release of 23.03 ### Quality Assurance Notes & Workarounds _No response_ ### Sub-Tasks & Estimates _No response_
defect
can t push publish content with a block editor field parent issue no response problem statement when you try to push publish content on auth we re getting this error error generating bundle id an error occurred when processing dependencies on contentlet an error occurred when retrieving contentlet references from story block field unexpected character code expected a valid value json string number array object or token null true or false steps to reproduce not sure yet we need to ask fishsmith in the website channel acceptance criteria allow the user to push publish content with block editor field dotcms version proposed objective core features proposed priority priority show stopper external links slack conversations support tickets figma designs etc no response assumptions initiation needs this is a blocker for the release of quality assurance notes workarounds no response sub tasks estimates no response
1
475,485
13,711,177,415
IssuesEvent
2020-10-02 03:33:44
gadenbuie/rsthemes
https://api.github.com/repos/gadenbuie/rsthemes
closed
Style job pane progress bar
priority: medium status: planned type: new
```css .rstheme_toolbarWrapper [role="progressbar"] > div { background-color: pink; } ```
1.0
Style job pane progress bar - ```css .rstheme_toolbarWrapper [role="progressbar"] > div { background-color: pink; } ```
non_defect
style job pane progress bar css rstheme toolbarwrapper div background color pink
0
22,979
11,812,162,427
IssuesEvent
2020-03-19 19:36:21
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Reporting: add a "Remove" action in "Reporting" Management app
Feature:Reporting Team:Reporting Services enhancement
Currently, when we go to "Management" app, and then to "Reporting" section, in the reports list, we don't have a "remove" button in "Actions" column. This is quite annoying, when you generate a report (CSV, PDF), download it, and want to remove it immediately, because report is not useful anymore. This would be great to add a "Remove" action available in reports list section. ![image](https://user-images.githubusercontent.com/3750300/49533732-11bfec80-f8c0-11e8-9252-546e5e56cc2d.png)
1.0
Reporting: add a "Remove" action in "Reporting" Management app - Currently, when we go to "Management" app, and then to "Reporting" section, in the reports list, we don't have a "remove" button in "Actions" column. This is quite annoying, when you generate a report (CSV, PDF), download it, and want to remove it immediately, because report is not useful anymore. This would be great to add a "Remove" action available in reports list section. ![image](https://user-images.githubusercontent.com/3750300/49533732-11bfec80-f8c0-11e8-9252-546e5e56cc2d.png)
non_defect
reporting add a remove action in reporting management app currently when we go to management app and then to reporting section in the reports list we don t have a remove button in actions column this is quite annoying when you generate a report csv pdf download it and want to remove it immediately because report is not useful anymore this would be great to add a remove action available in reports list section
0
510,820
14,826,766,269
IssuesEvent
2021-01-15 00:48:23
jef/streetmerchant
https://api.github.com/repos/jef/streetmerchant
closed
Desktop notifications need appID set in Windows 10
priority: low status: needs triage status: stale type: bug
### Description Desktop notifications through node-notifier are not working in Windows 10. This seems to be due to the requirement to set an `appID`: https://www.npmjs.com/package/node-notifier#usage-windowstoaster ``` C:\Users\Hedova\Desktop\Programs\streetmerchant>npm run test:notification > test:notification > npm run build && node build/__test__/notification-test.js > build > tsc STREETMERCHANT 3.1.0 INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). [4:31:26 AM] info :: ✔ sound player found: cmdmp3 [4:31:27 AM] info :: ✔ desktop notification sent Notifications are disabled Reason: DisabledForUser Please make sure that the app id is set correctly. Command Line: C:\Users\Hedova\Desktop\Programs\streetmerchant\node_modules\node-notifier\vendor\snoreToast\snoretoast-x64.exe -appID streetmerchant -pipeName \\.\pipe\notifierPipe-efc1ad6d-7071-4038-966f-ffe741f32252 -m https://www.example.com/cartUrl -t "[4:31:29 AM] info :: ✔ played sound ```
1.0
Desktop notifications need appID set in Windows 10 - ### Description Desktop notifications through node-notifier are not working in Windows 10. This seems to be due to the requirement to set an `appID`: https://www.npmjs.com/package/node-notifier#usage-windowstoaster ``` C:\Users\Hedova\Desktop\Programs\streetmerchant>npm run test:notification > test:notification > npm run build && node build/__test__/notification-test.js > build > tsc STREETMERCHANT 3.1.0 INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). INFO: Could not find files for the given pattern(s). [4:31:26 AM] info :: ✔ sound player found: cmdmp3 [4:31:27 AM] info :: ✔ desktop notification sent Notifications are disabled Reason: DisabledForUser Please make sure that the app id is set correctly. Command Line: C:\Users\Hedova\Desktop\Programs\streetmerchant\node_modules\node-notifier\vendor\snoreToast\snoretoast-x64.exe -appID streetmerchant -pipeName \\.\pipe\notifierPipe-efc1ad6d-7071-4038-966f-ffe741f32252 -m https://www.example.com/cartUrl -t "[4:31:29 AM] info :: ✔ played sound ```
non_defect
desktop notifications need appid set in windows description desktop notifications through node notifier are not working in windows this seems to be due to the requirement to set an appid c users hedova desktop programs streetmerchant npm run test notification test notification npm run build node build test notification test js build tsc streetmerchant info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info could not find files for the given pattern s info ✔ sound player found info ✔ desktop notification sent notifications are disabled reason disabledforuser please make sure that the app id is set correctly command line c users hedova desktop programs streetmerchant node modules node notifier vendor snoretoast snoretoast exe appid streetmerchant pipename pipe notifierpipe m t info ✔ played sound
0
11,380
2,649,862,548
IssuesEvent
2015-03-15 11:09:46
bbk79/xbmc-glwiz-addon
https://api.github.com/repos/bbk79/xbmc-glwiz-addon
closed
Audio/Video sync when "Paid Account" enabled
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Enable "Paid Account" with an active subscription. 2. 3. What is the expected output? What do you see instead? 720Kbps quality expected and seen however framerate for Video is a bit low and audio is ahead 3-4 seconds! What version of the product are you using? On what operating system? .8 on XBMC eden ATV2 Please provide any additional information below. Normal quality works fine, but soon as you enable Paid Account the issue above pops up. It's a pity as the quality improvement in video is very obvious. Any clues/ideas on this at all? ``` Original issue reported on code.google.com by `paya...@gmail.com` on 22 Dec 2012 at 8:28
1.0
Audio/Video sync when "Paid Account" enabled - ``` What steps will reproduce the problem? 1. Enable "Paid Account" with an active subscription. 2. 3. What is the expected output? What do you see instead? 720Kbps quality expected and seen however framerate for Video is a bit low and audio is ahead 3-4 seconds! What version of the product are you using? On what operating system? .8 on XBMC eden ATV2 Please provide any additional information below. Normal quality works fine, but soon as you enable Paid Account the issue above pops up. It's a pity as the quality improvement in video is very obvious. Any clues/ideas on this at all? ``` Original issue reported on code.google.com by `paya...@gmail.com` on 22 Dec 2012 at 8:28
defect
audio video sync when paid account enabled what steps will reproduce the problem enable paid account with an active subscription what is the expected output what do you see instead quality expected and seen however framerate for video is a bit low and audio is ahead seconds what version of the product are you using on what operating system on xbmc eden please provide any additional information below normal quality works fine but soon as you enable paid account the issue above pops up it s a pity as the quality improvement in video is very obvious any clues ideas on this at all original issue reported on code google com by paya gmail com on dec at
1
13,323
2,753,736,103
IssuesEvent
2015-04-25 00:47:29
sma-bluetooth/sma-bluetooth
https://api.github.com/repos/sma-bluetooth/sma-bluetooth
closed
SB2500 Did not find string
auto-migrated Priority-Medium Type-Defect
``` The communication with SB2500 is not working fine as the format does not seem to be correct (also for SB1700) [18] 2011-07-19 20:39:03 Waiting for string [18] 2011-07-19 20:39:03 waiting for: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04 70 00 [18] 2011-07-19 20:39:03 Waiting for data on rfcomm Receiving 00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61 00000010: 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04 00000020: 70 00 01 00 00 00 00 01 00 00 00 rr=31 [18] 2011-07-19 20:39:03 looking for: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04 70 00 [18] 2011-07-19 20:39:03 received: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04 70 00 01 00 00 00 00 01 00 00 00 [18] 2011-07-19 20:39:03 Found string we are waiting for [19] 2011-07-19 20:39:03 Extracting extracting invcode=01 [20] 2011-07-19 20:39:03 Sending [20] 2011-07-19 20:39:03 sending: 00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61 00000010: 00 00 00 00 00 00 05 7e 1b 25 80 00 02 00 00 04 00000020: 70 00 01 00 00 00 00 01 00 00 00 cc=31 [21] 2011-07-19 20:39:03 Waiting for string [21] 2011-07-19 20:39:03 waiting for: 7e 22 00 5c 05 7e 1b 25 80 00 00 00 00 00 00 00 05 00 05 7e 1b 25 80 00 [21] 2011-07-19 20:39:03 Waiting for data on rfcomm Receiving 00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61 00000010: 05 7e 1b 25 80 00 00 00 00 00 00 00 0a 00 05 7e 00000020: 1b 25 80 00 01 2d 8d 00 7d 1a 00 rr=31 sum=80length change from 1f to 1e [21] 2011-07-19 20:39:03 looking for: 7e 22 00 5c 05 7e 1b 25 80 00 00 00 00 00 00 00 05 00 05 7e 1b 25 80 00 [21] 2011-07-19 20:39:03 received: 7e 1e 00 62 05 7e 1b 25 80 00 00 00 00 00 00 00 0a 00 05 7e 1b 25 80 00 01 2d 8d 00 3a 00 [21] 2011-07-19 20:39:03 Did not find string How can I solve this? Thanks ``` Original issue reported on code.google.com by `patrik.d...@gmail.com` on 19 Jul 2011 at 9:05 Attachments: * [smatool_20110719.log](https://storage.googleapis.com/google-code-attachments/sma-bluetooth/issue-35/comment-0/smatool_20110719.log)
1.0
SB2500 Did not find string - ``` The communication with SB2500 is not working fine as the format does not seem to be correct (also for SB1700) [18] 2011-07-19 20:39:03 Waiting for string [18] 2011-07-19 20:39:03 waiting for: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04 70 00 [18] 2011-07-19 20:39:03 Waiting for data on rfcomm Receiving 00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61 00000010: 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04 00000020: 70 00 01 00 00 00 00 01 00 00 00 rr=31 [18] 2011-07-19 20:39:03 looking for: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04 70 00 [18] 2011-07-19 20:39:03 received: 7e 1f 00 61 05 7e 1b 25 80 00 00 00 00 00 00 00 02 00 00 04 70 00 01 00 00 00 00 01 00 00 00 [18] 2011-07-19 20:39:03 Found string we are waiting for [19] 2011-07-19 20:39:03 Extracting extracting invcode=01 [20] 2011-07-19 20:39:03 Sending [20] 2011-07-19 20:39:03 sending: 00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61 00000010: 00 00 00 00 00 00 05 7e 1b 25 80 00 02 00 00 04 00000020: 70 00 01 00 00 00 00 01 00 00 00 cc=31 [21] 2011-07-19 20:39:03 Waiting for string [21] 2011-07-19 20:39:03 waiting for: 7e 22 00 5c 05 7e 1b 25 80 00 00 00 00 00 00 00 05 00 05 7e 1b 25 80 00 [21] 2011-07-19 20:39:03 Waiting for data on rfcomm Receiving 00000000: .. .. .. .. .. .. .. .. .. .. .. .. 7e 1f 00 61 00000010: 05 7e 1b 25 80 00 00 00 00 00 00 00 0a 00 05 7e 00000020: 1b 25 80 00 01 2d 8d 00 7d 1a 00 rr=31 sum=80length change from 1f to 1e [21] 2011-07-19 20:39:03 looking for: 7e 22 00 5c 05 7e 1b 25 80 00 00 00 00 00 00 00 05 00 05 7e 1b 25 80 00 [21] 2011-07-19 20:39:03 received: 7e 1e 00 62 05 7e 1b 25 80 00 00 00 00 00 00 00 0a 00 05 7e 1b 25 80 00 01 2d 8d 00 3a 00 [21] 2011-07-19 20:39:03 Did not find string How can I solve this? Thanks ``` Original issue reported on code.google.com by `patrik.d...@gmail.com` on 19 Jul 2011 at 9:05 Attachments: * [smatool_20110719.log](https://storage.googleapis.com/google-code-attachments/sma-bluetooth/issue-35/comment-0/smatool_20110719.log)
defect
did not find string the communication with is not working fine as the format does not seem to be correct also for waiting for string waiting for waiting for data on rfcomm receiving rr looking for received found string we are waiting for extracting extracting invcode sending sending cc waiting for string waiting for waiting for data on rfcomm receiving rr sum change from to looking for received did not find string how can i solve this thanks original issue reported on code google com by patrik d gmail com on jul at attachments
1
81,432
30,844,627,260
IssuesEvent
2023-08-02 13:00:19
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
opened
DataTyble: Empty aria-label generation in multiselect tables
:lady_beetle: defect :bangbang: needs-triage
### Describe the bug During our improvements / rewrites to make our webapp accessibly and fullfil the requirements of the official German BITV 2.0 test we came along a slightly problem with multi select [data tables](https://www.primefaces.org/showcase/ui/data/datatable/selection.xhtml?jfwid=8db7c). The tried to achieve that every checkbox has an (hidden) description like "select row with identifier X" so the screen reader (we use NVDA, which is also used in the official test) reads to a visual impaired person what row he selects, when ticking the checkbox. For this we tried to make use of the PF API attribute `ariaRowLabel` which descriptions sounds to do exactly this. But it seems that it does not as the screenreader keeps quiet, regardless of the value inside. So we used a pass-trough attribute which is picked up correctly by the screenreader, but which is not rendered at the checkbox, but at the `td` tag. So this code ``` <p:column headerText="Selecton" selectionMode="multiple" pt:aria-label="XXX" style="width:10ch; text-align:left"> </p:column> ``` generates this outcome in PF 12.00 (Not yet tested with PF 13, but release notes don't show any changes here and the sadly has no correspondening example :( ) ``` <td role="gridcell" style="width:10ch; text-align:left" class="ui-selection-column" aria-label="XXX"> <div class="ui-chkbox ui-widget"> <div id="frm_kleineSucheSuchErgebnisse:row_2:crd_2:tbl_suchErgebnisse_1621_checkbox" role="checkbox" tabindex="0" aria-label="" aria-checked="false" class="ui-chkbox-box ui-widget ui-corner-all ui-state-default" > <span class="ui-chkbox-icon ui-icon ui-icon-blank ui-c"></span> </div> </div> </td> ``` As you can see, this renderes an empty `aria-label` for the "checkbox-div". So why do I file an issue, when the screen reader handles it correctly? Two reasons: * I / We assume that the empty `aria-label` should have the value the `td` has, where the `td` should probably have none or the same? (Don't know if the same would cause trouble tbh) * The lighthouse report claims that the checkbox has no aria label and is therefore not accessible * What is the `ariaRowLabel` for, if not for this? > ARIA toggle fields do not have accessible names > > When a toggle field doesn't have an accessible name, screen readers announce it with a generic name, making it unusable > for users who rely on screen readers. > > div#frm_kleineSucheSuchErgebnisse:row_2:crd_2:tbl_suchErgebnisse_1621_checkbox. > ui-chkbox-box.ui-widget.ui-corner-all.ui-state-default Further notes * The lightous report always claimes that error, regardless of usage of a `pt` or the `ariaRowLabel` or none/bith of them * Using the `pt` makes the screenreader work, not using or only using the `ariaRowLabel` does not ### Reproducer _No response_ ### Expected behavior _No response_ ### PrimeFaces edition Community ### PrimeFaces version 12.0.0 ### Theme _No response_ ### JSF implementation MyFaces ### JSF version 2.2 ### Java version 8 ### Browser(s) _No response_
1.0
DataTyble: Empty aria-label generation in multiselect tables - ### Describe the bug During our improvements / rewrites to make our webapp accessibly and fullfil the requirements of the official German BITV 2.0 test we came along a slightly problem with multi select [data tables](https://www.primefaces.org/showcase/ui/data/datatable/selection.xhtml?jfwid=8db7c). The tried to achieve that every checkbox has an (hidden) description like "select row with identifier X" so the screen reader (we use NVDA, which is also used in the official test) reads to a visual impaired person what row he selects, when ticking the checkbox. For this we tried to make use of the PF API attribute `ariaRowLabel` which descriptions sounds to do exactly this. But it seems that it does not as the screenreader keeps quiet, regardless of the value inside. So we used a pass-trough attribute which is picked up correctly by the screenreader, but which is not rendered at the checkbox, but at the `td` tag. So this code ``` <p:column headerText="Selecton" selectionMode="multiple" pt:aria-label="XXX" style="width:10ch; text-align:left"> </p:column> ``` generates this outcome in PF 12.00 (Not yet tested with PF 13, but release notes don't show any changes here and the sadly has no correspondening example :( ) ``` <td role="gridcell" style="width:10ch; text-align:left" class="ui-selection-column" aria-label="XXX"> <div class="ui-chkbox ui-widget"> <div id="frm_kleineSucheSuchErgebnisse:row_2:crd_2:tbl_suchErgebnisse_1621_checkbox" role="checkbox" tabindex="0" aria-label="" aria-checked="false" class="ui-chkbox-box ui-widget ui-corner-all ui-state-default" > <span class="ui-chkbox-icon ui-icon ui-icon-blank ui-c"></span> </div> </div> </td> ``` As you can see, this renderes an empty `aria-label` for the "checkbox-div". So why do I file an issue, when the screen reader handles it correctly? Two reasons: * I / We assume that the empty `aria-label` should have the value the `td` has, where the `td` should probably have none or the same? (Don't know if the same would cause trouble tbh) * The lighthouse report claims that the checkbox has no aria label and is therefore not accessible * What is the `ariaRowLabel` for, if not for this? > ARIA toggle fields do not have accessible names > > When a toggle field doesn't have an accessible name, screen readers announce it with a generic name, making it unusable > for users who rely on screen readers. > > div#frm_kleineSucheSuchErgebnisse:row_2:crd_2:tbl_suchErgebnisse_1621_checkbox. > ui-chkbox-box.ui-widget.ui-corner-all.ui-state-default Further notes * The lightous report always claimes that error, regardless of usage of a `pt` or the `ariaRowLabel` or none/bith of them * Using the `pt` makes the screenreader work, not using or only using the `ariaRowLabel` does not ### Reproducer _No response_ ### Expected behavior _No response_ ### PrimeFaces edition Community ### PrimeFaces version 12.0.0 ### Theme _No response_ ### JSF implementation MyFaces ### JSF version 2.2 ### Java version 8 ### Browser(s) _No response_
defect
datatyble empty aria label generation in multiselect tables describe the bug during our improvements rewrites to make our webapp accessibly and fullfil the requirements of the official german bitv test we came along a slightly problem with multi select the tried to achieve that every checkbox has an hidden description like select row with identifier x so the screen reader we use nvda which is also used in the official test reads to a visual impaired person what row he selects when ticking the checkbox for this we tried to make use of the pf api attribute ariarowlabel which descriptions sounds to do exactly this but it seems that it does not as the screenreader keeps quiet regardless of the value inside so we used a pass trough attribute which is picked up correctly by the screenreader but which is not rendered at the checkbox but at the td tag so this code p column headertext selecton selectionmode multiple pt aria label xxx style width text align left generates this outcome in pf not yet tested with pf but release notes don t show any changes here and the sadly has no correspondening example td role gridcell style width text align left class ui selection column aria label xxx div class ui chkbox ui widget div id frm kleinesuchesuchergebnisse row crd tbl suchergebnisse checkbox role checkbox tabindex aria label aria checked false class ui chkbox box ui widget ui corner all ui state default as you can see this renderes an empty aria label for the checkbox div so why do i file an issue when the screen reader handles it correctly two reasons i we assume that the empty aria label should have the value the td has where the td should probably have none or the same don t know if the same would cause trouble tbh the lighthouse report claims that the checkbox has no aria label and is therefore not accessible what is the ariarowlabel for if not for this aria toggle fields do not have accessible names when a toggle field doesn t have an accessible name screen readers announce it with a generic name making it unusable for users who rely on screen readers div frm kleinesuchesuchergebnisse row crd tbl suchergebnisse checkbox ui chkbox box ui widget ui corner all ui state default further notes the lightous report always claimes that error regardless of usage of a pt or the ariarowlabel or none bith of them using the pt makes the screenreader work not using or only using the ariarowlabel does not reproducer no response expected behavior no response primefaces edition community primefaces version theme no response jsf implementation myfaces jsf version java version browser s no response
1
13,677
2,775,779,459
IssuesEvent
2015-05-04 18:02:12
douglasdrumond/macvim
https://api.github.com/repos/douglasdrumond/macvim
closed
No windows or files opening
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? Opening MacVim, either through application icon or mvim command. What is the expected output? What do you see instead? Expect a window to open with the desired file, or even a blank new file. Nothing happens, the app will start (as indicated by MacVim options in toolbar) but no windows will open, even when explicitly told to open a valid file through File->Open... What version of MacVim and OS X are you using (see "MacVim->About MacVim" and "Apple Menu->About This Mac" menu items, e.g. "Snapshot 40, 10.5.6 Intel")? MacVim Snapshot 70 OS X 10.10 Yosemite Intel Please provide any additional information below. Again, the app seems to be opening and running but I cannot open any files or windows. All I get is the toolbar and the icon in the Dock. ``` Original issue reported on code.google.com by `knole...@gmail.com` on 21 Oct 2014 at 3:09
1.0
No windows or files opening - ``` What steps will reproduce the problem? Opening MacVim, either through application icon or mvim command. What is the expected output? What do you see instead? Expect a window to open with the desired file, or even a blank new file. Nothing happens, the app will start (as indicated by MacVim options in toolbar) but no windows will open, even when explicitly told to open a valid file through File->Open... What version of MacVim and OS X are you using (see "MacVim->About MacVim" and "Apple Menu->About This Mac" menu items, e.g. "Snapshot 40, 10.5.6 Intel")? MacVim Snapshot 70 OS X 10.10 Yosemite Intel Please provide any additional information below. Again, the app seems to be opening and running but I cannot open any files or windows. All I get is the toolbar and the icon in the Dock. ``` Original issue reported on code.google.com by `knole...@gmail.com` on 21 Oct 2014 at 3:09
defect
no windows or files opening what steps will reproduce the problem opening macvim either through application icon or mvim command what is the expected output what do you see instead expect a window to open with the desired file or even a blank new file nothing happens the app will start as indicated by macvim options in toolbar but no windows will open even when explicitly told to open a valid file through file open what version of macvim and os x are you using see macvim about macvim and apple menu about this mac menu items e g snapshot intel macvim snapshot os x yosemite intel please provide any additional information below again the app seems to be opening and running but i cannot open any files or windows all i get is the toolbar and the icon in the dock original issue reported on code google com by knole gmail com on oct at
1
86,818
8,051,130,707
IssuesEvent
2018-08-01 15:15:51
syndesisio/syndesis
https://api.github.com/repos/syndesisio/syndesis
closed
Logout on firefox does not work correctly
cat/bug prio/p1 qe/logout-test
I tried new logout endpoint on firefox from anonymous window and it does not work. After I log out, I am on logout page. Then when I click on Login button, this happens: ![logout_error_firefox](https://user-images.githubusercontent.com/14313995/42498786-6bf51318-842d-11e8-950a-62c475d3d856.png) Console looked like this: ![logout_error_console](https://user-images.githubusercontent.com/14313995/42498806-76af477e-842d-11e8-80fe-1b8a28213d8e.png) @zregvart
1.0
Logout on firefox does not work correctly - I tried new logout endpoint on firefox from anonymous window and it does not work. After I log out, I am on logout page. Then when I click on Login button, this happens: ![logout_error_firefox](https://user-images.githubusercontent.com/14313995/42498786-6bf51318-842d-11e8-950a-62c475d3d856.png) Console looked like this: ![logout_error_console](https://user-images.githubusercontent.com/14313995/42498806-76af477e-842d-11e8-80fe-1b8a28213d8e.png) @zregvart
non_defect
logout on firefox does not work correctly i tried new logout endpoint on firefox from anonymous window and it does not work after i log out i am on logout page then when i click on login button this happens console looked like this zregvart
0
65,412
19,487,203,422
IssuesEvent
2021-12-26 16:11:03
dkfans/keeperfx
https://api.github.com/repos/dkfans/keeperfx
opened
Custom tilesets not visibible in straight view
Priority-Medium Type-Defect
To reproduce: 1) Make any map where part of the tileset is different from default. (so an .slx file is present) 2) Start the map, and observe the custom texture. Like big boobs on a snow map. 3) Switch to straight view -> notice the custom textures are not visible
1.0
Custom tilesets not visibible in straight view - To reproduce: 1) Make any map where part of the tileset is different from default. (so an .slx file is present) 2) Start the map, and observe the custom texture. Like big boobs on a snow map. 3) Switch to straight view -> notice the custom textures are not visible
defect
custom tilesets not visibible in straight view to reproduce make any map where part of the tileset is different from default so an slx file is present start the map and observe the custom texture like big boobs on a snow map switch to straight view notice the custom textures are not visible
1
54,529
23,290,449,903
IssuesEvent
2022-08-05 21:59:42
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Marshaller for treating char array as null-terminated string in source generation
area-System.Runtime.InteropServices
Arrays are - and should continue to be - treated as buffers in marshalling, but arrays of `char` are often used as a way to marshal a string between managed/unmanaged. Per https://github.com/dotnet/runtime/issues/69809#issuecomment-1137708142, we can let users opt-in to treating a `ReadOnlySpan<char>` or `char[]` as a null-terminated string. We can provide a marshaller that can be used with `MarshalUsing` at the parameter level.
1.0
Marshaller for treating char array as null-terminated string in source generation - Arrays are - and should continue to be - treated as buffers in marshalling, but arrays of `char` are often used as a way to marshal a string between managed/unmanaged. Per https://github.com/dotnet/runtime/issues/69809#issuecomment-1137708142, we can let users opt-in to treating a `ReadOnlySpan<char>` or `char[]` as a null-terminated string. We can provide a marshaller that can be used with `MarshalUsing` at the parameter level.
non_defect
marshaller for treating char array as null terminated string in source generation arrays are and should continue to be treated as buffers in marshalling but arrays of char are often used as a way to marshal a string between managed unmanaged per we can let users opt in to treating a readonlyspan or char as a null terminated string we can provide a marshaller that can be used with marshalusing at the parameter level
0