Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
17,900
23,874,099,870
IssuesEvent
2022-09-07 17:17:23
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
K8S Processor: Include more tags
processor/k8sattributes
Currently, the processor supports following tags: - namespace - podName - deployment - cluster - node - startTime It would be great to include more tags, such as: - hostName - containerName - daemonSetName - serviceName - statefulSetName Additionally, it would be good to have an ability to include all labels and annotations
1.0
K8S Processor: Include more tags - Currently, the processor supports following tags: - namespace - podName - deployment - cluster - node - startTime It would be great to include more tags, such as: - hostName - containerName - daemonSetName - serviceName - statefulSetName Additionally, it would be good to have an ability to include all labels and annotations
process
processor include more tags currently the processor supports following tags namespace podname deployment cluster node starttime it would be great to include more tags such as hostname containername daemonsetname servicename statefulsetname additionally it would be good to have an ability to include all labels and annotations
1
374
2,815,784,384
IssuesEvent
2015-05-19 07:39:32
mattesCZ/mtbmap
https://api.github.com/repos/mattesCZ/mtbmap
opened
Add hiking trails with red:red:white_bar
data_processing map_style
OSMC:symbol red:red:white_bar is used in Austria, Italy, France... It should be rendered as Czech red:white:red_bar. There are only inverted foreground and background colors.
1.0
Add hiking trails with red:red:white_bar - OSMC:symbol red:red:white_bar is used in Austria, Italy, France... It should be rendered as Czech red:white:red_bar. There are only inverted foreground and background colors.
process
add hiking trails with red red white bar osmc symbol red red white bar is used in austria italy france it should be rendered as czech red white red bar there are only inverted foreground and background colors
1
478
2,911,386,423
IssuesEvent
2015-06-22 09:14:03
haskell-distributed/distributed-process-simplelocalnet
https://api.github.com/repos/haskell-distributed/distributed-process-simplelocalnet
opened
Re-export runProcess and forkProcess in SimpleLocalnet
distributed-process-simplelocalnet Feature Request
_From @edsko on October 23, 2012 14:51_ so that startSlave and startMaster are truly optional. _Copied from original issue: haskell-distributed/distributed-process#57_
1.0
Re-export runProcess and forkProcess in SimpleLocalnet - _From @edsko on October 23, 2012 14:51_ so that startSlave and startMaster are truly optional. _Copied from original issue: haskell-distributed/distributed-process#57_
process
re export runprocess and forkprocess in simplelocalnet from edsko on october so that startslave and startmaster are truly optional copied from original issue haskell distributed distributed process
1
17,693
23,540,085,296
IssuesEvent
2022-08-20 08:43:53
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Request in QGIS ([processing][needs-docs] Adjust status of controls executing an algorithm dialog )
Processing 3.14
### Request for documentation From pull request QGIS/qgis#34928 Author: @gacarrillor QGIS version: 3.14 **[processing][needs-docs] Adjust status of controls executing an algorithm dialog ** ### PR Description: ## Description Currently, after you run a Processing algorithm, the `Log` tab is visible and the `Run` button is enabled, so you can run the algorithm again with a single click. From a user perspective, it shouldn't be the case. If you want to run the algorithm again, you should at least have a look at the `Parameters` tab once more. This PR adds the following changes to such behaviour: + `Run` button is not shown anymore in the `Log` tab, therefore, you can only run algorithms from the `Parameters` tab. + While running an algorithm, the `Parameters` panel is now disabled, as well as `Run as Batch Process...` and `Run as Single Process...` buttons. + When an algorithm execution finishes (either successfully or not), a new button `Change Parameters` is shown in the `Log` tab. + There is now consistency among algorithm dialogs: Before this PR, `Batch Algorithm Dialog` was the only dialog blocking parameter widgets. These changes were applied to the `Algorithm Dialog` and `Batch Algorithm Dialog`, and work on `Edit in place` dialogs as well. This PR also takes into account cancelling an algorithm execution or obtaining failures in the execution. (More details and preliminary discussion in issue https://github.com/qgis/QGIS/issues/34484) ## Screencasts ### Algorithm dialog ![pr_algorithm_dialog_v3](https://user-images.githubusercontent.com/652785/76163646-f4f31400-6115-11ea-9b69-1af3517a4ee7.gif) ### Batch algorithm dialog ![pr_batch_algorithm_dialog_v3](https://user-images.githubusercontent.com/652785/76163648-f8869b00-6115-11ea-84b2-8f1b104df33d.gif) ### Cancelling algorithm execution ![pr_cancelling_algorithm_dialog_v3](https://user-images.githubusercontent.com/652785/76163650-fc1a2200-6115-11ea-82f0-3eb452b4966a.gif) ---------------------------- Fix #34484 Since this PR implies updating QGIS docs, I could give a hand adjusting the corresponding section. ### Commits tagged with [need-docs] or [FEATURE] "[processing]\n\n\n\n Adjust status of controls in algorithm dialog\n\n + Run button is not shown anymore in the Log tab, therefore, you can only run algorithms from the Parameters tab.\n + While running an algorithm, the Parameters tab is now blocked.\n + When an algorithm execution finishes (either successfully or not), a new button Change Parameters is shown in the Log tab.\n + The Batch Algorithm Dialog is now consistent with the described behavior (before, it blocked the Parameters panel, but not the tab; and it was the only dialog blocking parameters widgets).\n\nThese changes were applied to the Algorithm Dialog and Batch Algorithm Dialog, and work on Edit in place dialogs as well."
1.0
Request in QGIS ([processing][needs-docs] Adjust status of controls executing an algorithm dialog ) - ### Request for documentation From pull request QGIS/qgis#34928 Author: @gacarrillor QGIS version: 3.14 **[processing][needs-docs] Adjust status of controls executing an algorithm dialog ** ### PR Description: ## Description Currently, after you run a Processing algorithm, the `Log` tab is visible and the `Run` button is enabled, so you can run the algorithm again with a single click. From a user perspective, it shouldn't be the case. If you want to run the algorithm again, you should at least have a look at the `Parameters` tab once more. This PR adds the following changes to such behaviour: + `Run` button is not shown anymore in the `Log` tab, therefore, you can only run algorithms from the `Parameters` tab. + While running an algorithm, the `Parameters` panel is now disabled, as well as `Run as Batch Process...` and `Run as Single Process...` buttons. + When an algorithm execution finishes (either successfully or not), a new button `Change Parameters` is shown in the `Log` tab. + There is now consistency among algorithm dialogs: Before this PR, `Batch Algorithm Dialog` was the only dialog blocking parameter widgets. These changes were applied to the `Algorithm Dialog` and `Batch Algorithm Dialog`, and work on `Edit in place` dialogs as well. This PR also takes into account cancelling an algorithm execution or obtaining failures in the execution. (More details and preliminary discussion in issue https://github.com/qgis/QGIS/issues/34484) ## Screencasts ### Algorithm dialog ![pr_algorithm_dialog_v3](https://user-images.githubusercontent.com/652785/76163646-f4f31400-6115-11ea-9b69-1af3517a4ee7.gif) ### Batch algorithm dialog ![pr_batch_algorithm_dialog_v3](https://user-images.githubusercontent.com/652785/76163648-f8869b00-6115-11ea-84b2-8f1b104df33d.gif) ### Cancelling algorithm execution ![pr_cancelling_algorithm_dialog_v3](https://user-images.githubusercontent.com/652785/76163650-fc1a2200-6115-11ea-82f0-3eb452b4966a.gif) ---------------------------- Fix #34484 Since this PR implies updating QGIS docs, I could give a hand adjusting the corresponding section. ### Commits tagged with [need-docs] or [FEATURE] "[processing]\n\n\n\n Adjust status of controls in algorithm dialog\n\n + Run button is not shown anymore in the Log tab, therefore, you can only run algorithms from the Parameters tab.\n + While running an algorithm, the Parameters tab is now blocked.\n + When an algorithm execution finishes (either successfully or not), a new button Change Parameters is shown in the Log tab.\n + The Batch Algorithm Dialog is now consistent with the described behavior (before, it blocked the Parameters panel, but not the tab; and it was the only dialog blocking parameters widgets).\n\nThese changes were applied to the Algorithm Dialog and Batch Algorithm Dialog, and work on Edit in place dialogs as well."
process
request in qgis adjust status of controls executing an algorithm dialog request for documentation from pull request qgis qgis author gacarrillor qgis version adjust status of controls executing an algorithm dialog pr description description currently after you run a processing algorithm the log tab is visible and the run button is enabled so you can run the algorithm again with a single click from a user perspective it shouldn t be the case if you want to run the algorithm again you should at least have a look at the parameters tab once more this pr adds the following changes to such behaviour run button is not shown anymore in the log tab therefore you can only run algorithms from the parameters tab while running an algorithm the parameters panel is now disabled as well as run as batch process and run as single process buttons when an algorithm execution finishes either successfully or not a new button change parameters is shown in the log tab there is now consistency among algorithm dialogs before this pr batch algorithm dialog was the only dialog blocking parameter widgets these changes were applied to the algorithm dialog and batch algorithm dialog and work on edit in place dialogs as well this pr also takes into account cancelling an algorithm execution or obtaining failures in the execution more details and preliminary discussion in issue screencasts algorithm dialog batch algorithm dialog cancelling algorithm execution fix since this pr implies updating qgis docs i could give a hand adjusting the corresponding section commits tagged with or n n n n adjust status of controls in algorithm dialog n n run button is not shown anymore in the log tab therefore you can only run algorithms from the parameters tab n while running an algorithm the parameters tab is now blocked n when an algorithm execution finishes either successfully or not a new button change parameters is shown in the log tab n the batch algorithm dialog is now consistent with the described behavior before it blocked the parameters panel but not the tab and it was the only dialog blocking parameters widgets n nthese changes were applied to the algorithm dialog and batch algorithm dialog and work on edit in place dialogs as well
1
2,714
5,580,379,221
IssuesEvent
2017-03-28 16:28:30
TEAMMATES/teammates
https://api.github.com/repos/TEAMMATES/teammates
closed
Explore using mentionbot to identify potential reviewers
a-Process c.DevOps d.Committers p.Medium
Seems easy to set up and could be useful https://github.com/facebook/mention-bot
1.0
Explore using mentionbot to identify potential reviewers - Seems easy to set up and could be useful https://github.com/facebook/mention-bot
process
explore using mentionbot to identify potential reviewers seems easy to set up and could be useful
1
22,504
31,558,669,704
IssuesEvent
2023-09-03 01:13:53
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
opened
[Mirror] uber/hermetic_cc_toolchain v2.1.0
P2 type: process team-OSS mirror request
### Please list the URLs of the archives you'd like to mirror: https://github.com/uber/hermetic_cc_toolchain/releases/download/v2.1.0/hermetic_cc_toolchain-v2.1.0.tar.gz
1.0
[Mirror] uber/hermetic_cc_toolchain v2.1.0 - ### Please list the URLs of the archives you'd like to mirror: https://github.com/uber/hermetic_cc_toolchain/releases/download/v2.1.0/hermetic_cc_toolchain-v2.1.0.tar.gz
process
uber hermetic cc toolchain please list the urls of the archives you d like to mirror
1
109,398
16,843,685,085
IssuesEvent
2021-06-19 02:50:48
bharathirajatut/fitbit-api-example-java2
https://api.github.com/repos/bharathirajatut/fitbit-api-example-java2
opened
CVE-2016-9879 (High) detected in spring-security-web-4.1.1.RELEASE.jar
security vulnerability
## CVE-2016-9879 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-4.1.1.RELEASE.jar</b></p></summary> <p>spring-security-web</p> <p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p> <p>Path to dependency file: fitbit-api-example-java2/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/4.1.1.RELEASE/spring-security-web-4.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-security-1.4.0.RELEASE.jar (Root Library) - :x: **spring-security-web-4.1.1.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/bharathirajatut/fitbit-api-example-java2/commits/8c153ad064e8f07a4ddade35ac13a9b485ca3dac">8c153ad064e8f07a4ddade35ac13a9b485ca3dac</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Pivotal Spring Security before 3.2.10, 4.1.x before 4.1.4, and 4.2.x before 4.2.1. Spring Security does not consider URL path parameters when processing security constraints. By adding a URL path parameter with an encoded "/" to a request, an attacker may be able to bypass a security constraint. The root cause of this issue is a lack of clarity regarding the handling of path parameters in the Servlet Specification. Some Servlet containers include path parameters in the value returned for getPathInfo() and some do not. Spring Security uses the value returned by getPathInfo() as part of the process of mapping requests to security constraints. The unexpected presence of path parameters can cause a constraint to be bypassed. Users of Apache Tomcat (all current versions) are not affected by this vulnerability since Tomcat follows the guidance previously provided by the Servlet Expert group and strips path parameters from the value returned by getContextPath(), getServletPath(), and getPathInfo(). Users of other Servlet containers based on Apache Tomcat may or may not be affected depending on whether or not the handling of path parameters has been modified. Users of IBM WebSphere Application Server 8.5.x are known to be affected. Users of other containers that implement the Servlet specification may be affected. <p>Publish Date: 2017-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9879>CVE-2016-9879</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9879">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9879</a></p> <p>Release Date: 2017-01-06</p> <p>Fix Resolution: 3.2.10,4.1.4,4.2.1.</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-9879 (High) detected in spring-security-web-4.1.1.RELEASE.jar - ## CVE-2016-9879 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-4.1.1.RELEASE.jar</b></p></summary> <p>spring-security-web</p> <p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p> <p>Path to dependency file: fitbit-api-example-java2/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/4.1.1.RELEASE/spring-security-web-4.1.1.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-security-1.4.0.RELEASE.jar (Root Library) - :x: **spring-security-web-4.1.1.RELEASE.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/bharathirajatut/fitbit-api-example-java2/commits/8c153ad064e8f07a4ddade35ac13a9b485ca3dac">8c153ad064e8f07a4ddade35ac13a9b485ca3dac</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Pivotal Spring Security before 3.2.10, 4.1.x before 4.1.4, and 4.2.x before 4.2.1. Spring Security does not consider URL path parameters when processing security constraints. By adding a URL path parameter with an encoded "/" to a request, an attacker may be able to bypass a security constraint. The root cause of this issue is a lack of clarity regarding the handling of path parameters in the Servlet Specification. Some Servlet containers include path parameters in the value returned for getPathInfo() and some do not. Spring Security uses the value returned by getPathInfo() as part of the process of mapping requests to security constraints. The unexpected presence of path parameters can cause a constraint to be bypassed. Users of Apache Tomcat (all current versions) are not affected by this vulnerability since Tomcat follows the guidance previously provided by the Servlet Expert group and strips path parameters from the value returned by getContextPath(), getServletPath(), and getPathInfo(). Users of other Servlet containers based on Apache Tomcat may or may not be affected depending on whether or not the handling of path parameters has been modified. Users of IBM WebSphere Application Server 8.5.x are known to be affected. Users of other containers that implement the Servlet specification may be affected. <p>Publish Date: 2017-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9879>CVE-2016-9879</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9879">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9879</a></p> <p>Release Date: 2017-01-06</p> <p>Fix Resolution: 3.2.10,4.1.4,4.2.1.</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in spring security web release jar cve high severity vulnerability vulnerable library spring security web release jar spring security web library home page a href path to dependency file fitbit api example pom xml path to vulnerable library home wss scanner repository org springframework security spring security web release spring security web release jar dependency hierarchy spring boot starter security release jar root library x spring security web release jar vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in pivotal spring security before x before and x before spring security does not consider url path parameters when processing security constraints by adding a url path parameter with an encoded to a request an attacker may be able to bypass a security constraint the root cause of this issue is a lack of clarity regarding the handling of path parameters in the servlet specification some servlet containers include path parameters in the value returned for getpathinfo and some do not spring security uses the value returned by getpathinfo as part of the process of mapping requests to security constraints the unexpected presence of path parameters can cause a constraint to be bypassed users of apache tomcat all current versions are not affected by this vulnerability since tomcat follows the guidance previously provided by the servlet expert group and strips path parameters from the value returned by getcontextpath getservletpath and getpathinfo users of other servlet containers based on apache tomcat may or may not be affected depending on whether or not the handling of path parameters has been modified users of ibm websphere application server x are known to be affected users of other containers that implement the servlet specification may be affected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
136,529
5,284,300,172
IssuesEvent
2017-02-07 23:49:53
openshift/origin
https://api.github.com/repos/openshift/origin
closed
v2 gear to v3 pod conversion
component/image kind/enhancement priority/P2
Hello, In today's community meeting, I spoke with @smarterclayton about what, if anything, could be done in the form of tooling for ose/origin v2 gears that can then be converted to some form of usable template/pod in the v3 world. In my "napkin design" it seems like the process will follow something similar to: 1. take a gear backup.zip 2. break it down to discover components 3. If something is in the data dir create PVC (with default size, at some point allow for custom size to be requested) 4. If DB discovered create the appropriate template for the needed DB (inquire about persistence vs not, add another PVC with default sizing) 5. Separate out the code portion and allow user to plug in URL to their SC system (github, local gitlab, whatever) and push it out. Add to template with new source location 6. Go ahead and provision a project and create the services needed 7. attempt DB data import/tell user they have to import their DB data, here are some links on how to do that.
1.0
v2 gear to v3 pod conversion - Hello, In today's community meeting, I spoke with @smarterclayton about what, if anything, could be done in the form of tooling for ose/origin v2 gears that can then be converted to some form of usable template/pod in the v3 world. In my "napkin design" it seems like the process will follow something similar to: 1. take a gear backup.zip 2. break it down to discover components 3. If something is in the data dir create PVC (with default size, at some point allow for custom size to be requested) 4. If DB discovered create the appropriate template for the needed DB (inquire about persistence vs not, add another PVC with default sizing) 5. Separate out the code portion and allow user to plug in URL to their SC system (github, local gitlab, whatever) and push it out. Add to template with new source location 6. Go ahead and provision a project and create the services needed 7. attempt DB data import/tell user they have to import their DB data, here are some links on how to do that.
non_process
gear to pod conversion hello in today s community meeting i spoke with smarterclayton about what if anything could be done in the form of tooling for ose origin gears that can then be converted to some form of usable template pod in the world in my napkin design it seems like the process will follow something similar to take a gear backup zip break it down to discover components if something is in the data dir create pvc with default size at some point allow for custom size to be requested if db discovered create the appropriate template for the needed db inquire about persistence vs not add another pvc with default sizing separate out the code portion and allow user to plug in url to their sc system github local gitlab whatever and push it out add to template with new source location go ahead and provision a project and create the services needed attempt db data import tell user they have to import their db data here are some links on how to do that
0
244,477
7,875,420,697
IssuesEvent
2018-06-25 20:22:27
gctools-outilsgc/gcconnex
https://api.github.com/repos/gctools-outilsgc/gcconnex
closed
Polls: Ability to Close polls to retain and display results after poll periode has ended
Priority: Low Project: Legacy Tools Type: Enhancement [zube]: Inbox
Users only have the option to delete a poll once their desired poll period has ended preventing them from retaining the poll results for other users to view. Request: Allow users to close polls in order to be able to share the results once the poll period has ended. while retaining the ability to delete the poll once its use has expired. In addition to this, it would be beneficial to have the ability to not only select “All”, “Mine” or “Colleagues” when searching for polls, but also the ability to choose “Open”, “Closed” or “All” as a sub-selection to each of the aforementioned selection options. this will provide more flexibility in the use of the Poll feature.
1.0
Polls: Ability to Close polls to retain and display results after poll periode has ended - Users only have the option to delete a poll once their desired poll period has ended preventing them from retaining the poll results for other users to view. Request: Allow users to close polls in order to be able to share the results once the poll period has ended. while retaining the ability to delete the poll once its use has expired. In addition to this, it would be beneficial to have the ability to not only select “All”, “Mine” or “Colleagues” when searching for polls, but also the ability to choose “Open”, “Closed” or “All” as a sub-selection to each of the aforementioned selection options. this will provide more flexibility in the use of the Poll feature.
non_process
polls ability to close polls to retain and display results after poll periode has ended users only have the option to delete a poll once their desired poll period has ended preventing them from retaining the poll results for other users to view request allow users to close polls in order to be able to share the results once the poll period has ended while retaining the ability to delete the poll once its use has expired in addition to this it would be beneficial to have the ability to not only select “all” “mine” or “colleagues” when searching for polls but also the ability to choose “open” “closed” or “all” as a sub selection to each of the aforementioned selection options this will provide more flexibility in the use of the poll feature
0
125,657
16,823,983,439
IssuesEvent
2021-06-17 16:05:04
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
[Variant Test] Product Checklist - Sorting Facilities
product vaos-product-design
# Summary ## Tasks - [ ] Summary [documentation in GitHub](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/health-care/appointments/va-online-scheduling/initiatives/variant-tests/2021-sorting-facilities.md) - [ ] Designs - [Sketch file](https://www.sketch.com/s/439713c6-cb08-4b65-94ab-d97dcac0beba) (In progress) - [ ] Dev ticket(s) - #24280 (dev) - #26106 (events for analytics) - [ ] [Results readout slides](https://docs.google.com/presentation/d/1SvwwoiDa6OrzAuTWwetqAz2WUH3j4z0Q73muJXfwipk/edit?usp=sharing) - [ ] Benchmark - [ ] Outcomes
1.0
[Variant Test] Product Checklist - Sorting Facilities - # Summary ## Tasks - [ ] Summary [documentation in GitHub](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/health-care/appointments/va-online-scheduling/initiatives/variant-tests/2021-sorting-facilities.md) - [ ] Designs - [Sketch file](https://www.sketch.com/s/439713c6-cb08-4b65-94ab-d97dcac0beba) (In progress) - [ ] Dev ticket(s) - #24280 (dev) - #26106 (events for analytics) - [ ] [Results readout slides](https://docs.google.com/presentation/d/1SvwwoiDa6OrzAuTWwetqAz2WUH3j4z0Q73muJXfwipk/edit?usp=sharing) - [ ] Benchmark - [ ] Outcomes
non_process
product checklist sorting facilities summary tasks summary designs in progress dev ticket s dev events for analytics benchmark outcomes
0
9,341
12,342,808,451
IssuesEvent
2020-05-15 01:59:57
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `Rpad` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `Rpad` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `Rpad` from TiDB - ## Description Port the scalar function `Rpad` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @andylokandy ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function rpad from tidb description port the scalar function rpad from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
1
169,776
13,160,672,205
IssuesEvent
2020-08-10 18:02:46
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
The table auto refresh after canceling the file import
:gear: tables 🧪 testing
**Storage Explorer Version:** 1.15.0-dev **Build**: 20200806.6 **Branch**: master **Platform/OS:** Windows 10/ CentOS 7.6.1810 (Core) / MacOS Catalina **Architecture**: ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one storage account -> Tables. 2. Create a new table -> Click 'Import' action in the toolbar -> Click 'Cancel' button in the file explorer. 3. Check the result. **Expect Experience:** The table doesn't refresh. (Unnecessary auto refreshing will make the table loading a long time when one table has a large number of entities). **Actual Experience:** The table auto refresh.
1.0
The table auto refresh after canceling the file import - **Storage Explorer Version:** 1.15.0-dev **Build**: 20200806.6 **Branch**: master **Platform/OS:** Windows 10/ CentOS 7.6.1810 (Core) / MacOS Catalina **Architecture**: ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one storage account -> Tables. 2. Create a new table -> Click 'Import' action in the toolbar -> Click 'Cancel' button in the file explorer. 3. Check the result. **Expect Experience:** The table doesn't refresh. (Unnecessary auto refreshing will make the table loading a long time when one table has a large number of entities). **Actual Experience:** The table auto refresh.
non_process
the table auto refresh after canceling the file import storage explorer version dev build branch master platform os windows centos core macos catalina architecture regression from not a regression steps to reproduce expand one storage account tables create a new table click import action in the toolbar click cancel button in the file explorer check the result expect experience the table doesn t refresh unnecessary auto refreshing will make the table loading a long time when one table has a large number of entities actual experience the table auto refresh
0
416,343
12,143,028,358
IssuesEvent
2020-04-24 03:34:46
FStarLang/FStar
https://api.github.com/repos/FStarLang/FStar
reopened
Proper escaping for --dep
component/dependency-analysis kind/enhancement priority/low
If you include a directory with spaces in it, then the dependencies generated by `--dep full` are not quoted, and thus it generates a broken Makefile. To reproduce: ``` mkdir "a b" cd "a b" cat > A.fst <<EOF module A EOF cd .. cat > B.fst <<EOF module B open A EOF fstar.exe --include "a b" --dep full B.fst ``` One of the parts causing the issue: ``` /tmp/asd/a b/A.fst.checked: /tmp/asd/a b/A.fst \ /home/jay/everest/FStar/ulib/FStar.Pervasives.fst.checked\ /home/jay/everest/FStar/ulib/prims.fst.checked ``` Expected output: ``` /tmp/asd/a\ b/A.fst.checked: /tmp/asd/a\ b/A.fst \ /home/jay/everest/FStar/ulib/FStar.Pervasives.fst.checked\ /home/jay/everest/FStar/ulib/prims.fst.checked ``` Basically, all spaces in filenames need to be replaced with `\ ` instead. It does indeed seem to do this in one place (in particular, the first rule), but then stops doing so. Not a very high priority issue, but I thought it best to have it documented.
1.0
Proper escaping for --dep - If you include a directory with spaces in it, then the dependencies generated by `--dep full` are not quoted, and thus it generates a broken Makefile. To reproduce: ``` mkdir "a b" cd "a b" cat > A.fst <<EOF module A EOF cd .. cat > B.fst <<EOF module B open A EOF fstar.exe --include "a b" --dep full B.fst ``` One of the parts causing the issue: ``` /tmp/asd/a b/A.fst.checked: /tmp/asd/a b/A.fst \ /home/jay/everest/FStar/ulib/FStar.Pervasives.fst.checked\ /home/jay/everest/FStar/ulib/prims.fst.checked ``` Expected output: ``` /tmp/asd/a\ b/A.fst.checked: /tmp/asd/a\ b/A.fst \ /home/jay/everest/FStar/ulib/FStar.Pervasives.fst.checked\ /home/jay/everest/FStar/ulib/prims.fst.checked ``` Basically, all spaces in filenames need to be replaced with `\ ` instead. It does indeed seem to do this in one place (in particular, the first rule), but then stops doing so. Not a very high priority issue, but I thought it best to have it documented.
non_process
proper escaping for dep if you include a directory with spaces in it then the dependencies generated by dep full are not quoted and thus it generates a broken makefile to reproduce mkdir a b cd a b cat a fst eof module a eof cd cat b fst eof module b open a eof fstar exe include a b dep full b fst one of the parts causing the issue tmp asd a b a fst checked tmp asd a b a fst home jay everest fstar ulib fstar pervasives fst checked home jay everest fstar ulib prims fst checked expected output tmp asd a b a fst checked tmp asd a b a fst home jay everest fstar ulib fstar pervasives fst checked home jay everest fstar ulib prims fst checked basically all spaces in filenames need to be replaced with instead it does indeed seem to do this in one place in particular the first rule but then stops doing so not a very high priority issue but i thought it best to have it documented
0
212,216
7,229,417,644
IssuesEvent
2018-02-11 19:38:04
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
[Coverity CID: 182769] Error handling issues in /subsys/bluetooth/host/mesh/beacon.c
Coverity area: Bluetooth bug priority: medium
Static code scan issues seen in File: /subsys/bluetooth/host/mesh/beacon.c Category: Error handling issues Function: unprovisioned_beacon_send Component: Bluetooth CID: 182769 Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996
1.0
[Coverity CID: 182769] Error handling issues in /subsys/bluetooth/host/mesh/beacon.c - Static code scan issues seen in File: /subsys/bluetooth/host/mesh/beacon.c Category: Error handling issues Function: unprovisioned_beacon_send Component: Bluetooth CID: 182769 Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996
non_process
error handling issues in subsys bluetooth host mesh beacon c static code scan issues seen in file subsys bluetooth host mesh beacon c category error handling issues function unprovisioned beacon send component bluetooth cid please fix or provide comments to square it off in coverity in the link
0
561,763
16,623,450,024
IssuesEvent
2021-06-03 06:29:30
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Custom filter expression does not show editor properly on subsequent click
.Frontend .Reproduced Priority:P2 Querying/Notebook Type:Bug
**Describe the bug** In the notebook query editor, if a Filter is added with a Custom expression, and added, then that clause is clicked again, the editor that appears is not the correct one. **Logs** N/A **To Reproduce** Steps to reproduce the behavior: 1. Go to a Notebook question editor 2. Click on the + under Filter, in order to add a new filter clause 3. Select Custom expression from the dropdown, and enter some expression (ex: `between([ID], 96, 97)`); see screenshot 1 4. Click Done to add the clause 5. Click on the clause again (ex: to make a change) 6. Notice the dialog that appears is blank (doesn't show the expression value); see screenshot 2 **Expected behavior** The same dialog that appears when adding the custom filter expression (ex: from step 3 above) should appear, and the custom expression that was already populated should be prefilled. As a workaround, you can click the left/back arrow shown in the dialog header, then select Custom expression again, and the expression will again be editable there. **Screenshots** 1. <img width="495" alt="Screen Shot 2021-02-09 at 1 35 43 PM" src="https://user-images.githubusercontent.com/3521562/107419654-a6948500-6add-11eb-830e-5a509f6f8e1b.png"> 2. <img width="332" alt="Screen Shot 2021-02-09 at 1 35 48 PM" src="https://user-images.githubusercontent.com/3521562/107419665-aac0a280-6add-11eb-8737-3f4099df6cd6.png"> **Information about your Metabase Installation:** You can get this information by going to Admin -> Troubleshooting. - Your browser and the version: Chrome 88.0.4324.150 - Your operating system: OS X 11.1 - Your databases: Redshift - Metabase version: `0.38.1-SNAPSHOT` (src_hash=`6d77f0992f5f51345554bc34338f05786f26c54b`) - Metabase hosting environment: N/A - Metabase internal database: H2 **Severity** Minor. **Additional context** N/A
1.0
Custom filter expression does not show editor properly on subsequent click - **Describe the bug** In the notebook query editor, if a Filter is added with a Custom expression, and added, then that clause is clicked again, the editor that appears is not the correct one. **Logs** N/A **To Reproduce** Steps to reproduce the behavior: 1. Go to a Notebook question editor 2. Click on the + under Filter, in order to add a new filter clause 3. Select Custom expression from the dropdown, and enter some expression (ex: `between([ID], 96, 97)`); see screenshot 1 4. Click Done to add the clause 5. Click on the clause again (ex: to make a change) 6. Notice the dialog that appears is blank (doesn't show the expression value); see screenshot 2 **Expected behavior** The same dialog that appears when adding the custom filter expression (ex: from step 3 above) should appear, and the custom expression that was already populated should be prefilled. As a workaround, you can click the left/back arrow shown in the dialog header, then select Custom expression again, and the expression will again be editable there. **Screenshots** 1. <img width="495" alt="Screen Shot 2021-02-09 at 1 35 43 PM" src="https://user-images.githubusercontent.com/3521562/107419654-a6948500-6add-11eb-830e-5a509f6f8e1b.png"> 2. <img width="332" alt="Screen Shot 2021-02-09 at 1 35 48 PM" src="https://user-images.githubusercontent.com/3521562/107419665-aac0a280-6add-11eb-8737-3f4099df6cd6.png"> **Information about your Metabase Installation:** You can get this information by going to Admin -> Troubleshooting. - Your browser and the version: Chrome 88.0.4324.150 - Your operating system: OS X 11.1 - Your databases: Redshift - Metabase version: `0.38.1-SNAPSHOT` (src_hash=`6d77f0992f5f51345554bc34338f05786f26c54b`) - Metabase hosting environment: N/A - Metabase internal database: H2 **Severity** Minor. **Additional context** N/A
non_process
custom filter expression does not show editor properly on subsequent click describe the bug in the notebook query editor if a filter is added with a custom expression and added then that clause is clicked again the editor that appears is not the correct one logs n a to reproduce steps to reproduce the behavior go to a notebook question editor click on the under filter in order to add a new filter clause select custom expression from the dropdown and enter some expression ex between see screenshot click done to add the clause click on the clause again ex to make a change notice the dialog that appears is blank doesn t show the expression value see screenshot expected behavior the same dialog that appears when adding the custom filter expression ex from step above should appear and the custom expression that was already populated should be prefilled as a workaround you can click the left back arrow shown in the dialog header then select custom expression again and the expression will again be editable there screenshots img width alt screen shot at pm src img width alt screen shot at pm src information about your metabase installation you can get this information by going to admin troubleshooting your browser and the version chrome your operating system os x your databases redshift metabase version snapshot src hash metabase hosting environment n a metabase internal database severity minor additional context n a
0
813,391
30,455,816,738
IssuesEvent
2023-07-16 21:48:56
GoogleCloudPlatform/java-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/java-docs-samples
closed
The build failed
type: bug priority: p1 samples flakybot: issue
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: da33713f879e447658cf452dea5f17a0827cddd6 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/4d0d6834-0b98-4f5f-a020-d9af8ddcff7f), [Sponge](http://sponge2/4d0d6834-0b98-4f5f-a020-d9af8ddcff7f) status: failed <details><summary>Test output</summary><br><pre>java.lang.RuntimeException: Base URL not found. at com.example.filesystem.ApplicationTests.setup(ApplicationTests.java:111) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.springframework.test.context.junit4.statements.RunBeforeTestClassCallbacks.evaluate(RunBeforeTestClassCallbacks.java:61) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.springframework.test.context.junit4.statements.RunAfterTestClassCallbacks.evaluate(RunAfterTestClassCallbacks.java:70) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.run(SpringJUnit4ClassRunner.java:191) at org.junit.runner.JUnitCore.run(JUnitCore.java:137) at org.junit.runner.JUnitCore.run(JUnitCore.java:115) at org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42) at org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80) at org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:102) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:54) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86) at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86) at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:188) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:128) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) </pre></details>
1.0
The build failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: da33713f879e447658cf452dea5f17a0827cddd6 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/4d0d6834-0b98-4f5f-a020-d9af8ddcff7f), [Sponge](http://sponge2/4d0d6834-0b98-4f5f-a020-d9af8ddcff7f) status: failed <details><summary>Test output</summary><br><pre>java.lang.RuntimeException: Base URL not found. at com.example.filesystem.ApplicationTests.setup(ApplicationTests.java:111) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.base/java.lang.reflect.Method.invoke(Method.java:568) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33) at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24) at org.springframework.test.context.junit4.statements.RunBeforeTestClassCallbacks.evaluate(RunBeforeTestClassCallbacks.java:61) at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) at org.springframework.test.context.junit4.statements.RunAfterTestClassCallbacks.evaluate(RunAfterTestClassCallbacks.java:70) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.run(SpringJUnit4ClassRunner.java:191) at org.junit.runner.JUnitCore.run(JUnitCore.java:137) at org.junit.runner.JUnitCore.run(JUnitCore.java:115) at org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42) at org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80) at org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:102) at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:54) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114) at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86) at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86) at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:188) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154) at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:128) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) </pre></details>
non_process
the build failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output java lang runtimeexception base url not found at com example filesystem applicationtests setup applicationtests java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements runbefores invokemethod runbefores java at org junit internal runners statements runbefores evaluate runbefores java at org springframework test context statements runbeforetestclasscallbacks evaluate runbeforetestclasscallbacks java at org junit internal runners statements runafters evaluate runafters java at org springframework test context statements runaftertestclasscallbacks evaluate runaftertestclasscallbacks java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org springframework test context run java at org junit runner junitcore run junitcore java at org junit runner junitcore run junitcore java at org junit vintage engine execution runnerexecutor execute runnerexecutor java at org junit vintage engine vintagetestengine executeallchildren vintagetestengine java at org junit vintage engine vintagetestengine execute vintagetestengine java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator lambda execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator withinterceptedstreams engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlaunchersession delegatinglauncher execute defaultlaunchersession java at org junit platform launcher core sessionperrequestlauncher execute sessionperrequestlauncher java at org apache maven surefire junitplatform junitplatformprovider execute junitplatformprovider java at org apache maven surefire junitplatform junitplatformprovider invokealltests junitplatformprovider java at org apache maven surefire junitplatform junitplatformprovider invoke junitplatformprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java
0
117,698
11,954,253,671
IssuesEvent
2020-04-03 23:00:48
rdmtc/node-red-contrib-sun-position
https://api.github.com/repos/rdmtc/node-red-contrib-sun-position
closed
Node installation not working
👎 wontfix 📖 documentation
Hi, I like to install your node on my RPI4. But I get errors Steps to reproduce the behavior: 1. Searched in Node Red 2. Installed your node 3. See error The error: node-red-contrib-sun-position Cannot read property 'nodes' of null ![grafik](https://user-images.githubusercontent.com/15846090/78336766-37323880-7590-11ea-8580-8a5e7027bc57.png) System information - Node Version [ 1.0.16] - Node-Red Version [ 0.20.8] - Hardware [e.g. RaspberryPi3, TinkerboardS, etc.] Any help is highly appriciated. best regards René
1.0
Node installation not working - Hi, I like to install your node on my RPI4. But I get errors Steps to reproduce the behavior: 1. Searched in Node Red 2. Installed your node 3. See error The error: node-red-contrib-sun-position Cannot read property 'nodes' of null ![grafik](https://user-images.githubusercontent.com/15846090/78336766-37323880-7590-11ea-8580-8a5e7027bc57.png) System information - Node Version [ 1.0.16] - Node-Red Version [ 0.20.8] - Hardware [e.g. RaspberryPi3, TinkerboardS, etc.] Any help is highly appriciated. best regards René
non_process
node installation not working hi i like to install your node on my but i get errors steps to reproduce the behavior searched in node red installed your node see error the error node red contrib sun position cannot read property nodes of null system information node version node red version hardware any help is highly appriciated best regards rené
0
259,296
8,196,769,256
IssuesEvent
2018-08-31 11:00:59
threefoldfoundation/www_threefold.io
https://api.github.com/repos/threefoldfoundation/www_threefold.io
closed
Join and Follow Page: clean up
priority_major
On the current page, there are multiple links. If we are to clean and rationalise we can follow this decision tree: A. Are we really dissatisfied with the page? B. If we are, is it us that is dissatisfied? Or do our audience not like it? C. Do we want to poll a fully representative sample of them to ask? D. If they give us an answer we don't like, are we sure we If none of the above: E. Are we satisfied with simply making the links smaller? If not: F. Is this actually an issue that we do not like having to have (and maintain) multiple social platforms? G. If we do not like F. Is this only our view - or that of our audience? H. Do we want to poll a fully representative sample of them to ask? I. If they give us an answer we don't like, are we sure we If none of the above it implies we want to get rid of some or all of the links. J. Which do we want to keep? K. Once this is decided. Action point J.
1.0
Join and Follow Page: clean up - On the current page, there are multiple links. If we are to clean and rationalise we can follow this decision tree: A. Are we really dissatisfied with the page? B. If we are, is it us that is dissatisfied? Or do our audience not like it? C. Do we want to poll a fully representative sample of them to ask? D. If they give us an answer we don't like, are we sure we If none of the above: E. Are we satisfied with simply making the links smaller? If not: F. Is this actually an issue that we do not like having to have (and maintain) multiple social platforms? G. If we do not like F. Is this only our view - or that of our audience? H. Do we want to poll a fully representative sample of them to ask? I. If they give us an answer we don't like, are we sure we If none of the above it implies we want to get rid of some or all of the links. J. Which do we want to keep? K. Once this is decided. Action point J.
non_process
join and follow page clean up on the current page there are multiple links if we are to clean and rationalise we can follow this decision tree a are we really dissatisfied with the page b if we are is it us that is dissatisfied or do our audience not like it c do we want to poll a fully representative sample of them to ask d if they give us an answer we don t like are we sure we if none of the above e are we satisfied with simply making the links smaller if not f is this actually an issue that we do not like having to have and maintain multiple social platforms g if we do not like f is this only our view or that of our audience h do we want to poll a fully representative sample of them to ask i if they give us an answer we don t like are we sure we if none of the above it implies we want to get rid of some or all of the links j which do we want to keep k once this is decided action point j
0
14,306
17,294,755,166
IssuesEvent
2021-07-25 13:51:56
googleapis/python-bigquery
https://api.github.com/repos/googleapis/python-bigquery
closed
test: System tests create and destroy way too many datasets
api: bigquery testing type: process
Every system tests creates and destroys a temporary dataset in it's setUp and tearDown. - Some system tests don't use the dataset created. - Some system tests create and destroy their own datasets, seemingly oblivious that temporary datasets were already created for them. This makes system tests take far longer than they should. ~10 minutes Proposed fix: - Don't create datasets by default. Make tests elect to create them. - Move toward pytest fixtures to elect creating datasets. - Use session fixtures to avoid creating a dataset for each test.
1.0
test: System tests create and destroy way too many datasets - Every system tests creates and destroys a temporary dataset in it's setUp and tearDown. - Some system tests don't use the dataset created. - Some system tests create and destroy their own datasets, seemingly oblivious that temporary datasets were already created for them. This makes system tests take far longer than they should. ~10 minutes Proposed fix: - Don't create datasets by default. Make tests elect to create them. - Move toward pytest fixtures to elect creating datasets. - Use session fixtures to avoid creating a dataset for each test.
process
test system tests create and destroy way too many datasets every system tests creates and destroys a temporary dataset in it s setup and teardown some system tests don t use the dataset created some system tests create and destroy their own datasets seemingly oblivious that temporary datasets were already created for them this makes system tests take far longer than they should minutes proposed fix don t create datasets by default make tests elect to create them move toward pytest fixtures to elect creating datasets use session fixtures to avoid creating a dataset for each test
1
1,775
4,489,075,931
IssuesEvent
2016-08-30 09:40:54
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
opened
Add in missing relationship for 'viral release from host cell'
multiorganism processes Other term-related request quick fix viruses
A reminder to add in missing relationship: viral release from host cell ; GO:0019076 is_a: exit from host cell ; GO:0035891
1.0
Add in missing relationship for 'viral release from host cell' - A reminder to add in missing relationship: viral release from host cell ; GO:0019076 is_a: exit from host cell ; GO:0035891
process
add in missing relationship for viral release from host cell a reminder to add in missing relationship viral release from host cell go is a exit from host cell go
1
8,614
11,770,629,953
IssuesEvent
2020-03-15 20:05:34
pawn-lang/compiler
https://api.github.com/repos/pawn-lang/compiler
closed
#pragma option limited to 31 characters
area: pre-processor good first issue state: stale
<!-- pawn-lang compiler issue template --> <!-- Please ensure that your issue is directly related to the pawn-lang/compiler. Issues seeking help with scripting or doubts about SA-MP are not relevant to this project. --> ### Issue description: <!-- Information that could be useful: - how to reproduce the problem - observed behavior - expected behavior - other relevant information --> Using an option over 31 characters (I assume 32 with NULL) crashes the compiler. ### Minimal complete verifiable example (MCVE): <!-- Please provide a minimal complete verifiable code which compiles and reproduces the problem. If the compiler is showing undefined (random) behavior, provide the code which invokes such behavior. Learn more about writing MCVE from [StackOverflow](https://stackoverflow.com/help/mcve). If this section is not relevant, feel free to remove this section from your issue. --> ```pawn // It doesn't even matter if this exists or not. #pragma option -idependencies\y_thirdpartyinclude ``` <!-- comments on the code, if any --> ### Workspace Information: * Compiler version: * Command line arguments provided (or sampctl version): * Operating System:
1.0
#pragma option limited to 31 characters - <!-- pawn-lang compiler issue template --> <!-- Please ensure that your issue is directly related to the pawn-lang/compiler. Issues seeking help with scripting or doubts about SA-MP are not relevant to this project. --> ### Issue description: <!-- Information that could be useful: - how to reproduce the problem - observed behavior - expected behavior - other relevant information --> Using an option over 31 characters (I assume 32 with NULL) crashes the compiler. ### Minimal complete verifiable example (MCVE): <!-- Please provide a minimal complete verifiable code which compiles and reproduces the problem. If the compiler is showing undefined (random) behavior, provide the code which invokes such behavior. Learn more about writing MCVE from [StackOverflow](https://stackoverflow.com/help/mcve). If this section is not relevant, feel free to remove this section from your issue. --> ```pawn // It doesn't even matter if this exists or not. #pragma option -idependencies\y_thirdpartyinclude ``` <!-- comments on the code, if any --> ### Workspace Information: * Compiler version: * Command line arguments provided (or sampctl version): * Operating System:
process
pragma option limited to characters please ensure that your issue is directly related to the pawn lang compiler issues seeking help with scripting or doubts about sa mp are not relevant to this project issue description information that could be useful how to reproduce the problem observed behavior expected behavior other relevant information using an option over characters i assume with null crashes the compiler minimal complete verifiable example mcve please provide a minimal complete verifiable code which compiles and reproduces the problem if the compiler is showing undefined random behavior provide the code which invokes such behavior learn more about writing mcve from if this section is not relevant feel free to remove this section from your issue pawn it doesn t even matter if this exists or not pragma option idependencies y thirdpartyinclude workspace information compiler version command line arguments provided or sampctl version operating system
1
22,634
31,881,989,680
IssuesEvent
2023-09-16 13:37:49
pex-gl/pex-renderer
https://api.github.com/repos/pex-gl/pex-renderer
closed
View frustum culling
type/feat scope/post-processsing
Crucial for optimizing rendering of big scenes Lighthouse3D tutorial on View Frustum Culling http://www.lighthouse3d.com/tutorials/view-frustum-culling/
1.0
View frustum culling - Crucial for optimizing rendering of big scenes Lighthouse3D tutorial on View Frustum Culling http://www.lighthouse3d.com/tutorials/view-frustum-culling/
process
view frustum culling crucial for optimizing rendering of big scenes tutorial on view frustum culling
1
45,141
5,907,922,663
IssuesEvent
2017-05-19 18:57:02
ludumdare/ludumdare
https://api.github.com/repos/ludumdare/ludumdare
opened
List Making
core design enhancement
This would be **SUPER USEFUL** to have. I have a few cases where I've needed to users to provide me list. If this was an actual feature, I could actually run queries on the raw lists I'm provided.
1.0
List Making - This would be **SUPER USEFUL** to have. I have a few cases where I've needed to users to provide me list. If this was an actual feature, I could actually run queries on the raw lists I'm provided.
non_process
list making this would be super useful to have i have a few cases where i ve needed to users to provide me list if this was an actual feature i could actually run queries on the raw lists i m provided
0
92,066
15,856,856,255
IssuesEvent
2021-04-08 03:20:57
DaniOlivares/test-truffle-oz-tutorial-token
https://api.github.com/repos/DaniOlivares/test-truffle-oz-tutorial-token
opened
CVE-2018-20676 (Medium) detected in bootstrap-3.3.7.min.js
security vulnerability
## CVE-2018-20676 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: test-truffle-oz-tutorial-token/src/index.html</p> <p>Path to vulnerable library: test-truffle-oz-tutorial-token/src/js/bootstrap.min.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20676 (Medium) detected in bootstrap-3.3.7.min.js - ## CVE-2018-20676 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: test-truffle-oz-tutorial-token/src/index.html</p> <p>Path to vulnerable library: test-truffle-oz-tutorial-token/src/js/bootstrap.min.js</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: bootstrap - 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file test truffle oz tutorial token src index html path to vulnerable library test truffle oz tutorial token src js bootstrap min js dependency hierarchy x bootstrap min js vulnerable library vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap step up your open source security game with whitesource
0
22,115
30,644,964,447
IssuesEvent
2023-07-25 03:18:17
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[MLv2] [Bug] `join-alias` is only added to the first condition's field reference
.Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
It seems like the `join-alias` is only added to the first condition's RHS field reference. When adding more conditions to the same join, the `join-alias` is missing. Here's an example query with two conditions: <details> <summary>Expand snippet</summary> ```js { database: 1, type: "query", query: { "source-table": 2, joins: [ { fields: "all", alias: "Products", condition: [ "and", [ "=", ["field", 13, { "base-type": "type/Integer" }], [ "field", 5, { "base-type": "type/BigInteger", "join-alias": "Products" }, ], ], [ "=", ["field", 14, { "base-type": "type/DateTime" }], ["field", 7, { "base-type": "type/DateTime" }], ], ], "source-table": 1, }, ], }, } ``` </details> ### To reproduce 1. Start a query from the sample Orders table 2. Join the Products table on Orders.PRODUCT_ID = Products.ID (using `Lib.joinConditionClause` + `Lib.join`) 3. Use `Lib.withJoinConditions` + `Lib.replaceClause` and add another join condition (e.g. Orders.CREATED_AT = Products.CREATED_AT) 4. Inspect the second condition's field reference **Expected:** RHS field reference has a `join-alias` **Actual**: RHS field reference only has a common set of options and gets dropped as invalid
1.0
[MLv2] [Bug] `join-alias` is only added to the first condition's field reference - It seems like the `join-alias` is only added to the first condition's RHS field reference. When adding more conditions to the same join, the `join-alias` is missing. Here's an example query with two conditions: <details> <summary>Expand snippet</summary> ```js { database: 1, type: "query", query: { "source-table": 2, joins: [ { fields: "all", alias: "Products", condition: [ "and", [ "=", ["field", 13, { "base-type": "type/Integer" }], [ "field", 5, { "base-type": "type/BigInteger", "join-alias": "Products" }, ], ], [ "=", ["field", 14, { "base-type": "type/DateTime" }], ["field", 7, { "base-type": "type/DateTime" }], ], ], "source-table": 1, }, ], }, } ``` </details> ### To reproduce 1. Start a query from the sample Orders table 2. Join the Products table on Orders.PRODUCT_ID = Products.ID (using `Lib.joinConditionClause` + `Lib.join`) 3. Use `Lib.withJoinConditions` + `Lib.replaceClause` and add another join condition (e.g. Orders.CREATED_AT = Products.CREATED_AT) 4. Inspect the second condition's field reference **Expected:** RHS field reference has a `join-alias` **Actual**: RHS field reference only has a common set of options and gets dropped as invalid
process
join alias is only added to the first condition s field reference it seems like the join alias is only added to the first condition s rhs field reference when adding more conditions to the same join the join alias is missing here s an example query with two conditions expand snippet js database type query query source table joins fields all alias products condition and field base type type biginteger join alias products source table to reproduce start a query from the sample orders table join the products table on orders product id products id using lib joinconditionclause lib join use lib withjoinconditions lib replaceclause and add another join condition e g orders created at products created at inspect the second condition s field reference expected rhs field reference has a join alias actual rhs field reference only has a common set of options and gets dropped as invalid
1
20,183
26,744,720,521
IssuesEvent
2023-01-30 15:16:33
polarby/render
https://api.github.com/repos/polarby/render
closed
Server side rendering
question render processing waiting for response
Hello, is there a way to render the widget to an image on the server side, maybe with headless flutter?
1.0
Server side rendering - Hello, is there a way to render the widget to an image on the server side, maybe with headless flutter?
process
server side rendering hello is there a way to render the widget to an image on the server side maybe with headless flutter
1
16,618
2,920,435,340
IssuesEvent
2015-06-24 18:55:24
ashanbh/chrome-rest-client
https://api.github.com/repos/ashanbh/chrome-rest-client
closed
add a feature to generate a curl command from the request
auto-migrated Priority-Medium Type-Defect Usability
``` Feature - todo ``` Original issue reported on code.google.com by `pawel.ps...@kalicinscy.com` on 12 Dec 2012 at 10:29
1.0
add a feature to generate a curl command from the request - ``` Feature - todo ``` Original issue reported on code.google.com by `pawel.ps...@kalicinscy.com` on 12 Dec 2012 at 10:29
non_process
add a feature to generate a curl command from the request feature todo original issue reported on code google com by pawel ps kalicinscy com on dec at
0
14,860
11,204,336,028
IssuesEvent
2020-01-05 04:39:27
MultiMC/MultiMC5
https://api.github.com/repos/MultiMC/MultiMC5
closed
MultiMC should be possible to translate before a stable release
multimc: feature multimc: infrastructure multimc: translations
I have been translating to spanish using weblate, but I can´t find some still pending strings using the search function in weblate, for example: "Close", "Help" buttons in edit instance form "Installation de Forge..." button in edit instance form
1.0
MultiMC should be possible to translate before a stable release - I have been translating to spanish using weblate, but I can´t find some still pending strings using the search function in weblate, for example: "Close", "Help" buttons in edit instance form "Installation de Forge..." button in edit instance form
non_process
multimc should be possible to translate before a stable release i have been translating to spanish using weblate but i can´t find some still pending strings using the search function in weblate for example close help buttons in edit instance form installation de forge button in edit instance form
0
12,498
14,961,464,570
IssuesEvent
2021-01-27 07:48:43
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
Japanese character is garbled.
bug mail processing verified
<!-- Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓 Since november 15th we handle all requests, except real bugs, at our community board. Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21 Please post: - Feature requests - Development questions - Technical questions on the board -> https://community.zammad.org ! If you think you hit a bug, please continue: - Search existing issues and the CHANGELOG.md for your issue - there might be a solution already - Make sure to use the latest version of Zammad if possible - Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it! - Please write the issue in english - Don't remove the template - otherwise we will close the issue without further comments - Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted). * The upper textblock will be removed automatically when you submit your issue * --> ### Infos: * Used Zammad version: 3.6.x * Installation method (source, package, ..): apt install zammad * Operating system: ubuntu 20.04 * Database + version: psql (PostgreSQL) 12.5 (Ubuntu 12.5-0ubuntu0.20.04.1) * Elasticsearch version: sorry I couldn't find * Browser + version: FireFox 84.0.2 ### Expected behavior: e-mail at Japanese-encoding displayed by Japanese character. ### Actual behavior: Some e-mail at Japanese-encoding displayed garbled character. (subject of e-mail displayed correctlly.only main text of mail is garbled.) Similar problem like Japanese encoding 7-bit iso-2022-jp is processed incorrectly #2498 but above issue is closed. so I make this one. ### Steps to reproduce the behavior: ・I will send eml file to support. Yes I'm sure this is a bug and no feature request or a general question.
1.0
Japanese character is garbled. - <!-- Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓 Since november 15th we handle all requests, except real bugs, at our community board. Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21 Please post: - Feature requests - Development questions - Technical questions on the board -> https://community.zammad.org ! If you think you hit a bug, please continue: - Search existing issues and the CHANGELOG.md for your issue - there might be a solution already - Make sure to use the latest version of Zammad if possible - Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it! - Please write the issue in english - Don't remove the template - otherwise we will close the issue without further comments - Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted). * The upper textblock will be removed automatically when you submit your issue * --> ### Infos: * Used Zammad version: 3.6.x * Installation method (source, package, ..): apt install zammad * Operating system: ubuntu 20.04 * Database + version: psql (PostgreSQL) 12.5 (Ubuntu 12.5-0ubuntu0.20.04.1) * Elasticsearch version: sorry I couldn't find * Browser + version: FireFox 84.0.2 ### Expected behavior: e-mail at Japanese-encoding displayed by Japanese character. ### Actual behavior: Some e-mail at Japanese-encoding displayed garbled character. (subject of e-mail displayed correctlly.only main text of mail is garbled.) Similar problem like Japanese encoding 7-bit iso-2022-jp is processed incorrectly #2498 but above issue is closed. so I make this one. ### Steps to reproduce the behavior: ・I will send eml file to support. Yes I'm sure this is a bug and no feature request or a general question.
process
japanese character is garbled hi there thanks for filing an issue please ensure the following things before creating an issue thank you 🤓 since november we handle all requests except real bugs at our community board full explanation please post feature requests development questions technical questions on the board if you think you hit a bug please continue search existing issues and the changelog md for your issue there might be a solution already make sure to use the latest version of zammad if possible add the log production log file from your system attention make sure no confidential data is in it please write the issue in english don t remove the template otherwise we will close the issue without further comments ask questions about zammad configuration and usage at our mailinglist see note we always do our best unfortunately sometimes there are too many requests and we can t handle everything at once if you want to prioritize escalate your issue you can do so by means of a support contract see the upper textblock will be removed automatically when you submit your issue infos used zammad version x installation method source package apt install zammad operating system ubuntu database version psql postgresql ubuntu elasticsearch version sorry i couldn t find browser version firefox expected behavior e mail at japanese encoding displayed by japanese character actual behavior some e mail at japanese encoding displayed garbled character subject of e mail displayed correctlly only main text of mail is garbled similar problem like japanese encoding bit iso jp is processed incorrectly but above issue is closed so i make this one steps to reproduce the behavior ・i will send eml file to support yes i m sure this is a bug and no feature request or a general question
1
9,405
12,404,115,311
IssuesEvent
2020-05-21 15:00:55
checkifcovid/data-science-experiments
https://api.github.com/repos/checkifcovid/data-science-experiments
closed
Update preprocessing to allow ingestion of new schema for location data.
preprocessing
Update preprocessing to allow ingestion of new schema for location data. **Current:** ``` user: { ... postcode: string ... } ``` **New:** ``` user: { ... location: { postcode: string, street: string, coordinates: { lat: float, lon: float } } ... } ```
1.0
Update preprocessing to allow ingestion of new schema for location data. - Update preprocessing to allow ingestion of new schema for location data. **Current:** ``` user: { ... postcode: string ... } ``` **New:** ``` user: { ... location: { postcode: string, street: string, coordinates: { lat: float, lon: float } } ... } ```
process
update preprocessing to allow ingestion of new schema for location data update preprocessing to allow ingestion of new schema for location data current user postcode string new user location postcode string street string coordinates lat float lon float
1
270,066
8,446,036,717
IssuesEvent
2018-10-19 00:15:04
robot-lab/judyst-main-web-service
https://api.github.com/repos/robot-lab/judyst-main-web-service
opened
Страница настроек организации
area/front-end priority/high type/feature type/task
# Task request ## Цель задачи Страница с настройками для организации. На основании feature #82 и feature #86 ## Решение задачи Страница доступна только администратору организации и содержит поля необходимые в #82. Имеет ссылку на страницу организации. ## Дополнительный контекст или ссылки на связанные с данной задачей issues
1.0
Страница настроек организации - # Task request ## Цель задачи Страница с настройками для организации. На основании feature #82 и feature #86 ## Решение задачи Страница доступна только администратору организации и содержит поля необходимые в #82. Имеет ссылку на страницу организации. ## Дополнительный контекст или ссылки на связанные с данной задачей issues
non_process
страница настроек организации task request цель задачи страница с настройками для организации на основании feature и feature решение задачи страница доступна только администратору организации и содержит поля необходимые в имеет ссылку на страницу организации дополнительный контекст или ссылки на связанные с данной задачей issues
0
828,143
31,813,367,980
IssuesEvent
2023-09-13 18:30:23
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] Sequential scan and Index scan on primary index cost differently
kind/enhancement area/ysql priority/medium
Jira Link: [DB-7871](https://yugabyte.atlassian.net/browse/DB-7871) ### Description We see that sequential scans on a table and index scans on the table's primary index cost very differently even though they are effectively the same operation. Consider the following setup: ``` create table p100000(c1 int, primary key(c1 asc)); insert into p100000 select generate_series(1,100000); analyze p100000; set yb_enable_base_scans_cost_model to true; explain select * from p100000; QUERY PLAN ---------------------------------------------------------------- Seq Scan on p100000 (cost=7.20..51995.53 rows=100000 width=4) (1 row) set enable_seqscan to false; explain select * from p100000 order by c1 asc; QUERY PLAN -------------------------------------------------------------------------------------- Index Scan using p100000_pkey on p100000 (cost=7.20..64570.53 rows=100000 width=4) (1 row) ``` We see that these two queries that effectively do the same thing are costed very differently. Because of this, the cost model seems to believe that even sorting the sequential scan is more efficient than directly reading the desired ordering from the primary index in the following query: ``` set enable_seqscan to true; explain select * from p100000 order by c1 asc; QUERY PLAN ---------------------------------------------------------------------- Sort (cost=60300.35..60550.35 rows=100000 width=4) Sort Key: c1 -> Seq Scan on p100000 (cost=7.20..51995.53 rows=100000 width=4) (3 rows) ``` This needs to be investigated. ### Warning: Please confirm that this issue does not contain any sensitive information - [X] I confirm this issue does not contain any sensitive information. [DB-7871]: https://yugabyte.atlassian.net/browse/DB-7871?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
1.0
[YSQL] Sequential scan and Index scan on primary index cost differently - Jira Link: [DB-7871](https://yugabyte.atlassian.net/browse/DB-7871) ### Description We see that sequential scans on a table and index scans on the table's primary index cost very differently even though they are effectively the same operation. Consider the following setup: ``` create table p100000(c1 int, primary key(c1 asc)); insert into p100000 select generate_series(1,100000); analyze p100000; set yb_enable_base_scans_cost_model to true; explain select * from p100000; QUERY PLAN ---------------------------------------------------------------- Seq Scan on p100000 (cost=7.20..51995.53 rows=100000 width=4) (1 row) set enable_seqscan to false; explain select * from p100000 order by c1 asc; QUERY PLAN -------------------------------------------------------------------------------------- Index Scan using p100000_pkey on p100000 (cost=7.20..64570.53 rows=100000 width=4) (1 row) ``` We see that these two queries that effectively do the same thing are costed very differently. Because of this, the cost model seems to believe that even sorting the sequential scan is more efficient than directly reading the desired ordering from the primary index in the following query: ``` set enable_seqscan to true; explain select * from p100000 order by c1 asc; QUERY PLAN ---------------------------------------------------------------------- Sort (cost=60300.35..60550.35 rows=100000 width=4) Sort Key: c1 -> Seq Scan on p100000 (cost=7.20..51995.53 rows=100000 width=4) (3 rows) ``` This needs to be investigated. ### Warning: Please confirm that this issue does not contain any sensitive information - [X] I confirm this issue does not contain any sensitive information. [DB-7871]: https://yugabyte.atlassian.net/browse/DB-7871?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
non_process
sequential scan and index scan on primary index cost differently jira link description we see that sequential scans on a table and index scans on the table s primary index cost very differently even though they are effectively the same operation consider the following setup create table int primary key asc insert into select generate series analyze set yb enable base scans cost model to true explain select from query plan seq scan on cost rows width row set enable seqscan to false explain select from order by asc query plan index scan using pkey on cost rows width row we see that these two queries that effectively do the same thing are costed very differently because of this the cost model seems to believe that even sorting the sequential scan is more efficient than directly reading the desired ordering from the primary index in the following query set enable seqscan to true explain select from order by asc query plan sort cost rows width sort key seq scan on cost rows width rows this needs to be investigated warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information
0
15,318
19,425,956,620
IssuesEvent
2021-12-21 05:32:09
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
All GDAL algorithms in QGIS 3.23.0-Master: "Wrong or missing parameter value" error
Processing Regression Bug
### What is the bug or the crash? Just opening any GDAL algorithm's window using QGIS 3.23.0-Master (93399e33a3 - package qgis-dev-3.23.0-137-93399e33a3-1 from OSGeo4W v2) on Windows 10 the following error is displayed in the algorithm window and in the Log Messages panel: `WARNING Wrong or missing parameter value:` followed by the name of one of the missing parameters. ![image](https://user-images.githubusercontent.com/16253859/146534205-03219bbd-ed36-4789-a175-e1fa6db4c336.png) ### Steps to reproduce the issue 1. Open any GDAL algorithm from the Processing Toolbox panel ### Versions QGIS version 3.23.0-Master QGIS code revision 93399e33a3 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.0 PROJ version 8.2.0 EPSG Registry database version v10.038 (2021-10-21) GEOS version 3.10.0-CAPI-1.16.0 SQLite version 3.35.2 PDAL version 2.3.0 PostgreSQL client version unknown SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 1903 This copy of QGIS writes debugging output. Active Python plugins db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context The issue doesn't occur in QGIS 3.16.14 and QGIS 3.22.1 (on Windows 10 from OSGeo4W v2).
1.0
All GDAL algorithms in QGIS 3.23.0-Master: "Wrong or missing parameter value" error - ### What is the bug or the crash? Just opening any GDAL algorithm's window using QGIS 3.23.0-Master (93399e33a3 - package qgis-dev-3.23.0-137-93399e33a3-1 from OSGeo4W v2) on Windows 10 the following error is displayed in the algorithm window and in the Log Messages panel: `WARNING Wrong or missing parameter value:` followed by the name of one of the missing parameters. ![image](https://user-images.githubusercontent.com/16253859/146534205-03219bbd-ed36-4789-a175-e1fa6db4c336.png) ### Steps to reproduce the issue 1. Open any GDAL algorithm from the Processing Toolbox panel ### Versions QGIS version 3.23.0-Master QGIS code revision 93399e33a3 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.0 PROJ version 8.2.0 EPSG Registry database version v10.038 (2021-10-21) GEOS version 3.10.0-CAPI-1.16.0 SQLite version 3.35.2 PDAL version 2.3.0 PostgreSQL client version unknown SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 1903 This copy of QGIS writes debugging output. Active Python plugins db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context The issue doesn't occur in QGIS 3.16.14 and QGIS 3.22.1 (on Windows 10 from OSGeo4W v2).
process
all gdal algorithms in qgis master wrong or missing parameter value error what is the bug or the crash just opening any gdal algorithm s window using qgis master package qgis dev from on windows the following error is displayed in the algorithm window and in the log messages panel warning wrong or missing parameter value followed by the name of one of the missing parameters steps to reproduce the issue open any gdal algorithm from the processing toolbox panel versions qgis version master qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version windows version this copy of qgis writes debugging output active python plugins db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context the issue doesn t occur in qgis and qgis on windows from
1
15,780
27,884,229,631
IssuesEvent
2023-03-21 22:05:53
adrrf/acme-life-long-learning
https://api.github.com/repos/adrrf/acme-life-long-learning
opened
d3-s3-fr-15 - assistant operations on sessions
develop functional_requirement mandatory
Operations by assistants on sessions: - [ ] List the sessions in their tutorials. - [ ] Show the details of their sessions. - [ ] Create a new session in a tutorial, as long as it has been published. - [ ] Update or delete the sessions in their tutorials, as long as they have not been published.
1.0
d3-s3-fr-15 - assistant operations on sessions - Operations by assistants on sessions: - [ ] List the sessions in their tutorials. - [ ] Show the details of their sessions. - [ ] Create a new session in a tutorial, as long as it has been published. - [ ] Update or delete the sessions in their tutorials, as long as they have not been published.
non_process
fr assistant operations on sessions operations by assistants on sessions list the sessions in their tutorials show the details of their sessions create a new session in a tutorial as long as it has been published update or delete the sessions in their tutorials as long as they have not been published
0
1,919
4,756,042,340
IssuesEvent
2016-10-24 12:53:39
opentrials/opentrials
https://api.github.com/repos/opentrials/opentrials
closed
Should we use some data pipeline framework in our processors?
Collectors Processors question refactoring
This issue is mostly to discuss our alternatives. We're starting to create a few tasks in our processors' pipeline that will be used by many processors (e.g. downloading PDFs and pushing to S3 for the FDA data, https://github.com/opentrials/opentrials/issues/250). We can save them in a "helpers" folder for reusing, but looking forward it seems we would benefit by using a data pipeline framework. This would help us in things like not rerunning code that was already ran, keeping track of the status of the data pipeline, retrying steps that didn't work, etc. There're quite a few options. The most interesting for us IMHO are [Airbnb's Airflow](http://nerds.airbnb.com/airflow/) (in incubation to become an Apache project), [Spotify's Luigi](https://github.com/spotify/luigi) and [Pinterest's Pinball](https://github.com/pinterest/pinball). I found a couple blog posts comparing them [here](https://www.michaelcho.me/article/data-pipelines-airflow-vs-pinball-vs-luigi) and [here](http://bytepawn.com/luigi-airflow-pinball.html) ([discussion on HN](https://news.ycombinator.com/item?id=11049684)). @roll @akariv Does any of you have experience with these tools? Any thoughts on using them?
1.0
Should we use some data pipeline framework in our processors? - This issue is mostly to discuss our alternatives. We're starting to create a few tasks in our processors' pipeline that will be used by many processors (e.g. downloading PDFs and pushing to S3 for the FDA data, https://github.com/opentrials/opentrials/issues/250). We can save them in a "helpers" folder for reusing, but looking forward it seems we would benefit by using a data pipeline framework. This would help us in things like not rerunning code that was already ran, keeping track of the status of the data pipeline, retrying steps that didn't work, etc. There're quite a few options. The most interesting for us IMHO are [Airbnb's Airflow](http://nerds.airbnb.com/airflow/) (in incubation to become an Apache project), [Spotify's Luigi](https://github.com/spotify/luigi) and [Pinterest's Pinball](https://github.com/pinterest/pinball). I found a couple blog posts comparing them [here](https://www.michaelcho.me/article/data-pipelines-airflow-vs-pinball-vs-luigi) and [here](http://bytepawn.com/luigi-airflow-pinball.html) ([discussion on HN](https://news.ycombinator.com/item?id=11049684)). @roll @akariv Does any of you have experience with these tools? Any thoughts on using them?
process
should we use some data pipeline framework in our processors this issue is mostly to discuss our alternatives we re starting to create a few tasks in our processors pipeline that will be used by many processors e g downloading pdfs and pushing to for the fda data we can save them in a helpers folder for reusing but looking forward it seems we would benefit by using a data pipeline framework this would help us in things like not rerunning code that was already ran keeping track of the status of the data pipeline retrying steps that didn t work etc there re quite a few options the most interesting for us imho are in incubation to become an apache project and i found a couple blog posts comparing them and roll akariv does any of you have experience with these tools any thoughts on using them
1
4,519
7,361,236,341
IssuesEvent
2018-03-11 04:26:09
shirou/gopsutil
https://api.github.com/repos/shirou/gopsutil
closed
fillFromStat causes significant CPU usage.
os:linux package:process
While using Nomad I observed that CPU usage to monitor processes was significant in certain situations. Using this [application](https://gist.github.com/maihde/30cacbc40003f6c59f625e6c0f901cb3) I profiled the calls to `process.Times()` and determined that a significant amount of the CPU usage was due to the calls by `getTerminalMap()` within the `fillFromStat()` method. Given that the `terminal` return value is usually thrown away it's better if `getTerminalMap()` is only called within the `process.Terminal()` method itself. See this [Nomad PR](https://github.com/hashicorp/nomad/pull/3883) for resolution and this [Nomad Issue](https://github.com/hashicorp/nomad/issues/3881) for more details.
1.0
fillFromStat causes significant CPU usage. - While using Nomad I observed that CPU usage to monitor processes was significant in certain situations. Using this [application](https://gist.github.com/maihde/30cacbc40003f6c59f625e6c0f901cb3) I profiled the calls to `process.Times()` and determined that a significant amount of the CPU usage was due to the calls by `getTerminalMap()` within the `fillFromStat()` method. Given that the `terminal` return value is usually thrown away it's better if `getTerminalMap()` is only called within the `process.Terminal()` method itself. See this [Nomad PR](https://github.com/hashicorp/nomad/pull/3883) for resolution and this [Nomad Issue](https://github.com/hashicorp/nomad/issues/3881) for more details.
process
fillfromstat causes significant cpu usage while using nomad i observed that cpu usage to monitor processes was significant in certain situations using this i profiled the calls to process times and determined that a significant amount of the cpu usage was due to the calls by getterminalmap within the fillfromstat method given that the terminal return value is usually thrown away it s better if getterminalmap is only called within the process terminal method itself see this for resolution and this for more details
1
16,078
20,249,228,402
IssuesEvent
2022-02-14 16:20:38
Bone008/orbiteye
https://api.github.com/repos/Bone008/orbiteye
opened
Implement timeline
data processing Front End/UI
- [ ] Create timeline component - [ ] Hook up timeline to global app state to affect visualizations
1.0
Implement timeline - - [ ] Create timeline component - [ ] Hook up timeline to global app state to affect visualizations
process
implement timeline create timeline component hook up timeline to global app state to affect visualizations
1
241,496
18,459,993,955
IssuesEvent
2021-10-15 22:49:54
bounswe/2021SpringGroup9
https://api.github.com/repos/bounswe/2021SpringGroup9
opened
Think about new requirements and how to improve the project plan
documentation priority: medium wiki difficulty: medium
Try to improve project requirements by thinking about what requirements we can add to the project. Also try to find ways to improve project plan.
1.0
Think about new requirements and how to improve the project plan - Try to improve project requirements by thinking about what requirements we can add to the project. Also try to find ways to improve project plan.
non_process
think about new requirements and how to improve the project plan try to improve project requirements by thinking about what requirements we can add to the project also try to find ways to improve project plan
0
13,278
15,759,440,103
IssuesEvent
2021-03-31 07:56:01
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
Process restoration with quick C-g
process:command
From @vspinu in #1108. > Actually not, on very quick C-g C-g the proc buffer is still not returned :(
1.0
Process restoration with quick C-g - From @vspinu in #1108. > Actually not, on very quick C-g C-g the proc buffer is still not returned :(
process
process restoration with quick c g from vspinu in actually not on very quick c g c g the proc buffer is still not returned
1
14,683
17,798,386,681
IssuesEvent
2021-09-01 02:57:22
jim-king-2000/IndustryCamera
https://api.github.com/repos/jim-king-2000/IndustryCamera
closed
[bug]: 设备中心 UI
bug processing C
### 问题描述 - 列展示太多的情况下,避免标题换行增加滑动条 - 设备区域未显示完整目录 ### 您预期的行为 - 不出现标题文字重叠 - 设备区域显示完整目录,例 风场A/风塔1/层1 ### 系统表现的行为 ![image](https://user-images.githubusercontent.com/89175659/131308042-0d6b9086-0e10-41de-afd3-6745e4cded32.png) ### 复现路径 <!-- 如何重现bug --> ### 辅助信息 - 浏览器版本:Edge/Chrome 92 - 固件版本:v1.0
1.0
[bug]: 设备中心 UI - ### 问题描述 - 列展示太多的情况下,避免标题换行增加滑动条 - 设备区域未显示完整目录 ### 您预期的行为 - 不出现标题文字重叠 - 设备区域显示完整目录,例 风场A/风塔1/层1 ### 系统表现的行为 ![image](https://user-images.githubusercontent.com/89175659/131308042-0d6b9086-0e10-41de-afd3-6745e4cded32.png) ### 复现路径 <!-- 如何重现bug --> ### 辅助信息 - 浏览器版本:Edge/Chrome 92 - 固件版本:v1.0
process
设备中心 ui 问题描述 列展示太多的情况下,避免标题换行增加滑动条 设备区域未显示完整目录 您预期的行为 不出现标题文字重叠 设备区域显示完整目录,例 风场a 系统表现的行为 复现路径 辅助信息 浏览器版本:edge chrome 固件版本:
1
261,489
27,809,782,829
IssuesEvent
2023-03-18 01:43:09
madhans23/linux-4.1.15
https://api.github.com/repos/madhans23/linux-4.1.15
closed
CVE-2020-25668 (High) detected in linux-stable-rtv4.1.33 - autoclosed
Mend: dependency security vulnerability
## CVE-2020-25668 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/vt_ioctl.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/vt_ioctl.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in Linux Kernel because access to the global variable fg_console is not properly synchronized leading to a use after free in con_font_op. <p>Publish Date: 2021-05-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25668>CVE-2020-25668</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25668">https://www.linuxkernelcves.com/cves/CVE-2020-25668</a></p> <p>Release Date: 2021-05-26</p> <p>Fix Resolution: v4.14.204,v4.19.155,v5.4.75,v5.9.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-25668 (High) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2020-25668 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/vt_ioctl.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/vt_ioctl.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw was found in Linux Kernel because access to the global variable fg_console is not properly synchronized leading to a use after free in con_font_op. <p>Publish Date: 2021-05-26 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25668>CVE-2020-25668</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25668">https://www.linuxkernelcves.com/cves/CVE-2020-25668</a></p> <p>Release Date: 2021-05-26</p> <p>Fix Resolution: v4.14.204,v4.19.155,v5.4.75,v5.9.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linux stable autoclosed cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers tty vt vt ioctl c drivers tty vt vt ioctl c vulnerability details a flaw was found in linux kernel because access to the global variable fg console is not properly synchronized leading to a use after free in con font op publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
562,782
16,669,747,613
IssuesEvent
2021-06-07 09:22:11
AyeCode/userswp
https://api.github.com/repos/AyeCode/userswp
opened
Doc request for new multiple registration forms functionality
Docs Priority: High Type: Needs Docs
Please create a doc with the multiple registration forms functionality with form settings and how it works for the account and profile page. Let me know where my help needed.
1.0
Doc request for new multiple registration forms functionality - Please create a doc with the multiple registration forms functionality with form settings and how it works for the account and profile page. Let me know where my help needed.
non_process
doc request for new multiple registration forms functionality please create a doc with the multiple registration forms functionality with form settings and how it works for the account and profile page let me know where my help needed
0
10,296
13,148,136,951
IssuesEvent
2020-08-08 19:37:46
unisonweb/unison
https://api.github.com/repos/unisonweb/unison
opened
pull-request.create is awkward when the output is bigger than a screen
pr-process
What are others doing when the `pull-request.create` output is longer than their screen size? For me the output opens in `less` (I think?), and I don't know of a convenient way to copy the entirety of the text to paste in a GitHub issue. I tried doing `echo 'pull-request.create git@github.com:unisonweb/base:.trunk git@github.com:ceedubs/unison-dev:.prs._base.nelfoldl' | ucm | pbcopy`, but this ends up capturing a lot of ANSI escape characters. I could be missing something obvious, but it seemed worth capturing in an issue since this might trip up others.
1.0
pull-request.create is awkward when the output is bigger than a screen - What are others doing when the `pull-request.create` output is longer than their screen size? For me the output opens in `less` (I think?), and I don't know of a convenient way to copy the entirety of the text to paste in a GitHub issue. I tried doing `echo 'pull-request.create git@github.com:unisonweb/base:.trunk git@github.com:ceedubs/unison-dev:.prs._base.nelfoldl' | ucm | pbcopy`, but this ends up capturing a lot of ANSI escape characters. I could be missing something obvious, but it seemed worth capturing in an issue since this might trip up others.
process
pull request create is awkward when the output is bigger than a screen what are others doing when the pull request create output is longer than their screen size for me the output opens in less i think and i don t know of a convenient way to copy the entirety of the text to paste in a github issue i tried doing echo pull request create git github com unisonweb base trunk git github com ceedubs unison dev prs base nelfoldl ucm pbcopy but this ends up capturing a lot of ansi escape characters i could be missing something obvious but it seemed worth capturing in an issue since this might trip up others
1
4,132
7,088,571,440
IssuesEvent
2018-01-11 22:02:06
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
An empty sequence is not allowed as the value of parameter $href [DOT 2.3.1]
bug preprocess
Using one of my client's samples I obtain an error when publishing to XHTML-based outputs, the error is something like this: ``` maplink: [maplink] Error on line 418 of maplinkImpl.xsl: [maplink] XPTY0004: An empty sequence is not allowed as the value of parameter $href [maplink] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/maplinkImpl.xsl#410) [maplink] processing /map/reltable[1]/relheader[1]/relcolspec[1]/topichead[1] [maplink] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/maplinkImpl.xsl#335) ``` and it occurs at line: ``` <xsl:template match="*[contains(@class, ' map/topicref ')]" mode="grab-group-title" as="xs:string?"> <xsl:variable name="file-origin"> <xsl:call-template name="get-file-uri"> <xsl:with-param name="href" select="@href"/> ``` so somehow the topicref does not have an @href attribute set to it and the publishing breaks. I added an xsl:message and it breaks when processing this topichead: ``` <topichead xmlns:ditaarch="http://dita.oasis-open.org/architecture/2005/" xmlns:dita-ot="http://dita-ot.sourceforge.net/ns/201007/dita-ot" class="+ map/topicref mapgroup-d/topichead " xtrc="topichead:1;28:28" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap" toc="no" print="no"><topicmeta class="- map/topicmeta " xtrc="topicmeta:2;29:32" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap"><navtitle class="- topic/navtitle " xtrc="navtitle:1;30:35" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap">See also</navtitle></topicmeta></topichead> ```
1.0
An empty sequence is not allowed as the value of parameter $href [DOT 2.3.1] - Using one of my client's samples I obtain an error when publishing to XHTML-based outputs, the error is something like this: ``` maplink: [maplink] Error on line 418 of maplinkImpl.xsl: [maplink] XPTY0004: An empty sequence is not allowed as the value of parameter $href [maplink] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/maplinkImpl.xsl#410) [maplink] processing /map/reltable[1]/relheader[1]/relcolspec[1]/topichead[1] [maplink] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/maplinkImpl.xsl#335) ``` and it occurs at line: ``` <xsl:template match="*[contains(@class, ' map/topicref ')]" mode="grab-group-title" as="xs:string?"> <xsl:variable name="file-origin"> <xsl:call-template name="get-file-uri"> <xsl:with-param name="href" select="@href"/> ``` so somehow the topicref does not have an @href attribute set to it and the publishing breaks. I added an xsl:message and it breaks when processing this topichead: ``` <topichead xmlns:ditaarch="http://dita.oasis-open.org/architecture/2005/" xmlns:dita-ot="http://dita-ot.sourceforge.net/ns/201007/dita-ot" class="+ map/topicref mapgroup-d/topichead " xtrc="topichead:1;28:28" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap" toc="no" print="no"><topicmeta class="- map/topicmeta " xtrc="topicmeta:2;29:32" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap"><navtitle class="- topic/navtitle " xtrc="navtitle:1;30:35" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap">See also</navtitle></topicmeta></topichead> ```
process
an empty sequence is not allowed as the value of parameter href using one of my client s samples i obtain an error when publishing to xhtml based outputs the error is something like this maplink error on line of maplinkimpl xsl an empty sequence is not allowed as the value of parameter href at xsl apply templates file d projects exml frameworks dita dita x xsl preprocess maplinkimpl xsl processing map reltable relheader relcolspec topichead at xsl apply templates file d projects exml frameworks dita dita x xsl preprocess maplinkimpl xsl and it occurs at line so somehow the topicref does not have an href attribute set to it and the publishing breaks i added an xsl message and it breaks when processing this topichead see also
1
3,055
6,046,755,959
IssuesEvent
2017-06-12 12:58:32
itsyouonline/identityserver
https://api.github.com/repos/itsyouonline/identityserver
closed
Can create 2 accounts with same email and phone number
process_wontfix
- Please prevent creating more than one account with same information ![22_user11](https://cloud.githubusercontent.com/assets/15011431/19729879/200809b8-9ba1-11e6-8a18-5ecb4b9f0548.png) ![11_user2](https://cloud.githubusercontent.com/assets/15011431/19729861/184e8d0a-9ba1-11e6-8287-ccb44545c614.png)
1.0
Can create 2 accounts with same email and phone number - - Please prevent creating more than one account with same information ![22_user11](https://cloud.githubusercontent.com/assets/15011431/19729879/200809b8-9ba1-11e6-8a18-5ecb4b9f0548.png) ![11_user2](https://cloud.githubusercontent.com/assets/15011431/19729861/184e8d0a-9ba1-11e6-8287-ccb44545c614.png)
process
can create accounts with same email and phone number please prevent creating more than one account with same information
1
19,050
25,060,336,037
IssuesEvent
2022-11-07 00:20:32
ethereum/EIPs
https://api.github.com/repos/ethereum/EIPs
closed
EIP registries
w-stale enhancement r-process r-eips
Some EIPs require registries - lists of mappings relevant to the EIP. Examples include EIP 191 and EIP 634. Currently, there's no good solution for this: changes to an EIP are prohibited after it's final, so the registry can't be kept inside the EIP itself, and external links are likewise prohibited, so the registry can't be kept elsewhere. The IETF maintains a list of registries; each RFC can define one or more new named registries, and the registries are maintained separately from the (final) RFC. We should define a best practice for this with EIPs.
1.0
EIP registries - Some EIPs require registries - lists of mappings relevant to the EIP. Examples include EIP 191 and EIP 634. Currently, there's no good solution for this: changes to an EIP are prohibited after it's final, so the registry can't be kept inside the EIP itself, and external links are likewise prohibited, so the registry can't be kept elsewhere. The IETF maintains a list of registries; each RFC can define one or more new named registries, and the registries are maintained separately from the (final) RFC. We should define a best practice for this with EIPs.
process
eip registries some eips require registries lists of mappings relevant to the eip examples include eip and eip currently there s no good solution for this changes to an eip are prohibited after it s final so the registry can t be kept inside the eip itself and external links are likewise prohibited so the registry can t be kept elsewhere the ietf maintains a list of registries each rfc can define one or more new named registries and the registries are maintained separately from the final rfc we should define a best practice for this with eips
1
157,205
5,996,453,946
IssuesEvent
2017-06-03 14:29:41
universAAL/tools.eclipse-plugins
https://api.github.com/repos/universAAL/tools.eclipse-plugins
closed
when transforming .OWL files it requests .xml file
bug imported priority 4
_Originally Opened: @amedranogil (2013-01-11 10:30:15_) _Originally Closed: 2013-01-18 11:12:47_ when I click on an owl file (i'm using this one: http://purl.obolibrary.org/obo/doid.owl ) the tool seems to do a first reading of the OWL and the second reading seems it is trying to read the .xml (non-existent). Type... UML2 Reading...C:/Users/amedrano/workspaceOnt/ont.disease/Disease.owl Read DONE C:/Users/amedrano/workspaceOnt/ont.disease/Disease.owl Read base name.................http://purl.org/obo/owl java.io.FileNotFoundException: C:\Users\amedrano\workspaceOnt\ont.disease\Disease.xml (El sistema no puede encontrar el archivo especificado) at java.io.FileInputStream.open(Native Method) at java.io.FileInputStream.&lt;init&gt;(Unknown Source) at java.io.FileInputStream.&lt;init&gt;(Unknown Source) at sun.net.www.protocol.file.FileURLConnection.connect(Unknown Source) at sun.net.www.protocol.file.FileURLConnection.getInputStream(Unknown Source) at org.apache.xerces.impl.XMLEntityManager.setupCurrentEntity(Unknown Source) at org.apache.xerces.impl.XMLVersionDetector.determineDocVersion(Unknown Source) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XMLParser.parse(Unknown Source) at org.apache.xerces.parsers.DOMParser.parse(Unknown Source) at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source) at javax.xml.parsers.DocumentBuilder.parse(Unknown Source) at org.universaal.tools.owl2uml.uml2.UML2Factory.readXML(UML2Factory.java:781) at org.universaal.tools.owl2uml.uml2.UML2Factory.&lt;init&gt;(UML2Factory.java:91) at org.universaal.tools.owl2uml.core.UML2Parser.loadOntology(UML2Parser.java:113) at org.universaal.tools.owl2uml.OWL2UML.main(OWL2UML.java:39) at org.universaal.tools.owl2uml.handlers.SampleHandler.execute(SampleHandler.java:79) at org.eclipse.ui.internal.handlers.HandlerProxy.execute(HandlerProxy.java:293) at org.eclipse.core.commands.Command.executeWithChecks(Command.java:476) at org.eclipse.core.commands.ParameterizedCommand.executeWithChecks(ParameterizedCommand.java:508) at org.eclipse.ui.internal.handlers.HandlerService.executeCommand(HandlerService.java:169) at org.eclipse.ui.internal.handlers.SlaveHandlerService.executeCommand(SlaveHandlerService.java:241) at org.eclipse.ui.internal.handlers.SlaveHandlerService.executeCommand(SlaveHandlerService.java:241) at org.eclipse.ui.menus.CommandContributionItem.handleWidgetSelection(CommandContributionItem.java:829) at org.eclipse.ui.menus.CommandContributionItem.access$19(CommandContributionItem.java:815) at org.eclipse.ui.menus.CommandContributionItem$5.handleEvent(CommandContributionItem.java:805) at org.eclipse.swt.widgets.EventTable.sendEvent(EventTable.java:84) at org.eclipse.swt.widgets.Widget.sendEvent(Widget.java:1053) at org.eclipse.swt.widgets.Display.runDeferredEvents(Display.java:4165) at org.eclipse.swt.widgets.Display.readAndDispatch(Display.java:3754) at org.eclipse.ui.internal.Workbench.runEventLoop(Workbench.java:2701) at org.eclipse.ui.internal.Workbench.runUI(Workbench.java:2665) at org.eclipse.ui.internal.Workbench.access$4(Workbench.java:2499) at org.eclipse.ui.internal.Workbench$7.run(Workbench.java:679) at org.eclipse.core.databinding.observable.Realm.runWithDefault(Realm.java:332) at org.eclipse.ui.internal.Workbench.createAndRunWorkbench(Workbench.java:668) at org.eclipse.ui.PlatformUI.createAndRunWorkbench(PlatformUI.java:149) at org.eclipse.ui.internal.ide.application.IDEApplication.start(IDEApplication.java:123) at org.eclipse.equinox.internal.app.EclipseAppHandle.run(EclipseAppHandle.java:196) at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.runApplication(EclipseAppLauncher.java:110) at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.start(EclipseAppLauncher.java:79) at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:344) at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:179) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.eclipse.equinox.launcher.Main.invokeFramework(Main.java:622) at org.eclipse.equinox.launcher.Main.basicRun(Main.java:577) at org.eclipse.equinox.launcher.Main.run(Main.java:1410) -- From: _this issue has been automatically imported from our old issue tracker_
1.0
when transforming .OWL files it requests .xml file - _Originally Opened: @amedranogil (2013-01-11 10:30:15_) _Originally Closed: 2013-01-18 11:12:47_ when I click on an owl file (i'm using this one: http://purl.obolibrary.org/obo/doid.owl ) the tool seems to do a first reading of the OWL and the second reading seems it is trying to read the .xml (non-existent). Type... UML2 Reading...C:/Users/amedrano/workspaceOnt/ont.disease/Disease.owl Read DONE C:/Users/amedrano/workspaceOnt/ont.disease/Disease.owl Read base name.................http://purl.org/obo/owl java.io.FileNotFoundException: C:\Users\amedrano\workspaceOnt\ont.disease\Disease.xml (El sistema no puede encontrar el archivo especificado) at java.io.FileInputStream.open(Native Method) at java.io.FileInputStream.&lt;init&gt;(Unknown Source) at java.io.FileInputStream.&lt;init&gt;(Unknown Source) at sun.net.www.protocol.file.FileURLConnection.connect(Unknown Source) at sun.net.www.protocol.file.FileURLConnection.getInputStream(Unknown Source) at org.apache.xerces.impl.XMLEntityManager.setupCurrentEntity(Unknown Source) at org.apache.xerces.impl.XMLVersionDetector.determineDocVersion(Unknown Source) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source) at org.apache.xerces.parsers.XMLParser.parse(Unknown Source) at org.apache.xerces.parsers.DOMParser.parse(Unknown Source) at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source) at javax.xml.parsers.DocumentBuilder.parse(Unknown Source) at org.universaal.tools.owl2uml.uml2.UML2Factory.readXML(UML2Factory.java:781) at org.universaal.tools.owl2uml.uml2.UML2Factory.&lt;init&gt;(UML2Factory.java:91) at org.universaal.tools.owl2uml.core.UML2Parser.loadOntology(UML2Parser.java:113) at org.universaal.tools.owl2uml.OWL2UML.main(OWL2UML.java:39) at org.universaal.tools.owl2uml.handlers.SampleHandler.execute(SampleHandler.java:79) at org.eclipse.ui.internal.handlers.HandlerProxy.execute(HandlerProxy.java:293) at org.eclipse.core.commands.Command.executeWithChecks(Command.java:476) at org.eclipse.core.commands.ParameterizedCommand.executeWithChecks(ParameterizedCommand.java:508) at org.eclipse.ui.internal.handlers.HandlerService.executeCommand(HandlerService.java:169) at org.eclipse.ui.internal.handlers.SlaveHandlerService.executeCommand(SlaveHandlerService.java:241) at org.eclipse.ui.internal.handlers.SlaveHandlerService.executeCommand(SlaveHandlerService.java:241) at org.eclipse.ui.menus.CommandContributionItem.handleWidgetSelection(CommandContributionItem.java:829) at org.eclipse.ui.menus.CommandContributionItem.access$19(CommandContributionItem.java:815) at org.eclipse.ui.menus.CommandContributionItem$5.handleEvent(CommandContributionItem.java:805) at org.eclipse.swt.widgets.EventTable.sendEvent(EventTable.java:84) at org.eclipse.swt.widgets.Widget.sendEvent(Widget.java:1053) at org.eclipse.swt.widgets.Display.runDeferredEvents(Display.java:4165) at org.eclipse.swt.widgets.Display.readAndDispatch(Display.java:3754) at org.eclipse.ui.internal.Workbench.runEventLoop(Workbench.java:2701) at org.eclipse.ui.internal.Workbench.runUI(Workbench.java:2665) at org.eclipse.ui.internal.Workbench.access$4(Workbench.java:2499) at org.eclipse.ui.internal.Workbench$7.run(Workbench.java:679) at org.eclipse.core.databinding.observable.Realm.runWithDefault(Realm.java:332) at org.eclipse.ui.internal.Workbench.createAndRunWorkbench(Workbench.java:668) at org.eclipse.ui.PlatformUI.createAndRunWorkbench(PlatformUI.java:149) at org.eclipse.ui.internal.ide.application.IDEApplication.start(IDEApplication.java:123) at org.eclipse.equinox.internal.app.EclipseAppHandle.run(EclipseAppHandle.java:196) at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.runApplication(EclipseAppLauncher.java:110) at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.start(EclipseAppLauncher.java:79) at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:344) at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:179) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source) at java.lang.reflect.Method.invoke(Unknown Source) at org.eclipse.equinox.launcher.Main.invokeFramework(Main.java:622) at org.eclipse.equinox.launcher.Main.basicRun(Main.java:577) at org.eclipse.equinox.launcher.Main.run(Main.java:1410) -- From: _this issue has been automatically imported from our old issue tracker_
non_process
when transforming owl files it requests xml file originally opened amedranogil originally closed when i click on an owl file i m using this one the tool seems to do a first reading of the owl and the second reading seems it is trying to read the xml non existent type reading c users amedrano workspaceont ont disease disease owl read done c users amedrano workspaceont ont disease disease owl read base name java io filenotfoundexception c users amedrano workspaceont ont disease disease xml el sistema no puede encontrar el archivo especificado at java io fileinputstream open native method at java io fileinputstream lt init gt unknown source at java io fileinputstream lt init gt unknown source at sun net source at sun net source at org apache xerces impl xmlentitymanager setupcurrententity unknown source at org apache xerces impl xmlversiondetector determinedocversion unknown source at org apache xerces parsers parse unknown source at org apache xerces parsers parse unknown source at org apache xerces parsers xmlparser parse unknown source at org apache xerces parsers domparser parse unknown source at org apache xerces jaxp documentbuilderimpl parse unknown source at javax xml parsers documentbuilder parse unknown source at org universaal tools readxml java at org universaal tools lt init gt java at org universaal tools core loadontology java at org universaal tools main java at org universaal tools handlers samplehandler execute samplehandler java at org eclipse ui internal handlers handlerproxy execute handlerproxy java at org eclipse core commands command executewithchecks command java at org eclipse core commands parameterizedcommand executewithchecks parameterizedcommand java at org eclipse ui internal handlers handlerservice executecommand handlerservice java at org eclipse ui internal handlers slavehandlerservice executecommand slavehandlerservice java at org eclipse ui internal handlers slavehandlerservice executecommand slavehandlerservice java at org eclipse ui menus commandcontributionitem handlewidgetselection commandcontributionitem java at org eclipse ui menus commandcontributionitem access commandcontributionitem java at org eclipse ui menus commandcontributionitem handleevent commandcontributionitem java at org eclipse swt widgets eventtable sendevent eventtable java at org eclipse swt widgets widget sendevent widget java at org eclipse swt widgets display rundeferredevents display java at org eclipse swt widgets display readanddispatch display java at org eclipse ui internal workbench runeventloop workbench java at org eclipse ui internal workbench runui workbench java at org eclipse ui internal workbench access workbench java at org eclipse ui internal workbench run workbench java at org eclipse core databinding observable realm runwithdefault realm java at org eclipse ui internal workbench createandrunworkbench workbench java at org eclipse ui platformui createandrunworkbench platformui java at org eclipse ui internal ide application ideapplication start ideapplication java at org eclipse equinox internal app eclipseapphandle run eclipseapphandle java at org eclipse core runtime internal adaptor eclipseapplauncher runapplication eclipseapplauncher java at org eclipse core runtime internal adaptor eclipseapplauncher start eclipseapplauncher java at org eclipse core runtime adaptor eclipsestarter run eclipsestarter java at org eclipse core runtime adaptor eclipsestarter run eclipsestarter java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org eclipse equinox launcher main invokeframework main java at org eclipse equinox launcher main basicrun main java at org eclipse equinox launcher main run main java from this issue has been automatically imported from our old issue tracker
0
18,187
24,236,165,089
IssuesEvent
2022-09-26 23:36:58
opensearch-project/data-prepper
https://api.github.com/repos/opensearch-project/data-prepper
closed
Support parsing JSON
enhancement plugin - processor
**Is your feature request related to a problem? Please describe.** Data Prepper events may have JSON values inside Event fields. Data Prepper should be able to parse these JSON strings and create fields directly in the Event from the JSON. **Describe the solution you'd like** Provide a JSON parsing processor - `parse_json`. It should be able to parse a JSON string from a field and set the values in the Event object. This processor will automatically support nesting. **Example** Given the following configuration: ``` processor: parse_json: source: my_field ``` Given this input event: ``` "my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}" ``` The input event is changed to: ``` "my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}" "key1" : "value1" "key2" : "value2" ``` **Example with Nesting** Given this input event: ``` "my_field" : "{\"key1\" : \"value1\", \"key2\" : { \"key2child\" : \"innerValue\" }}" ``` The input event is changed to: ``` "my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}" "key1" : "value1" "key2" : { "key2child" : "innerValue" } ``` **Configurations** `source` - the field with JSON `target` - the field to set the values in; by default this is the root object
1.0
Support parsing JSON - **Is your feature request related to a problem? Please describe.** Data Prepper events may have JSON values inside Event fields. Data Prepper should be able to parse these JSON strings and create fields directly in the Event from the JSON. **Describe the solution you'd like** Provide a JSON parsing processor - `parse_json`. It should be able to parse a JSON string from a field and set the values in the Event object. This processor will automatically support nesting. **Example** Given the following configuration: ``` processor: parse_json: source: my_field ``` Given this input event: ``` "my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}" ``` The input event is changed to: ``` "my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}" "key1" : "value1" "key2" : "value2" ``` **Example with Nesting** Given this input event: ``` "my_field" : "{\"key1\" : \"value1\", \"key2\" : { \"key2child\" : \"innerValue\" }}" ``` The input event is changed to: ``` "my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}" "key1" : "value1" "key2" : { "key2child" : "innerValue" } ``` **Configurations** `source` - the field with JSON `target` - the field to set the values in; by default this is the root object
process
support parsing json is your feature request related to a problem please describe data prepper events may have json values inside event fields data prepper should be able to parse these json strings and create fields directly in the event from the json describe the solution you d like provide a json parsing processor parse json it should be able to parse a json string from a field and set the values in the event object this processor will automatically support nesting example given the following configuration processor parse json source my field given this input event my field the input event is changed to my field example with nesting given this input event my field innervalue the input event is changed to my field innervalue configurations source the field with json target the field to set the values in by default this is the root object
1
134,882
18,513,051,764
IssuesEvent
2021-10-20 06:56:12
mgh3326/createandread
https://api.github.com/repos/mgh3326/createandread
opened
CVE-2021-37713 (High) detected in tar-4.4.13.tgz
security vulnerability
## CVE-2021-37713 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p>Path to dependency file: createandread/package.json</p> <p>Path to vulnerable library: createandread/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - webpack-dev-server-3.9.0.tgz (Root Library) - chokidar-2.1.8.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - :x: **tar-4.4.13.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mgh3326/createandread/commit/da66d826bfaeb3ccfb4d58ebf3cd7d5c1af76a48">da66d826bfaeb3ccfb4d58ebf3cd7d5c1af76a48</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted. This is, in part, accomplished by sanitizing absolute paths of entries within the archive, skipping archive entries that contain `..` path portions, and resolving the sanitized paths against the extraction target directory. This logic was insufficient on Windows systems when extracting tar files that contained a path that was not an absolute path, but specified a drive letter different from the extraction target, such as `C:some\path`. If the drive letter does not match the extraction target, for example `D:\extraction\dir`, then the result of `path.resolve(extractionDirectory, entryPath)` would resolve against the current working directory on the `C:` drive, rather than the extraction target directory. Additionally, a `..` portion of the path could occur immediately after the drive letter, such as `C:../foo`, and was not properly sanitized by the logic that checked for `..` within the normalized and split portions of the path. This only affects users of `node-tar` on Windows systems. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. There is no reasonable way to work around this issue without performing the same path normalization procedures that node-tar now does. Users are encouraged to upgrade to the latest patched versions of node-tar, rather than attempt to sanitize paths themselves. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37713>CVE-2021-37713</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh">https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.18, 5.0.10, 6.1.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37713 (High) detected in tar-4.4.13.tgz - ## CVE-2021-37713 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.13.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p> <p>Path to dependency file: createandread/package.json</p> <p>Path to vulnerable library: createandread/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - webpack-dev-server-3.9.0.tgz (Root Library) - chokidar-2.1.8.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - :x: **tar-4.4.13.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mgh3326/createandread/commit/da66d826bfaeb3ccfb4d58ebf3cd7d5c1af76a48">da66d826bfaeb3ccfb4d58ebf3cd7d5c1af76a48</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted. This is, in part, accomplished by sanitizing absolute paths of entries within the archive, skipping archive entries that contain `..` path portions, and resolving the sanitized paths against the extraction target directory. This logic was insufficient on Windows systems when extracting tar files that contained a path that was not an absolute path, but specified a drive letter different from the extraction target, such as `C:some\path`. If the drive letter does not match the extraction target, for example `D:\extraction\dir`, then the result of `path.resolve(extractionDirectory, entryPath)` would resolve against the current working directory on the `C:` drive, rather than the extraction target directory. Additionally, a `..` portion of the path could occur immediately after the drive letter, such as `C:../foo`, and was not properly sanitized by the logic that checked for `..` within the normalized and split portions of the path. This only affects users of `node-tar` on Windows systems. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. There is no reasonable way to work around this issue without performing the same path normalization procedures that node-tar now does. Users are encouraged to upgrade to the latest patched versions of node-tar, rather than attempt to sanitize paths themselves. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37713>CVE-2021-37713</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh">https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.18, 5.0.10, 6.1.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file createandread package json path to vulnerable library createandread node modules tar package json dependency hierarchy webpack dev server tgz root library chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in head commit a href vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted this is in part accomplished by sanitizing absolute paths of entries within the archive skipping archive entries that contain path portions and resolving the sanitized paths against the extraction target directory this logic was insufficient on windows systems when extracting tar files that contained a path that was not an absolute path but specified a drive letter different from the extraction target such as c some path if the drive letter does not match the extraction target for example d extraction dir then the result of path resolve extractiondirectory entrypath would resolve against the current working directory on the c drive rather than the extraction target directory additionally a portion of the path could occur immediately after the drive letter such as c foo and was not properly sanitized by the logic that checked for within the normalized and split portions of the path this only affects users of node tar on windows systems these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar there is no reasonable way to work around this issue without performing the same path normalization procedures that node tar now does users are encouraged to upgrade to the latest patched versions of node tar rather than attempt to sanitize paths themselves publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource
0
13,584
16,135,484,256
IssuesEvent
2021-04-29 11:15:39
sanzcarlos/CiscoAXL
https://api.github.com/repos/sanzcarlos/CiscoAXL
closed
Add new field in CSV
enhancement process
**Is your feature request related to a problem? Please describe.** We need add the main DDI of site and create a Translation Pattern to Hunt Pilot **Describe the solution you'd like** N/A **Additional context** N/A
1.0
Add new field in CSV - **Is your feature request related to a problem? Please describe.** We need add the main DDI of site and create a Translation Pattern to Hunt Pilot **Describe the solution you'd like** N/A **Additional context** N/A
process
add new field in csv is your feature request related to a problem please describe we need add the main ddi of site and create a translation pattern to hunt pilot describe the solution you d like n a additional context n a
1
15,881
20,070,639,759
IssuesEvent
2022-02-04 06:06:50
codeanit/til
https://api.github.com/repos/codeanit/til
opened
Vanguard Method - Understanding variation in demand
wip process
The Vanguard Method is one framework for understanding variation in demand. This takes a data driven approach of understanding what the customer wanted to achieve. Here are the steps: 1. Define the purpose of the system from customer perspective. 2. Study the nature of demand coming towards the system 3. Learn how the system responds to the demands. 4. Understand why this happens. 5. Identify what policies or measures cause problems in the flow of work. 6. Recognize the thinking behind the design and management of the system
1.0
Vanguard Method - Understanding variation in demand - The Vanguard Method is one framework for understanding variation in demand. This takes a data driven approach of understanding what the customer wanted to achieve. Here are the steps: 1. Define the purpose of the system from customer perspective. 2. Study the nature of demand coming towards the system 3. Learn how the system responds to the demands. 4. Understand why this happens. 5. Identify what policies or measures cause problems in the flow of work. 6. Recognize the thinking behind the design and management of the system
process
vanguard method understanding variation in demand the vanguard method is one framework for understanding variation in demand this takes a data driven approach of understanding what the customer wanted to achieve here are the steps define the purpose of the system from customer perspective study the nature of demand coming towards the system learn how the system responds to the demands understand why this happens identify what policies or measures cause problems in the flow of work recognize the thinking behind the design and management of the system
1
639,809
20,766,675,465
IssuesEvent
2022-03-15 21:23:23
SIMPLE-AstroDB/SIMPLE-db
https://api.github.com/repos/SIMPLE-AstroDB/SIMPLE-db
closed
Add tests for Gaia and 2MASS data
enhancement low priority
I'm not quite sure how to write these, but they would be useful to have: - If 2MASS designation in Names, 2MASS photometry should exist - If 2MASS photometry, 2MASS designation should be in Names - If Gaia designation in Names, Gaia phot and astrometry should exist - If Gaia phot, Gaia designation should be in Names - If Gaia pm, Gaia designation should be in Names - If Gaia parallax, Gaia designation should be in Names
1.0
Add tests for Gaia and 2MASS data - I'm not quite sure how to write these, but they would be useful to have: - If 2MASS designation in Names, 2MASS photometry should exist - If 2MASS photometry, 2MASS designation should be in Names - If Gaia designation in Names, Gaia phot and astrometry should exist - If Gaia phot, Gaia designation should be in Names - If Gaia pm, Gaia designation should be in Names - If Gaia parallax, Gaia designation should be in Names
non_process
add tests for gaia and data i m not quite sure how to write these but they would be useful to have if designation in names photometry should exist if photometry designation should be in names if gaia designation in names gaia phot and astrometry should exist if gaia phot gaia designation should be in names if gaia pm gaia designation should be in names if gaia parallax gaia designation should be in names
0
4,451
7,319,320,758
IssuesEvent
2018-03-02 00:11:45
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Outdated documentation
active-directory cxp in-process triaged
There is no such option as Single Sign On in azure ad now. Atleast not in my directory. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: f97b43d9-4a25-1552-c719-f0f0b418b0a8 * Version Independent ID: 13445782-bdb8-b9fa-b130-630bb770187c * [Content](https://docs.microsoft.com/en-us/azure/active-directory/active-directory-enterprise-apps-manage-sso) * [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/active-directory-enterprise-apps-manage-sso.md) * Service: active-directory
1.0
Outdated documentation - There is no such option as Single Sign On in azure ad now. Atleast not in my directory. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: f97b43d9-4a25-1552-c719-f0f0b418b0a8 * Version Independent ID: 13445782-bdb8-b9fa-b130-630bb770187c * [Content](https://docs.microsoft.com/en-us/azure/active-directory/active-directory-enterprise-apps-manage-sso) * [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/active-directory-enterprise-apps-manage-sso.md) * Service: active-directory
process
outdated documentation there is no such option as single sign on in azure ad now atleast not in my directory document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id service active directory
1
211,980
23,856,895,146
IssuesEvent
2022-09-07 01:14:24
CartoonFan/node-jscs
https://api.github.com/repos/CartoonFan/node-jscs
opened
WS-2021-0638 (High) detected in mocha-2.5.3.tgz
security vulnerability
## WS-2021-0638 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mocha-2.5.3.tgz</b></p></summary> <p>simple, flexible, fun test framework</p> <p>Library home page: <a href="https://registry.npmjs.org/mocha/-/mocha-2.5.3.tgz">https://registry.npmjs.org/mocha/-/mocha-2.5.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/mocha/package.json</p> <p> Dependency Hierarchy: - :x: **mocha-2.5.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/CartoonFan/node-jscs/commit/d7995bed5d8b00cdb4811d5cb4631eb63ac6f70f">d7995bed5d8b00cdb4811d5cb4631eb63ac6f70f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is regular Expression Denial of Service (ReDoS) vulnerability in mocha. It allows cause a denial of service when stripping crafted invalid function definition from strs. <p>Publish Date: 2021-09-18 <p>URL: <a href=https://github.com/mochajs/mocha/commit/61b4b9209c2c64b32c8d48b1761c3b9384d411ea>WS-2021-0638</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/1d8a3d95-d199-4129-a6ad-8eafe5e77b9e/">https://huntr.dev/bounties/1d8a3d95-d199-4129-a6ad-8eafe5e77b9e/</a></p> <p>Release Date: 2021-09-18</p> <p>Fix Resolution: 3.0.0-0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0638 (High) detected in mocha-2.5.3.tgz - ## WS-2021-0638 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mocha-2.5.3.tgz</b></p></summary> <p>simple, flexible, fun test framework</p> <p>Library home page: <a href="https://registry.npmjs.org/mocha/-/mocha-2.5.3.tgz">https://registry.npmjs.org/mocha/-/mocha-2.5.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/mocha/package.json</p> <p> Dependency Hierarchy: - :x: **mocha-2.5.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/CartoonFan/node-jscs/commit/d7995bed5d8b00cdb4811d5cb4631eb63ac6f70f">d7995bed5d8b00cdb4811d5cb4631eb63ac6f70f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is regular Expression Denial of Service (ReDoS) vulnerability in mocha. It allows cause a denial of service when stripping crafted invalid function definition from strs. <p>Publish Date: 2021-09-18 <p>URL: <a href=https://github.com/mochajs/mocha/commit/61b4b9209c2c64b32c8d48b1761c3b9384d411ea>WS-2021-0638</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/1d8a3d95-d199-4129-a6ad-8eafe5e77b9e/">https://huntr.dev/bounties/1d8a3d95-d199-4129-a6ad-8eafe5e77b9e/</a></p> <p>Release Date: 2021-09-18</p> <p>Fix Resolution: 3.0.0-0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws high detected in mocha tgz ws high severity vulnerability vulnerable library mocha tgz simple flexible fun test framework library home page a href path to dependency file package json path to vulnerable library node modules mocha package json dependency hierarchy x mocha tgz vulnerable library found in head commit a href found in base branch master vulnerability details there is regular expression denial of service redos vulnerability in mocha it allows cause a denial of service when stripping crafted invalid function definition from strs publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
19,582
25,906,137,007
IssuesEvent
2022-12-15 10:21:17
CSE201-project/PaperFriend-desktop-app
https://api.github.com/repos/CSE201-project/PaperFriend-desktop-app
closed
Work on EntryRecap : Best/Worst day (Entry), Text (String), Average mood (Double)
file processing frontend analysis
Class used to show the recap over week/month/year to the user Also add another parameter (int type) that will tell the frontend team if it's week/month/year to show it in different colors.
1.0
Work on EntryRecap : Best/Worst day (Entry), Text (String), Average mood (Double) - Class used to show the recap over week/month/year to the user Also add another parameter (int type) that will tell the frontend team if it's week/month/year to show it in different colors.
process
work on entryrecap best worst day entry text string average mood double class used to show the recap over week month year to the user also add another parameter int type that will tell the frontend team if it s week month year to show it in different colors
1
509,177
14,723,296,470
IssuesEvent
2021-01-06 00:09:23
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
opened
"Creator" button needed in Tools-Customize to display at search results
Bug Collection Type - Cultural Collections Function-SearchOrDownload Priority-Critical
Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html **Describe the bug** Maker --> Creator resulted in customized search results showing no data for Maker. There is no creator field to add in the customize box. Please change maker to creator in the customize box so we can add this field to search results and see who made our stuff! **To Reproduce** Steps to reproduce the behavior: 1. Go to https://arctos.database.museum/saved/Firearms 2. <img width="1196" alt="Screen Shot 2021-01-05 at 2 47 16 PM" src="https://user-images.githubusercontent.com/17605945/103712898-6b93b480-4f67-11eb-9736-157f792ae3b8.png"> 3. Open record and see there is a creator listed: <img width="1173" alt="Screen Shot 2021-01-05 at 2 51 33 PM" src="https://user-images.githubusercontent.com/17605945/103712925-7bab9400-4f67-11eb-89ba-859ddcfafcd8.png"> 4. See that there is no creator in customize tool box: <img width="726" alt="Screen Shot 2021-01-05 at 2 47 02 PM" src="https://user-images.githubusercontent.com/17605945/103712958-8f56fa80-4f67-11eb-8690-49b1aba72cc7.png"> **Expected behavior** The information that previously displayed in Maker column should be accessible via Creator, which needs to be added to the customize box. **Priority** Super high priority! Makes us look like we don't know our stuff!
1.0
"Creator" button needed in Tools-Customize to display at search results - Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html **Describe the bug** Maker --> Creator resulted in customized search results showing no data for Maker. There is no creator field to add in the customize box. Please change maker to creator in the customize box so we can add this field to search results and see who made our stuff! **To Reproduce** Steps to reproduce the behavior: 1. Go to https://arctos.database.museum/saved/Firearms 2. <img width="1196" alt="Screen Shot 2021-01-05 at 2 47 16 PM" src="https://user-images.githubusercontent.com/17605945/103712898-6b93b480-4f67-11eb-9736-157f792ae3b8.png"> 3. Open record and see there is a creator listed: <img width="1173" alt="Screen Shot 2021-01-05 at 2 51 33 PM" src="https://user-images.githubusercontent.com/17605945/103712925-7bab9400-4f67-11eb-89ba-859ddcfafcd8.png"> 4. See that there is no creator in customize tool box: <img width="726" alt="Screen Shot 2021-01-05 at 2 47 02 PM" src="https://user-images.githubusercontent.com/17605945/103712958-8f56fa80-4f67-11eb-8690-49b1aba72cc7.png"> **Expected behavior** The information that previously displayed in Maker column should be accessible via Creator, which needs to be added to the customize box. **Priority** Super high priority! Makes us look like we don't know our stuff!
non_process
creator button needed in tools customize to display at search results issue documentation is describe the bug maker creator resulted in customized search results showing no data for maker there is no creator field to add in the customize box please change maker to creator in the customize box so we can add this field to search results and see who made our stuff to reproduce steps to reproduce the behavior go to img width alt screen shot at pm src open record and see there is a creator listed img width alt screen shot at pm src see that there is no creator in customize tool box img width alt screen shot at pm src expected behavior the information that previously displayed in maker column should be accessible via creator which needs to be added to the customize box priority super high priority makes us look like we don t know our stuff
0
8,879
11,980,553,695
IssuesEvent
2020-04-07 09:32:08
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Investigate origin of `AWS Lambda timed out` (#1754)
bug/2-confirmed kind/bug process/candidate topic: deployment-platforms
I found a solution to #1754 [reproduction is here](https://github.com/UWellFail/prisma2--1754) https://github.com/prisma/prisma2/issues/1754#issuecomment-601271228 ``` // Set to false to send the response right away when the callback executes, instead of waiting for the Node.js event loop to be empty. context.callbackWaitsForEmptyEventLoop = false; ``` But this means that the Node.js event loop is not empty and it should be empty. ### Without `context.callbackWaitsForEmptyEventLoop = false` A POST request works perfectly. Only the GET requests in this example have a non empty event loop. We could start investigating with: - Try out pure Prisma Client in lambda without the apollo server overhead, get the GET and POST event running - Instead of the callback based approach, try using a promise and play with async: true as described here https://github.com/serverless/serverless/issues/3171#issuecomment-540101244 - Try to get serverless offline running with a simple prisma client example https://serverless.com/plugins/serverless-offline/
1.0
Investigate origin of `AWS Lambda timed out` (#1754) - I found a solution to #1754 [reproduction is here](https://github.com/UWellFail/prisma2--1754) https://github.com/prisma/prisma2/issues/1754#issuecomment-601271228 ``` // Set to false to send the response right away when the callback executes, instead of waiting for the Node.js event loop to be empty. context.callbackWaitsForEmptyEventLoop = false; ``` But this means that the Node.js event loop is not empty and it should be empty. ### Without `context.callbackWaitsForEmptyEventLoop = false` A POST request works perfectly. Only the GET requests in this example have a non empty event loop. We could start investigating with: - Try out pure Prisma Client in lambda without the apollo server overhead, get the GET and POST event running - Instead of the callback based approach, try using a promise and play with async: true as described here https://github.com/serverless/serverless/issues/3171#issuecomment-540101244 - Try to get serverless offline running with a simple prisma client example https://serverless.com/plugins/serverless-offline/
process
investigate origin of aws lambda timed out i found a solution to set to false to send the response right away when the callback executes instead of waiting for the node js event loop to be empty context callbackwaitsforemptyeventloop false but this means that the node js event loop is not empty and it should be empty without context callbackwaitsforemptyeventloop false a post request works perfectly only the get requests in this example have a non empty event loop we could start investigating with try out pure prisma client in lambda without the apollo server overhead get the get and post event running instead of the callback based approach try using a promise and play with async true as described here try to get serverless offline running with a simple prisma client example
1
1,510
4,103,283,526
IssuesEvent
2016-06-04 15:32:37
ProgrammingLife2016/PL2-2016
https://api.github.com/repos/ProgrammingLife2016/PL2-2016
opened
Edge drawing optimization
data processing development enhancement gui optimization
Investigate if it is possible to draw the edges more efficiently, or otherwise draw fewer edges. The application zooms/scrolls very fast when the edges are not drawn. When the edges are drawn this is not always the case (especially for larger datasets), because there can be 100's of thousands of edges at the same time and currently a separate JavaFX Line object is created for each edge. This can possibly be done more efficient if something like a Canvas is used, where it is not necessary to create a new object for every drawn line.
1.0
Edge drawing optimization - Investigate if it is possible to draw the edges more efficiently, or otherwise draw fewer edges. The application zooms/scrolls very fast when the edges are not drawn. When the edges are drawn this is not always the case (especially for larger datasets), because there can be 100's of thousands of edges at the same time and currently a separate JavaFX Line object is created for each edge. This can possibly be done more efficient if something like a Canvas is used, where it is not necessary to create a new object for every drawn line.
process
edge drawing optimization investigate if it is possible to draw the edges more efficiently or otherwise draw fewer edges the application zooms scrolls very fast when the edges are not drawn when the edges are drawn this is not always the case especially for larger datasets because there can be s of thousands of edges at the same time and currently a separate javafx line object is created for each edge this can possibly be done more efficient if something like a canvas is used where it is not necessary to create a new object for every drawn line
1
550,977
16,135,506,786
IssuesEvent
2021-04-29 11:17:28
bounswe/2021SpringGroup9
https://api.github.com/repos/bounswe/2021SpringGroup9
opened
Research about UML Use Case, Class and Sequence Diagrams
priority: high research status: in progress
I start doing research about UML diagrams and in particular how to use the arrows in specific cases such as inheritance and so on.
1.0
Research about UML Use Case, Class and Sequence Diagrams - I start doing research about UML diagrams and in particular how to use the arrows in specific cases such as inheritance and so on.
non_process
research about uml use case class and sequence diagrams i start doing research about uml diagrams and in particular how to use the arrows in specific cases such as inheritance and so on
0
17,903
23,877,390,484
IssuesEvent
2022-09-07 20:30:16
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Requested files vs new visitors a day, discrepency in hits and visitors
question log-processing
Hi, I am unsure if it is a bug, but when I see the panel of unquie visitors a day including spiders it shows around 1400 hits and 220 visitors. But if I look at requested URLS, it shows what I believe to be more accurate of almost the same number of visitors and hits. ![image](https://user-images.githubusercontent.com/7340380/175856190-85d94b26-34a7-4e1b-80e9-1427f6421477.png) So not sure since the domain ends in .to if there is a regex issue? The command I use to make this report is `cd /var/log/nginx && zcat access.log.*.gz | goaccess --4xx-to-unique-count access.log access.log.1 - -o report.html --log-format=COMBINED` Any clarification this would help
1.0
Requested files vs new visitors a day, discrepency in hits and visitors - Hi, I am unsure if it is a bug, but when I see the panel of unquie visitors a day including spiders it shows around 1400 hits and 220 visitors. But if I look at requested URLS, it shows what I believe to be more accurate of almost the same number of visitors and hits. ![image](https://user-images.githubusercontent.com/7340380/175856190-85d94b26-34a7-4e1b-80e9-1427f6421477.png) So not sure since the domain ends in .to if there is a regex issue? The command I use to make this report is `cd /var/log/nginx && zcat access.log.*.gz | goaccess --4xx-to-unique-count access.log access.log.1 - -o report.html --log-format=COMBINED` Any clarification this would help
process
requested files vs new visitors a day discrepency in hits and visitors hi i am unsure if it is a bug but when i see the panel of unquie visitors a day including spiders it shows around hits and visitors but if i look at requested urls it shows what i believe to be more accurate of almost the same number of visitors and hits so not sure since the domain ends in to if there is a regex issue the command i use to make this report is cd var log nginx zcat access log gz goaccess to unique count access log access log o report html log format combined any clarification this would help
1
238,446
26,112,059,459
IssuesEvent
2022-12-27 21:43:33
microsoft/ebpf-for-windows
https://api.github.com/repos/microsoft/ebpf-for-windows
closed
AddressSanitizer: heap-buffer-overflow in btf_parse_line_information
bug triaged security
https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129 ``` ================================================================= ==5[976](https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129#step:16:977)==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x126a01034074 at pc 0x7ff72c0cbe6a bp 0x00f6bbb6cc80 sp 0x00f6bbb6cc80 READ of size 4 at 0x126a01034074 thread T0 #0 0x7ff72c0cbe69 in btf_parse_line_information(class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::function<(class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, unsigned int)>) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14021be69) #1 0x7ff72c041eeb in read_elf(class std::basic_istream<char, struct std::char_traits<char>> &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, struct ebpf_verifier_options_t const *, struct ebpf_platform_t const *) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140191eeb) #2 0x7ff72bec2fc6 in _ebpf_api_elf_verify_section_from_stream D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:516 #3 0x7ff72bec4227 in _verify_section_from_string D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:605 #4 0x7ff72bec4631 in ebpf_api_elf_verify_section_from_memory D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:637 #5 0x7ff72bf9c302 in LLVMFuzzerTestOneInput D:\a\ebpf-for-windows\ebpf-for-windows\tests\libfuzzer\verifier\libfuzz_harness.cpp:18 #6 0x7ff72bff5f9f in fuzzer::Fuzzer::ExecuteCallback(unsigned char const *, unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:559 #7 0x7ff72bff80fe in fuzzer::Fuzzer::RunOne(unsigned char const *, unsigned __int64, bool, struct fuzzer::InputInfo *, bool *) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:471 #8 0x7ff72bff6bbd in fuzzer::Fuzzer::MutateAndTestOne(void) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:702 #9 0x7ff72bff6683 in fuzzer::Fuzzer::Loop(class std::vector<struct fuzzer::SizedFile, class fuzzer::fuzzer_allocator<struct fuzzer::SizedFile>> &) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:838 #10 0x7ff72bfeafe9 in fuzzer::FuzzerDriver(int *, char ***, int (__cdecl *)(unsigned char const *, unsigned __int64)) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerDriver.cpp:853 #11 0x7ff72bfdfe82 in main D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerMain.cpp:20 #12 0x7ff72bfdac7b in invoke_main D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78 #13 0x7ff72bfdac7b in __scrt_common_main_seh D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288 #14 0x7ffc1a1a4ecf (C:\Windows\System32\KERNEL32.DLL+0x180014ecf) #15 0x7ffc1a7be40a (C:\Windows\SYSTEM32\ntdll.dll+0x18007e40a) 0x126a01034076 is located 0 bytes to the right of 54-byte region [0x126a01034040,0x126a01034076) allocated by thread T0 here: #0 0x7ff72bfd407a in operator new(unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\asan\asan_win_new_scalar_thunk.cpp:41 #1 0x7ff72bf3065e in std::allocator<unsigned char>::allocate C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\xmemory:835 #2 0x7ff72bf3065e in std::vector<unsigned char, class std::allocator<unsigned char>>::_Buy_raw(unsigned __int64) C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector:2052 #3 0x7ff72c02e050 in std::vector<unsigned char, class std::allocator<unsigned char>>::_Construct_n<unsigned char *, unsigned char *>(unsigned __int64, unsigned char *&&, unsigned char *&&) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14017e050) #4 0x7ff72c02bf3d in std::vector<unsigned char, class std::allocator<unsigned char>>::vector<unsigned char, class std::allocator<unsigned char>><unsigned char *, 0>(unsigned char *, unsigned char *, class std::allocator<unsigned char> const &) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14017bf3d) #5 0x7ff72c034705 in std::_Tree<class std::_Tset_traits<unsigned short, struct std::less<unsigned short>, class std::allocator<unsigned short>, 0>>::insert<0, 0>(unsigned short &&) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140184705) #6 0x7ff72c041ec2 in read_elf(class std::basic_istream<char, struct std::char_traits<char>> &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, struct ebpf_verifier_options_t const *, struct ebpf_platform_t const *) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140191ec2) #7 0x7ff72bec2fc6 in _ebpf_api_elf_verify_section_from_stream D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:516 #8 0x7ff72bec4227 in _verify_section_from_string D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:605 #9 0x7ff72bec4631 in ebpf_api_elf_verify_section_from_memory D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:637 #10 0x7ff72bf9c302 in LLVMFuzzerTestOneInput D:\a\ebpf-for-windows\ebpf-for-windows\tests\libfuzzer\verifier\libfuzz_harness.cpp:18 #11 0x7ff72bff5f9f in fuzzer::Fuzzer::ExecuteCallback(unsigned char const *, unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:559 #12 0x7ff72bff80fe in fuzzer::Fuzzer::RunOne(unsigned char const *, unsigned __int64, bool, struct fuzzer::InputInfo *, bool *) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:471 #13 0x7ff72bff6bbd in fuzzer::Fuzzer::MutateAndTestOne(void) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:702 #14 0x7ff72bff6683 in fuzzer::Fuzzer::Loop(class std::vector<struct fuzzer::SizedFile, class fuzzer::fuzzer_allocator<struct fuzzer::SizedFile>> &) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:838 #15 0x7ff72bfeafe9 in fuzzer::FuzzerDriver(int *, char ***, int (__cdecl *)(unsigned char const *, unsigned __int64)) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerDriver.cpp:853 #16 0x7ff72bfdfe82 in main D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerMain.cpp:20 #17 0x7ff72bfdac7b in invoke_main D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78 #18 0x7ff72bfdac7b in __scrt_common_main_seh D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288 #19 0x7ffc1a1a4ecf (C:\Windows\System32\KERNEL32.DLL+0x180014ecf) #20 0x7ffc1a7be40a (C:\Windows\SYSTEM32\ntdll.dll+0x18007e40a) SUMMARY: AddressSanitizer: heap-buffer-overflow (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14021be69) in btf_parse_line_information(class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::function<(class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, unsigned int)>) Shadow bytes around the buggy address: 0x04ab410067b0: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa 0x04ab410067c0: fd fd fd fd fd fd fd fa fa fa fa fa fd fd fd fd 0x04ab410067d0: fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd 0x04ab410067e0: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa 0x04ab410067f0: fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd =>0x04ab41006800: fd fd fd fd fa fa fa fa 00 00 00 00 00 00[06]fa 0x04ab41006810: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x04ab41006820: fa fa fa fa fa fa fa fa fa fa fa fa fd fd fd fd 0x04ab41006830: fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd 0x04ab41006840: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa 0x04ab41006850: fd fd fd fd fd fd fd fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==5976==ABORTING MS: 3 ChangeASCIIInt-ChangeByte-CMP- DE: "6\x00\x00\x00\x00\x00\x00\x00"-; base unit: 8e6ec735553a53dafa51712487c621a5e42321b6 artifact_prefix='Artifacts\'; Test unit written to Artifacts\crash-b7220d69a1e24129[984](https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129#step:16:985)c0a1d855e5cd1146f6938 ```
True
AddressSanitizer: heap-buffer-overflow in btf_parse_line_information - https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129 ``` ================================================================= ==5[976](https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129#step:16:977)==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x126a01034074 at pc 0x7ff72c0cbe6a bp 0x00f6bbb6cc80 sp 0x00f6bbb6cc80 READ of size 4 at 0x126a01034074 thread T0 #0 0x7ff72c0cbe69 in btf_parse_line_information(class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::function<(class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, unsigned int)>) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14021be69) #1 0x7ff72c041eeb in read_elf(class std::basic_istream<char, struct std::char_traits<char>> &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, struct ebpf_verifier_options_t const *, struct ebpf_platform_t const *) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140191eeb) #2 0x7ff72bec2fc6 in _ebpf_api_elf_verify_section_from_stream D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:516 #3 0x7ff72bec4227 in _verify_section_from_string D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:605 #4 0x7ff72bec4631 in ebpf_api_elf_verify_section_from_memory D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:637 #5 0x7ff72bf9c302 in LLVMFuzzerTestOneInput D:\a\ebpf-for-windows\ebpf-for-windows\tests\libfuzzer\verifier\libfuzz_harness.cpp:18 #6 0x7ff72bff5f9f in fuzzer::Fuzzer::ExecuteCallback(unsigned char const *, unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:559 #7 0x7ff72bff80fe in fuzzer::Fuzzer::RunOne(unsigned char const *, unsigned __int64, bool, struct fuzzer::InputInfo *, bool *) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:471 #8 0x7ff72bff6bbd in fuzzer::Fuzzer::MutateAndTestOne(void) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:702 #9 0x7ff72bff6683 in fuzzer::Fuzzer::Loop(class std::vector<struct fuzzer::SizedFile, class fuzzer::fuzzer_allocator<struct fuzzer::SizedFile>> &) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:838 #10 0x7ff72bfeafe9 in fuzzer::FuzzerDriver(int *, char ***, int (__cdecl *)(unsigned char const *, unsigned __int64)) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerDriver.cpp:853 #11 0x7ff72bfdfe82 in main D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerMain.cpp:20 #12 0x7ff72bfdac7b in invoke_main D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78 #13 0x7ff72bfdac7b in __scrt_common_main_seh D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288 #14 0x7ffc1a1a4ecf (C:\Windows\System32\KERNEL32.DLL+0x180014ecf) #15 0x7ffc1a7be40a (C:\Windows\SYSTEM32\ntdll.dll+0x18007e40a) 0x126a01034076 is located 0 bytes to the right of 54-byte region [0x126a01034040,0x126a01034076) allocated by thread T0 here: #0 0x7ff72bfd407a in operator new(unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\asan\asan_win_new_scalar_thunk.cpp:41 #1 0x7ff72bf3065e in std::allocator<unsigned char>::allocate C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\xmemory:835 #2 0x7ff72bf3065e in std::vector<unsigned char, class std::allocator<unsigned char>>::_Buy_raw(unsigned __int64) C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector:2052 #3 0x7ff72c02e050 in std::vector<unsigned char, class std::allocator<unsigned char>>::_Construct_n<unsigned char *, unsigned char *>(unsigned __int64, unsigned char *&&, unsigned char *&&) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14017e050) #4 0x7ff72c02bf3d in std::vector<unsigned char, class std::allocator<unsigned char>>::vector<unsigned char, class std::allocator<unsigned char>><unsigned char *, 0>(unsigned char *, unsigned char *, class std::allocator<unsigned char> const &) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14017bf3d) #5 0x7ff72c034705 in std::_Tree<class std::_Tset_traits<unsigned short, struct std::less<unsigned short>, class std::allocator<unsigned short>, 0>>::insert<0, 0>(unsigned short &&) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140184705) #6 0x7ff72c041ec2 in read_elf(class std::basic_istream<char, struct std::char_traits<char>> &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, struct ebpf_verifier_options_t const *, struct ebpf_platform_t const *) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140191ec2) #7 0x7ff72bec2fc6 in _ebpf_api_elf_verify_section_from_stream D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:516 #8 0x7ff72bec4227 in _verify_section_from_string D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:605 #9 0x7ff72bec4631 in ebpf_api_elf_verify_section_from_memory D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:637 #10 0x7ff72bf9c302 in LLVMFuzzerTestOneInput D:\a\ebpf-for-windows\ebpf-for-windows\tests\libfuzzer\verifier\libfuzz_harness.cpp:18 #11 0x7ff72bff5f9f in fuzzer::Fuzzer::ExecuteCallback(unsigned char const *, unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:559 #12 0x7ff72bff80fe in fuzzer::Fuzzer::RunOne(unsigned char const *, unsigned __int64, bool, struct fuzzer::InputInfo *, bool *) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:471 #13 0x7ff72bff6bbd in fuzzer::Fuzzer::MutateAndTestOne(void) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:702 #14 0x7ff72bff6683 in fuzzer::Fuzzer::Loop(class std::vector<struct fuzzer::SizedFile, class fuzzer::fuzzer_allocator<struct fuzzer::SizedFile>> &) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:838 #15 0x7ff72bfeafe9 in fuzzer::FuzzerDriver(int *, char ***, int (__cdecl *)(unsigned char const *, unsigned __int64)) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerDriver.cpp:853 #16 0x7ff72bfdfe82 in main D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerMain.cpp:20 #17 0x7ff72bfdac7b in invoke_main D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78 #18 0x7ff72bfdac7b in __scrt_common_main_seh D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288 #19 0x7ffc1a1a4ecf (C:\Windows\System32\KERNEL32.DLL+0x180014ecf) #20 0x7ffc1a7be40a (C:\Windows\SYSTEM32\ntdll.dll+0x18007e40a) SUMMARY: AddressSanitizer: heap-buffer-overflow (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14021be69) in btf_parse_line_information(class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::function<(class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, unsigned int)>) Shadow bytes around the buggy address: 0x04ab410067b0: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa 0x04ab410067c0: fd fd fd fd fd fd fd fa fa fa fa fa fd fd fd fd 0x04ab410067d0: fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd 0x04ab410067e0: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa 0x04ab410067f0: fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd =>0x04ab41006800: fd fd fd fd fa fa fa fa 00 00 00 00 00 00[06]fa 0x04ab41006810: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa 0x04ab41006820: fa fa fa fa fa fa fa fa fa fa fa fa fd fd fd fd 0x04ab41006830: fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd 0x04ab41006840: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa 0x04ab41006850: fd fd fd fd fd fd fd fa fa fa fa fa fa fa fa fa Shadow byte legend (one shadow byte represents 8 application bytes): Addressable: 00 Partially addressable: 01 02 03 04 05 06 07 Heap left redzone: fa Freed heap region: fd Stack left redzone: f1 Stack mid redzone: f2 Stack right redzone: f3 Stack after return: f5 Stack use after scope: f8 Global redzone: f9 Global init order: f6 Poisoned by user: f7 Container overflow: fc Array cookie: ac Intra object redzone: bb ASan internal: fe Left alloca redzone: ca Right alloca redzone: cb Shadow gap: cc ==5976==ABORTING MS: 3 ChangeASCIIInt-ChangeByte-CMP- DE: "6\x00\x00\x00\x00\x00\x00\x00"-; base unit: 8e6ec735553a53dafa51712487c621a5e42321b6 artifact_prefix='Artifacts\'; Test unit written to Artifacts\crash-b7220d69a1e24129[984](https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129#step:16:985)c0a1d855e5cd1146f6938 ```
non_process
addresssanitizer heap buffer overflow in btf parse line information addresssanitizer heap buffer overflow on address at pc bp sp read of size at thread in btf parse line information class std vector const class std vector const class std function class std allocator const unsigned int class std basic string class std allocator const class std basic string class std allocator const unsigned int unsigned int d a ebpf for windows ebpf for windows release verifier fuzzer exe in read elf class std basic istream class std basic string class std allocator const class std basic string class std allocator const struct ebpf verifier options t const struct ebpf platform t const d a ebpf for windows ebpf for windows release verifier fuzzer exe in ebpf api elf verify section from stream d a ebpf for windows ebpf for windows libs api verifier cpp in verify section from string d a ebpf for windows ebpf for windows libs api verifier cpp in ebpf api elf verify section from memory d a ebpf for windows ebpf for windows libs api verifier cpp in llvmfuzzertestoneinput d a ebpf for windows ebpf for windows tests libfuzzer verifier libfuzz harness cpp in fuzzer fuzzer executecallback unsigned char const unsigned d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer runone unsigned char const unsigned bool struct fuzzer inputinfo bool d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer mutateandtestone void d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer loop class std vector d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzerdriver int char int cdecl unsigned char const unsigned d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerdriver cpp in main d a work s src vctools asan llvm compiler rt lib fuzzer fuzzermain cpp in invoke main d a work s src vctools crt vcstartup src startup exe common inl in scrt common main seh d a work s src vctools crt vcstartup src startup exe common inl c windows dll c windows ntdll dll is located bytes to the right of byte region allocated by thread here in operator new unsigned d a work s src vctools asan llvm compiler rt lib asan asan win new scalar thunk cpp in std allocator allocate c program files microsoft visual studio enterprise vc tools msvc include xmemory in std vector buy raw unsigned c program files microsoft visual studio enterprise vc tools msvc include vector in std vector construct n unsigned unsigned char unsigned char d a ebpf for windows ebpf for windows release verifier fuzzer exe in std vector vector unsigned char unsigned char class std allocator const d a ebpf for windows ebpf for windows release verifier fuzzer exe in std tree class std allocator insert unsigned short d a ebpf for windows ebpf for windows release verifier fuzzer exe in read elf class std basic istream class std basic string class std allocator const class std basic string class std allocator const struct ebpf verifier options t const struct ebpf platform t const d a ebpf for windows ebpf for windows release verifier fuzzer exe in ebpf api elf verify section from stream d a ebpf for windows ebpf for windows libs api verifier cpp in verify section from string d a ebpf for windows ebpf for windows libs api verifier cpp in ebpf api elf verify section from memory d a ebpf for windows ebpf for windows libs api verifier cpp in llvmfuzzertestoneinput d a ebpf for windows ebpf for windows tests libfuzzer verifier libfuzz harness cpp in fuzzer fuzzer executecallback unsigned char const unsigned d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer runone unsigned char const unsigned bool struct fuzzer inputinfo bool d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer mutateandtestone void d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer loop class std vector d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzerdriver int char int cdecl unsigned char const unsigned d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerdriver cpp in main d a work s src vctools asan llvm compiler rt lib fuzzer fuzzermain cpp in invoke main d a work s src vctools crt vcstartup src startup exe common inl in scrt common main seh d a work s src vctools crt vcstartup src startup exe common inl c windows dll c windows ntdll dll summary addresssanitizer heap buffer overflow d a ebpf for windows ebpf for windows release verifier fuzzer exe in btf parse line information class std vector const class std vector const class std function class std allocator const unsigned int class std basic string class std allocator const class std basic string class std allocator const unsigned int unsigned int shadow bytes around the buggy address fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fa fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fa fa fa fa fa fa fa fa fa shadow byte legend one shadow byte represents application bytes addressable partially addressable heap left redzone fa freed heap region fd stack left redzone stack mid redzone stack right redzone stack after return stack use after scope global redzone global init order poisoned by user container overflow fc array cookie ac intra object redzone bb asan internal fe left alloca redzone ca right alloca redzone cb shadow gap cc aborting ms changeasciiint changebyte cmp de base unit artifact prefix artifacts test unit written to artifacts crash
0
231,524
18,772,411,328
IssuesEvent
2021-11-07 03:30:02
ProjectStarlight/StarlightRiver
https://api.github.com/repos/ProjectStarlight/StarlightRiver
closed
scrapshot grapple not working as intended when the snek enemy teleports mid-grapple
demo Checked - Test again
## Issue Description Expected behavior: If a grappled Snek were to teleport, I would expect the grapple to break. Actual behavior: The player is dragged to the top of the map by the Snek. Context: I was grappling Sneks with the Scrapshot, and suddenly I was on top of the map. Error message (if applicable): The first time this happened I was given a System.OutOfMemoryException, however on subsequent attempts to reproduce the game did not crash. ## Reproduction Mod list: Starlight River Config settings: No custom mod config. Game settings: All settings default with the exception of custom keybinds and volume. Gameplay events happening: Fighting a Snek with the Scrapshot and a set of Breacher armor. Your inputs: All you need to do to recreate this is to time the grapple of the Scrapshot so that the Snek teleports before you reach it. OS: Windows 10 Pro 64 bit ## Additional context and log files **[client.log](https://github.com/ProjectStarlight/StarlightRiver/files/7341039/client.log)**
1.0
scrapshot grapple not working as intended when the snek enemy teleports mid-grapple - ## Issue Description Expected behavior: If a grappled Snek were to teleport, I would expect the grapple to break. Actual behavior: The player is dragged to the top of the map by the Snek. Context: I was grappling Sneks with the Scrapshot, and suddenly I was on top of the map. Error message (if applicable): The first time this happened I was given a System.OutOfMemoryException, however on subsequent attempts to reproduce the game did not crash. ## Reproduction Mod list: Starlight River Config settings: No custom mod config. Game settings: All settings default with the exception of custom keybinds and volume. Gameplay events happening: Fighting a Snek with the Scrapshot and a set of Breacher armor. Your inputs: All you need to do to recreate this is to time the grapple of the Scrapshot so that the Snek teleports before you reach it. OS: Windows 10 Pro 64 bit ## Additional context and log files **[client.log](https://github.com/ProjectStarlight/StarlightRiver/files/7341039/client.log)**
non_process
scrapshot grapple not working as intended when the snek enemy teleports mid grapple issue description expected behavior if a grappled snek were to teleport i would expect the grapple to break actual behavior the player is dragged to the top of the map by the snek context i was grappling sneks with the scrapshot and suddenly i was on top of the map error message if applicable the first time this happened i was given a system outofmemoryexception however on subsequent attempts to reproduce the game did not crash reproduction mod list starlight river config settings no custom mod config game settings all settings default with the exception of custom keybinds and volume gameplay events happening fighting a snek with the scrapshot and a set of breacher armor your inputs all you need to do to recreate this is to time the grapple of the scrapshot so that the snek teleports before you reach it os windows pro bit additional context and log files
0
4,581
23,793,802,947
IssuesEvent
2022-09-02 17:08:43
Vivelin/SMZ3Randomizer
https://api.github.com/repos/Vivelin/SMZ3Randomizer
opened
Split out configs into a separate project
:wrench: maintainability
The current config system can't be accessed in the randomizer project. Because of that, the configs should be added to a unique project so that they can be accessed globally.
True
Split out configs into a separate project - The current config system can't be accessed in the randomizer project. Because of that, the configs should be added to a unique project so that they can be accessed globally.
non_process
split out configs into a separate project the current config system can t be accessed in the randomizer project because of that the configs should be added to a unique project so that they can be accessed globally
0
18,882
24,820,798,387
IssuesEvent
2022-10-25 16:16:19
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/resourcedetection] Flaky test - TestDetectResource_Parallel
bug help wanted good first issue priority:p3 flaky test processor/resourcedetection
Seen in https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6580882988?check_suite_focus=true ``` === RUN TestDetectResource_Parallel ================== WARNING: DATA RACE Read at 0x00c000216ba0 by goroutine 20: go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0x78 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Previous write at 0x00c000216ba0 by goroutine 19: reflect.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0 internal/reflectlite.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Goroutine 20 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 Goroutine 19 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 ================== ================== WARNING: DATA RACE Read at 0x00c000216b80 by goroutine 21: go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0xd4 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:113 +0x1f35 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:127 +0xf79 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.DeepEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:237 +0x3f9 github.com/stretchr/testify/assert.ObjectsAreEqual() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:65 +0x184 github.com/stretchr/testify/assert.Equal() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:343 +0x212 github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:232 +0x16a Previous write at 0x00c000216b90 by goroutine 19: reflect.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0 internal/reflectlite.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Goroutine 20 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 Goroutine 19 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 ================== testing.go:1312: race detected during execution of test --- FAIL: TestDetectResource_Parallel (0.01s)=== RUN TestDetectResource_Parallel ================== WARNING: DATA RACE Read at 0x00c000216ba0 by goroutine 20: go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0x78 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Previous write at 0x00c000216ba0 by goroutine 19: reflect.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0 internal/reflectlite.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Goroutine 20 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 Goroutine 19 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 ================== ================== WARNING: DATA RACE Read at 0x00c000216b80 by goroutine 21: go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0xd4 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:113 +0x1f35 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:127 +0xf79 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.DeepEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:237 +0x3f9 github.com/stretchr/testify/assert.ObjectsAreEqual() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:65 +0x184 github.com/stretchr/testify/assert.Equal() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:343 +0x212 github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:232 +0x16a Previous write at 0x00c000216b90 by goroutine 19: reflect.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0 internal/reflectlite.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Goroutine 20 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 Goroutine 19 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 ================== testing.go:1312: race detected during execution of test --- FAIL: TestDetectResource_Parallel (0.01s) ```
1.0
[processor/resourcedetection] Flaky test - TestDetectResource_Parallel - Seen in https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6580882988?check_suite_focus=true ``` === RUN TestDetectResource_Parallel ================== WARNING: DATA RACE Read at 0x00c000216ba0 by goroutine 20: go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0x78 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Previous write at 0x00c000216ba0 by goroutine 19: reflect.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0 internal/reflectlite.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Goroutine 20 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 Goroutine 19 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 ================== ================== WARNING: DATA RACE Read at 0x00c000216b80 by goroutine 21: go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0xd4 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:113 +0x1f35 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:127 +0xf79 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.DeepEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:237 +0x3f9 github.com/stretchr/testify/assert.ObjectsAreEqual() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:65 +0x184 github.com/stretchr/testify/assert.Equal() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:343 +0x212 github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:232 +0x16a Previous write at 0x00c000216b90 by goroutine 19: reflect.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0 internal/reflectlite.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Goroutine 20 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 Goroutine 19 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 ================== testing.go:1312: race detected during execution of test --- FAIL: TestDetectResource_Parallel (0.01s)=== RUN TestDetectResource_Parallel ================== WARNING: DATA RACE Read at 0x00c000216ba0 by goroutine 20: go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0x78 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Previous write at 0x00c000216ba0 by goroutine 19: reflect.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0 internal/reflectlite.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Goroutine 20 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 Goroutine 19 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 ================== ================== WARNING: DATA RACE Read at 0x00c000216b80 by goroutine 21: go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0xd4 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:113 +0x1f35 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:127 +0xf79 reflect.deepValueEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b reflect.DeepEqual() /opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:237 +0x3f9 github.com/stretchr/testify/assert.ObjectsAreEqual() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:65 +0x184 github.com/stretchr/testify/assert.Equal() /home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:343 +0x212 github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:232 +0x16a Previous write at 0x00c000216b90 by goroutine 19: reflect.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0 internal/reflectlite.typedmemmove() /opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18 sort.insertionSort_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94 sort.stable_func() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c sort.SliceStable() /opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae go.opentelemetry.io/collector/pdata/internal.Map.Sort() /home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133 Goroutine 20 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 Goroutine 19 (running) created at: github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel() /home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1 testing.tRunner() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213 testing.(*T).Run.func1() /opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47 ================== testing.go:1312: race detected during execution of test --- FAIL: TestDetectResource_Parallel (0.01s) ```
process
flaky test testdetectresource parallel seen in run testdetectresource parallel warning data race read at by goroutine go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go previous write at by goroutine reflect typedmemmove opt hostedtoolcache go src runtime mbarrier go internal reflectlite typedmemmove opt hostedtoolcache go src runtime mbarrier go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go warning data race read at by goroutine go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepequal opt hostedtoolcache go src reflect deepequal go github com stretchr testify assert objectsareequal home runner go pkg mod github com stretchr testify assert assertions go github com stretchr testify assert equal home runner go pkg mod github com stretchr testify assert assertions go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go previous write at by goroutine reflect typedmemmove opt hostedtoolcache go src runtime mbarrier go internal reflectlite typedmemmove opt hostedtoolcache go src runtime mbarrier go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go testing go race detected during execution of test fail testdetectresource parallel run testdetectresource parallel warning data race read at by goroutine go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go previous write at by goroutine reflect typedmemmove opt hostedtoolcache go src runtime mbarrier go internal reflectlite typedmemmove opt hostedtoolcache go src runtime mbarrier go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go warning data race read at by goroutine go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepequal opt hostedtoolcache go src reflect deepequal go github com stretchr testify assert objectsareequal home runner go pkg mod github com stretchr testify assert assertions go github com stretchr testify assert equal home runner go pkg mod github com stretchr testify assert assertions go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go previous write at by goroutine reflect typedmemmove opt hostedtoolcache go src runtime mbarrier go internal reflectlite typedmemmove opt hostedtoolcache go src runtime mbarrier go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go testing go race detected during execution of test fail testdetectresource parallel
1
18,391
24,522,416,044
IssuesEvent
2022-10-11 10:31:34
streamnative/flink
https://api.github.com/repos/streamnative/flink
closed
[Enhancement][Stream] Remove pulsar.consumer.autoUpdatePartitionsIntervalSeconds
compute/data-processing type/enhancement
This config option is useless, we should remove it now.
1.0
[Enhancement][Stream] Remove pulsar.consumer.autoUpdatePartitionsIntervalSeconds - This config option is useless, we should remove it now.
process
remove pulsar consumer autoupdatepartitionsintervalseconds this config option is useless we should remove it now
1
15,753
19,911,722,699
IssuesEvent
2022-01-25 17:50:11
input-output-hk/high-assurance-legacy
https://api.github.com/repos/input-output-hk/high-assurance-legacy
closed
Generalize the proofs of the basic bisimilarity core laws
type: enhancement language: isabelle topic: process calculus
We have implemented a number of core laws about the basic strong bisimilarity relation, namely compatibility (preservation) laws, scope extension laws, and neutrality, associativity, and commutativity of parallel composition. These laws have counterparts for the basic _weak_ bisimilarity relation, but these haven’t been proved yet. Incidentally, the proofs of the core laws for the strong case can be easily generalized to also cover the weak case. This is because the weak analogs of all basic transition rules hold (see #76). The generalization works by replacing all applications of basic transition rules with applications of fact parameters that can be specialized to either the basic transition rules or their weak analogs. Our goal is to generalize the proofs in this way.
1.0
Generalize the proofs of the basic bisimilarity core laws - We have implemented a number of core laws about the basic strong bisimilarity relation, namely compatibility (preservation) laws, scope extension laws, and neutrality, associativity, and commutativity of parallel composition. These laws have counterparts for the basic _weak_ bisimilarity relation, but these haven’t been proved yet. Incidentally, the proofs of the core laws for the strong case can be easily generalized to also cover the weak case. This is because the weak analogs of all basic transition rules hold (see #76). The generalization works by replacing all applications of basic transition rules with applications of fact parameters that can be specialized to either the basic transition rules or their weak analogs. Our goal is to generalize the proofs in this way.
process
generalize the proofs of the basic bisimilarity core laws we have implemented a number of core laws about the basic strong bisimilarity relation namely compatibility preservation laws scope extension laws and neutrality associativity and commutativity of parallel composition these laws have counterparts for the basic weak bisimilarity relation but these haven’t been proved yet incidentally the proofs of the core laws for the strong case can be easily generalized to also cover the weak case this is because the weak analogs of all basic transition rules hold see the generalization works by replacing all applications of basic transition rules with applications of fact parameters that can be specialized to either the basic transition rules or their weak analogs our goal is to generalize the proofs in this way
1
244,667
18,764,946,547
IssuesEvent
2021-11-05 21:51:28
vmware-tanzu/carvel-ytt
https://api.github.com/repos/vmware-tanzu/carvel-ytt
closed
Clarify what "data values" are in the docs
enhancement documentation
The doc page for https://carvel.dev/ytt/docs/latest/ytt-data-values/ is self referential :) Also the https://carvel.dev/ytt/docs/latest/how-to-use-data-values/ page seems to reference "Data Values" as a proper qouted term, but doesnt seem to really explain it either. ``` Overview ¶ A ytt run can be configured by supplying custom Data Values. ``` Can we have `Data Values` definition on that page, where we sorta dumb it down... i.e. It can be a little confusing if you dont use YTT alot, to try to differentiate the data, template, and value...
1.0
Clarify what "data values" are in the docs - The doc page for https://carvel.dev/ytt/docs/latest/ytt-data-values/ is self referential :) Also the https://carvel.dev/ytt/docs/latest/how-to-use-data-values/ page seems to reference "Data Values" as a proper qouted term, but doesnt seem to really explain it either. ``` Overview ¶ A ytt run can be configured by supplying custom Data Values. ``` Can we have `Data Values` definition on that page, where we sorta dumb it down... i.e. It can be a little confusing if you dont use YTT alot, to try to differentiate the data, template, and value...
non_process
clarify what data values are in the docs the doc page for is self referential also the page seems to reference data values as a proper qouted term but doesnt seem to really explain it either overview ¶ a ytt run can be configured by supplying custom data values can we have data values definition on that page where we sorta dumb it down i e it can be a little confusing if you dont use ytt alot to try to differentiate the data template and value
0
11,711
14,546,460,665
IssuesEvent
2020-12-15 21:16:20
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Extends template does not render parameters
Pri2 devops-cicd-process/tech devops/prod product-feedback
I have tried to insert runtime param from simple-param.yml file same as [MS DOC](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#passing-parameters), The **"yesNo"** parameter in simple-param.yml file not render/view() into AzureDevOps ![Azure DevOps Build pipeline](https://user-images.githubusercontent.com/14990589/95014918-d674dd00-0652-11eb-9cc8-0776f153a6ed.png) ``` # File: simple-param.yml parameters: - name: yesNo # name of the parameter; required type: boolean # data type of the parameter; required default: false steps: - script: echo ${{ parameters.yesNo }} ``` ``` # File: azure-popelines.yml name: DEMO parameters: - name: demo-name# name of the parameter; required type: string # data type of the parameter; required default: xxx extends: template: simple-param.yml parameters: yesNo: false # set to a non-boolean value to have the build fail ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66 * Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065 * Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops) * Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/templates.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Extends template does not render parameters - I have tried to insert runtime param from simple-param.yml file same as [MS DOC](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#passing-parameters), The **"yesNo"** parameter in simple-param.yml file not render/view() into AzureDevOps ![Azure DevOps Build pipeline](https://user-images.githubusercontent.com/14990589/95014918-d674dd00-0652-11eb-9cc8-0776f153a6ed.png) ``` # File: simple-param.yml parameters: - name: yesNo # name of the parameter; required type: boolean # data type of the parameter; required default: false steps: - script: echo ${{ parameters.yesNo }} ``` ``` # File: azure-popelines.yml name: DEMO parameters: - name: demo-name# name of the parameter; required type: string # data type of the parameter; required default: xxx extends: template: simple-param.yml parameters: yesNo: false # set to a non-boolean value to have the build fail ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66 * Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065 * Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops) * Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/templates.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
extends template does not render parameters i have tried to insert runtime param from simple param yml file same as the yesno parameter in simple param yml file not render view into azuredevops file simple param yml parameters name yesno name of the parameter required type boolean data type of the parameter required default false steps script echo parameters yesno file azure popelines yml name demo parameters name demo name name of the parameter required type string data type of the parameter required default xxx extends template simple param yml parameters yesno false set to a non boolean value to have the build fail document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
315,168
27,051,499,129
IssuesEvent
2023-02-13 13:33:46
aquaaerobicsystem/aerationandmixing
https://api.github.com/repos/aquaaerobicsystem/aerationandmixing
closed
One slight change to the report for “sloped basin.” I just want to have this look cleaner on the report.
testing
![image](https://user-images.githubusercontent.com/46786408/217541160-cffe93f5-f996-48c3-87ea-6f5c4fd689ab.png)
1.0
One slight change to the report for “sloped basin.” I just want to have this look cleaner on the report. - ![image](https://user-images.githubusercontent.com/46786408/217541160-cffe93f5-f996-48c3-87ea-6f5c4fd689ab.png)
non_process
one slight change to the report for “sloped basin ” i just want to have this look cleaner on the report
0
94,378
11,866,630,990
IssuesEvent
2020-03-26 04:23:26
openstates/openstates.org
https://api.github.com/repos/openstates/openstates.org
opened
voter boxes are different heights based on matched or not
design web
<img width="1214" alt="Screen Shot 2020-03-26 at 12 23 05 AM" src="https://user-images.githubusercontent.com/8034/77610044-fdaa5f00-6ef7-11ea-9cbc-f468663c5f1a.png">
1.0
voter boxes are different heights based on matched or not - <img width="1214" alt="Screen Shot 2020-03-26 at 12 23 05 AM" src="https://user-images.githubusercontent.com/8034/77610044-fdaa5f00-6ef7-11ea-9cbc-f468663c5f1a.png">
non_process
voter boxes are different heights based on matched or not img width alt screen shot at am src
0
83,907
10,342,548,029
IssuesEvent
2019-09-04 06:46:42
wavespectra/wavespectra
https://api.github.com/repos/wavespectra/wavespectra
opened
Automated doc
documentation
shpynx stuff have not been updated since generated a few years back. The main issue with automated building was due to the fortran compilation.
1.0
Automated doc - shpynx stuff have not been updated since generated a few years back. The main issue with automated building was due to the fortran compilation.
non_process
automated doc shpynx stuff have not been updated since generated a few years back the main issue with automated building was due to the fortran compilation
0
10,334
13,163,456,879
IssuesEvent
2020-08-11 00:29:03
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Flaky parallel/test-child-process-fork-args on Windows
CI / flaky test child_process windows
Saw this test crashing in `node-test-binary-windows-js-suites` on https://ci.nodejs.org/job/node-test-pull-request/30753/: ``` 12:59:01 not ok 70 parallel/test-child-process-fork-args 12:59:01 --- 12:59:01 duration_ms: 0.305 12:59:01 severity: crashed 12:59:01 exitcode: -1073741819 12:59:01 stack: |- 12:59:01 ... ```
1.0
Flaky parallel/test-child-process-fork-args on Windows - Saw this test crashing in `node-test-binary-windows-js-suites` on https://ci.nodejs.org/job/node-test-pull-request/30753/: ``` 12:59:01 not ok 70 parallel/test-child-process-fork-args 12:59:01 --- 12:59:01 duration_ms: 0.305 12:59:01 severity: crashed 12:59:01 exitcode: -1073741819 12:59:01 stack: |- 12:59:01 ... ```
process
flaky parallel test child process fork args on windows saw this test crashing in node test binary windows js suites on not ok parallel test child process fork args duration ms severity crashed exitcode stack
1
71,311
7,241,738,759
IssuesEvent
2018-02-14 03:03:11
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
We need a test case for standalone kubelet
area/kubelet area/reliability area/test lifecycle/rotten priority/awaiting-more-evidence sig/node
We have a scenario of kubelet being used in standalone mode without a master which is the containervm. In that case, kubelet is started with a `--manifest_url` flag pointing to the URL of a YAML file with instructions of which containers to start. This case is getting broken often, particularly as we seem to be testing for `nil` and in some cases the pointer vs. interface nil breaks our checks which are not being caught by any of the tests... I'd say we need to include a test for `--manifest_url` in open source Kubernetes so that we don't get these bugs piling up until we try to build a new containervm with a version of Kubelet. @dchen1107 @jbeda @brendanburns @satnam6502 @lavalamp
1.0
We need a test case for standalone kubelet - We have a scenario of kubelet being used in standalone mode without a master which is the containervm. In that case, kubelet is started with a `--manifest_url` flag pointing to the URL of a YAML file with instructions of which containers to start. This case is getting broken often, particularly as we seem to be testing for `nil` and in some cases the pointer vs. interface nil breaks our checks which are not being caught by any of the tests... I'd say we need to include a test for `--manifest_url` in open source Kubernetes so that we don't get these bugs piling up until we try to build a new containervm with a version of Kubelet. @dchen1107 @jbeda @brendanburns @satnam6502 @lavalamp
non_process
we need a test case for standalone kubelet we have a scenario of kubelet being used in standalone mode without a master which is the containervm in that case kubelet is started with a manifest url flag pointing to the url of a yaml file with instructions of which containers to start this case is getting broken often particularly as we seem to be testing for nil and in some cases the pointer vs interface nil breaks our checks which are not being caught by any of the tests i d say we need to include a test for manifest url in open source kubernetes so that we don t get these bugs piling up until we try to build a new containervm with a version of kubelet jbeda brendanburns lavalamp
0
247,320
18,857,598,727
IssuesEvent
2021-11-12 08:48:23
DystoriaX/pe
https://api.github.com/repos/DystoriaX/pe
opened
Help Message: Inconsistent Formatting
severity.VeryLow type.DocumentationBug
Some noticable inconsistent formatting: - The parameter for `find` and `task find`, the additional `KEYWORD`. - Parameter of user-keyed-in input is inconsistent with that of UG (i.e. should be capitalized according to UG) <!--session: 1636702907050-1a839069-12c6-4666-ae30-5fe0011d2628--> <!--Version: Web v3.4.1-->
1.0
Help Message: Inconsistent Formatting - Some noticable inconsistent formatting: - The parameter for `find` and `task find`, the additional `KEYWORD`. - Parameter of user-keyed-in input is inconsistent with that of UG (i.e. should be capitalized according to UG) <!--session: 1636702907050-1a839069-12c6-4666-ae30-5fe0011d2628--> <!--Version: Web v3.4.1-->
non_process
help message inconsistent formatting some noticable inconsistent formatting the parameter for find and task find the additional keyword parameter of user keyed in input is inconsistent with that of ug i e should be capitalized according to ug
0
29,484
14,141,820,732
IssuesEvent
2020-11-10 13:15:44
microsoft/vscode-pull-request-github
https://api.github.com/repos/microsoft/vscode-pull-request-github
closed
Checkout is slow on Windows
performance
Checkout is taking like 5+ seconds, here's a log: ``` Review> swtich to Pull Requet #55776 > git merge-base 1e45ce6add2ac86c2a6e214aa84c85281df1814b 44837ff8323d2a9e769a4157f5a62068518132f8 > git diff --no-ext-diff --exit-code > git symbolic-ref --short HEAD Review> there is already one local branch origin/tyriar/auto_switch_renderer associated with Pull Request #55776 > git fetch origin > git rev-parse pr/segevfiner/55194 > git rev-parse --symbolic-full-name pr/segevfiner/55194@{u} > git rev-list --left-right pr/segevfiner/55194...refs/remotes/segevfiner/cli-env-inheritance-fix > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git config --local --get remote.njkevlani.github-pr-remote > git config --local --get remote.origin.github-pr-remote > git config --local --get remote.segevfiner.github-pr-remote > git config --local --get branch.pr/segevfiner/55194.github-pr-owner-number > git rev-parse tyriar/auto_switch_renderer > git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u} > git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer > git checkout tyriar/auto_switch_renderer > git symbolic-ref --short HEAD > git rev-parse pr/segevfiner/55194 > git rev-parse --symbolic-full-name pr/segevfiner/55194@{u} > git rev-list --left-right pr/segevfiner/55194...refs/remotes/segevfiner/cli-env-inheritance-fix > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git config --local --get remote.njkevlani.github-pr-remote GitHelper> associate tyriar/auto_switch_renderer with Pull Request #55776 > git config --local branch.tyriar/auto_switch_renderer.github-pr-owner-number Microsoft#vscode#55776 > git config --local --get remote.origin.github-pr-remote > git symbolic-ref --short HEAD > git config --local --get remote.segevfiner.github-pr-remote > git rev-parse tyriar/auto_switch_renderer > git config --local --get branch.pr/segevfiner/55194.github-pr-owner-number > git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u} > git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git config --local --get remote.njkevlani.github-pr-remote > git config --local --get remote.origin.github-pr-remote > git config --local --get remote.segevfiner.github-pr-remote > git symbolic-ref --short HEAD > git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number > git rev-parse tyriar/auto_switch_renderer Review> current branch tyriar/auto_switch_renderer is associated with pull request #55776 > git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u} > git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git merge-base 1e45ce6add2ac86c2a6e214aa84c85281df1814b 44837ff8323d2a9e769a4157f5a62068518132f8 > git cat-file -e 7f5beafe6e7bedec2fa309b6429c293523dee6b2:src/vs/workbench/parts/terminal/electron-browser/terminalInstance.ts Review> register comments provider Review> display pull request status bar indicator and refresh pull request tree view. > git config --local --get remote.origin.github-pr-remote > git symbolic-ref --short HEAD > git rev-parse tyriar/auto_switch_renderer > git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u} > git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git config --local --get remote.njkevlani.github-pr-remote > git config --local --get remote.origin.github-pr-remote > git config --local --get remote.segevfiner.github-pr-remote > git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number > git merge-base a957654db2cbf8df37f7d0139e103a7ecffdf926 487b6f745a93c9a344676ef2203e643421dbdf89 > git cat-file -e a957654db2cbf8df37f7d0139e103a7ecffdf926:src/vs/workbench/parts/terminal/common/terminalColorRegistry.ts > git cat-file -e a957654db2cbf8df37f7d0139e103a7ecffdf926:src/vs/workbench/parts/terminal/electron-browser/terminalPanel.ts > git merge-base 957a16690d156774c918fb447ff5d63b120f503e 945e6f81a9cbe01f12568dd973b8e33cbecfac53 > git cat-file -e 7b7c7dad6f740d73d685b3622b4fe528de7d5804:src/vs/workbench/parts/terminal/electron-browser/terminal.contribution.ts > git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number ```
True
Checkout is slow on Windows - Checkout is taking like 5+ seconds, here's a log: ``` Review> swtich to Pull Requet #55776 > git merge-base 1e45ce6add2ac86c2a6e214aa84c85281df1814b 44837ff8323d2a9e769a4157f5a62068518132f8 > git diff --no-ext-diff --exit-code > git symbolic-ref --short HEAD Review> there is already one local branch origin/tyriar/auto_switch_renderer associated with Pull Request #55776 > git fetch origin > git rev-parse pr/segevfiner/55194 > git rev-parse --symbolic-full-name pr/segevfiner/55194@{u} > git rev-list --left-right pr/segevfiner/55194...refs/remotes/segevfiner/cli-env-inheritance-fix > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git config --local --get remote.njkevlani.github-pr-remote > git config --local --get remote.origin.github-pr-remote > git config --local --get remote.segevfiner.github-pr-remote > git config --local --get branch.pr/segevfiner/55194.github-pr-owner-number > git rev-parse tyriar/auto_switch_renderer > git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u} > git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer > git checkout tyriar/auto_switch_renderer > git symbolic-ref --short HEAD > git rev-parse pr/segevfiner/55194 > git rev-parse --symbolic-full-name pr/segevfiner/55194@{u} > git rev-list --left-right pr/segevfiner/55194...refs/remotes/segevfiner/cli-env-inheritance-fix > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git config --local --get remote.njkevlani.github-pr-remote GitHelper> associate tyriar/auto_switch_renderer with Pull Request #55776 > git config --local branch.tyriar/auto_switch_renderer.github-pr-owner-number Microsoft#vscode#55776 > git config --local --get remote.origin.github-pr-remote > git symbolic-ref --short HEAD > git config --local --get remote.segevfiner.github-pr-remote > git rev-parse tyriar/auto_switch_renderer > git config --local --get branch.pr/segevfiner/55194.github-pr-owner-number > git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u} > git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git config --local --get remote.njkevlani.github-pr-remote > git config --local --get remote.origin.github-pr-remote > git config --local --get remote.segevfiner.github-pr-remote > git symbolic-ref --short HEAD > git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number > git rev-parse tyriar/auto_switch_renderer Review> current branch tyriar/auto_switch_renderer is associated with pull request #55776 > git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u} > git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git merge-base 1e45ce6add2ac86c2a6e214aa84c85281df1814b 44837ff8323d2a9e769a4157f5a62068518132f8 > git cat-file -e 7f5beafe6e7bedec2fa309b6429c293523dee6b2:src/vs/workbench/parts/terminal/electron-browser/terminalInstance.ts Review> register comments provider Review> display pull request status bar indicator and refresh pull request tree view. > git config --local --get remote.origin.github-pr-remote > git symbolic-ref --short HEAD > git rev-parse tyriar/auto_switch_renderer > git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u} > git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer > git for-each-ref --format %(refname) %(objectname) --sort -committerdate > git remote --verbose > git config --local --get remote.njkevlani.github-pr-remote > git config --local --get remote.origin.github-pr-remote > git config --local --get remote.segevfiner.github-pr-remote > git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number > git merge-base a957654db2cbf8df37f7d0139e103a7ecffdf926 487b6f745a93c9a344676ef2203e643421dbdf89 > git cat-file -e a957654db2cbf8df37f7d0139e103a7ecffdf926:src/vs/workbench/parts/terminal/common/terminalColorRegistry.ts > git cat-file -e a957654db2cbf8df37f7d0139e103a7ecffdf926:src/vs/workbench/parts/terminal/electron-browser/terminalPanel.ts > git merge-base 957a16690d156774c918fb447ff5d63b120f503e 945e6f81a9cbe01f12568dd973b8e33cbecfac53 > git cat-file -e 7b7c7dad6f740d73d685b3622b4fe528de7d5804:src/vs/workbench/parts/terminal/electron-browser/terminal.contribution.ts > git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number ```
non_process
checkout is slow on windows checkout is taking like seconds here s a log review swtich to pull requet git merge base git diff no ext diff exit code git symbolic ref short head review there is already one local branch origin tyriar auto switch renderer associated with pull request git fetch origin git rev parse pr segevfiner git rev parse symbolic full name pr segevfiner u git rev list left right pr segevfiner refs remotes segevfiner cli env inheritance fix git for each ref format refname objectname sort committerdate git remote verbose git config local get remote njkevlani github pr remote git config local get remote origin github pr remote git config local get remote segevfiner github pr remote git config local get branch pr segevfiner github pr owner number git rev parse tyriar auto switch renderer git rev parse symbolic full name tyriar auto switch renderer u git rev list left right tyriar auto switch renderer refs remotes origin tyriar auto switch renderer git checkout tyriar auto switch renderer git symbolic ref short head git rev parse pr segevfiner git rev parse symbolic full name pr segevfiner u git rev list left right pr segevfiner refs remotes segevfiner cli env inheritance fix git for each ref format refname objectname sort committerdate git remote verbose git config local get remote njkevlani github pr remote githelper associate tyriar auto switch renderer with pull request git config local branch tyriar auto switch renderer github pr owner number microsoft vscode git config local get remote origin github pr remote git symbolic ref short head git config local get remote segevfiner github pr remote git rev parse tyriar auto switch renderer git config local get branch pr segevfiner github pr owner number git rev parse symbolic full name tyriar auto switch renderer u git rev list left right tyriar auto switch renderer refs remotes origin tyriar auto switch renderer git for each ref format refname objectname sort committerdate git remote verbose git config local get remote njkevlani github pr remote git config local get remote origin github pr remote git config local get remote segevfiner github pr remote git symbolic ref short head git config local get branch tyriar auto switch renderer github pr owner number git rev parse tyriar auto switch renderer review current branch tyriar auto switch renderer is associated with pull request git rev parse symbolic full name tyriar auto switch renderer u git rev list left right tyriar auto switch renderer refs remotes origin tyriar auto switch renderer git for each ref format refname objectname sort committerdate git remote verbose git merge base git cat file e src vs workbench parts terminal electron browser terminalinstance ts review register comments provider review display pull request status bar indicator and refresh pull request tree view git config local get remote origin github pr remote git symbolic ref short head git rev parse tyriar auto switch renderer git rev parse symbolic full name tyriar auto switch renderer u git rev list left right tyriar auto switch renderer refs remotes origin tyriar auto switch renderer git for each ref format refname objectname sort committerdate git remote verbose git config local get remote njkevlani github pr remote git config local get remote origin github pr remote git config local get remote segevfiner github pr remote git config local get branch tyriar auto switch renderer github pr owner number git merge base git cat file e src vs workbench parts terminal common terminalcolorregistry ts git cat file e src vs workbench parts terminal electron browser terminalpanel ts git merge base git cat file e src vs workbench parts terminal electron browser terminal contribution ts git config local get branch tyriar auto switch renderer github pr owner number
0
603,543
18,668,845,316
IssuesEvent
2021-10-30 10:01:49
AY2122S1-CS2103T-W13-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-W13-2/tp
closed
[PE-D] Unable to view long fields
type.Bug priority.High severity.Medium
![Screenshot 2021-10-29 at 4.31.52 PM.png](https://raw.githubusercontent.com/s7u4rt99/ped/main/files/8bd4a34b-3cc4-4a34-907b-e35b5da1095c.png) If any of the fields are very long for some reason (name, email, address, etc), we may not be able to see it, especially if it is longer than the width of our entire screen. Maybe you can add a horizontal scroll bar so that even though it may be long, we are able to scroll horizontally to see it. <!--session: 1635494539368-6e4ca0fb-c787-43c4-a16e-48a16480861f--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.FeatureFlaw` original: s7u4rt99/ped#2
1.0
[PE-D] Unable to view long fields - ![Screenshot 2021-10-29 at 4.31.52 PM.png](https://raw.githubusercontent.com/s7u4rt99/ped/main/files/8bd4a34b-3cc4-4a34-907b-e35b5da1095c.png) If any of the fields are very long for some reason (name, email, address, etc), we may not be able to see it, especially if it is longer than the width of our entire screen. Maybe you can add a horizontal scroll bar so that even though it may be long, we are able to scroll horizontally to see it. <!--session: 1635494539368-6e4ca0fb-c787-43c4-a16e-48a16480861f--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.FeatureFlaw` original: s7u4rt99/ped#2
non_process
unable to view long fields if any of the fields are very long for some reason name email address etc we may not be able to see it especially if it is longer than the width of our entire screen maybe you can add a horizontal scroll bar so that even though it may be long we are able to scroll horizontally to see it labels severity low type featureflaw original ped
0
20,999
27,866,169,100
IssuesEvent
2023-03-21 10:23:56
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Ability to replace current Node process with another
child_process feature request stale
**Edit:** If someone can come up with a better shim for `execve` for Windows, that'd be *far* better. The form below is *very* expensive and *very* horrible. **Edit 2:** Linked relevant [SO question](https://stackoverflow.com/questions/51185115/what-is-the-ideal-way-to-emulate-process-replacement-on-windows). **Edit 3:** Clarify FS changes **Edit 4:** Here's the text from that SO question as of July 6, 2018 (so you don't have to search for it), where I asked about how to do the Windows part. <details> <summary>Click to show (warning: lots of text)</summary> So, in a [feature request I filed against Node.js](https://github.com/nodejs/node/issues/21664), I was looking for a way to replace the current Node process with another. In Linux and friends (really, any POSIX-compliant system), this is easy: use [`execve`](http://man7.org/linux/man-pages/man2/execve.2.html) and friends and call it a day. But obviously, that won't work on Windows, since it only has `CreateProcess` (which `execve` and friends delegate to, [complete with async behavior](https://stackoverflow.com/questions/49736973/blocking-version-of-execvp-windows)). And it's not like [people](https://stackoverflow.com/questions/35111313/windows-exec-equivalent) [haven't](https://stackoverflow.com/questions/6743567/replace-current-process-with-invocation-of-subprocess) [wanted](https://stackoverflow.com/questions/7198666/strategies-for-replacing-program-executable-in-windows) [to](https://stackoverflow.com/questions/198122/how-can-i-replace-the-current-java-process-like-a-unix-style-exec) [do](https://stackoverflow.com/questions/5450147/how-to-replace-the-current-java-process-in-windows-using-jna-jni) [similar](https://stackoverflow.com/questions/45607959/restart-windows-process-inplace-preserving-process-id-and-handles), leading to [numerous duplicate questions on this site](https://www.google.com/search?q=windows+replace+current+process+site:stackoverflow.com). (This isn't a duplicate because it's explicitly seeking a workaround given certain constraints, not just asking for direct replacement.) Process replacement has several facets that have to addressed: 1. All console I/O streams have to be forwarded to the new process. 1. All signals need transparently forwarded to the new process. 1. The data from the old process have to be destroyed, with as many resources reclaimed as possible. 1. All pre-existing threads and child processes should be destroyed. 1. All pre-existing handles should be destroyed apart from open file descriptors and named pipes/etc. 1. Optimally, the old process's memory should be kept to a minimum after the process is created. 1. For my particular use case, retaining the process ID is not important. And for my particular case, there are a few constraints: 1. I can control the initial process's startup as well as the location of my "process replacement" function. 1. I could load arbitrary native code via add-ons at potentially any stack offset. - Implication: I can't even dream of tracking `malloc` calls, handles, thread manipulation, or process manipulation to track and free them all, since DLL rewriting isn't exactly practical. 1. I have no control over *when* my "process replacement" is called. It could be called through an add-on, which could've been called through either interpreted code via FFI or even another add-on recursively. It could even be called during add-on initialization. - Implication: I would have no ability to know what's in the stack, even if I perfectly instrumented my side. And rewriting all their `call`s and `push`es is far from practical, and would just be all-around slow for obvious reasons. So, here's the gist of what I was thinking: use something similar to a pseudo-trampoline. 1. Statically allocate the following: 1. A single pointer for the stack pointer. 1. `MAX_PATH + 1` chars for the application path + `'\0'`. 1. `MAX_PATH + 1` chars for the current working directory path + `'\0'`. 1. 32768 chars for the arguments + `'\0'`. 1. 32768 chars for the environment + `'\0'`. 1. On entry, set the global stack pointer reference to the stack pointer. 1. On "replacement": 1. Do relevant process cleanup and lock/release everything you can. 1. Set the stack pointer to the stored original global one. 1. Terminate each child thread. 1. Kill each child process. 1. Free [each open handle](https://stackoverflow.com/questions/733384/how-to-enumerate-process-handles). 1. If possible (i.e. not in a UWP program), [For each heap](https://docs.microsoft.com/en-us/windows/desktop/api/heapapi/nf-heapapi-getprocessheaps), [destroy it](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapdestroy) if it's not the [default heap](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-getprocessheap) or the temporary heap (if it exists). 1. If possible, close [each open handle](https://stackoverflow.com/questions/733384/how-to-enumerate-process-handles). 1. If possible, [walk](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapwalk) the default heap and [free](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapfree) each segment associated with it. 1. Create a new process with the statically allocated file/arguments/environment/etc. with no new window created. 1. Proxy all future received signals, exceptions, etc. without modification to this process somehow. [The standard signals are easy](https://docs.microsoft.com/en-us/windows/console/setconsolectrlhandler), but not so much with the exceptions. 1. Wait for the process to end. 1. Return with [the process's exit code](https://docs.microsoft.com/en-us/windows/desktop/api/processthreadsapi/nf-processthreadsapi-getexitcodeprocess). The idea here is to use a process-based trampoline and drop the current process size to an absolute minimum while the newly created one is started. But where I'm not very familiar with Windows, I probably made quite a few mistakes here. Also, the above seems *extremely* inefficient and to an extent it just feels horribly wrong for something a kernel could just release a few memory pages, deallocate a bunch of memory handles, and move some memory around for the next process. So, to summarize, what's the ideal way to emulate process replacement on Windows with the fewest limitations? </details> ----- I would like a means to "replace" the current Node process with another, keeping the same process ID. It would be something morally similar to [this function](https://github.com/isiahmeadows/thallium/blob/master/lib/cli/util.js#L92-L111), but it wouldn't return. This would be most useful for conditionally replacing Node flags in a startup script - for example, if someone wants to enable modules and your behavior needs to change non-trivially in the presence of them (like if you need to install a default loader), you'll want to respawn the process with `--experimental-modules --loader <file>` so you can install the loader. This is also for scenarios when you want to run a module as a `main` module. If you want to do logic after the process ends, you should be using `child_process.spawn` regardless - you shouldn't be attempting to "replace" it in any capacity. Here's what I propose: - `child_process.replaceSpawn(command [ , args] [ , options ])` - `command` is the path to the new command. - `args` is the args to replace the arguments with. This defaults to the empty array. - `options` is for the various options for replacing the process. This defaults to an empty object. - `options.cwd` is the new cwd to use. (Default: `process.cwd()`) - `options.env` is the new environment to use. (Default: `process.env`) - `options.argv0` is the binary to spawn as. (Default: `command`) - `child_process.replaceFork(mainPath [ , args] [ , options ])` works similarly to above. - `mainPath` is the path to the new `require.main`. - `options.execPath` is the new binary to spawn as. (Default: `process.execPath`) - `options.execArgv` are the new Node flags to spawn with. (Default: `process.execArgv`) - `options.argv0` is the binary to spawn as. (Default: `process.argv0`) - The command is the original binary itself. - Add a `napi_terminating` member for `napi_status` to represent `try_catch.HasTerminated()` and the result of each call after replacement termination. - Add a `napi_set_terminate_hook(napi_env env, void (*fun)(void*), void* data)` function to register a callback called on termination, to make it easier to clean up resources. Internally, there are two cases you need to cover, and the simulated part for Windows is where it gets really hairy due to all the edge cases. Here's pseudocode for the basic algorithm (I'm not really familiar with Node internals, so take this as a rough guideline): 1. Stop the main event loop. 1. Go through the standard shutdown routine. 1. Destroy any open libuv handles and cancel any remaining event loop tasks. 1. If we're on a platform that supports process replacement (like Linux or Mac): 1. Invoke `execve` or equivalent with the new process path, arguments, and environment. 1. Else, if we're on Windows (the only supported OS that doesn't), we have to simulate it entirely: 1. Terminate execution via `v8::V8::TerminateExecution()`. All N-API callbacks should return `napi_terminated` during this step. 1. For each loaded native module: 1. If the native module has a terminate hook, call it. 1. Unload the native module's DLL. 1. Close the event loop. 1. Dispose the isolate. 1. Do the rest according to whatever happens to [this SO question](https://stackoverflow.com/questions/51185115/what-is-the-ideal-way-to-emulate-process-replacement-on-windows). 1. Else, on other OSs without a process replacement function, it'd look similar to Windows. In addition, file system requests will have to generally create each file descriptor with `O_CLOEXEC`. As for precedent where this could be used immediately: - [Liftoff](https://www.npmjs.com/package/liftoff) works very similarly, just with a little extra opinionated sugar, and that's used natively in Gulp. This kind of thing would speed that up quite a bit. - I do [very similar](https://github.com/isiahmeadows/thallium/blob/master/cli.js#L126-L150) to transparently pass through unknown Node flags. - Babel [attempts to use `kexec`](https://github.com/babel/babel/blob/master/packages/babel-node/src/babel-node.js#L87-L88) where available, which [is a POSIX-only module that replaces the process literally](https://www.npmjs.com/package/kexec). Absent that, it falls back to [its own implementation](https://github.com/babel/babel/blob/master/packages/babel-node/src/babel-node.js#L90-L109) that works like the other two examples.
1.0
Ability to replace current Node process with another - **Edit:** If someone can come up with a better shim for `execve` for Windows, that'd be *far* better. The form below is *very* expensive and *very* horrible. **Edit 2:** Linked relevant [SO question](https://stackoverflow.com/questions/51185115/what-is-the-ideal-way-to-emulate-process-replacement-on-windows). **Edit 3:** Clarify FS changes **Edit 4:** Here's the text from that SO question as of July 6, 2018 (so you don't have to search for it), where I asked about how to do the Windows part. <details> <summary>Click to show (warning: lots of text)</summary> So, in a [feature request I filed against Node.js](https://github.com/nodejs/node/issues/21664), I was looking for a way to replace the current Node process with another. In Linux and friends (really, any POSIX-compliant system), this is easy: use [`execve`](http://man7.org/linux/man-pages/man2/execve.2.html) and friends and call it a day. But obviously, that won't work on Windows, since it only has `CreateProcess` (which `execve` and friends delegate to, [complete with async behavior](https://stackoverflow.com/questions/49736973/blocking-version-of-execvp-windows)). And it's not like [people](https://stackoverflow.com/questions/35111313/windows-exec-equivalent) [haven't](https://stackoverflow.com/questions/6743567/replace-current-process-with-invocation-of-subprocess) [wanted](https://stackoverflow.com/questions/7198666/strategies-for-replacing-program-executable-in-windows) [to](https://stackoverflow.com/questions/198122/how-can-i-replace-the-current-java-process-like-a-unix-style-exec) [do](https://stackoverflow.com/questions/5450147/how-to-replace-the-current-java-process-in-windows-using-jna-jni) [similar](https://stackoverflow.com/questions/45607959/restart-windows-process-inplace-preserving-process-id-and-handles), leading to [numerous duplicate questions on this site](https://www.google.com/search?q=windows+replace+current+process+site:stackoverflow.com). (This isn't a duplicate because it's explicitly seeking a workaround given certain constraints, not just asking for direct replacement.) Process replacement has several facets that have to addressed: 1. All console I/O streams have to be forwarded to the new process. 1. All signals need transparently forwarded to the new process. 1. The data from the old process have to be destroyed, with as many resources reclaimed as possible. 1. All pre-existing threads and child processes should be destroyed. 1. All pre-existing handles should be destroyed apart from open file descriptors and named pipes/etc. 1. Optimally, the old process's memory should be kept to a minimum after the process is created. 1. For my particular use case, retaining the process ID is not important. And for my particular case, there are a few constraints: 1. I can control the initial process's startup as well as the location of my "process replacement" function. 1. I could load arbitrary native code via add-ons at potentially any stack offset. - Implication: I can't even dream of tracking `malloc` calls, handles, thread manipulation, or process manipulation to track and free them all, since DLL rewriting isn't exactly practical. 1. I have no control over *when* my "process replacement" is called. It could be called through an add-on, which could've been called through either interpreted code via FFI or even another add-on recursively. It could even be called during add-on initialization. - Implication: I would have no ability to know what's in the stack, even if I perfectly instrumented my side. And rewriting all their `call`s and `push`es is far from practical, and would just be all-around slow for obvious reasons. So, here's the gist of what I was thinking: use something similar to a pseudo-trampoline. 1. Statically allocate the following: 1. A single pointer for the stack pointer. 1. `MAX_PATH + 1` chars for the application path + `'\0'`. 1. `MAX_PATH + 1` chars for the current working directory path + `'\0'`. 1. 32768 chars for the arguments + `'\0'`. 1. 32768 chars for the environment + `'\0'`. 1. On entry, set the global stack pointer reference to the stack pointer. 1. On "replacement": 1. Do relevant process cleanup and lock/release everything you can. 1. Set the stack pointer to the stored original global one. 1. Terminate each child thread. 1. Kill each child process. 1. Free [each open handle](https://stackoverflow.com/questions/733384/how-to-enumerate-process-handles). 1. If possible (i.e. not in a UWP program), [For each heap](https://docs.microsoft.com/en-us/windows/desktop/api/heapapi/nf-heapapi-getprocessheaps), [destroy it](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapdestroy) if it's not the [default heap](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-getprocessheap) or the temporary heap (if it exists). 1. If possible, close [each open handle](https://stackoverflow.com/questions/733384/how-to-enumerate-process-handles). 1. If possible, [walk](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapwalk) the default heap and [free](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapfree) each segment associated with it. 1. Create a new process with the statically allocated file/arguments/environment/etc. with no new window created. 1. Proxy all future received signals, exceptions, etc. without modification to this process somehow. [The standard signals are easy](https://docs.microsoft.com/en-us/windows/console/setconsolectrlhandler), but not so much with the exceptions. 1. Wait for the process to end. 1. Return with [the process's exit code](https://docs.microsoft.com/en-us/windows/desktop/api/processthreadsapi/nf-processthreadsapi-getexitcodeprocess). The idea here is to use a process-based trampoline and drop the current process size to an absolute minimum while the newly created one is started. But where I'm not very familiar with Windows, I probably made quite a few mistakes here. Also, the above seems *extremely* inefficient and to an extent it just feels horribly wrong for something a kernel could just release a few memory pages, deallocate a bunch of memory handles, and move some memory around for the next process. So, to summarize, what's the ideal way to emulate process replacement on Windows with the fewest limitations? </details> ----- I would like a means to "replace" the current Node process with another, keeping the same process ID. It would be something morally similar to [this function](https://github.com/isiahmeadows/thallium/blob/master/lib/cli/util.js#L92-L111), but it wouldn't return. This would be most useful for conditionally replacing Node flags in a startup script - for example, if someone wants to enable modules and your behavior needs to change non-trivially in the presence of them (like if you need to install a default loader), you'll want to respawn the process with `--experimental-modules --loader <file>` so you can install the loader. This is also for scenarios when you want to run a module as a `main` module. If you want to do logic after the process ends, you should be using `child_process.spawn` regardless - you shouldn't be attempting to "replace" it in any capacity. Here's what I propose: - `child_process.replaceSpawn(command [ , args] [ , options ])` - `command` is the path to the new command. - `args` is the args to replace the arguments with. This defaults to the empty array. - `options` is for the various options for replacing the process. This defaults to an empty object. - `options.cwd` is the new cwd to use. (Default: `process.cwd()`) - `options.env` is the new environment to use. (Default: `process.env`) - `options.argv0` is the binary to spawn as. (Default: `command`) - `child_process.replaceFork(mainPath [ , args] [ , options ])` works similarly to above. - `mainPath` is the path to the new `require.main`. - `options.execPath` is the new binary to spawn as. (Default: `process.execPath`) - `options.execArgv` are the new Node flags to spawn with. (Default: `process.execArgv`) - `options.argv0` is the binary to spawn as. (Default: `process.argv0`) - The command is the original binary itself. - Add a `napi_terminating` member for `napi_status` to represent `try_catch.HasTerminated()` and the result of each call after replacement termination. - Add a `napi_set_terminate_hook(napi_env env, void (*fun)(void*), void* data)` function to register a callback called on termination, to make it easier to clean up resources. Internally, there are two cases you need to cover, and the simulated part for Windows is where it gets really hairy due to all the edge cases. Here's pseudocode for the basic algorithm (I'm not really familiar with Node internals, so take this as a rough guideline): 1. Stop the main event loop. 1. Go through the standard shutdown routine. 1. Destroy any open libuv handles and cancel any remaining event loop tasks. 1. If we're on a platform that supports process replacement (like Linux or Mac): 1. Invoke `execve` or equivalent with the new process path, arguments, and environment. 1. Else, if we're on Windows (the only supported OS that doesn't), we have to simulate it entirely: 1. Terminate execution via `v8::V8::TerminateExecution()`. All N-API callbacks should return `napi_terminated` during this step. 1. For each loaded native module: 1. If the native module has a terminate hook, call it. 1. Unload the native module's DLL. 1. Close the event loop. 1. Dispose the isolate. 1. Do the rest according to whatever happens to [this SO question](https://stackoverflow.com/questions/51185115/what-is-the-ideal-way-to-emulate-process-replacement-on-windows). 1. Else, on other OSs without a process replacement function, it'd look similar to Windows. In addition, file system requests will have to generally create each file descriptor with `O_CLOEXEC`. As for precedent where this could be used immediately: - [Liftoff](https://www.npmjs.com/package/liftoff) works very similarly, just with a little extra opinionated sugar, and that's used natively in Gulp. This kind of thing would speed that up quite a bit. - I do [very similar](https://github.com/isiahmeadows/thallium/blob/master/cli.js#L126-L150) to transparently pass through unknown Node flags. - Babel [attempts to use `kexec`](https://github.com/babel/babel/blob/master/packages/babel-node/src/babel-node.js#L87-L88) where available, which [is a POSIX-only module that replaces the process literally](https://www.npmjs.com/package/kexec). Absent that, it falls back to [its own implementation](https://github.com/babel/babel/blob/master/packages/babel-node/src/babel-node.js#L90-L109) that works like the other two examples.
process
ability to replace current node process with another edit if someone can come up with a better shim for execve for windows that d be far better the form below is very expensive and very horrible edit linked relevant edit clarify fs changes edit here s the text from that so question as of july so you don t have to search for it where i asked about how to do the windows part click to show warning lots of text so in a i was looking for a way to replace the current node process with another in linux and friends really any posix compliant system this is easy use and friends and call it a day but obviously that won t work on windows since it only has createprocess which execve and friends delegate to and it s not like leading to this isn t a duplicate because it s explicitly seeking a workaround given certain constraints not just asking for direct replacement process replacement has several facets that have to addressed all console i o streams have to be forwarded to the new process all signals need transparently forwarded to the new process the data from the old process have to be destroyed with as many resources reclaimed as possible all pre existing threads and child processes should be destroyed all pre existing handles should be destroyed apart from open file descriptors and named pipes etc optimally the old process s memory should be kept to a minimum after the process is created for my particular use case retaining the process id is not important and for my particular case there are a few constraints i can control the initial process s startup as well as the location of my process replacement function i could load arbitrary native code via add ons at potentially any stack offset implication i can t even dream of tracking malloc calls handles thread manipulation or process manipulation to track and free them all since dll rewriting isn t exactly practical i have no control over when my process replacement is called it could be called through an add on which could ve been called through either interpreted code via ffi or even another add on recursively it could even be called during add on initialization implication i would have no ability to know what s in the stack even if i perfectly instrumented my side and rewriting all their call s and push es is far from practical and would just be all around slow for obvious reasons so here s the gist of what i was thinking use something similar to a pseudo trampoline statically allocate the following a single pointer for the stack pointer max path chars for the application path max path chars for the current working directory path chars for the arguments chars for the environment on entry set the global stack pointer reference to the stack pointer on replacement do relevant process cleanup and lock release everything you can set the stack pointer to the stored original global one terminate each child thread kill each child process free if possible i e not in a uwp program if it s not the or the temporary heap if it exists if possible close if possible the default heap and each segment associated with it create a new process with the statically allocated file arguments environment etc with no new window created proxy all future received signals exceptions etc without modification to this process somehow but not so much with the exceptions wait for the process to end return with the idea here is to use a process based trampoline and drop the current process size to an absolute minimum while the newly created one is started but where i m not very familiar with windows i probably made quite a few mistakes here also the above seems extremely inefficient and to an extent it just feels horribly wrong for something a kernel could just release a few memory pages deallocate a bunch of memory handles and move some memory around for the next process so to summarize what s the ideal way to emulate process replacement on windows with the fewest limitations i would like a means to replace the current node process with another keeping the same process id it would be something morally similar to but it wouldn t return this would be most useful for conditionally replacing node flags in a startup script for example if someone wants to enable modules and your behavior needs to change non trivially in the presence of them like if you need to install a default loader you ll want to respawn the process with experimental modules loader so you can install the loader this is also for scenarios when you want to run a module as a main module if you want to do logic after the process ends you should be using child process spawn regardless you shouldn t be attempting to replace it in any capacity here s what i propose child process replacespawn command command is the path to the new command args is the args to replace the arguments with this defaults to the empty array options is for the various options for replacing the process this defaults to an empty object options cwd is the new cwd to use default process cwd options env is the new environment to use default process env options is the binary to spawn as default command child process replacefork mainpath works similarly to above mainpath is the path to the new require main options execpath is the new binary to spawn as default process execpath options execargv are the new node flags to spawn with default process execargv options is the binary to spawn as default process the command is the original binary itself add a napi terminating member for napi status to represent try catch hasterminated and the result of each call after replacement termination add a napi set terminate hook napi env env void fun void void data function to register a callback called on termination to make it easier to clean up resources internally there are two cases you need to cover and the simulated part for windows is where it gets really hairy due to all the edge cases here s pseudocode for the basic algorithm i m not really familiar with node internals so take this as a rough guideline stop the main event loop go through the standard shutdown routine destroy any open libuv handles and cancel any remaining event loop tasks if we re on a platform that supports process replacement like linux or mac invoke execve or equivalent with the new process path arguments and environment else if we re on windows the only supported os that doesn t we have to simulate it entirely terminate execution via terminateexecution all n api callbacks should return napi terminated during this step for each loaded native module if the native module has a terminate hook call it unload the native module s dll close the event loop dispose the isolate do the rest according to whatever happens to else on other oss without a process replacement function it d look similar to windows in addition file system requests will have to generally create each file descriptor with o cloexec as for precedent where this could be used immediately works very similarly just with a little extra opinionated sugar and that s used natively in gulp this kind of thing would speed that up quite a bit i do to transparently pass through unknown node flags babel where available which absent that it falls back to that works like the other two examples
1
17,034
22,409,207,055
IssuesEvent
2022-06-18 13:05:51
RobertCraigie/prisma-client-py
https://api.github.com/repos/RobertCraigie/prisma-client-py
opened
Do not override already set env variables from `.env`
bug/2-confirmed kind/bug process/candidate topic: client level/beginner priority/high
## Problem <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Currently the `.env` variables take precedence over the system environment variables, this can cause issues as the Prisma CLI will use the system environment variables instead which could lead to migrations being applied to a different database if you have two different connection strings set. ## Suggested solution <!-- A clear and concise description of what you want to happen. --> System environment variables should take priority.
1.0
Do not override already set env variables from `.env` - ## Problem <!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] --> Currently the `.env` variables take precedence over the system environment variables, this can cause issues as the Prisma CLI will use the system environment variables instead which could lead to migrations being applied to a different database if you have two different connection strings set. ## Suggested solution <!-- A clear and concise description of what you want to happen. --> System environment variables should take priority.
process
do not override already set env variables from env problem currently the env variables take precedence over the system environment variables this can cause issues as the prisma cli will use the system environment variables instead which could lead to migrations being applied to a different database if you have two different connection strings set suggested solution system environment variables should take priority
1
11,492
14,366,340,810
IssuesEvent
2020-12-01 04:08:07
aodn/imos-toolbox
https://api.github.com/repos/aodn/imos-toolbox
opened
New PP/QC routines from CSIRO
Unit:Processing Unit:QC workshop day
This is just a placeholder for the merge of the PP/QC routines from https://github.com/aodn/imos-toolbox/pull/683
1.0
New PP/QC routines from CSIRO - This is just a placeholder for the merge of the PP/QC routines from https://github.com/aodn/imos-toolbox/pull/683
process
new pp qc routines from csiro this is just a placeholder for the merge of the pp qc routines from
1
22,466
31,273,166,196
IssuesEvent
2023-08-22 02:52:34
q191201771/lal
https://api.github.com/repos/q191201771/lal
closed
重复的Access-Control-Allow-Origin
*In process #Opt
通过hls拉流m3u8 接口会默认添加Access-Control-Allow-Origin:*, 但是我通过ngingx转发也会给这个接口加上自己的Access-Control-Allow-Origin。会导致头部重复,请问如何去掉lal自带的Access-Control-Allow-Origin。 <img width="979" alt="image" src="https://github.com/q191201771/lal/assets/9177693/8078f246-e722-4b57-8ec5-801f9d011d02"> ![image](https://github.com/q191201771/lal/assets/9177693/3b3fbcd7-f3aa-47e8-8725-c4f8b7619979)
1.0
重复的Access-Control-Allow-Origin - 通过hls拉流m3u8 接口会默认添加Access-Control-Allow-Origin:*, 但是我通过ngingx转发也会给这个接口加上自己的Access-Control-Allow-Origin。会导致头部重复,请问如何去掉lal自带的Access-Control-Allow-Origin。 <img width="979" alt="image" src="https://github.com/q191201771/lal/assets/9177693/8078f246-e722-4b57-8ec5-801f9d011d02"> ![image](https://github.com/q191201771/lal/assets/9177693/3b3fbcd7-f3aa-47e8-8725-c4f8b7619979)
process
重复的access control allow origin 接口会默认添加access control allow origin: , 但是我通过ngingx转发也会给这个接口加上自己的access control allow origin。会导致头部重复,请问如何去掉lal自带的access control allow origin。 img width alt image src
1
56,009
31,475,376,779
IssuesEvent
2023-08-30 10:19:00
GeopJr/Tuba
https://api.github.com/repos/GeopJr/Tuba
closed
[Feature request]: Status widget needs to be refactored
bug status performance
### Describe the bug The more I think about `Widgets.Status`, the more I agree that its architecture has become questionable. Currently, each `Widgets.Status` has a pre-instanced `Widgets.VoteBox`, `Widgets.AttachmentBox`, and emoji `Gtk.FlowBox`. These widgets stay even if their status entity doesn't really need them, so they just hang there being invisible. There really is no need to create so many useless widgets. Ideally, the entire content box should be generated dynamically on demand (when bound to a `API.Status`) to avoid creating a larger memory footprint. Some widgets can be skipped altogether if the account backend doesn't support some features (like emoji reactions). To make things worse, Notification widgets derive from `Widget.Status`, so they all succumb to this behavior as well (even though it makes no sense for a follow request to have a `VoteBox` at its disposal). ### Steps To Reproduce Inspect https://github.com/GeopJr/Tooth/blob/main/data/ui/widgets/status.ui#L220-L226 and https://github.com/GeopJr/Tooth/blob/main/data/ui/widgets/status.ui#L257-L264 ### Logs and/or Screenshots _No response_ ### Instance Backend Mastodon ### Operating System Pop!_OS 22.04 LTS ### Package Flatpak ### Troubleshooting information flatpak: false version: main-15b8f4a (development) gtk: 4.8.4 (4.8.4) libadwaita: 1.2.1 (1.2.1) libsoup: 2.74.2 (2.74.2) ### Additional Context _No response_
True
[Feature request]: Status widget needs to be refactored - ### Describe the bug The more I think about `Widgets.Status`, the more I agree that its architecture has become questionable. Currently, each `Widgets.Status` has a pre-instanced `Widgets.VoteBox`, `Widgets.AttachmentBox`, and emoji `Gtk.FlowBox`. These widgets stay even if their status entity doesn't really need them, so they just hang there being invisible. There really is no need to create so many useless widgets. Ideally, the entire content box should be generated dynamically on demand (when bound to a `API.Status`) to avoid creating a larger memory footprint. Some widgets can be skipped altogether if the account backend doesn't support some features (like emoji reactions). To make things worse, Notification widgets derive from `Widget.Status`, so they all succumb to this behavior as well (even though it makes no sense for a follow request to have a `VoteBox` at its disposal). ### Steps To Reproduce Inspect https://github.com/GeopJr/Tooth/blob/main/data/ui/widgets/status.ui#L220-L226 and https://github.com/GeopJr/Tooth/blob/main/data/ui/widgets/status.ui#L257-L264 ### Logs and/or Screenshots _No response_ ### Instance Backend Mastodon ### Operating System Pop!_OS 22.04 LTS ### Package Flatpak ### Troubleshooting information flatpak: false version: main-15b8f4a (development) gtk: 4.8.4 (4.8.4) libadwaita: 1.2.1 (1.2.1) libsoup: 2.74.2 (2.74.2) ### Additional Context _No response_
non_process
status widget needs to be refactored describe the bug the more i think about widgets status the more i agree that its architecture has become questionable currently each widgets status has a pre instanced widgets votebox widgets attachmentbox and emoji gtk flowbox these widgets stay even if their status entity doesn t really need them so they just hang there being invisible there really is no need to create so many useless widgets ideally the entire content box should be generated dynamically on demand when bound to a api status to avoid creating a larger memory footprint some widgets can be skipped altogether if the account backend doesn t support some features like emoji reactions to make things worse notification widgets derive from widget status so they all succumb to this behavior as well even though it makes no sense for a follow request to have a votebox at its disposal steps to reproduce inspect and logs and or screenshots no response instance backend mastodon operating system pop os lts package flatpak troubleshooting information flatpak false version main development gtk libadwaita libsoup additional context no response
0
6,955
10,113,956,276
IssuesEvent
2019-07-30 18:00:10
material-components/material-components-ios
https://api.github.com/repos/material-components/material-components-ios
closed
Internal issue: b/114705309
[schemes] type:Process
This is an internal issue. If you are a Googler, please visit [b/114705309](http://b/114705309) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/114705309](http://b/114705309)
1.0
Internal issue: b/114705309 - This is an internal issue. If you are a Googler, please visit [b/114705309](http://b/114705309) for more details. <!-- Auto-generated content below, do not modify --> --- #### Internal data - Associated internal bug: [b/114705309](http://b/114705309)
process
internal issue b this is an internal issue if you are a googler please visit for more details internal data associated internal bug
1
51,965
3,016,118,709
IssuesEvent
2015-07-29 23:31:29
archon810/androidpolice
https://api.github.com/repos/archon810/androidpolice
closed
no pushbullet subscriptions buttons are visible in new designed page
auto-migrated Priority-Medium Type-Enhancement
``` follow the summary ``` Original issue reported on code.google.com by `gopinathms2012` on 1 Feb 2015 at 7:15
1.0
no pushbullet subscriptions buttons are visible in new designed page - ``` follow the summary ``` Original issue reported on code.google.com by `gopinathms2012` on 1 Feb 2015 at 7:15
non_process
no pushbullet subscriptions buttons are visible in new designed page follow the summary original issue reported on code google com by on feb at
0
130,374
12,427,465,055
IssuesEvent
2020-05-25 02:31:30
guilherme-adesouza/fretz-freedom
https://api.github.com/repos/guilherme-adesouza/fretz-freedom
closed
Criar apresentação de testes de segurança
documentation
Criar apresentação de slides explicando os testes de segurança e a(s) ferramenta(s) e medidas utilizadas.
1.0
Criar apresentação de testes de segurança - Criar apresentação de slides explicando os testes de segurança e a(s) ferramenta(s) e medidas utilizadas.
non_process
criar apresentação de testes de segurança criar apresentação de slides explicando os testes de segurança e a s ferramenta s e medidas utilizadas
0
81,769
15,800,929,544
IssuesEvent
2021-04-03 02:01:18
PyCQA/flake8
https://api.github.com/repos/PyCQA/flake8
closed
noqa is not respected for E241
bug:invalid component:docs component:pycodestyle
In GitLab by @gforcada on Jun 3, 2015, 08:53 If you have something like this: ``` dictionary = { 'key1' : 'value1', # noqa 'long_key2': 'value2', } ``` The ``# noqa`` comment is ignored and a E241 is reported.
1.0
noqa is not respected for E241 - In GitLab by @gforcada on Jun 3, 2015, 08:53 If you have something like this: ``` dictionary = { 'key1' : 'value1', # noqa 'long_key2': 'value2', } ``` The ``# noqa`` comment is ignored and a E241 is reported.
non_process
noqa is not respected for in gitlab by gforcada on jun if you have something like this dictionary noqa long the noqa comment is ignored and a is reported
0
14,418
17,467,088,028
IssuesEvent
2021-08-06 18:32:11
2i2c-org/pilot-hubs
https://api.github.com/repos/2i2c-org/pilot-hubs
opened
Improve our understanding of expected workloads per hub
type: enhancement :label: team-process :label: CI/CD
# Summary IN https://github.com/2i2c-org/pilot-hubs/issues/524 we found that the hub's workload scaled to a higher degree than we were expecting when setting up the hub. We should collect better information from Hub Representatives about the expected usage for their hubs, so that we can better estimate what resources to plan ahead for them. # User Stories - As a Hub Representative and as a Hub Engineer, I want to know that the hub's environment/resources have been provisioned according to what I'll actually use, so that nothing unexpected happens when we scale. # Tasks to complete - [ ] Understand what information we are missing with our current "New Hub Template" - [ ] Update the template so that we capture this extra information
1.0
Improve our understanding of expected workloads per hub - # Summary IN https://github.com/2i2c-org/pilot-hubs/issues/524 we found that the hub's workload scaled to a higher degree than we were expecting when setting up the hub. We should collect better information from Hub Representatives about the expected usage for their hubs, so that we can better estimate what resources to plan ahead for them. # User Stories - As a Hub Representative and as a Hub Engineer, I want to know that the hub's environment/resources have been provisioned according to what I'll actually use, so that nothing unexpected happens when we scale. # Tasks to complete - [ ] Understand what information we are missing with our current "New Hub Template" - [ ] Update the template so that we capture this extra information
process
improve our understanding of expected workloads per hub summary in we found that the hub s workload scaled to a higher degree than we were expecting when setting up the hub we should collect better information from hub representatives about the expected usage for their hubs so that we can better estimate what resources to plan ahead for them user stories as a hub representative and as a hub engineer i want to know that the hub s environment resources have been provisioned according to what i ll actually use so that nothing unexpected happens when we scale tasks to complete understand what information we are missing with our current new hub template update the template so that we capture this extra information
1
20,871
27,658,688,896
IssuesEvent
2023-03-12 09:09:00
polarismesh/polaris
https://api.github.com/repos/polarismesh/polaris
closed
web管理端无法删除openapi接口发布的配置文件
bug config in processed
**Describe the bug** web管理端无法删除openapi接口发布的配置文件 ![image](https://user-images.githubusercontent.com/60428924/224010951-b4090bbc-fb17-42b1-8c13-f396c047f181.png) **To Reproduce** ```shell # 创建新的配置文件 curl -X "POST" "http://localhost:8090/config/v1/configfiles" \ -H 'X-Polaris-Token: nu/0WRA4EqSR1FagrjRj0fZwPXuGlMpX+zCuWu4uMqy8xr1vRjisSbA25aAC3mtU8MeeRsKhQiDAynUR09I=' \ -H 'Content-Type: application/json; charset=utf-8' \ -d $'{ "group": "test", "namespace": "default", "content": "redis.cache.age=10", "name": "app/properties.yml" }' # 发布该配置文件 curl -X "POST" "http://localhost:8090/config/v1/configfiles/release" \ -H 'X-Polaris-Token: nu/0WRA4EqSR1FagrjRj0fZwPXuGlMpX+zCuWu4uMqy8xr1vRjisSbA25aAC3mtU8MeeRsKhQiDAynUR09I=' \ -H 'Content-Type: application/json; charset=utf-8' \ -d $'{ "namespace": "default", "group": "test", "fileName": "app/properties.yml" }' ``` **Expected behavior** ```text // polaris-default-error.log 2023-03-09T11:42:11.582147Z error http/handler.go:167 code:<value:500001 > info:<value:"store layer exception" > {"request-id": ""} ``` **Environment** - Version: polarismesh/polaris-server-standalone:latest - OS: docker. **Additional context** Add any other context about the problem here.
1.0
web管理端无法删除openapi接口发布的配置文件 - **Describe the bug** web管理端无法删除openapi接口发布的配置文件 ![image](https://user-images.githubusercontent.com/60428924/224010951-b4090bbc-fb17-42b1-8c13-f396c047f181.png) **To Reproduce** ```shell # 创建新的配置文件 curl -X "POST" "http://localhost:8090/config/v1/configfiles" \ -H 'X-Polaris-Token: nu/0WRA4EqSR1FagrjRj0fZwPXuGlMpX+zCuWu4uMqy8xr1vRjisSbA25aAC3mtU8MeeRsKhQiDAynUR09I=' \ -H 'Content-Type: application/json; charset=utf-8' \ -d $'{ "group": "test", "namespace": "default", "content": "redis.cache.age=10", "name": "app/properties.yml" }' # 发布该配置文件 curl -X "POST" "http://localhost:8090/config/v1/configfiles/release" \ -H 'X-Polaris-Token: nu/0WRA4EqSR1FagrjRj0fZwPXuGlMpX+zCuWu4uMqy8xr1vRjisSbA25aAC3mtU8MeeRsKhQiDAynUR09I=' \ -H 'Content-Type: application/json; charset=utf-8' \ -d $'{ "namespace": "default", "group": "test", "fileName": "app/properties.yml" }' ``` **Expected behavior** ```text // polaris-default-error.log 2023-03-09T11:42:11.582147Z error http/handler.go:167 code:<value:500001 > info:<value:"store layer exception" > {"request-id": ""} ``` **Environment** - Version: polarismesh/polaris-server-standalone:latest - OS: docker. **Additional context** Add any other context about the problem here.
process
web管理端无法删除openapi接口发布的配置文件 describe the bug web管理端无法删除openapi接口发布的配置文件 to reproduce shell 创建新的配置文件 curl x post h x polaris token nu h content type application json charset utf d group test namespace default content redis cache age name app properties yml 发布该配置文件 curl x post h x polaris token nu h content type application json charset utf d namespace default group test filename app properties yml expected behavior text polaris default error log error http handler go code info request id environment version polarismesh polaris server standalone latest os docker additional context add any other context about the problem here
1
259,622
27,700,065,207
IssuesEvent
2023-03-14 07:12:06
UrielProd/big-data-plugin
https://api.github.com/repos/UrielProd/big-data-plugin
opened
commons-beanutils-1.9.3.jar: 2 vulnerabilities (highest severity is: 7.3)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p></summary> <p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p> <p>Path to dependency file: /impl/cluster/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (commons-beanutils version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-10086](https://www.mend.io/vulnerability-database/CVE-2019-10086) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.9.3.jar | Direct | 1.9.4 | &#9989; | | [CVE-2014-0114](https://www.mend.io/vulnerability-database/CVE-2014-0114) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.9.3.jar | Direct | 1.9.4 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10086</summary> ### Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p> <p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p> <p>Path to dependency file: /impl/cluster/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p> <p> Dependency Hierarchy: - :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10086>CVE-2019-10086</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-08-20</p> <p>Fix Resolution: 1.9.4</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2014-0114</summary> ### Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p> <p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p> <p>Path to dependency file: /impl/cluster/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p> <p> Dependency Hierarchy: - :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1. <p>Publish Date: 2014-04-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-0114>CVE-2014-0114</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p> <p>Release Date: 2014-04-30</p> <p>Fix Resolution: 1.9.4</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
True
commons-beanutils-1.9.3.jar: 2 vulnerabilities (highest severity is: 7.3) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p></summary> <p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p> <p>Path to dependency file: /impl/cluster/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p> <p> <p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (commons-beanutils version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-10086](https://www.mend.io/vulnerability-database/CVE-2019-10086) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.9.3.jar | Direct | 1.9.4 | &#9989; | | [CVE-2014-0114](https://www.mend.io/vulnerability-database/CVE-2014-0114) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.9.3.jar | Direct | 1.9.4 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10086</summary> ### Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p> <p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p> <p>Path to dependency file: /impl/cluster/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p> <p> Dependency Hierarchy: - :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean. <p>Publish Date: 2019-08-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10086>CVE-2019-10086</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-08-20</p> <p>Fix Resolution: 1.9.4</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2014-0114</summary> ### Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p> <p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p> <p>Path to dependency file: /impl/cluster/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p> <p> Dependency Hierarchy: - :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1. <p>Publish Date: 2014-04-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-0114>CVE-2014-0114</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p> <p>Release Date: 2014-04-30</p> <p>Fix Resolution: 1.9.4</p> </p> <p></p> :rescue_worker_helmet: Automatic Remediation is available for this issue </details> *** <p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
non_process
commons beanutils jar vulnerabilities highest severity is vulnerable library commons beanutils jar apache commons beanutils provides an easy to use but flexible wrapper around reflection and introspection library home page a href path to dependency file impl cluster pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in commons beanutils version remediation available high commons beanutils jar direct high commons beanutils jar direct details cve vulnerable library commons beanutils jar apache commons beanutils provides an easy to use but flexible wrapper around reflection and introspection library home page a href path to dependency file impl cluster pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy x commons beanutils jar vulnerable library found in head commit a href found in base branch master vulnerability details in apache commons beanutils a special beanintrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all java objects we however were not using this by default characteristic of the propertyutilsbean publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library commons beanutils jar apache commons beanutils provides an easy to use but flexible wrapper around reflection and introspection library home page a href path to dependency file impl cluster pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy x commons beanutils jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
0
435,620
30,510,349,185
IssuesEvent
2023-07-18 20:18:13
cilium/cilium
https://api.github.com/repos/cilium/cilium
opened
Documentation style: follow K8s guidance on capitalizing API objects
area/documentation
Opened after reviewing #25945. I think Cilium docs should follow the [Kubernetes style guide](https://kubernetes.io/docs/contribute/style/style-guide/#use-upper-camel-case-for-api-objects) for when to (not) capitalize API objects. Specifically: > When you refer specifically to interacting with an API object, use [UpperCamelCase](https://en.wikipedia.org/wiki/Camel_case), also known as Pascal case. And: > When you are generally discussing an API object, use [sentence-style capitalization](https://docs.microsoft.com/en-us/style-guide/text-formatting/using-type/use-sentence-style-capitalization). In the case of #25945, I'd like to see "Gateway API" capitalized, "Gateway" when talking about an API object as an entity, and "gateway" when talking about a specific instance. For example, this line would read: ```rst Cilium only programs gateways with the class ``cilium``. ``` This issue is non-blocking for #25945. I mention it specifically because it's a good example of a larger phenomenon I see in Cilium API docs in general. @qmonnet 👋🏻 I'd love your feedback here cc @learnitall for visibility
1.0
Documentation style: follow K8s guidance on capitalizing API objects - Opened after reviewing #25945. I think Cilium docs should follow the [Kubernetes style guide](https://kubernetes.io/docs/contribute/style/style-guide/#use-upper-camel-case-for-api-objects) for when to (not) capitalize API objects. Specifically: > When you refer specifically to interacting with an API object, use [UpperCamelCase](https://en.wikipedia.org/wiki/Camel_case), also known as Pascal case. And: > When you are generally discussing an API object, use [sentence-style capitalization](https://docs.microsoft.com/en-us/style-guide/text-formatting/using-type/use-sentence-style-capitalization). In the case of #25945, I'd like to see "Gateway API" capitalized, "Gateway" when talking about an API object as an entity, and "gateway" when talking about a specific instance. For example, this line would read: ```rst Cilium only programs gateways with the class ``cilium``. ``` This issue is non-blocking for #25945. I mention it specifically because it's a good example of a larger phenomenon I see in Cilium API docs in general. @qmonnet 👋🏻 I'd love your feedback here cc @learnitall for visibility
non_process
documentation style follow guidance on capitalizing api objects opened after reviewing i think cilium docs should follow the for when to not capitalize api objects specifically when you refer specifically to interacting with an api object use also known as pascal case and when you are generally discussing an api object use in the case of i d like to see gateway api capitalized gateway when talking about an api object as an entity and gateway when talking about a specific instance for example this line would read rst cilium only programs gateways with the class cilium this issue is non blocking for i mention it specifically because it s a good example of a larger phenomenon i see in cilium api docs in general qmonnet 👋🏻 i d love your feedback here cc learnitall for visibility
0
23,172
10,852,343,693
IssuesEvent
2019-11-13 12:37:07
ChenLuigi/TestingPOM
https://api.github.com/repos/ChenLuigi/TestingPOM
opened
CVE-2017-5648 (High) detected in tomcat-catalina-7.0.42.jar, tomcat-embed-core-8.0.18.jar
security vulnerability
## CVE-2017-5648 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-catalina-7.0.42.jar</b>, <b>tomcat-embed-core-8.0.18.jar</b></p></summary> <p> <details><summary><b>tomcat-catalina-7.0.42.jar</b></p></summary> <p>Tomcat Servlet Engine Core Classes and Standard implementations</p> <p> Dependency Hierarchy: - :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library) </details> <details><summary><b>tomcat-embed-core-8.0.18.jar</b></p></summary> <p>Core Tomcat implementation</p> <p> Dependency Hierarchy: - :x: **tomcat-embed-core-8.0.18.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> While investigating bug 60718, it was noticed that some calls to application listeners in Apache Tomcat 9.0.0.M1 to 9.0.0.M17, 8.5.0 to 8.5.11, 8.0.0.RC1 to 8.0.41, and 7.0.0 to 7.0.75 did not use the appropriate facade object. When running an untrusted application under a SecurityManager, it was therefore possible for that untrusted application to retain a reference to the request or response object and thereby access and/or modify information associated with another web application. <p>Publish Date: 2017-04-17 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648>CVE-2017-5648</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648</a></p> <p>Release Date: 2017-04-17</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M18,8.5.12,8.0.42,7.0.76,org.apache.tomcat:tomcat-catalina:9.0.0.M18,8.5.12,8.0.42,7.0.76</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-5648 (High) detected in tomcat-catalina-7.0.42.jar, tomcat-embed-core-8.0.18.jar - ## CVE-2017-5648 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-catalina-7.0.42.jar</b>, <b>tomcat-embed-core-8.0.18.jar</b></p></summary> <p> <details><summary><b>tomcat-catalina-7.0.42.jar</b></p></summary> <p>Tomcat Servlet Engine Core Classes and Standard implementations</p> <p> Dependency Hierarchy: - :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library) </details> <details><summary><b>tomcat-embed-core-8.0.18.jar</b></p></summary> <p>Core Tomcat implementation</p> <p> Dependency Hierarchy: - :x: **tomcat-embed-core-8.0.18.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> While investigating bug 60718, it was noticed that some calls to application listeners in Apache Tomcat 9.0.0.M1 to 9.0.0.M17, 8.5.0 to 8.5.11, 8.0.0.RC1 to 8.0.41, and 7.0.0 to 7.0.75 did not use the appropriate facade object. When running an untrusted application under a SecurityManager, it was therefore possible for that untrusted application to retain a reference to the request or response object and thereby access and/or modify information associated with another web application. <p>Publish Date: 2017-04-17 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648>CVE-2017-5648</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648</a></p> <p>Release Date: 2017-04-17</p> <p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M18,8.5.12,8.0.42,7.0.76,org.apache.tomcat:tomcat-catalina:9.0.0.M18,8.5.12,8.0.42,7.0.76</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tomcat catalina jar tomcat embed core jar cve high severity vulnerability vulnerable libraries tomcat catalina jar tomcat embed core jar tomcat catalina jar tomcat servlet engine core classes and standard implementations dependency hierarchy x tomcat catalina jar vulnerable library tomcat embed core jar core tomcat implementation dependency hierarchy x tomcat embed core jar vulnerable library found in head commit a href vulnerability details while investigating bug it was noticed that some calls to application listeners in apache tomcat to to to and to did not use the appropriate facade object when running an untrusted application under a securitymanager it was therefore possible for that untrusted application to retain a reference to the request or response object and thereby access and or modify information associated with another web application publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource
0
19,334
25,472,602,873
IssuesEvent
2022-11-25 11:31:27
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[IDP][PM][MFA Enabled] Admin is not authenticated with 2 factor authentication even though MFA is enabled
Bug Blocker P0 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
**Pre-condition:** IDP and MFA should be enabled in the PM **Steps:** 1. Add organizational user in the participant manager 2. Complete set up your account process 3. Sign in with registered credentials and Verify **AR:** Admin is not authenticated with 2-factor authentication even though MFA is enabled **ER:** Admin should be authenticated with 2-factor authentication if MFA is enabled
3.0
[IDP][PM][MFA Enabled] Admin is not authenticated with 2 factor authentication even though MFA is enabled - **Pre-condition:** IDP and MFA should be enabled in the PM **Steps:** 1. Add organizational user in the participant manager 2. Complete set up your account process 3. Sign in with registered credentials and Verify **AR:** Admin is not authenticated with 2-factor authentication even though MFA is enabled **ER:** Admin should be authenticated with 2-factor authentication if MFA is enabled
process
admin is not authenticated with factor authentication even though mfa is enabled pre condition idp and mfa should be enabled in the pm steps add organizational user in the participant manager complete set up your account process sign in with registered credentials and verify ar admin is not authenticated with factor authentication even though mfa is enabled er admin should be authenticated with factor authentication if mfa is enabled
1
42,868
5,478,204,975
IssuesEvent
2017-03-12 16:02:58
mattstone22133/JavaNetworkLibrary
https://api.github.com/repos/mattstone22133/JavaNetworkLibrary
opened
Network Black Box Test - Test Mode Changing
medium length story test
testModeChanging - these tests will take some time because multiple disonnects are required startClient_changeToServer_Listen_ChangeToClient_Join startServer_changeToClient_Join_Disconnect_ChangeToServer_listenAndAccept_changeBackToClient_repeat multiple changes (including duplicates to test for realistic behavior Server -> Server -> client -> client -> server -> client client -> server -> client -> client -> server -> server
1.0
Network Black Box Test - Test Mode Changing - testModeChanging - these tests will take some time because multiple disonnects are required startClient_changeToServer_Listen_ChangeToClient_Join startServer_changeToClient_Join_Disconnect_ChangeToServer_listenAndAccept_changeBackToClient_repeat multiple changes (including duplicates to test for realistic behavior Server -> Server -> client -> client -> server -> client client -> server -> client -> client -> server -> server
non_process
network black box test test mode changing testmodechanging these tests will take some time because multiple disonnects are required startclient changetoserver listen changetoclient join startserver changetoclient join disconnect changetoserver listenandaccept changebacktoclient repeat multiple changes including duplicates to test for realistic behavior server server client client server client client server client client server server
0
295,866
22,275,213,890
IssuesEvent
2022-06-10 15:56:52
aws/aws-cdk
https://api.github.com/repos/aws/aws-cdk
closed
(aws-autoscaling): Lambda autoscaling example uses an alias to latestVersion, which doesn't work
good first issue p2 @aws-cdk/aws-autoscaling effort/small documentation
### link to reference doc page https://docs.aws.amazon.com/cdk/api/v2/docs/aws-cdk-lib.aws_lambda.AutoScalingOptions.html ### Describe your issue? When following the example in the docs to setup auto-scaling provisioned concurrency on a lambda function, the docs explicitly state in the example to first add an alias on the lambda function's `latestVersion`. I followed the example, setup an alias to `latestVersion`, added a basic `scaleOnUtilization` target, and then deployed the stack. The deploy went through without errors. After deploy, I checked the lambda in the AWS Console and noticed there was no provisioned concurrency setup for it. I drilled down into the alias associated with the lambda that I created from the example and promptly saw: ![image](https://user-images.githubusercontent.com/8990195/155265769-551cb85f-c6b1-4950-a131-548dc94601f2.png) Did I misunderstand something in the docs or is this a bad example? Looking at the docs for `Version` I see it's example uses the `Function` method `addVersion` - and the `addVersion` method docs tell me is deprecated and points me to use `function.currentVersion`. If this is not a misunderstanding on my part, can these be updated to whatever the current best practice for setting up basic provisioned concurrency auto-scaling on the most recently deployed lambda version is?
1.0
(aws-autoscaling): Lambda autoscaling example uses an alias to latestVersion, which doesn't work - ### link to reference doc page https://docs.aws.amazon.com/cdk/api/v2/docs/aws-cdk-lib.aws_lambda.AutoScalingOptions.html ### Describe your issue? When following the example in the docs to setup auto-scaling provisioned concurrency on a lambda function, the docs explicitly state in the example to first add an alias on the lambda function's `latestVersion`. I followed the example, setup an alias to `latestVersion`, added a basic `scaleOnUtilization` target, and then deployed the stack. The deploy went through without errors. After deploy, I checked the lambda in the AWS Console and noticed there was no provisioned concurrency setup for it. I drilled down into the alias associated with the lambda that I created from the example and promptly saw: ![image](https://user-images.githubusercontent.com/8990195/155265769-551cb85f-c6b1-4950-a131-548dc94601f2.png) Did I misunderstand something in the docs or is this a bad example? Looking at the docs for `Version` I see it's example uses the `Function` method `addVersion` - and the `addVersion` method docs tell me is deprecated and points me to use `function.currentVersion`. If this is not a misunderstanding on my part, can these be updated to whatever the current best practice for setting up basic provisioned concurrency auto-scaling on the most recently deployed lambda version is?
non_process
aws autoscaling lambda autoscaling example uses an alias to latestversion which doesn t work link to reference doc page describe your issue when following the example in the docs to setup auto scaling provisioned concurrency on a lambda function the docs explicitly state in the example to first add an alias on the lambda function s latestversion i followed the example setup an alias to latestversion added a basic scaleonutilization target and then deployed the stack the deploy went through without errors after deploy i checked the lambda in the aws console and noticed there was no provisioned concurrency setup for it i drilled down into the alias associated with the lambda that i created from the example and promptly saw did i misunderstand something in the docs or is this a bad example looking at the docs for version i see it s example uses the function method addversion and the addversion method docs tell me is deprecated and points me to use function currentversion if this is not a misunderstanding on my part can these be updated to whatever the current best practice for setting up basic provisioned concurrency auto scaling on the most recently deployed lambda version is
0
322,875
27,640,896,897
IssuesEvent
2023-03-10 17:57:40
Roukys/HHauto
https://api.github.com/repos/Roukys/HHauto
closed
5.6.136 = Testing auto boss bang option.
new feature to be tested
![Screenshot_1](https://user-images.githubusercontent.com/86404525/178144722-9591248d-c23e-4aea-b1f2-57d7e1f8d168.jpg) no autofill girls here - we can do it by myself just an option start fight, end fight, start new and etc for the end - the end of HP girls or victory and winning event girl here.
1.0
5.6.136 = Testing auto boss bang option. - ![Screenshot_1](https://user-images.githubusercontent.com/86404525/178144722-9591248d-c23e-4aea-b1f2-57d7e1f8d168.jpg) no autofill girls here - we can do it by myself just an option start fight, end fight, start new and etc for the end - the end of HP girls or victory and winning event girl here.
non_process
testing auto boss bang option no autofill girls here we can do it by myself just an option start fight end fight start new and etc for the end the end of hp girls or victory and winning event girl here
0
174,143
14,448,182,588
IssuesEvent
2020-12-08 05:41:35
zmkfirmware/zmk
https://api.github.com/repos/zmkfirmware/zmk
closed
CMake issue when using 3.19+
bug core documentation good first issue
The build docs specifically mention that our cmake installs shouldn't be "old", but it's only a lower limit at 3.15. But after I installed a fresh cmake v3.19.1, it failed the build saying something along the lines of: `in function 'exit': exit.c:(.text.exit+0x60): undefined reference to '_exit'` Tracked this to a [related bug report](https://github.com/zephyrproject-rtos/zephyr/issues/22060) where the consensus seemed to be that this is caused by cmake 3.19 (either 3.19.0 or 3.19.1). Downgrading to 3.18.5 did indeed help. This might be worth noting in the docs @mcrosson...
1.0
CMake issue when using 3.19+ - The build docs specifically mention that our cmake installs shouldn't be "old", but it's only a lower limit at 3.15. But after I installed a fresh cmake v3.19.1, it failed the build saying something along the lines of: `in function 'exit': exit.c:(.text.exit+0x60): undefined reference to '_exit'` Tracked this to a [related bug report](https://github.com/zephyrproject-rtos/zephyr/issues/22060) where the consensus seemed to be that this is caused by cmake 3.19 (either 3.19.0 or 3.19.1). Downgrading to 3.18.5 did indeed help. This might be worth noting in the docs @mcrosson...
non_process
cmake issue when using the build docs specifically mention that our cmake installs shouldn t be old but it s only a lower limit at but after i installed a fresh cmake it failed the build saying something along the lines of in function exit exit c text exit undefined reference to exit tracked this to a where the consensus seemed to be that this is caused by cmake either or downgrading to did indeed help this might be worth noting in the docs mcrosson
0
176,851
6,565,797,706
IssuesEvent
2017-09-08 09:47:26
hobbit-project/platform
https://api.github.com/repos/hobbit-project/platform
closed
Queue status not shown when controller creates new experiment
component: controller priority: high type: bug
## Problem It is possible that the queue status is not shown if the controller is currently starting a new experiment. ![hobbit gui](https://user-images.githubusercontent.com/6122307/27992752-a01e667e-649b-11e7-9a7c-13bc738bffab.png) This small window in which the controller is not responding is even extended, when the controller starts to prefetch docker images. ## Cause The problem might be caused by the controller trying to access the currently running experiment data which is secured by a mutex. When the controller is creating a new experiment, this mutex might be owned by a different thread.
1.0
Queue status not shown when controller creates new experiment - ## Problem It is possible that the queue status is not shown if the controller is currently starting a new experiment. ![hobbit gui](https://user-images.githubusercontent.com/6122307/27992752-a01e667e-649b-11e7-9a7c-13bc738bffab.png) This small window in which the controller is not responding is even extended, when the controller starts to prefetch docker images. ## Cause The problem might be caused by the controller trying to access the currently running experiment data which is secured by a mutex. When the controller is creating a new experiment, this mutex might be owned by a different thread.
non_process
queue status not shown when controller creates new experiment problem it is possible that the queue status is not shown if the controller is currently starting a new experiment this small window in which the controller is not responding is even extended when the controller starts to prefetch docker images cause the problem might be caused by the controller trying to access the currently running experiment data which is secured by a mutex when the controller is creating a new experiment this mutex might be owned by a different thread
0
132,577
18,751,850,972
IssuesEvent
2021-11-05 03:46:17
hackforla/food-oasis
https://api.github.com/repos/hackforla/food-oasis
reopened
Experiment with creating a widget to embed Food Oasis in a web site
Release Note: Enhancement Role: Design UI/UX Role: Front-end 3pt
### Overview We have had several requests to provide a Food Oasis widget that could be embedded in an external organization's web site. Explore the options and implications ### Action Items - [ ] Look into the functionality provided by an IFrame or other implementation. - Can the behavior be tailored to the site by customizing the query, allowing some styling, etc.? - How can this be linked to attribute Food Oasis and re-direct to Food Oasis properly - [ ] How can we design to work with the Mapbox license agreement to avoid absorbing the charges incurred on the host site by mapbox requests? - [ ] Can we limit the number of map calls (which cost money) per embedded site so that we don't go broke hosting maps on other sites. - [ ] Can we generate a static snapshot (ideally hyperlinked) map that links to our site for cost efficiency that can be put on external sites. ### Resources/Instructions Requests have been made by Julia Blanton (Food Oasis Santa Barbara) Nancy Beyda (Food Cycle LA)
1.0
Experiment with creating a widget to embed Food Oasis in a web site - ### Overview We have had several requests to provide a Food Oasis widget that could be embedded in an external organization's web site. Explore the options and implications ### Action Items - [ ] Look into the functionality provided by an IFrame or other implementation. - Can the behavior be tailored to the site by customizing the query, allowing some styling, etc.? - How can this be linked to attribute Food Oasis and re-direct to Food Oasis properly - [ ] How can we design to work with the Mapbox license agreement to avoid absorbing the charges incurred on the host site by mapbox requests? - [ ] Can we limit the number of map calls (which cost money) per embedded site so that we don't go broke hosting maps on other sites. - [ ] Can we generate a static snapshot (ideally hyperlinked) map that links to our site for cost efficiency that can be put on external sites. ### Resources/Instructions Requests have been made by Julia Blanton (Food Oasis Santa Barbara) Nancy Beyda (Food Cycle LA)
non_process
experiment with creating a widget to embed food oasis in a web site overview we have had several requests to provide a food oasis widget that could be embedded in an external organization s web site explore the options and implications action items look into the functionality provided by an iframe or other implementation can the behavior be tailored to the site by customizing the query allowing some styling etc how can this be linked to attribute food oasis and re direct to food oasis properly how can we design to work with the mapbox license agreement to avoid absorbing the charges incurred on the host site by mapbox requests can we limit the number of map calls which cost money per embedded site so that we don t go broke hosting maps on other sites can we generate a static snapshot ideally hyperlinked map that links to our site for cost efficiency that can be put on external sites resources instructions requests have been made by julia blanton food oasis santa barbara nancy beyda food cycle la
0
43,165
2,885,717,061
IssuesEvent
2015-06-12 00:32:32
parallaxinc/ChromeIDE
https://api.github.com/repos/parallaxinc/ChromeIDE
closed
File dialog should set focus on filename field automatically
Priority 1
The New File dialog should put focus on filename field automatically once opened. Request: A very slight delay (after dialog appearance) followed by a focus-event animation (perhaps an expanding and translucent-fading circle on the field) to simulate the appearance of a user click/touch on the field, followed by the existing "filename" label slide-shrink animation. If it's possible for a previous entry to appear (like what happens after an Esc key press on a previous dialog, then the cursor should appear at the end of the field's filename text.
1.0
File dialog should set focus on filename field automatically - The New File dialog should put focus on filename field automatically once opened. Request: A very slight delay (after dialog appearance) followed by a focus-event animation (perhaps an expanding and translucent-fading circle on the field) to simulate the appearance of a user click/touch on the field, followed by the existing "filename" label slide-shrink animation. If it's possible for a previous entry to appear (like what happens after an Esc key press on a previous dialog, then the cursor should appear at the end of the field's filename text.
non_process
file dialog should set focus on filename field automatically the new file dialog should put focus on filename field automatically once opened request a very slight delay after dialog appearance followed by a focus event animation perhaps an expanding and translucent fading circle on the field to simulate the appearance of a user click touch on the field followed by the existing filename label slide shrink animation if it s possible for a previous entry to appear like what happens after an esc key press on a previous dialog then the cursor should appear at the end of the field s filename text
0
42,479
2,870,709,729
IssuesEvent
2015-06-07 12:50:43
bobbybee/uPresent
https://api.github.com/repos/bobbybee/uPresent
closed
Combining bold and italics _**like this**_ won't compile
bug medium-priority
It's worth noting `_ **text** _` works
1.0
Combining bold and italics _**like this**_ won't compile - It's worth noting `_ **text** _` works
non_process
combining bold and italics like this won t compile it s worth noting text works
0
25,255
12,231,359,090
IssuesEvent
2020-05-04 07:35:45
hashicorp/consul
https://api.github.com/repos/hashicorp/consul
closed
Support server selection priorities / weights
theme/service-metadata type/enhancement
If we support priories, then you can in support cases like a server in a remote region (for backup / quorum purposes). The priorities would disable client routing through them unless necessary (higher priority servers have failed / are unreachable).
1.0
Support server selection priorities / weights - If we support priories, then you can in support cases like a server in a remote region (for backup / quorum purposes). The priorities would disable client routing through them unless necessary (higher priority servers have failed / are unreachable).
non_process
support server selection priorities weights if we support priories then you can in support cases like a server in a remote region for backup quorum purposes the priorities would disable client routing through them unless necessary higher priority servers have failed are unreachable
0