Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
17,900
| 23,874,099,870
|
IssuesEvent
|
2022-09-07 17:17:23
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
K8S Processor: Include more tags
|
processor/k8sattributes
|
Currently, the processor supports following tags:
- namespace
- podName
- deployment
- cluster
- node
- startTime
It would be great to include more tags, such as:
- hostName
- containerName
- daemonSetName
- serviceName
- statefulSetName
Additionally, it would be good to have an ability to include all labels and annotations
|
1.0
|
K8S Processor: Include more tags - Currently, the processor supports following tags:
- namespace
- podName
- deployment
- cluster
- node
- startTime
It would be great to include more tags, such as:
- hostName
- containerName
- daemonSetName
- serviceName
- statefulSetName
Additionally, it would be good to have an ability to include all labels and annotations
|
process
|
processor include more tags currently the processor supports following tags namespace podname deployment cluster node starttime it would be great to include more tags such as hostname containername daemonsetname servicename statefulsetname additionally it would be good to have an ability to include all labels and annotations
| 1
|
374
| 2,815,784,384
|
IssuesEvent
|
2015-05-19 07:39:32
|
mattesCZ/mtbmap
|
https://api.github.com/repos/mattesCZ/mtbmap
|
opened
|
Add hiking trails with red:red:white_bar
|
data_processing map_style
|
OSMC:symbol red:red:white_bar is used in Austria, Italy, France... It should be rendered as Czech red:white:red_bar. There are only inverted foreground and background colors.
|
1.0
|
Add hiking trails with red:red:white_bar - OSMC:symbol red:red:white_bar is used in Austria, Italy, France... It should be rendered as Czech red:white:red_bar. There are only inverted foreground and background colors.
|
process
|
add hiking trails with red red white bar osmc symbol red red white bar is used in austria italy france it should be rendered as czech red white red bar there are only inverted foreground and background colors
| 1
|
478
| 2,911,386,423
|
IssuesEvent
|
2015-06-22 09:14:03
|
haskell-distributed/distributed-process-simplelocalnet
|
https://api.github.com/repos/haskell-distributed/distributed-process-simplelocalnet
|
opened
|
Re-export runProcess and forkProcess in SimpleLocalnet
|
distributed-process-simplelocalnet Feature Request
|
_From @edsko on October 23, 2012 14:51_
so that startSlave and startMaster are truly optional.
_Copied from original issue: haskell-distributed/distributed-process#57_
|
1.0
|
Re-export runProcess and forkProcess in SimpleLocalnet - _From @edsko on October 23, 2012 14:51_
so that startSlave and startMaster are truly optional.
_Copied from original issue: haskell-distributed/distributed-process#57_
|
process
|
re export runprocess and forkprocess in simplelocalnet from edsko on october so that startslave and startmaster are truly optional copied from original issue haskell distributed distributed process
| 1
|
17,693
| 23,540,085,296
|
IssuesEvent
|
2022-08-20 08:43:53
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
Request in QGIS ([processing][needs-docs] Adjust status of controls executing an algorithm dialog )
|
Processing 3.14
|
### Request for documentation
From pull request QGIS/qgis#34928
Author: @gacarrillor
QGIS version: 3.14
**[processing][needs-docs] Adjust status of controls executing an algorithm dialog **
### PR Description:
## Description
Currently, after you run a Processing algorithm, the `Log` tab is visible and the `Run` button is enabled, so you can run the algorithm again with a single click. From a user perspective, it shouldn't be the case. If you want to run the algorithm again, you should at least have a look at the `Parameters` tab once more.
This PR adds the following changes to such behaviour:
+ `Run` button is not shown anymore in the `Log` tab, therefore, you can only run algorithms from the `Parameters` tab.
+ While running an algorithm, the `Parameters` panel is now disabled, as well as `Run as Batch Process...` and `Run as Single Process...` buttons.
+ When an algorithm execution finishes (either successfully or not), a new button `Change Parameters` is shown in the `Log` tab.
+ There is now consistency among algorithm dialogs: Before this PR, `Batch Algorithm Dialog` was the only dialog blocking parameter widgets.
These changes were applied to the `Algorithm Dialog` and `Batch Algorithm Dialog`, and work on `Edit in place` dialogs as well. This PR also takes into account cancelling an algorithm execution or obtaining failures in the execution.
(More details and preliminary discussion in issue https://github.com/qgis/QGIS/issues/34484)
## Screencasts
### Algorithm dialog

### Batch algorithm dialog

### Cancelling algorithm execution

----------------------------
Fix #34484
Since this PR implies updating QGIS docs, I could give a hand adjusting the corresponding section.
### Commits tagged with [need-docs] or [FEATURE]
"[processing]\n\n\n\n Adjust status of controls in algorithm dialog\n\n + Run button is not shown anymore in the Log tab, therefore, you can only run algorithms from the Parameters tab.\n + While running an algorithm, the Parameters tab is now blocked.\n + When an algorithm execution finishes (either successfully or not), a new button Change Parameters is shown in the Log tab.\n + The Batch Algorithm Dialog is now consistent with the described behavior (before, it blocked the Parameters panel, but not the tab; and it was the only dialog blocking parameters widgets).\n\nThese changes were applied to the Algorithm Dialog and Batch Algorithm Dialog, and work on Edit in place dialogs as well."
|
1.0
|
Request in QGIS ([processing][needs-docs] Adjust status of controls executing an algorithm dialog ) - ### Request for documentation
From pull request QGIS/qgis#34928
Author: @gacarrillor
QGIS version: 3.14
**[processing][needs-docs] Adjust status of controls executing an algorithm dialog **
### PR Description:
## Description
Currently, after you run a Processing algorithm, the `Log` tab is visible and the `Run` button is enabled, so you can run the algorithm again with a single click. From a user perspective, it shouldn't be the case. If you want to run the algorithm again, you should at least have a look at the `Parameters` tab once more.
This PR adds the following changes to such behaviour:
+ `Run` button is not shown anymore in the `Log` tab, therefore, you can only run algorithms from the `Parameters` tab.
+ While running an algorithm, the `Parameters` panel is now disabled, as well as `Run as Batch Process...` and `Run as Single Process...` buttons.
+ When an algorithm execution finishes (either successfully or not), a new button `Change Parameters` is shown in the `Log` tab.
+ There is now consistency among algorithm dialogs: Before this PR, `Batch Algorithm Dialog` was the only dialog blocking parameter widgets.
These changes were applied to the `Algorithm Dialog` and `Batch Algorithm Dialog`, and work on `Edit in place` dialogs as well. This PR also takes into account cancelling an algorithm execution or obtaining failures in the execution.
(More details and preliminary discussion in issue https://github.com/qgis/QGIS/issues/34484)
## Screencasts
### Algorithm dialog

### Batch algorithm dialog

### Cancelling algorithm execution

----------------------------
Fix #34484
Since this PR implies updating QGIS docs, I could give a hand adjusting the corresponding section.
### Commits tagged with [need-docs] or [FEATURE]
"[processing]\n\n\n\n Adjust status of controls in algorithm dialog\n\n + Run button is not shown anymore in the Log tab, therefore, you can only run algorithms from the Parameters tab.\n + While running an algorithm, the Parameters tab is now blocked.\n + When an algorithm execution finishes (either successfully or not), a new button Change Parameters is shown in the Log tab.\n + The Batch Algorithm Dialog is now consistent with the described behavior (before, it blocked the Parameters panel, but not the tab; and it was the only dialog blocking parameters widgets).\n\nThese changes were applied to the Algorithm Dialog and Batch Algorithm Dialog, and work on Edit in place dialogs as well."
|
process
|
request in qgis adjust status of controls executing an algorithm dialog request for documentation from pull request qgis qgis author gacarrillor qgis version adjust status of controls executing an algorithm dialog pr description description currently after you run a processing algorithm the log tab is visible and the run button is enabled so you can run the algorithm again with a single click from a user perspective it shouldn t be the case if you want to run the algorithm again you should at least have a look at the parameters tab once more this pr adds the following changes to such behaviour run button is not shown anymore in the log tab therefore you can only run algorithms from the parameters tab while running an algorithm the parameters panel is now disabled as well as run as batch process and run as single process buttons when an algorithm execution finishes either successfully or not a new button change parameters is shown in the log tab there is now consistency among algorithm dialogs before this pr batch algorithm dialog was the only dialog blocking parameter widgets these changes were applied to the algorithm dialog and batch algorithm dialog and work on edit in place dialogs as well this pr also takes into account cancelling an algorithm execution or obtaining failures in the execution more details and preliminary discussion in issue screencasts algorithm dialog batch algorithm dialog cancelling algorithm execution fix since this pr implies updating qgis docs i could give a hand adjusting the corresponding section commits tagged with or n n n n adjust status of controls in algorithm dialog n n run button is not shown anymore in the log tab therefore you can only run algorithms from the parameters tab n while running an algorithm the parameters tab is now blocked n when an algorithm execution finishes either successfully or not a new button change parameters is shown in the log tab n the batch algorithm dialog is now consistent with the described behavior before it blocked the parameters panel but not the tab and it was the only dialog blocking parameters widgets n nthese changes were applied to the algorithm dialog and batch algorithm dialog and work on edit in place dialogs as well
| 1
|
2,714
| 5,580,379,221
|
IssuesEvent
|
2017-03-28 16:28:30
|
TEAMMATES/teammates
|
https://api.github.com/repos/TEAMMATES/teammates
|
closed
|
Explore using mentionbot to identify potential reviewers
|
a-Process c.DevOps d.Committers p.Medium
|
Seems easy to set up and could be useful
https://github.com/facebook/mention-bot
|
1.0
|
Explore using mentionbot to identify potential reviewers - Seems easy to set up and could be useful
https://github.com/facebook/mention-bot
|
process
|
explore using mentionbot to identify potential reviewers seems easy to set up and could be useful
| 1
|
22,504
| 31,558,669,704
|
IssuesEvent
|
2023-09-03 01:13:53
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
opened
|
[Mirror] uber/hermetic_cc_toolchain v2.1.0
|
P2 type: process team-OSS mirror request
|
### Please list the URLs of the archives you'd like to mirror:
https://github.com/uber/hermetic_cc_toolchain/releases/download/v2.1.0/hermetic_cc_toolchain-v2.1.0.tar.gz
|
1.0
|
[Mirror] uber/hermetic_cc_toolchain v2.1.0 - ### Please list the URLs of the archives you'd like to mirror:
https://github.com/uber/hermetic_cc_toolchain/releases/download/v2.1.0/hermetic_cc_toolchain-v2.1.0.tar.gz
|
process
|
uber hermetic cc toolchain please list the urls of the archives you d like to mirror
| 1
|
109,398
| 16,843,685,085
|
IssuesEvent
|
2021-06-19 02:50:48
|
bharathirajatut/fitbit-api-example-java2
|
https://api.github.com/repos/bharathirajatut/fitbit-api-example-java2
|
opened
|
CVE-2016-9879 (High) detected in spring-security-web-4.1.1.RELEASE.jar
|
security vulnerability
|
## CVE-2016-9879 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-4.1.1.RELEASE.jar</b></p></summary>
<p>spring-security-web</p>
<p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p>
<p>Path to dependency file: fitbit-api-example-java2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/4.1.1.RELEASE/spring-security-web-4.1.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-security-1.4.0.RELEASE.jar (Root Library)
- :x: **spring-security-web-4.1.1.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/bharathirajatut/fitbit-api-example-java2/commits/8c153ad064e8f07a4ddade35ac13a9b485ca3dac">8c153ad064e8f07a4ddade35ac13a9b485ca3dac</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in Pivotal Spring Security before 3.2.10, 4.1.x before 4.1.4, and 4.2.x before 4.2.1. Spring Security does not consider URL path parameters when processing security constraints. By adding a URL path parameter with an encoded "/" to a request, an attacker may be able to bypass a security constraint. The root cause of this issue is a lack of clarity regarding the handling of path parameters in the Servlet Specification. Some Servlet containers include path parameters in the value returned for getPathInfo() and some do not. Spring Security uses the value returned by getPathInfo() as part of the process of mapping requests to security constraints. The unexpected presence of path parameters can cause a constraint to be bypassed. Users of Apache Tomcat (all current versions) are not affected by this vulnerability since Tomcat follows the guidance previously provided by the Servlet Expert group and strips path parameters from the value returned by getContextPath(), getServletPath(), and getPathInfo(). Users of other Servlet containers based on Apache Tomcat may or may not be affected depending on whether or not the handling of path parameters has been modified. Users of IBM WebSphere Application Server 8.5.x are known to be affected. Users of other containers that implement the Servlet specification may be affected.
<p>Publish Date: 2017-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9879>CVE-2016-9879</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9879">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9879</a></p>
<p>Release Date: 2017-01-06</p>
<p>Fix Resolution: 3.2.10,4.1.4,4.2.1.</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-9879 (High) detected in spring-security-web-4.1.1.RELEASE.jar - ## CVE-2016-9879 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-security-web-4.1.1.RELEASE.jar</b></p></summary>
<p>spring-security-web</p>
<p>Library home page: <a href="http://spring.io/spring-security">http://spring.io/spring-security</a></p>
<p>Path to dependency file: fitbit-api-example-java2/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/security/spring-security-web/4.1.1.RELEASE/spring-security-web-4.1.1.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-security-1.4.0.RELEASE.jar (Root Library)
- :x: **spring-security-web-4.1.1.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/bharathirajatut/fitbit-api-example-java2/commits/8c153ad064e8f07a4ddade35ac13a9b485ca3dac">8c153ad064e8f07a4ddade35ac13a9b485ca3dac</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in Pivotal Spring Security before 3.2.10, 4.1.x before 4.1.4, and 4.2.x before 4.2.1. Spring Security does not consider URL path parameters when processing security constraints. By adding a URL path parameter with an encoded "/" to a request, an attacker may be able to bypass a security constraint. The root cause of this issue is a lack of clarity regarding the handling of path parameters in the Servlet Specification. Some Servlet containers include path parameters in the value returned for getPathInfo() and some do not. Spring Security uses the value returned by getPathInfo() as part of the process of mapping requests to security constraints. The unexpected presence of path parameters can cause a constraint to be bypassed. Users of Apache Tomcat (all current versions) are not affected by this vulnerability since Tomcat follows the guidance previously provided by the Servlet Expert group and strips path parameters from the value returned by getContextPath(), getServletPath(), and getPathInfo(). Users of other Servlet containers based on Apache Tomcat may or may not be affected depending on whether or not the handling of path parameters has been modified. Users of IBM WebSphere Application Server 8.5.x are known to be affected. Users of other containers that implement the Servlet specification may be affected.
<p>Publish Date: 2017-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9879>CVE-2016-9879</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9879">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-9879</a></p>
<p>Release Date: 2017-01-06</p>
<p>Fix Resolution: 3.2.10,4.1.4,4.2.1.</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in spring security web release jar cve high severity vulnerability vulnerable library spring security web release jar spring security web library home page a href path to dependency file fitbit api example pom xml path to vulnerable library home wss scanner repository org springframework security spring security web release spring security web release jar dependency hierarchy spring boot starter security release jar root library x spring security web release jar vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in pivotal spring security before x before and x before spring security does not consider url path parameters when processing security constraints by adding a url path parameter with an encoded to a request an attacker may be able to bypass a security constraint the root cause of this issue is a lack of clarity regarding the handling of path parameters in the servlet specification some servlet containers include path parameters in the value returned for getpathinfo and some do not spring security uses the value returned by getpathinfo as part of the process of mapping requests to security constraints the unexpected presence of path parameters can cause a constraint to be bypassed users of apache tomcat all current versions are not affected by this vulnerability since tomcat follows the guidance previously provided by the servlet expert group and strips path parameters from the value returned by getcontextpath getservletpath and getpathinfo users of other servlet containers based on apache tomcat may or may not be affected depending on whether or not the handling of path parameters has been modified users of ibm websphere application server x are known to be affected users of other containers that implement the servlet specification may be affected publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
136,529
| 5,284,300,172
|
IssuesEvent
|
2017-02-07 23:49:53
|
openshift/origin
|
https://api.github.com/repos/openshift/origin
|
closed
|
v2 gear to v3 pod conversion
|
component/image kind/enhancement priority/P2
|
Hello,
In today's community meeting, I spoke with @smarterclayton about what, if anything, could be done in the form of tooling for ose/origin v2 gears that can then be converted to some form of usable template/pod in the v3 world.
In my "napkin design" it seems like the process will follow something similar to:
1. take a gear backup.zip
2. break it down to discover components
3. If something is in the data dir create PVC (with default size, at some point allow for custom size to be requested)
4. If DB discovered create the appropriate template for the needed DB (inquire about persistence vs not, add another PVC with default sizing)
5. Separate out the code portion and allow user to plug in URL to their SC system (github, local gitlab, whatever) and push it out. Add to template with new source location
6. Go ahead and provision a project and create the services needed
7. attempt DB data import/tell user they have to import their DB data, here are some links on how to do that.
|
1.0
|
v2 gear to v3 pod conversion - Hello,
In today's community meeting, I spoke with @smarterclayton about what, if anything, could be done in the form of tooling for ose/origin v2 gears that can then be converted to some form of usable template/pod in the v3 world.
In my "napkin design" it seems like the process will follow something similar to:
1. take a gear backup.zip
2. break it down to discover components
3. If something is in the data dir create PVC (with default size, at some point allow for custom size to be requested)
4. If DB discovered create the appropriate template for the needed DB (inquire about persistence vs not, add another PVC with default sizing)
5. Separate out the code portion and allow user to plug in URL to their SC system (github, local gitlab, whatever) and push it out. Add to template with new source location
6. Go ahead and provision a project and create the services needed
7. attempt DB data import/tell user they have to import their DB data, here are some links on how to do that.
|
non_process
|
gear to pod conversion hello in today s community meeting i spoke with smarterclayton about what if anything could be done in the form of tooling for ose origin gears that can then be converted to some form of usable template pod in the world in my napkin design it seems like the process will follow something similar to take a gear backup zip break it down to discover components if something is in the data dir create pvc with default size at some point allow for custom size to be requested if db discovered create the appropriate template for the needed db inquire about persistence vs not add another pvc with default sizing separate out the code portion and allow user to plug in url to their sc system github local gitlab whatever and push it out add to template with new source location go ahead and provision a project and create the services needed attempt db data import tell user they have to import their db data here are some links on how to do that
| 0
|
244,477
| 7,875,420,697
|
IssuesEvent
|
2018-06-25 20:22:27
|
gctools-outilsgc/gcconnex
|
https://api.github.com/repos/gctools-outilsgc/gcconnex
|
closed
|
Polls: Ability to Close polls to retain and display results after poll periode has ended
|
Priority: Low Project: Legacy Tools Type: Enhancement [zube]: Inbox
|
Users only have the option to delete a poll once their desired poll period has ended preventing them from retaining the poll results for other users to view.
Request: Allow users to close polls in order to be able to share the results once the poll period has ended. while retaining the ability to delete the poll once its use has expired.
In addition to this, it would be beneficial to have the ability to not only select “All”, “Mine” or “Colleagues” when searching for polls, but also the ability to choose “Open”, “Closed” or “All” as a sub-selection to each of the aforementioned selection options.
this will provide more flexibility in the use of the Poll feature.
|
1.0
|
Polls: Ability to Close polls to retain and display results after poll periode has ended - Users only have the option to delete a poll once their desired poll period has ended preventing them from retaining the poll results for other users to view.
Request: Allow users to close polls in order to be able to share the results once the poll period has ended. while retaining the ability to delete the poll once its use has expired.
In addition to this, it would be beneficial to have the ability to not only select “All”, “Mine” or “Colleagues” when searching for polls, but also the ability to choose “Open”, “Closed” or “All” as a sub-selection to each of the aforementioned selection options.
this will provide more flexibility in the use of the Poll feature.
|
non_process
|
polls ability to close polls to retain and display results after poll periode has ended users only have the option to delete a poll once their desired poll period has ended preventing them from retaining the poll results for other users to view request allow users to close polls in order to be able to share the results once the poll period has ended while retaining the ability to delete the poll once its use has expired in addition to this it would be beneficial to have the ability to not only select “all” “mine” or “colleagues” when searching for polls but also the ability to choose “open” “closed” or “all” as a sub selection to each of the aforementioned selection options this will provide more flexibility in the use of the poll feature
| 0
|
125,657
| 16,823,983,439
|
IssuesEvent
|
2021-06-17 16:05:04
|
department-of-veterans-affairs/va.gov-team
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
|
closed
|
[Variant Test] Product Checklist - Sorting Facilities
|
product vaos-product-design
|
# Summary
## Tasks
- [ ] Summary [documentation in GitHub](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/health-care/appointments/va-online-scheduling/initiatives/variant-tests/2021-sorting-facilities.md)
- [ ] Designs
- [Sketch file](https://www.sketch.com/s/439713c6-cb08-4b65-94ab-d97dcac0beba) (In progress)
- [ ] Dev ticket(s)
- #24280 (dev)
- #26106 (events for analytics)
- [ ] [Results readout slides](https://docs.google.com/presentation/d/1SvwwoiDa6OrzAuTWwetqAz2WUH3j4z0Q73muJXfwipk/edit?usp=sharing)
- [ ] Benchmark
- [ ] Outcomes
|
1.0
|
[Variant Test] Product Checklist - Sorting Facilities - # Summary
## Tasks
- [ ] Summary [documentation in GitHub](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/health-care/appointments/va-online-scheduling/initiatives/variant-tests/2021-sorting-facilities.md)
- [ ] Designs
- [Sketch file](https://www.sketch.com/s/439713c6-cb08-4b65-94ab-d97dcac0beba) (In progress)
- [ ] Dev ticket(s)
- #24280 (dev)
- #26106 (events for analytics)
- [ ] [Results readout slides](https://docs.google.com/presentation/d/1SvwwoiDa6OrzAuTWwetqAz2WUH3j4z0Q73muJXfwipk/edit?usp=sharing)
- [ ] Benchmark
- [ ] Outcomes
|
non_process
|
product checklist sorting facilities summary tasks summary designs in progress dev ticket s dev events for analytics benchmark outcomes
| 0
|
9,341
| 12,342,808,451
|
IssuesEvent
|
2020-05-15 01:59:57
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `Rpad` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `Rpad` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `Rpad` from TiDB -
## Description
Port the scalar function `Rpad` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @andylokandy
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function rpad from tidb description port the scalar function rpad from tidb to coprocessor score mentor s andylokandy recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
169,776
| 13,160,672,205
|
IssuesEvent
|
2020-08-10 18:02:46
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
The table auto refresh after canceling the file import
|
:gear: tables 🧪 testing
|
**Storage Explorer Version:** 1.15.0-dev
**Build**: 20200806.6
**Branch**: master
**Platform/OS:** Windows 10/ CentOS 7.6.1810 (Core) / MacOS Catalina
**Architecture**: ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Expand one storage account -> Tables.
2. Create a new table -> Click 'Import' action in the toolbar -> Click 'Cancel' button in the file explorer.
3. Check the result.
**Expect Experience:**
The table doesn't refresh.
(Unnecessary auto refreshing will make the table loading a long time when one table has a large number of entities).
**Actual Experience:**
The table auto refresh.
|
1.0
|
The table auto refresh after canceling the file import - **Storage Explorer Version:** 1.15.0-dev
**Build**: 20200806.6
**Branch**: master
**Platform/OS:** Windows 10/ CentOS 7.6.1810 (Core) / MacOS Catalina
**Architecture**: ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Expand one storage account -> Tables.
2. Create a new table -> Click 'Import' action in the toolbar -> Click 'Cancel' button in the file explorer.
3. Check the result.
**Expect Experience:**
The table doesn't refresh.
(Unnecessary auto refreshing will make the table loading a long time when one table has a large number of entities).
**Actual Experience:**
The table auto refresh.
|
non_process
|
the table auto refresh after canceling the file import storage explorer version dev build branch master platform os windows centos core macos catalina architecture regression from not a regression steps to reproduce expand one storage account tables create a new table click import action in the toolbar click cancel button in the file explorer check the result expect experience the table doesn t refresh unnecessary auto refreshing will make the table loading a long time when one table has a large number of entities actual experience the table auto refresh
| 0
|
416,343
| 12,143,028,358
|
IssuesEvent
|
2020-04-24 03:34:46
|
FStarLang/FStar
|
https://api.github.com/repos/FStarLang/FStar
|
reopened
|
Proper escaping for --dep
|
component/dependency-analysis kind/enhancement priority/low
|
If you include a directory with spaces in it, then the dependencies generated by `--dep full` are not quoted, and thus it generates a broken Makefile.
To reproduce:
```
mkdir "a b"
cd "a b"
cat > A.fst <<EOF
module A
EOF
cd ..
cat > B.fst <<EOF
module B
open A
EOF
fstar.exe --include "a b" --dep full B.fst
```
One of the parts causing the issue:
```
/tmp/asd/a b/A.fst.checked: /tmp/asd/a b/A.fst \
/home/jay/everest/FStar/ulib/FStar.Pervasives.fst.checked\
/home/jay/everest/FStar/ulib/prims.fst.checked
```
Expected output:
```
/tmp/asd/a\ b/A.fst.checked: /tmp/asd/a\ b/A.fst \
/home/jay/everest/FStar/ulib/FStar.Pervasives.fst.checked\
/home/jay/everest/FStar/ulib/prims.fst.checked
```
Basically, all spaces in filenames need to be replaced with `\ ` instead. It does indeed seem to do this in one place (in particular, the first rule), but then stops doing so.
Not a very high priority issue, but I thought it best to have it documented.
|
1.0
|
Proper escaping for --dep - If you include a directory with spaces in it, then the dependencies generated by `--dep full` are not quoted, and thus it generates a broken Makefile.
To reproduce:
```
mkdir "a b"
cd "a b"
cat > A.fst <<EOF
module A
EOF
cd ..
cat > B.fst <<EOF
module B
open A
EOF
fstar.exe --include "a b" --dep full B.fst
```
One of the parts causing the issue:
```
/tmp/asd/a b/A.fst.checked: /tmp/asd/a b/A.fst \
/home/jay/everest/FStar/ulib/FStar.Pervasives.fst.checked\
/home/jay/everest/FStar/ulib/prims.fst.checked
```
Expected output:
```
/tmp/asd/a\ b/A.fst.checked: /tmp/asd/a\ b/A.fst \
/home/jay/everest/FStar/ulib/FStar.Pervasives.fst.checked\
/home/jay/everest/FStar/ulib/prims.fst.checked
```
Basically, all spaces in filenames need to be replaced with `\ ` instead. It does indeed seem to do this in one place (in particular, the first rule), but then stops doing so.
Not a very high priority issue, but I thought it best to have it documented.
|
non_process
|
proper escaping for dep if you include a directory with spaces in it then the dependencies generated by dep full are not quoted and thus it generates a broken makefile to reproduce mkdir a b cd a b cat a fst eof module a eof cd cat b fst eof module b open a eof fstar exe include a b dep full b fst one of the parts causing the issue tmp asd a b a fst checked tmp asd a b a fst home jay everest fstar ulib fstar pervasives fst checked home jay everest fstar ulib prims fst checked expected output tmp asd a b a fst checked tmp asd a b a fst home jay everest fstar ulib fstar pervasives fst checked home jay everest fstar ulib prims fst checked basically all spaces in filenames need to be replaced with instead it does indeed seem to do this in one place in particular the first rule but then stops doing so not a very high priority issue but i thought it best to have it documented
| 0
|
212,216
| 7,229,417,644
|
IssuesEvent
|
2018-02-11 19:38:04
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
[Coverity CID: 182769] Error handling issues in /subsys/bluetooth/host/mesh/beacon.c
|
Coverity area: Bluetooth bug priority: medium
|
Static code scan issues seen in File: /subsys/bluetooth/host/mesh/beacon.c
Category: Error handling issues
Function: unprovisioned_beacon_send
Component: Bluetooth
CID: 182769
Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996
|
1.0
|
[Coverity CID: 182769] Error handling issues in /subsys/bluetooth/host/mesh/beacon.c - Static code scan issues seen in File: /subsys/bluetooth/host/mesh/beacon.c
Category: Error handling issues
Function: unprovisioned_beacon_send
Component: Bluetooth
CID: 182769
Please fix or provide comments to square it off in coverity in the link: https://scan9.coverity.com/reports.htm#v32951/p12996
|
non_process
|
error handling issues in subsys bluetooth host mesh beacon c static code scan issues seen in file subsys bluetooth host mesh beacon c category error handling issues function unprovisioned beacon send component bluetooth cid please fix or provide comments to square it off in coverity in the link
| 0
|
561,763
| 16,623,450,024
|
IssuesEvent
|
2021-06-03 06:29:30
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Custom filter expression does not show editor properly on subsequent click
|
.Frontend .Reproduced Priority:P2 Querying/Notebook Type:Bug
|
**Describe the bug**
In the notebook query editor, if a Filter is added with a Custom expression, and added, then that clause is clicked again, the editor that appears is not the correct one.
**Logs**
N/A
**To Reproduce**
Steps to reproduce the behavior:
1. Go to a Notebook question editor
2. Click on the + under Filter, in order to add a new filter clause
3. Select Custom expression from the dropdown, and enter some expression (ex: `between([ID], 96, 97)`); see screenshot 1
4. Click Done to add the clause
5. Click on the clause again (ex: to make a change)
6. Notice the dialog that appears is blank (doesn't show the expression value); see screenshot 2
**Expected behavior**
The same dialog that appears when adding the custom filter expression (ex: from step 3 above) should appear, and the custom expression that was already populated should be prefilled.
As a workaround, you can click the left/back arrow shown in the dialog header, then select Custom expression again, and the expression will again be editable there.
**Screenshots**
1. <img width="495" alt="Screen Shot 2021-02-09 at 1 35 43 PM" src="https://user-images.githubusercontent.com/3521562/107419654-a6948500-6add-11eb-830e-5a509f6f8e1b.png">
2. <img width="332" alt="Screen Shot 2021-02-09 at 1 35 48 PM" src="https://user-images.githubusercontent.com/3521562/107419665-aac0a280-6add-11eb-8737-3f4099df6cd6.png">
**Information about your Metabase Installation:**
You can get this information by going to Admin -> Troubleshooting.
- Your browser and the version: Chrome 88.0.4324.150
- Your operating system: OS X 11.1
- Your databases: Redshift
- Metabase version: `0.38.1-SNAPSHOT` (src_hash=`6d77f0992f5f51345554bc34338f05786f26c54b`)
- Metabase hosting environment: N/A
- Metabase internal database: H2
**Severity**
Minor.
**Additional context**
N/A
|
1.0
|
Custom filter expression does not show editor properly on subsequent click - **Describe the bug**
In the notebook query editor, if a Filter is added with a Custom expression, and added, then that clause is clicked again, the editor that appears is not the correct one.
**Logs**
N/A
**To Reproduce**
Steps to reproduce the behavior:
1. Go to a Notebook question editor
2. Click on the + under Filter, in order to add a new filter clause
3. Select Custom expression from the dropdown, and enter some expression (ex: `between([ID], 96, 97)`); see screenshot 1
4. Click Done to add the clause
5. Click on the clause again (ex: to make a change)
6. Notice the dialog that appears is blank (doesn't show the expression value); see screenshot 2
**Expected behavior**
The same dialog that appears when adding the custom filter expression (ex: from step 3 above) should appear, and the custom expression that was already populated should be prefilled.
As a workaround, you can click the left/back arrow shown in the dialog header, then select Custom expression again, and the expression will again be editable there.
**Screenshots**
1. <img width="495" alt="Screen Shot 2021-02-09 at 1 35 43 PM" src="https://user-images.githubusercontent.com/3521562/107419654-a6948500-6add-11eb-830e-5a509f6f8e1b.png">
2. <img width="332" alt="Screen Shot 2021-02-09 at 1 35 48 PM" src="https://user-images.githubusercontent.com/3521562/107419665-aac0a280-6add-11eb-8737-3f4099df6cd6.png">
**Information about your Metabase Installation:**
You can get this information by going to Admin -> Troubleshooting.
- Your browser and the version: Chrome 88.0.4324.150
- Your operating system: OS X 11.1
- Your databases: Redshift
- Metabase version: `0.38.1-SNAPSHOT` (src_hash=`6d77f0992f5f51345554bc34338f05786f26c54b`)
- Metabase hosting environment: N/A
- Metabase internal database: H2
**Severity**
Minor.
**Additional context**
N/A
|
non_process
|
custom filter expression does not show editor properly on subsequent click describe the bug in the notebook query editor if a filter is added with a custom expression and added then that clause is clicked again the editor that appears is not the correct one logs n a to reproduce steps to reproduce the behavior go to a notebook question editor click on the under filter in order to add a new filter clause select custom expression from the dropdown and enter some expression ex between see screenshot click done to add the clause click on the clause again ex to make a change notice the dialog that appears is blank doesn t show the expression value see screenshot expected behavior the same dialog that appears when adding the custom filter expression ex from step above should appear and the custom expression that was already populated should be prefilled as a workaround you can click the left back arrow shown in the dialog header then select custom expression again and the expression will again be editable there screenshots img width alt screen shot at pm src img width alt screen shot at pm src information about your metabase installation you can get this information by going to admin troubleshooting your browser and the version chrome your operating system os x your databases redshift metabase version snapshot src hash metabase hosting environment n a metabase internal database severity minor additional context n a
| 0
|
813,391
| 30,455,816,738
|
IssuesEvent
|
2023-07-16 21:48:56
|
GoogleCloudPlatform/java-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/java-docs-samples
|
closed
|
The build failed
|
type: bug priority: p1 samples flakybot: issue
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: da33713f879e447658cf452dea5f17a0827cddd6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/4d0d6834-0b98-4f5f-a020-d9af8ddcff7f), [Sponge](http://sponge2/4d0d6834-0b98-4f5f-a020-d9af8ddcff7f)
status: failed
<details><summary>Test output</summary><br><pre>java.lang.RuntimeException: Base URL not found.
at com.example.filesystem.ApplicationTests.setup(ApplicationTests.java:111)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:568)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at org.springframework.test.context.junit4.statements.RunBeforeTestClassCallbacks.evaluate(RunBeforeTestClassCallbacks.java:61)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.springframework.test.context.junit4.statements.RunAfterTestClassCallbacks.evaluate(RunAfterTestClassCallbacks.java:70)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.run(SpringJUnit4ClassRunner.java:191)
at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
at org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
at org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
at org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:102)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:188)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:128)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
</pre></details>
|
1.0
|
The build failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: da33713f879e447658cf452dea5f17a0827cddd6
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/4d0d6834-0b98-4f5f-a020-d9af8ddcff7f), [Sponge](http://sponge2/4d0d6834-0b98-4f5f-a020-d9af8ddcff7f)
status: failed
<details><summary>Test output</summary><br><pre>java.lang.RuntimeException: Base URL not found.
at com.example.filesystem.ApplicationTests.setup(ApplicationTests.java:111)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:568)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.RunBefores.invokeMethod(RunBefores.java:33)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at org.springframework.test.context.junit4.statements.RunBeforeTestClassCallbacks.evaluate(RunBeforeTestClassCallbacks.java:61)
at org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.springframework.test.context.junit4.statements.RunAfterTestClassCallbacks.evaluate(RunAfterTestClassCallbacks.java:70)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.springframework.test.context.junit4.SpringJUnit4ClassRunner.run(SpringJUnit4ClassRunner.java:191)
at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
at org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
at org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
at org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:102)
at org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:54)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
at org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
at org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
at org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.execute(JUnitPlatformProvider.java:188)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invokeAllTests(JUnitPlatformProvider.java:154)
at org.apache.maven.surefire.junitplatform.JUnitPlatformProvider.invoke(JUnitPlatformProvider.java:128)
at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428)
at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162)
at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
</pre></details>
|
non_process
|
the build failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output java lang runtimeexception base url not found at com example filesystem applicationtests setup applicationtests java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements runbefores invokemethod runbefores java at org junit internal runners statements runbefores evaluate runbefores java at org springframework test context statements runbeforetestclasscallbacks evaluate runbeforetestclasscallbacks java at org junit internal runners statements runafters evaluate runafters java at org springframework test context statements runaftertestclasscallbacks evaluate runaftertestclasscallbacks java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org springframework test context run java at org junit runner junitcore run junitcore java at org junit runner junitcore run junitcore java at org junit vintage engine execution runnerexecutor execute runnerexecutor java at org junit vintage engine vintagetestengine executeallchildren vintagetestengine java at org junit vintage engine vintagetestengine execute vintagetestengine java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator lambda execute engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator withinterceptedstreams engineexecutionorchestrator java at org junit platform launcher core engineexecutionorchestrator execute engineexecutionorchestrator java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlauncher execute defaultlauncher java at org junit platform launcher core defaultlaunchersession delegatinglauncher execute defaultlaunchersession java at org junit platform launcher core sessionperrequestlauncher execute sessionperrequestlauncher java at org apache maven surefire junitplatform junitplatformprovider execute junitplatformprovider java at org apache maven surefire junitplatform junitplatformprovider invokealltests junitplatformprovider java at org apache maven surefire junitplatform junitplatformprovider invoke junitplatformprovider java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java
| 0
|
117,698
| 11,954,253,671
|
IssuesEvent
|
2020-04-03 23:00:48
|
rdmtc/node-red-contrib-sun-position
|
https://api.github.com/repos/rdmtc/node-red-contrib-sun-position
|
closed
|
Node installation not working
|
👎 wontfix 📖 documentation
|
Hi, I like to install your node on my RPI4. But I get errors
Steps to reproduce the behavior:
1. Searched in Node Red
2. Installed your node
3. See error
The error:
node-red-contrib-sun-position
Cannot read property 'nodes' of null

System information
- Node Version [ 1.0.16]
- Node-Red Version [ 0.20.8]
- Hardware [e.g. RaspberryPi3, TinkerboardS, etc.]
Any help is highly appriciated.
best regards René
|
1.0
|
Node installation not working - Hi, I like to install your node on my RPI4. But I get errors
Steps to reproduce the behavior:
1. Searched in Node Red
2. Installed your node
3. See error
The error:
node-red-contrib-sun-position
Cannot read property 'nodes' of null

System information
- Node Version [ 1.0.16]
- Node-Red Version [ 0.20.8]
- Hardware [e.g. RaspberryPi3, TinkerboardS, etc.]
Any help is highly appriciated.
best regards René
|
non_process
|
node installation not working hi i like to install your node on my but i get errors steps to reproduce the behavior searched in node red installed your node see error the error node red contrib sun position cannot read property nodes of null system information node version node red version hardware any help is highly appriciated best regards rené
| 0
|
259,296
| 8,196,769,256
|
IssuesEvent
|
2018-08-31 11:00:59
|
threefoldfoundation/www_threefold.io
|
https://api.github.com/repos/threefoldfoundation/www_threefold.io
|
closed
|
Join and Follow Page: clean up
|
priority_major
|
On the current page, there are multiple links. If we are to clean and rationalise we can follow this decision tree:
A. Are we really dissatisfied with the page?
B. If we are, is it us that is dissatisfied? Or do our audience not like it?
C. Do we want to poll a fully representative sample of them to ask?
D. If they give us an answer we don't like, are we sure we
If none of the above:
E. Are we satisfied with simply making the links smaller?
If not:
F. Is this actually an issue that we do not like having to have (and maintain) multiple social platforms?
G. If we do not like F. Is this only our view - or that of our audience?
H. Do we want to poll a fully representative sample of them to ask?
I. If they give us an answer we don't like, are we sure we
If none of the above it implies we want to get rid of some or all of the links.
J. Which do we want to keep?
K. Once this is decided. Action point J.
|
1.0
|
Join and Follow Page: clean up - On the current page, there are multiple links. If we are to clean and rationalise we can follow this decision tree:
A. Are we really dissatisfied with the page?
B. If we are, is it us that is dissatisfied? Or do our audience not like it?
C. Do we want to poll a fully representative sample of them to ask?
D. If they give us an answer we don't like, are we sure we
If none of the above:
E. Are we satisfied with simply making the links smaller?
If not:
F. Is this actually an issue that we do not like having to have (and maintain) multiple social platforms?
G. If we do not like F. Is this only our view - or that of our audience?
H. Do we want to poll a fully representative sample of them to ask?
I. If they give us an answer we don't like, are we sure we
If none of the above it implies we want to get rid of some or all of the links.
J. Which do we want to keep?
K. Once this is decided. Action point J.
|
non_process
|
join and follow page clean up on the current page there are multiple links if we are to clean and rationalise we can follow this decision tree a are we really dissatisfied with the page b if we are is it us that is dissatisfied or do our audience not like it c do we want to poll a fully representative sample of them to ask d if they give us an answer we don t like are we sure we if none of the above e are we satisfied with simply making the links smaller if not f is this actually an issue that we do not like having to have and maintain multiple social platforms g if we do not like f is this only our view or that of our audience h do we want to poll a fully representative sample of them to ask i if they give us an answer we don t like are we sure we if none of the above it implies we want to get rid of some or all of the links j which do we want to keep k once this is decided action point j
| 0
|
14,306
| 17,294,755,166
|
IssuesEvent
|
2021-07-25 13:51:56
|
googleapis/python-bigquery
|
https://api.github.com/repos/googleapis/python-bigquery
|
closed
|
test: System tests create and destroy way too many datasets
|
api: bigquery testing type: process
|
Every system tests creates and destroys a temporary dataset in it's setUp and tearDown.
- Some system tests don't use the dataset created.
- Some system tests create and destroy their own datasets, seemingly oblivious that temporary datasets were already created for them.
This makes system tests take far longer than they should. ~10 minutes
Proposed fix:
- Don't create datasets by default. Make tests elect to create them.
- Move toward pytest fixtures to elect creating datasets.
- Use session fixtures to avoid creating a dataset for each test.
|
1.0
|
test: System tests create and destroy way too many datasets - Every system tests creates and destroys a temporary dataset in it's setUp and tearDown.
- Some system tests don't use the dataset created.
- Some system tests create and destroy their own datasets, seemingly oblivious that temporary datasets were already created for them.
This makes system tests take far longer than they should. ~10 minutes
Proposed fix:
- Don't create datasets by default. Make tests elect to create them.
- Move toward pytest fixtures to elect creating datasets.
- Use session fixtures to avoid creating a dataset for each test.
|
process
|
test system tests create and destroy way too many datasets every system tests creates and destroys a temporary dataset in it s setup and teardown some system tests don t use the dataset created some system tests create and destroy their own datasets seemingly oblivious that temporary datasets were already created for them this makes system tests take far longer than they should minutes proposed fix don t create datasets by default make tests elect to create them move toward pytest fixtures to elect creating datasets use session fixtures to avoid creating a dataset for each test
| 1
|
1,775
| 4,489,075,931
|
IssuesEvent
|
2016-08-30 09:40:54
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
opened
|
Add in missing relationship for 'viral release from host cell'
|
multiorganism processes Other term-related request quick fix viruses
|
A reminder to add in missing relationship:
viral release from host cell ; GO:0019076 is_a: exit from host cell ; GO:0035891
|
1.0
|
Add in missing relationship for 'viral release from host cell' - A reminder to add in missing relationship:
viral release from host cell ; GO:0019076 is_a: exit from host cell ; GO:0035891
|
process
|
add in missing relationship for viral release from host cell a reminder to add in missing relationship viral release from host cell go is a exit from host cell go
| 1
|
8,614
| 11,770,629,953
|
IssuesEvent
|
2020-03-15 20:05:34
|
pawn-lang/compiler
|
https://api.github.com/repos/pawn-lang/compiler
|
closed
|
#pragma option limited to 31 characters
|
area: pre-processor good first issue state: stale
|
<!-- pawn-lang compiler issue template -->
<!--
Please ensure that your issue is directly related to the pawn-lang/compiler. Issues
seeking help with scripting or doubts about SA-MP are not relevant to this project.
-->
### Issue description:
<!--
Information that could be useful:
- how to reproduce the problem
- observed behavior
- expected behavior
- other relevant information
-->
Using an option over 31 characters (I assume 32 with NULL) crashes the compiler.
### Minimal complete verifiable example (MCVE):
<!--
Please provide a minimal complete verifiable code which compiles
and reproduces the problem. If the compiler is showing undefined (random) behavior,
provide the code which invokes such behavior.
Learn more about writing MCVE from [StackOverflow](https://stackoverflow.com/help/mcve).
If this section is not relevant, feel free to remove this section from your issue.
-->
```pawn
// It doesn't even matter if this exists or not.
#pragma option -idependencies\y_thirdpartyinclude
```
<!-- comments on the code, if any -->
### Workspace Information:
* Compiler version:
* Command line arguments provided (or sampctl version):
* Operating System:
|
1.0
|
#pragma option limited to 31 characters - <!-- pawn-lang compiler issue template -->
<!--
Please ensure that your issue is directly related to the pawn-lang/compiler. Issues
seeking help with scripting or doubts about SA-MP are not relevant to this project.
-->
### Issue description:
<!--
Information that could be useful:
- how to reproduce the problem
- observed behavior
- expected behavior
- other relevant information
-->
Using an option over 31 characters (I assume 32 with NULL) crashes the compiler.
### Minimal complete verifiable example (MCVE):
<!--
Please provide a minimal complete verifiable code which compiles
and reproduces the problem. If the compiler is showing undefined (random) behavior,
provide the code which invokes such behavior.
Learn more about writing MCVE from [StackOverflow](https://stackoverflow.com/help/mcve).
If this section is not relevant, feel free to remove this section from your issue.
-->
```pawn
// It doesn't even matter if this exists or not.
#pragma option -idependencies\y_thirdpartyinclude
```
<!-- comments on the code, if any -->
### Workspace Information:
* Compiler version:
* Command line arguments provided (or sampctl version):
* Operating System:
|
process
|
pragma option limited to characters please ensure that your issue is directly related to the pawn lang compiler issues seeking help with scripting or doubts about sa mp are not relevant to this project issue description information that could be useful how to reproduce the problem observed behavior expected behavior other relevant information using an option over characters i assume with null crashes the compiler minimal complete verifiable example mcve please provide a minimal complete verifiable code which compiles and reproduces the problem if the compiler is showing undefined random behavior provide the code which invokes such behavior learn more about writing mcve from if this section is not relevant feel free to remove this section from your issue pawn it doesn t even matter if this exists or not pragma option idependencies y thirdpartyinclude workspace information compiler version command line arguments provided or sampctl version operating system
| 1
|
22,634
| 31,881,989,680
|
IssuesEvent
|
2023-09-16 13:37:49
|
pex-gl/pex-renderer
|
https://api.github.com/repos/pex-gl/pex-renderer
|
closed
|
View frustum culling
|
type/feat scope/post-processsing
|
Crucial for optimizing rendering of big scenes
Lighthouse3D tutorial on View Frustum Culling http://www.lighthouse3d.com/tutorials/view-frustum-culling/
|
1.0
|
View frustum culling - Crucial for optimizing rendering of big scenes
Lighthouse3D tutorial on View Frustum Culling http://www.lighthouse3d.com/tutorials/view-frustum-culling/
|
process
|
view frustum culling crucial for optimizing rendering of big scenes tutorial on view frustum culling
| 1
|
45,141
| 5,907,922,663
|
IssuesEvent
|
2017-05-19 18:57:02
|
ludumdare/ludumdare
|
https://api.github.com/repos/ludumdare/ludumdare
|
opened
|
List Making
|
core design enhancement
|
This would be **SUPER USEFUL** to have. I have a few cases where I've needed to users to provide me list. If this was an actual feature, I could actually run queries on the raw lists I'm provided.
|
1.0
|
List Making - This would be **SUPER USEFUL** to have. I have a few cases where I've needed to users to provide me list. If this was an actual feature, I could actually run queries on the raw lists I'm provided.
|
non_process
|
list making this would be super useful to have i have a few cases where i ve needed to users to provide me list if this was an actual feature i could actually run queries on the raw lists i m provided
| 0
|
92,066
| 15,856,856,255
|
IssuesEvent
|
2021-04-08 03:20:57
|
DaniOlivares/test-truffle-oz-tutorial-token
|
https://api.github.com/repos/DaniOlivares/test-truffle-oz-tutorial-token
|
opened
|
CVE-2018-20676 (Medium) detected in bootstrap-3.3.7.min.js
|
security vulnerability
|
## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: test-truffle-oz-tutorial-token/src/index.html</p>
<p>Path to vulnerable library: test-truffle-oz-tutorial-token/src/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-20676 (Medium) detected in bootstrap-3.3.7.min.js - ## CVE-2018-20676 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p>
<p>Path to dependency file: test-truffle-oz-tutorial-token/src/index.html</p>
<p>Path to vulnerable library: test-truffle-oz-tutorial-token/src/js/bootstrap.min.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.7.min.js** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 3.4.0, XSS is possible in the tooltip data-viewport attribute.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20676>CVE-2018-20676</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20676</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file test truffle oz tutorial token src index html path to vulnerable library test truffle oz tutorial token src js bootstrap min js dependency hierarchy x bootstrap min js vulnerable library vulnerability details in bootstrap before xss is possible in the tooltip data viewport attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap step up your open source security game with whitesource
| 0
|
22,115
| 30,644,964,447
|
IssuesEvent
|
2023-07-25 03:18:17
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
[MLv2] [Bug] `join-alias` is only added to the first condition's field reference
|
.Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench:
|
It seems like the `join-alias` is only added to the first condition's RHS field reference. When adding more conditions to the same join, the `join-alias` is missing.
Here's an example query with two conditions:
<details>
<summary>Expand snippet</summary>
```js
{
database: 1,
type: "query",
query: {
"source-table": 2,
joins: [
{
fields: "all",
alias: "Products",
condition: [
"and",
[
"=",
["field", 13, { "base-type": "type/Integer" }],
[
"field",
5,
{ "base-type": "type/BigInteger", "join-alias": "Products" },
],
],
[
"=",
["field", 14, { "base-type": "type/DateTime" }],
["field", 7, { "base-type": "type/DateTime" }],
],
],
"source-table": 1,
},
],
},
}
```
</details>
### To reproduce
1. Start a query from the sample Orders table
2. Join the Products table on Orders.PRODUCT_ID = Products.ID (using `Lib.joinConditionClause` + `Lib.join`)
3. Use `Lib.withJoinConditions` + `Lib.replaceClause` and add another join condition (e.g. Orders.CREATED_AT = Products.CREATED_AT)
4. Inspect the second condition's field reference
**Expected:** RHS field reference has a `join-alias`
**Actual**: RHS field reference only has a common set of options and gets dropped as invalid
|
1.0
|
[MLv2] [Bug] `join-alias` is only added to the first condition's field reference - It seems like the `join-alias` is only added to the first condition's RHS field reference. When adding more conditions to the same join, the `join-alias` is missing.
Here's an example query with two conditions:
<details>
<summary>Expand snippet</summary>
```js
{
database: 1,
type: "query",
query: {
"source-table": 2,
joins: [
{
fields: "all",
alias: "Products",
condition: [
"and",
[
"=",
["field", 13, { "base-type": "type/Integer" }],
[
"field",
5,
{ "base-type": "type/BigInteger", "join-alias": "Products" },
],
],
[
"=",
["field", 14, { "base-type": "type/DateTime" }],
["field", 7, { "base-type": "type/DateTime" }],
],
],
"source-table": 1,
},
],
},
}
```
</details>
### To reproduce
1. Start a query from the sample Orders table
2. Join the Products table on Orders.PRODUCT_ID = Products.ID (using `Lib.joinConditionClause` + `Lib.join`)
3. Use `Lib.withJoinConditions` + `Lib.replaceClause` and add another join condition (e.g. Orders.CREATED_AT = Products.CREATED_AT)
4. Inspect the second condition's field reference
**Expected:** RHS field reference has a `join-alias`
**Actual**: RHS field reference only has a common set of options and gets dropped as invalid
|
process
|
join alias is only added to the first condition s field reference it seems like the join alias is only added to the first condition s rhs field reference when adding more conditions to the same join the join alias is missing here s an example query with two conditions expand snippet js database type query query source table joins fields all alias products condition and field base type type biginteger join alias products source table to reproduce start a query from the sample orders table join the products table on orders product id products id using lib joinconditionclause lib join use lib withjoinconditions lib replaceclause and add another join condition e g orders created at products created at inspect the second condition s field reference expected rhs field reference has a join alias actual rhs field reference only has a common set of options and gets dropped as invalid
| 1
|
20,183
| 26,744,720,521
|
IssuesEvent
|
2023-01-30 15:16:33
|
polarby/render
|
https://api.github.com/repos/polarby/render
|
closed
|
Server side rendering
|
question render processing waiting for response
|
Hello, is there a way to render the widget to an image on the server side, maybe with headless flutter?
|
1.0
|
Server side rendering - Hello, is there a way to render the widget to an image on the server side, maybe with headless flutter?
|
process
|
server side rendering hello is there a way to render the widget to an image on the server side maybe with headless flutter
| 1
|
16,618
| 2,920,435,340
|
IssuesEvent
|
2015-06-24 18:55:24
|
ashanbh/chrome-rest-client
|
https://api.github.com/repos/ashanbh/chrome-rest-client
|
closed
|
add a feature to generate a curl command from the request
|
auto-migrated Priority-Medium Type-Defect Usability
|
```
Feature - todo
```
Original issue reported on code.google.com by `pawel.ps...@kalicinscy.com` on 12 Dec 2012 at 10:29
|
1.0
|
add a feature to generate a curl command from the request - ```
Feature - todo
```
Original issue reported on code.google.com by `pawel.ps...@kalicinscy.com` on 12 Dec 2012 at 10:29
|
non_process
|
add a feature to generate a curl command from the request feature todo original issue reported on code google com by pawel ps kalicinscy com on dec at
| 0
|
14,860
| 11,204,336,028
|
IssuesEvent
|
2020-01-05 04:39:27
|
MultiMC/MultiMC5
|
https://api.github.com/repos/MultiMC/MultiMC5
|
closed
|
MultiMC should be possible to translate before a stable release
|
multimc: feature multimc: infrastructure multimc: translations
|
I have been translating to spanish using weblate, but I can´t find some still pending strings using the search function in weblate, for example:
"Close", "Help" buttons in edit instance form
"Installation de Forge..." button in edit instance form
|
1.0
|
MultiMC should be possible to translate before a stable release - I have been translating to spanish using weblate, but I can´t find some still pending strings using the search function in weblate, for example:
"Close", "Help" buttons in edit instance form
"Installation de Forge..." button in edit instance form
|
non_process
|
multimc should be possible to translate before a stable release i have been translating to spanish using weblate but i can´t find some still pending strings using the search function in weblate for example close help buttons in edit instance form installation de forge button in edit instance form
| 0
|
12,498
| 14,961,464,570
|
IssuesEvent
|
2021-01-27 07:48:43
|
zammad/zammad
|
https://api.github.com/repos/zammad/zammad
|
closed
|
Japanese character is garbled.
|
bug mail processing verified
|
<!--
Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓
Since november 15th we handle all requests, except real bugs, at our community board.
Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21
Please post:
- Feature requests
- Development questions
- Technical questions
on the board -> https://community.zammad.org !
If you think you hit a bug, please continue:
- Search existing issues and the CHANGELOG.md for your issue - there might be a solution already
- Make sure to use the latest version of Zammad if possible
- Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it!
- Please write the issue in english
- Don't remove the template - otherwise we will close the issue without further comments
- Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate
Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted).
* The upper textblock will be removed automatically when you submit your issue *
-->
### Infos:
* Used Zammad version: 3.6.x
* Installation method (source, package, ..): apt install zammad
* Operating system: ubuntu 20.04
* Database + version: psql (PostgreSQL) 12.5 (Ubuntu 12.5-0ubuntu0.20.04.1)
* Elasticsearch version: sorry I couldn't find
* Browser + version: FireFox 84.0.2
### Expected behavior:
e-mail at Japanese-encoding displayed by Japanese character.
### Actual behavior:
Some e-mail at Japanese-encoding displayed garbled character.
(subject of e-mail displayed correctlly.only main text of mail is garbled.)
Similar problem like
Japanese encoding 7-bit iso-2022-jp is processed incorrectly #2498
but above issue is closed. so I make this one.
### Steps to reproduce the behavior:
・I will send eml file to support.
Yes I'm sure this is a bug and no feature request or a general question.
|
1.0
|
Japanese character is garbled. - <!--
Hi there - thanks for filing an issue. Please ensure the following things before creating an issue - thank you! 🤓
Since november 15th we handle all requests, except real bugs, at our community board.
Full explanation: https://community.zammad.org/t/major-change-regarding-github-issues-community-board/21
Please post:
- Feature requests
- Development questions
- Technical questions
on the board -> https://community.zammad.org !
If you think you hit a bug, please continue:
- Search existing issues and the CHANGELOG.md for your issue - there might be a solution already
- Make sure to use the latest version of Zammad if possible
- Add the `log/production.log` file from your system. Attention: Make sure no confidential data is in it!
- Please write the issue in english
- Don't remove the template - otherwise we will close the issue without further comments
- Ask questions about Zammad configuration and usage at our mailinglist. See: https://zammad.org/participate
Note: We always do our best. Unfortunately, sometimes there are too many requests and we can't handle everything at once. If you want to prioritize/escalate your issue, you can do so by means of a support contract (see https://zammad.com/pricing#selfhosted).
* The upper textblock will be removed automatically when you submit your issue *
-->
### Infos:
* Used Zammad version: 3.6.x
* Installation method (source, package, ..): apt install zammad
* Operating system: ubuntu 20.04
* Database + version: psql (PostgreSQL) 12.5 (Ubuntu 12.5-0ubuntu0.20.04.1)
* Elasticsearch version: sorry I couldn't find
* Browser + version: FireFox 84.0.2
### Expected behavior:
e-mail at Japanese-encoding displayed by Japanese character.
### Actual behavior:
Some e-mail at Japanese-encoding displayed garbled character.
(subject of e-mail displayed correctlly.only main text of mail is garbled.)
Similar problem like
Japanese encoding 7-bit iso-2022-jp is processed incorrectly #2498
but above issue is closed. so I make this one.
### Steps to reproduce the behavior:
・I will send eml file to support.
Yes I'm sure this is a bug and no feature request or a general question.
|
process
|
japanese character is garbled hi there thanks for filing an issue please ensure the following things before creating an issue thank you 🤓 since november we handle all requests except real bugs at our community board full explanation please post feature requests development questions technical questions on the board if you think you hit a bug please continue search existing issues and the changelog md for your issue there might be a solution already make sure to use the latest version of zammad if possible add the log production log file from your system attention make sure no confidential data is in it please write the issue in english don t remove the template otherwise we will close the issue without further comments ask questions about zammad configuration and usage at our mailinglist see note we always do our best unfortunately sometimes there are too many requests and we can t handle everything at once if you want to prioritize escalate your issue you can do so by means of a support contract see the upper textblock will be removed automatically when you submit your issue infos used zammad version x installation method source package apt install zammad operating system ubuntu database version psql postgresql ubuntu elasticsearch version sorry i couldn t find browser version firefox expected behavior e mail at japanese encoding displayed by japanese character actual behavior some e mail at japanese encoding displayed garbled character subject of e mail displayed correctlly only main text of mail is garbled similar problem like japanese encoding bit iso jp is processed incorrectly but above issue is closed so i make this one steps to reproduce the behavior ・i will send eml file to support yes i m sure this is a bug and no feature request or a general question
| 1
|
9,405
| 12,404,115,311
|
IssuesEvent
|
2020-05-21 15:00:55
|
checkifcovid/data-science-experiments
|
https://api.github.com/repos/checkifcovid/data-science-experiments
|
closed
|
Update preprocessing to allow ingestion of new schema for location data.
|
preprocessing
|
Update preprocessing to allow ingestion of new schema for location data.
**Current:**
```
user: {
...
postcode: string
...
}
```
**New:**
```
user: {
...
location: {
postcode: string,
street: string,
coordinates: {
lat: float,
lon: float
}
}
...
}
```
|
1.0
|
Update preprocessing to allow ingestion of new schema for location data. - Update preprocessing to allow ingestion of new schema for location data.
**Current:**
```
user: {
...
postcode: string
...
}
```
**New:**
```
user: {
...
location: {
postcode: string,
street: string,
coordinates: {
lat: float,
lon: float
}
}
...
}
```
|
process
|
update preprocessing to allow ingestion of new schema for location data update preprocessing to allow ingestion of new schema for location data current user postcode string new user location postcode string street string coordinates lat float lon float
| 1
|
270,066
| 8,446,036,717
|
IssuesEvent
|
2018-10-19 00:15:04
|
robot-lab/judyst-main-web-service
|
https://api.github.com/repos/robot-lab/judyst-main-web-service
|
opened
|
Страница настроек организации
|
area/front-end priority/high type/feature type/task
|
# Task request
## Цель задачи
Страница с настройками для организации.
На основании feature #82 и feature #86
## Решение задачи
Страница доступна только администратору организации и содержит поля необходимые в #82. Имеет ссылку на страницу организации.
## Дополнительный контекст или ссылки на связанные с данной задачей issues
|
1.0
|
Страница настроек организации - # Task request
## Цель задачи
Страница с настройками для организации.
На основании feature #82 и feature #86
## Решение задачи
Страница доступна только администратору организации и содержит поля необходимые в #82. Имеет ссылку на страницу организации.
## Дополнительный контекст или ссылки на связанные с данной задачей issues
|
non_process
|
страница настроек организации task request цель задачи страница с настройками для организации на основании feature и feature решение задачи страница доступна только администратору организации и содержит поля необходимые в имеет ссылку на страницу организации дополнительный контекст или ссылки на связанные с данной задачей issues
| 0
|
828,143
| 31,813,367,980
|
IssuesEvent
|
2023-09-13 18:30:23
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
closed
|
[YSQL] Sequential scan and Index scan on primary index cost differently
|
kind/enhancement area/ysql priority/medium
|
Jira Link: [DB-7871](https://yugabyte.atlassian.net/browse/DB-7871)
### Description
We see that sequential scans on a table and index scans on the table's primary index cost very differently even though they are effectively the same operation.
Consider the following setup:
```
create table p100000(c1 int, primary key(c1 asc));
insert into p100000 select generate_series(1,100000);
analyze p100000;
set yb_enable_base_scans_cost_model to true;
explain select * from p100000;
QUERY PLAN
----------------------------------------------------------------
Seq Scan on p100000 (cost=7.20..51995.53 rows=100000 width=4)
(1 row)
set enable_seqscan to false;
explain select * from p100000 order by c1 asc;
QUERY PLAN
--------------------------------------------------------------------------------------
Index Scan using p100000_pkey on p100000 (cost=7.20..64570.53 rows=100000 width=4)
(1 row)
```
We see that these two queries that effectively do the same thing are costed very differently. Because of this, the cost model seems to believe that even sorting the sequential scan is more efficient than directly reading the desired ordering from the primary index in the following query:
```
set enable_seqscan to true;
explain select * from p100000 order by c1 asc;
QUERY PLAN
----------------------------------------------------------------------
Sort (cost=60300.35..60550.35 rows=100000 width=4)
Sort Key: c1
-> Seq Scan on p100000 (cost=7.20..51995.53 rows=100000 width=4)
(3 rows)
```
This needs to be investigated.
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-7871]: https://yugabyte.atlassian.net/browse/DB-7871?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
|
1.0
|
[YSQL] Sequential scan and Index scan on primary index cost differently - Jira Link: [DB-7871](https://yugabyte.atlassian.net/browse/DB-7871)
### Description
We see that sequential scans on a table and index scans on the table's primary index cost very differently even though they are effectively the same operation.
Consider the following setup:
```
create table p100000(c1 int, primary key(c1 asc));
insert into p100000 select generate_series(1,100000);
analyze p100000;
set yb_enable_base_scans_cost_model to true;
explain select * from p100000;
QUERY PLAN
----------------------------------------------------------------
Seq Scan on p100000 (cost=7.20..51995.53 rows=100000 width=4)
(1 row)
set enable_seqscan to false;
explain select * from p100000 order by c1 asc;
QUERY PLAN
--------------------------------------------------------------------------------------
Index Scan using p100000_pkey on p100000 (cost=7.20..64570.53 rows=100000 width=4)
(1 row)
```
We see that these two queries that effectively do the same thing are costed very differently. Because of this, the cost model seems to believe that even sorting the sequential scan is more efficient than directly reading the desired ordering from the primary index in the following query:
```
set enable_seqscan to true;
explain select * from p100000 order by c1 asc;
QUERY PLAN
----------------------------------------------------------------------
Sort (cost=60300.35..60550.35 rows=100000 width=4)
Sort Key: c1
-> Seq Scan on p100000 (cost=7.20..51995.53 rows=100000 width=4)
(3 rows)
```
This needs to be investigated.
### Warning: Please confirm that this issue does not contain any sensitive information
- [X] I confirm this issue does not contain any sensitive information.
[DB-7871]: https://yugabyte.atlassian.net/browse/DB-7871?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ
|
non_process
|
sequential scan and index scan on primary index cost differently jira link description we see that sequential scans on a table and index scans on the table s primary index cost very differently even though they are effectively the same operation consider the following setup create table int primary key asc insert into select generate series analyze set yb enable base scans cost model to true explain select from query plan seq scan on cost rows width row set enable seqscan to false explain select from order by asc query plan index scan using pkey on cost rows width row we see that these two queries that effectively do the same thing are costed very differently because of this the cost model seems to believe that even sorting the sequential scan is more efficient than directly reading the desired ordering from the primary index in the following query set enable seqscan to true explain select from order by asc query plan sort cost rows width sort key seq scan on cost rows width rows this needs to be investigated warning please confirm that this issue does not contain any sensitive information i confirm this issue does not contain any sensitive information
| 0
|
15,318
| 19,425,956,620
|
IssuesEvent
|
2021-12-21 05:32:09
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
All GDAL algorithms in QGIS 3.23.0-Master: "Wrong or missing parameter value" error
|
Processing Regression Bug
|
### What is the bug or the crash?
Just opening any GDAL algorithm's window using QGIS 3.23.0-Master (93399e33a3 - package qgis-dev-3.23.0-137-93399e33a3-1 from OSGeo4W v2) on Windows 10 the following error is displayed in the algorithm window and in the Log Messages panel:
`WARNING Wrong or missing parameter value:` followed by the name of one of the missing parameters.

### Steps to reproduce the issue
1. Open any GDAL algorithm from the Processing Toolbox panel
### Versions
QGIS version
3.23.0-Master
QGIS code revision
93399e33a3
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.0
PROJ version
8.2.0
EPSG Registry database version
v10.038 (2021-10-21)
GEOS version
3.10.0-CAPI-1.16.0
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 1903
This copy of QGIS writes debugging output.
Active Python plugins
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
The issue doesn't occur in QGIS 3.16.14 and QGIS 3.22.1 (on Windows 10 from OSGeo4W v2).
|
1.0
|
All GDAL algorithms in QGIS 3.23.0-Master: "Wrong or missing parameter value" error - ### What is the bug or the crash?
Just opening any GDAL algorithm's window using QGIS 3.23.0-Master (93399e33a3 - package qgis-dev-3.23.0-137-93399e33a3-1 from OSGeo4W v2) on Windows 10 the following error is displayed in the algorithm window and in the Log Messages panel:
`WARNING Wrong or missing parameter value:` followed by the name of one of the missing parameters.

### Steps to reproduce the issue
1. Open any GDAL algorithm from the Processing Toolbox panel
### Versions
QGIS version
3.23.0-Master
QGIS code revision
93399e33a3
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.0
PROJ version
8.2.0
EPSG Registry database version
v10.038 (2021-10-21)
GEOS version
3.10.0-CAPI-1.16.0
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 1903
This copy of QGIS writes debugging output.
Active Python plugins
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
The issue doesn't occur in QGIS 3.16.14 and QGIS 3.22.1 (on Windows 10 from OSGeo4W v2).
|
process
|
all gdal algorithms in qgis master wrong or missing parameter value error what is the bug or the crash just opening any gdal algorithm s window using qgis master package qgis dev from on windows the following error is displayed in the algorithm window and in the log messages panel warning wrong or missing parameter value followed by the name of one of the missing parameters steps to reproduce the issue open any gdal algorithm from the processing toolbox panel versions qgis version master qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version windows version this copy of qgis writes debugging output active python plugins db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context the issue doesn t occur in qgis and qgis on windows from
| 1
|
15,780
| 27,884,229,631
|
IssuesEvent
|
2023-03-21 22:05:53
|
adrrf/acme-life-long-learning
|
https://api.github.com/repos/adrrf/acme-life-long-learning
|
opened
|
d3-s3-fr-15 - assistant operations on sessions
|
develop functional_requirement mandatory
|
Operations by assistants on sessions:
- [ ] List the sessions in their tutorials.
- [ ] Show the details of their sessions.
- [ ] Create a new session in a tutorial, as long as it has been published.
- [ ] Update or delete the sessions in their tutorials, as long as they have not been published.
|
1.0
|
d3-s3-fr-15 - assistant operations on sessions - Operations by assistants on sessions:
- [ ] List the sessions in their tutorials.
- [ ] Show the details of their sessions.
- [ ] Create a new session in a tutorial, as long as it has been published.
- [ ] Update or delete the sessions in their tutorials, as long as they have not been published.
|
non_process
|
fr assistant operations on sessions operations by assistants on sessions list the sessions in their tutorials show the details of their sessions create a new session in a tutorial as long as it has been published update or delete the sessions in their tutorials as long as they have not been published
| 0
|
1,919
| 4,756,042,340
|
IssuesEvent
|
2016-10-24 12:53:39
|
opentrials/opentrials
|
https://api.github.com/repos/opentrials/opentrials
|
closed
|
Should we use some data pipeline framework in our processors?
|
Collectors Processors question refactoring
|
This issue is mostly to discuss our alternatives.
We're starting to create a few tasks in our processors' pipeline that will be used by many processors (e.g. downloading PDFs and pushing to S3 for the FDA data, https://github.com/opentrials/opentrials/issues/250). We can save them in a "helpers" folder for reusing, but looking forward it seems we would benefit by using a data pipeline framework. This would help us in things like not rerunning code that was already ran, keeping track of the status of the data pipeline, retrying steps that didn't work, etc.
There're quite a few options. The most interesting for us IMHO are [Airbnb's Airflow](http://nerds.airbnb.com/airflow/) (in incubation to become an Apache project), [Spotify's Luigi](https://github.com/spotify/luigi) and [Pinterest's Pinball](https://github.com/pinterest/pinball). I found a couple blog posts comparing them [here](https://www.michaelcho.me/article/data-pipelines-airflow-vs-pinball-vs-luigi) and [here](http://bytepawn.com/luigi-airflow-pinball.html) ([discussion on HN](https://news.ycombinator.com/item?id=11049684)).
@roll @akariv Does any of you have experience with these tools? Any thoughts on using them?
|
1.0
|
Should we use some data pipeline framework in our processors? - This issue is mostly to discuss our alternatives.
We're starting to create a few tasks in our processors' pipeline that will be used by many processors (e.g. downloading PDFs and pushing to S3 for the FDA data, https://github.com/opentrials/opentrials/issues/250). We can save them in a "helpers" folder for reusing, but looking forward it seems we would benefit by using a data pipeline framework. This would help us in things like not rerunning code that was already ran, keeping track of the status of the data pipeline, retrying steps that didn't work, etc.
There're quite a few options. The most interesting for us IMHO are [Airbnb's Airflow](http://nerds.airbnb.com/airflow/) (in incubation to become an Apache project), [Spotify's Luigi](https://github.com/spotify/luigi) and [Pinterest's Pinball](https://github.com/pinterest/pinball). I found a couple blog posts comparing them [here](https://www.michaelcho.me/article/data-pipelines-airflow-vs-pinball-vs-luigi) and [here](http://bytepawn.com/luigi-airflow-pinball.html) ([discussion on HN](https://news.ycombinator.com/item?id=11049684)).
@roll @akariv Does any of you have experience with these tools? Any thoughts on using them?
|
process
|
should we use some data pipeline framework in our processors this issue is mostly to discuss our alternatives we re starting to create a few tasks in our processors pipeline that will be used by many processors e g downloading pdfs and pushing to for the fda data we can save them in a helpers folder for reusing but looking forward it seems we would benefit by using a data pipeline framework this would help us in things like not rerunning code that was already ran keeping track of the status of the data pipeline retrying steps that didn t work etc there re quite a few options the most interesting for us imho are in incubation to become an apache project and i found a couple blog posts comparing them and roll akariv does any of you have experience with these tools any thoughts on using them
| 1
|
4,519
| 7,361,236,341
|
IssuesEvent
|
2018-03-11 04:26:09
|
shirou/gopsutil
|
https://api.github.com/repos/shirou/gopsutil
|
closed
|
fillFromStat causes significant CPU usage.
|
os:linux package:process
|
While using Nomad I observed that CPU usage to monitor processes was significant in certain situations. Using this [application](https://gist.github.com/maihde/30cacbc40003f6c59f625e6c0f901cb3) I profiled the calls to `process.Times()` and determined that a significant amount of the CPU usage was due to the calls by `getTerminalMap()` within the `fillFromStat()` method. Given that the `terminal` return value is usually thrown away it's better if `getTerminalMap()` is only called within the `process.Terminal()` method itself.
See this [Nomad PR](https://github.com/hashicorp/nomad/pull/3883) for resolution and this [Nomad Issue](https://github.com/hashicorp/nomad/issues/3881) for more details.
|
1.0
|
fillFromStat causes significant CPU usage. - While using Nomad I observed that CPU usage to monitor processes was significant in certain situations. Using this [application](https://gist.github.com/maihde/30cacbc40003f6c59f625e6c0f901cb3) I profiled the calls to `process.Times()` and determined that a significant amount of the CPU usage was due to the calls by `getTerminalMap()` within the `fillFromStat()` method. Given that the `terminal` return value is usually thrown away it's better if `getTerminalMap()` is only called within the `process.Terminal()` method itself.
See this [Nomad PR](https://github.com/hashicorp/nomad/pull/3883) for resolution and this [Nomad Issue](https://github.com/hashicorp/nomad/issues/3881) for more details.
|
process
|
fillfromstat causes significant cpu usage while using nomad i observed that cpu usage to monitor processes was significant in certain situations using this i profiled the calls to process times and determined that a significant amount of the cpu usage was due to the calls by getterminalmap within the fillfromstat method given that the terminal return value is usually thrown away it s better if getterminalmap is only called within the process terminal method itself see this for resolution and this for more details
| 1
|
16,078
| 20,249,228,402
|
IssuesEvent
|
2022-02-14 16:20:38
|
Bone008/orbiteye
|
https://api.github.com/repos/Bone008/orbiteye
|
opened
|
Implement timeline
|
data processing Front End/UI
|
- [ ] Create timeline component
- [ ] Hook up timeline to global app state to affect visualizations
|
1.0
|
Implement timeline - - [ ] Create timeline component
- [ ] Hook up timeline to global app state to affect visualizations
|
process
|
implement timeline create timeline component hook up timeline to global app state to affect visualizations
| 1
|
241,496
| 18,459,993,955
|
IssuesEvent
|
2021-10-15 22:49:54
|
bounswe/2021SpringGroup9
|
https://api.github.com/repos/bounswe/2021SpringGroup9
|
opened
|
Think about new requirements and how to improve the project plan
|
documentation priority: medium wiki difficulty: medium
|
Try to improve project requirements by thinking about what requirements we can add to the project. Also try to find ways to improve project plan.
|
1.0
|
Think about new requirements and how to improve the project plan - Try to improve project requirements by thinking about what requirements we can add to the project. Also try to find ways to improve project plan.
|
non_process
|
think about new requirements and how to improve the project plan try to improve project requirements by thinking about what requirements we can add to the project also try to find ways to improve project plan
| 0
|
13,278
| 15,759,440,103
|
IssuesEvent
|
2021-03-31 07:56:01
|
emacs-ess/ESS
|
https://api.github.com/repos/emacs-ess/ESS
|
closed
|
Process restoration with quick C-g
|
process:command
|
From @vspinu in #1108.
> Actually not, on very quick C-g C-g the proc buffer is still not returned :(
|
1.0
|
Process restoration with quick C-g - From @vspinu in #1108.
> Actually not, on very quick C-g C-g the proc buffer is still not returned :(
|
process
|
process restoration with quick c g from vspinu in actually not on very quick c g c g the proc buffer is still not returned
| 1
|
14,683
| 17,798,386,681
|
IssuesEvent
|
2021-09-01 02:57:22
|
jim-king-2000/IndustryCamera
|
https://api.github.com/repos/jim-king-2000/IndustryCamera
|
closed
|
[bug]: 设备中心 UI
|
bug processing C
|
### 问题描述
- 列展示太多的情况下,避免标题换行增加滑动条
- 设备区域未显示完整目录
### 您预期的行为
- 不出现标题文字重叠
- 设备区域显示完整目录,例 风场A/风塔1/层1
### 系统表现的行为

### 复现路径
<!-- 如何重现bug -->
### 辅助信息
- 浏览器版本:Edge/Chrome 92
- 固件版本:v1.0
|
1.0
|
[bug]: 设备中心 UI - ### 问题描述
- 列展示太多的情况下,避免标题换行增加滑动条
- 设备区域未显示完整目录
### 您预期的行为
- 不出现标题文字重叠
- 设备区域显示完整目录,例 风场A/风塔1/层1
### 系统表现的行为

### 复现路径
<!-- 如何重现bug -->
### 辅助信息
- 浏览器版本:Edge/Chrome 92
- 固件版本:v1.0
|
process
|
设备中心 ui 问题描述 列展示太多的情况下,避免标题换行增加滑动条 设备区域未显示完整目录 您预期的行为 不出现标题文字重叠 设备区域显示完整目录,例 风场a 系统表现的行为 复现路径 辅助信息 浏览器版本:edge chrome 固件版本:
| 1
|
261,489
| 27,809,782,829
|
IssuesEvent
|
2023-03-18 01:43:09
|
madhans23/linux-4.1.15
|
https://api.github.com/repos/madhans23/linux-4.1.15
|
closed
|
CVE-2020-25668 (High) detected in linux-stable-rtv4.1.33 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2020-25668 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/vt_ioctl.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/vt_ioctl.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in Linux Kernel because access to the global variable fg_console is not properly synchronized leading to a use after free in con_font_op.
<p>Publish Date: 2021-05-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25668>CVE-2020-25668</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25668">https://www.linuxkernelcves.com/cves/CVE-2020-25668</a></p>
<p>Release Date: 2021-05-26</p>
<p>Fix Resolution: v4.14.204,v4.19.155,v5.4.75,v5.9.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-25668 (High) detected in linux-stable-rtv4.1.33 - autoclosed - ## CVE-2020-25668 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/madhans23/linux-4.1.15/commit/f9d19044b0eef1965f9bc412d7d9e579b74ec968">f9d19044b0eef1965f9bc412d7d9e579b74ec968</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/vt_ioctl.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/tty/vt/vt_ioctl.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in Linux Kernel because access to the global variable fg_console is not properly synchronized leading to a use after free in con_font_op.
<p>Publish Date: 2021-05-26
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-25668>CVE-2020-25668</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25668">https://www.linuxkernelcves.com/cves/CVE-2020-25668</a></p>
<p>Release Date: 2021-05-26</p>
<p>Fix Resolution: v4.14.204,v4.19.155,v5.4.75,v5.9.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in linux stable autoclosed cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers tty vt vt ioctl c drivers tty vt vt ioctl c vulnerability details a flaw was found in linux kernel because access to the global variable fg console is not properly synchronized leading to a use after free in con font op publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
562,782
| 16,669,747,613
|
IssuesEvent
|
2021-06-07 09:22:11
|
AyeCode/userswp
|
https://api.github.com/repos/AyeCode/userswp
|
opened
|
Doc request for new multiple registration forms functionality
|
Docs Priority: High Type: Needs Docs
|
Please create a doc with the multiple registration forms functionality with form settings and how it works for the account and profile page. Let me know where my help needed.
|
1.0
|
Doc request for new multiple registration forms functionality - Please create a doc with the multiple registration forms functionality with form settings and how it works for the account and profile page. Let me know where my help needed.
|
non_process
|
doc request for new multiple registration forms functionality please create a doc with the multiple registration forms functionality with form settings and how it works for the account and profile page let me know where my help needed
| 0
|
10,296
| 13,148,136,951
|
IssuesEvent
|
2020-08-08 19:37:46
|
unisonweb/unison
|
https://api.github.com/repos/unisonweb/unison
|
opened
|
pull-request.create is awkward when the output is bigger than a screen
|
pr-process
|
What are others doing when the `pull-request.create` output is longer than their screen size?
For me the output opens in `less` (I think?), and I don't know of a convenient way to copy the entirety of the text to paste in a GitHub issue.
I tried doing `echo 'pull-request.create git@github.com:unisonweb/base:.trunk git@github.com:ceedubs/unison-dev:.prs._base.nelfoldl' | ucm | pbcopy`, but this ends up capturing a lot of ANSI escape characters.
I could be missing something obvious, but it seemed worth capturing in an issue since this might trip up others.
|
1.0
|
pull-request.create is awkward when the output is bigger than a screen - What are others doing when the `pull-request.create` output is longer than their screen size?
For me the output opens in `less` (I think?), and I don't know of a convenient way to copy the entirety of the text to paste in a GitHub issue.
I tried doing `echo 'pull-request.create git@github.com:unisonweb/base:.trunk git@github.com:ceedubs/unison-dev:.prs._base.nelfoldl' | ucm | pbcopy`, but this ends up capturing a lot of ANSI escape characters.
I could be missing something obvious, but it seemed worth capturing in an issue since this might trip up others.
|
process
|
pull request create is awkward when the output is bigger than a screen what are others doing when the pull request create output is longer than their screen size for me the output opens in less i think and i don t know of a convenient way to copy the entirety of the text to paste in a github issue i tried doing echo pull request create git github com unisonweb base trunk git github com ceedubs unison dev prs base nelfoldl ucm pbcopy but this ends up capturing a lot of ansi escape characters i could be missing something obvious but it seemed worth capturing in an issue since this might trip up others
| 1
|
4,132
| 7,088,571,440
|
IssuesEvent
|
2018-01-11 22:02:06
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
An empty sequence is not allowed as the value of parameter $href [DOT 2.3.1]
|
bug preprocess
|
Using one of my client's samples I obtain an error when publishing to XHTML-based outputs, the error is something like this:
```
maplink:
[maplink] Error on line 418 of maplinkImpl.xsl:
[maplink] XPTY0004: An empty sequence is not allowed as the value of parameter $href
[maplink] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/maplinkImpl.xsl#410)
[maplink] processing /map/reltable[1]/relheader[1]/relcolspec[1]/topichead[1]
[maplink] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/maplinkImpl.xsl#335)
```
and it occurs at line:
```
<xsl:template match="*[contains(@class, ' map/topicref ')]" mode="grab-group-title" as="xs:string?">
<xsl:variable name="file-origin">
<xsl:call-template name="get-file-uri">
<xsl:with-param name="href" select="@href"/>
```
so somehow the topicref does not have an @href attribute set to it and the publishing breaks.
I added an xsl:message and it breaks when processing this topichead:
```
<topichead xmlns:ditaarch="http://dita.oasis-open.org/architecture/2005/" xmlns:dita-ot="http://dita-ot.sourceforge.net/ns/201007/dita-ot" class="+ map/topicref mapgroup-d/topichead " xtrc="topichead:1;28:28" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap" toc="no" print="no"><topicmeta class="- map/topicmeta " xtrc="topicmeta:2;29:32" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap"><navtitle class="- topic/navtitle " xtrc="navtitle:1;30:35" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap">See also</navtitle></topicmeta></topichead>
```
|
1.0
|
An empty sequence is not allowed as the value of parameter $href [DOT 2.3.1] - Using one of my client's samples I obtain an error when publishing to XHTML-based outputs, the error is something like this:
```
maplink:
[maplink] Error on line 418 of maplinkImpl.xsl:
[maplink] XPTY0004: An empty sequence is not allowed as the value of parameter $href
[maplink] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/maplinkImpl.xsl#410)
[maplink] processing /map/reltable[1]/relheader[1]/relcolspec[1]/topichead[1]
[maplink] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/maplinkImpl.xsl#335)
```
and it occurs at line:
```
<xsl:template match="*[contains(@class, ' map/topicref ')]" mode="grab-group-title" as="xs:string?">
<xsl:variable name="file-origin">
<xsl:call-template name="get-file-uri">
<xsl:with-param name="href" select="@href"/>
```
so somehow the topicref does not have an @href attribute set to it and the publishing breaks.
I added an xsl:message and it breaks when processing this topichead:
```
<topichead xmlns:ditaarch="http://dita.oasis-open.org/architecture/2005/" xmlns:dita-ot="http://dita-ot.sourceforge.net/ns/201007/dita-ot" class="+ map/topicref mapgroup-d/topichead " xtrc="topichead:1;28:28" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap" toc="no" print="no"><topicmeta class="- map/topicmeta " xtrc="topicmeta:2;29:32" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap"><navtitle class="- topic/navtitle " xtrc="navtitle:1;30:35" xtrf="file:/C:/Users/radu_coravu/Desktop/test/abbreviated/Producer_UG.ditamap">See also</navtitle></topicmeta></topichead>
```
|
process
|
an empty sequence is not allowed as the value of parameter href using one of my client s samples i obtain an error when publishing to xhtml based outputs the error is something like this maplink error on line of maplinkimpl xsl an empty sequence is not allowed as the value of parameter href at xsl apply templates file d projects exml frameworks dita dita x xsl preprocess maplinkimpl xsl processing map reltable relheader relcolspec topichead at xsl apply templates file d projects exml frameworks dita dita x xsl preprocess maplinkimpl xsl and it occurs at line so somehow the topicref does not have an href attribute set to it and the publishing breaks i added an xsl message and it breaks when processing this topichead see also
| 1
|
3,055
| 6,046,755,959
|
IssuesEvent
|
2017-06-12 12:58:32
|
itsyouonline/identityserver
|
https://api.github.com/repos/itsyouonline/identityserver
|
closed
|
Can create 2 accounts with same email and phone number
|
process_wontfix
|
- Please prevent creating more than one account with same information


|
1.0
|
Can create 2 accounts with same email and phone number - - Please prevent creating more than one account with same information


|
process
|
can create accounts with same email and phone number please prevent creating more than one account with same information
| 1
|
19,050
| 25,060,336,037
|
IssuesEvent
|
2022-11-07 00:20:32
|
ethereum/EIPs
|
https://api.github.com/repos/ethereum/EIPs
|
closed
|
EIP registries
|
w-stale enhancement r-process r-eips
|
Some EIPs require registries - lists of mappings relevant to the EIP. Examples include EIP 191 and EIP 634.
Currently, there's no good solution for this: changes to an EIP are prohibited after it's final, so the registry can't be kept inside the EIP itself, and external links are likewise prohibited, so the registry can't be kept elsewhere.
The IETF maintains a list of registries; each RFC can define one or more new named registries, and the registries are maintained separately from the (final) RFC. We should define a best practice for this with EIPs.
|
1.0
|
EIP registries - Some EIPs require registries - lists of mappings relevant to the EIP. Examples include EIP 191 and EIP 634.
Currently, there's no good solution for this: changes to an EIP are prohibited after it's final, so the registry can't be kept inside the EIP itself, and external links are likewise prohibited, so the registry can't be kept elsewhere.
The IETF maintains a list of registries; each RFC can define one or more new named registries, and the registries are maintained separately from the (final) RFC. We should define a best practice for this with EIPs.
|
process
|
eip registries some eips require registries lists of mappings relevant to the eip examples include eip and eip currently there s no good solution for this changes to an eip are prohibited after it s final so the registry can t be kept inside the eip itself and external links are likewise prohibited so the registry can t be kept elsewhere the ietf maintains a list of registries each rfc can define one or more new named registries and the registries are maintained separately from the final rfc we should define a best practice for this with eips
| 1
|
157,205
| 5,996,453,946
|
IssuesEvent
|
2017-06-03 14:29:41
|
universAAL/tools.eclipse-plugins
|
https://api.github.com/repos/universAAL/tools.eclipse-plugins
|
closed
|
when transforming .OWL files it requests .xml file
|
bug imported priority 4
|
_Originally Opened: @amedranogil (2013-01-11 10:30:15_)
_Originally Closed: 2013-01-18 11:12:47_
when I click on an owl file (i'm using this one: http://purl.obolibrary.org/obo/doid.owl ) the tool seems to do a first reading of the OWL and the second reading seems it is trying to read the .xml (non-existent).
Type... UML2
Reading...C:/Users/amedrano/workspaceOnt/ont.disease/Disease.owl
Read DONE C:/Users/amedrano/workspaceOnt/ont.disease/Disease.owl
Read base name.................http://purl.org/obo/owl
java.io.FileNotFoundException: C:\Users\amedrano\workspaceOnt\ont.disease\Disease.xml (El sistema no puede encontrar el archivo especificado)
at java.io.FileInputStream.open(Native Method)
at java.io.FileInputStream.<init>(Unknown Source)
at java.io.FileInputStream.<init>(Unknown Source)
at sun.net.www.protocol.file.FileURLConnection.connect(Unknown Source)
at sun.net.www.protocol.file.FileURLConnection.getInputStream(Unknown Source)
at org.apache.xerces.impl.XMLEntityManager.setupCurrentEntity(Unknown Source)
at org.apache.xerces.impl.XMLVersionDetector.determineDocVersion(Unknown Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
at javax.xml.parsers.DocumentBuilder.parse(Unknown Source)
at org.universaal.tools.owl2uml.uml2.UML2Factory.readXML(UML2Factory.java:781)
at org.universaal.tools.owl2uml.uml2.UML2Factory.<init>(UML2Factory.java:91)
at org.universaal.tools.owl2uml.core.UML2Parser.loadOntology(UML2Parser.java:113)
at org.universaal.tools.owl2uml.OWL2UML.main(OWL2UML.java:39)
at org.universaal.tools.owl2uml.handlers.SampleHandler.execute(SampleHandler.java:79)
at org.eclipse.ui.internal.handlers.HandlerProxy.execute(HandlerProxy.java:293)
at org.eclipse.core.commands.Command.executeWithChecks(Command.java:476)
at org.eclipse.core.commands.ParameterizedCommand.executeWithChecks(ParameterizedCommand.java:508)
at org.eclipse.ui.internal.handlers.HandlerService.executeCommand(HandlerService.java:169)
at org.eclipse.ui.internal.handlers.SlaveHandlerService.executeCommand(SlaveHandlerService.java:241)
at org.eclipse.ui.internal.handlers.SlaveHandlerService.executeCommand(SlaveHandlerService.java:241)
at org.eclipse.ui.menus.CommandContributionItem.handleWidgetSelection(CommandContributionItem.java:829)
at org.eclipse.ui.menus.CommandContributionItem.access$19(CommandContributionItem.java:815)
at org.eclipse.ui.menus.CommandContributionItem$5.handleEvent(CommandContributionItem.java:805)
at org.eclipse.swt.widgets.EventTable.sendEvent(EventTable.java:84)
at org.eclipse.swt.widgets.Widget.sendEvent(Widget.java:1053)
at org.eclipse.swt.widgets.Display.runDeferredEvents(Display.java:4165)
at org.eclipse.swt.widgets.Display.readAndDispatch(Display.java:3754)
at org.eclipse.ui.internal.Workbench.runEventLoop(Workbench.java:2701)
at org.eclipse.ui.internal.Workbench.runUI(Workbench.java:2665)
at org.eclipse.ui.internal.Workbench.access$4(Workbench.java:2499)
at org.eclipse.ui.internal.Workbench$7.run(Workbench.java:679)
at org.eclipse.core.databinding.observable.Realm.runWithDefault(Realm.java:332)
at org.eclipse.ui.internal.Workbench.createAndRunWorkbench(Workbench.java:668)
at org.eclipse.ui.PlatformUI.createAndRunWorkbench(PlatformUI.java:149)
at org.eclipse.ui.internal.ide.application.IDEApplication.start(IDEApplication.java:123)
at org.eclipse.equinox.internal.app.EclipseAppHandle.run(EclipseAppHandle.java:196)
at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.runApplication(EclipseAppLauncher.java:110)
at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.start(EclipseAppLauncher.java:79)
at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:344)
at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:179)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.eclipse.equinox.launcher.Main.invokeFramework(Main.java:622)
at org.eclipse.equinox.launcher.Main.basicRun(Main.java:577)
at org.eclipse.equinox.launcher.Main.run(Main.java:1410)
--
From: _this issue has been automatically imported from our old issue tracker_
|
1.0
|
when transforming .OWL files it requests .xml file - _Originally Opened: @amedranogil (2013-01-11 10:30:15_)
_Originally Closed: 2013-01-18 11:12:47_
when I click on an owl file (i'm using this one: http://purl.obolibrary.org/obo/doid.owl ) the tool seems to do a first reading of the OWL and the second reading seems it is trying to read the .xml (non-existent).
Type... UML2
Reading...C:/Users/amedrano/workspaceOnt/ont.disease/Disease.owl
Read DONE C:/Users/amedrano/workspaceOnt/ont.disease/Disease.owl
Read base name.................http://purl.org/obo/owl
java.io.FileNotFoundException: C:\Users\amedrano\workspaceOnt\ont.disease\Disease.xml (El sistema no puede encontrar el archivo especificado)
at java.io.FileInputStream.open(Native Method)
at java.io.FileInputStream.<init>(Unknown Source)
at java.io.FileInputStream.<init>(Unknown Source)
at sun.net.www.protocol.file.FileURLConnection.connect(Unknown Source)
at sun.net.www.protocol.file.FileURLConnection.getInputStream(Unknown Source)
at org.apache.xerces.impl.XMLEntityManager.setupCurrentEntity(Unknown Source)
at org.apache.xerces.impl.XMLVersionDetector.determineDocVersion(Unknown Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XML11Configuration.parse(Unknown Source)
at org.apache.xerces.parsers.XMLParser.parse(Unknown Source)
at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
at javax.xml.parsers.DocumentBuilder.parse(Unknown Source)
at org.universaal.tools.owl2uml.uml2.UML2Factory.readXML(UML2Factory.java:781)
at org.universaal.tools.owl2uml.uml2.UML2Factory.<init>(UML2Factory.java:91)
at org.universaal.tools.owl2uml.core.UML2Parser.loadOntology(UML2Parser.java:113)
at org.universaal.tools.owl2uml.OWL2UML.main(OWL2UML.java:39)
at org.universaal.tools.owl2uml.handlers.SampleHandler.execute(SampleHandler.java:79)
at org.eclipse.ui.internal.handlers.HandlerProxy.execute(HandlerProxy.java:293)
at org.eclipse.core.commands.Command.executeWithChecks(Command.java:476)
at org.eclipse.core.commands.ParameterizedCommand.executeWithChecks(ParameterizedCommand.java:508)
at org.eclipse.ui.internal.handlers.HandlerService.executeCommand(HandlerService.java:169)
at org.eclipse.ui.internal.handlers.SlaveHandlerService.executeCommand(SlaveHandlerService.java:241)
at org.eclipse.ui.internal.handlers.SlaveHandlerService.executeCommand(SlaveHandlerService.java:241)
at org.eclipse.ui.menus.CommandContributionItem.handleWidgetSelection(CommandContributionItem.java:829)
at org.eclipse.ui.menus.CommandContributionItem.access$19(CommandContributionItem.java:815)
at org.eclipse.ui.menus.CommandContributionItem$5.handleEvent(CommandContributionItem.java:805)
at org.eclipse.swt.widgets.EventTable.sendEvent(EventTable.java:84)
at org.eclipse.swt.widgets.Widget.sendEvent(Widget.java:1053)
at org.eclipse.swt.widgets.Display.runDeferredEvents(Display.java:4165)
at org.eclipse.swt.widgets.Display.readAndDispatch(Display.java:3754)
at org.eclipse.ui.internal.Workbench.runEventLoop(Workbench.java:2701)
at org.eclipse.ui.internal.Workbench.runUI(Workbench.java:2665)
at org.eclipse.ui.internal.Workbench.access$4(Workbench.java:2499)
at org.eclipse.ui.internal.Workbench$7.run(Workbench.java:679)
at org.eclipse.core.databinding.observable.Realm.runWithDefault(Realm.java:332)
at org.eclipse.ui.internal.Workbench.createAndRunWorkbench(Workbench.java:668)
at org.eclipse.ui.PlatformUI.createAndRunWorkbench(PlatformUI.java:149)
at org.eclipse.ui.internal.ide.application.IDEApplication.start(IDEApplication.java:123)
at org.eclipse.equinox.internal.app.EclipseAppHandle.run(EclipseAppHandle.java:196)
at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.runApplication(EclipseAppLauncher.java:110)
at org.eclipse.core.runtime.internal.adaptor.EclipseAppLauncher.start(EclipseAppLauncher.java:79)
at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:344)
at org.eclipse.core.runtime.adaptor.EclipseStarter.run(EclipseStarter.java:179)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at org.eclipse.equinox.launcher.Main.invokeFramework(Main.java:622)
at org.eclipse.equinox.launcher.Main.basicRun(Main.java:577)
at org.eclipse.equinox.launcher.Main.run(Main.java:1410)
--
From: _this issue has been automatically imported from our old issue tracker_
|
non_process
|
when transforming owl files it requests xml file originally opened amedranogil originally closed when i click on an owl file i m using this one the tool seems to do a first reading of the owl and the second reading seems it is trying to read the xml non existent type reading c users amedrano workspaceont ont disease disease owl read done c users amedrano workspaceont ont disease disease owl read base name java io filenotfoundexception c users amedrano workspaceont ont disease disease xml el sistema no puede encontrar el archivo especificado at java io fileinputstream open native method at java io fileinputstream lt init gt unknown source at java io fileinputstream lt init gt unknown source at sun net source at sun net source at org apache xerces impl xmlentitymanager setupcurrententity unknown source at org apache xerces impl xmlversiondetector determinedocversion unknown source at org apache xerces parsers parse unknown source at org apache xerces parsers parse unknown source at org apache xerces parsers xmlparser parse unknown source at org apache xerces parsers domparser parse unknown source at org apache xerces jaxp documentbuilderimpl parse unknown source at javax xml parsers documentbuilder parse unknown source at org universaal tools readxml java at org universaal tools lt init gt java at org universaal tools core loadontology java at org universaal tools main java at org universaal tools handlers samplehandler execute samplehandler java at org eclipse ui internal handlers handlerproxy execute handlerproxy java at org eclipse core commands command executewithchecks command java at org eclipse core commands parameterizedcommand executewithchecks parameterizedcommand java at org eclipse ui internal handlers handlerservice executecommand handlerservice java at org eclipse ui internal handlers slavehandlerservice executecommand slavehandlerservice java at org eclipse ui internal handlers slavehandlerservice executecommand slavehandlerservice java at org eclipse ui menus commandcontributionitem handlewidgetselection commandcontributionitem java at org eclipse ui menus commandcontributionitem access commandcontributionitem java at org eclipse ui menus commandcontributionitem handleevent commandcontributionitem java at org eclipse swt widgets eventtable sendevent eventtable java at org eclipse swt widgets widget sendevent widget java at org eclipse swt widgets display rundeferredevents display java at org eclipse swt widgets display readanddispatch display java at org eclipse ui internal workbench runeventloop workbench java at org eclipse ui internal workbench runui workbench java at org eclipse ui internal workbench access workbench java at org eclipse ui internal workbench run workbench java at org eclipse core databinding observable realm runwithdefault realm java at org eclipse ui internal workbench createandrunworkbench workbench java at org eclipse ui platformui createandrunworkbench platformui java at org eclipse ui internal ide application ideapplication start ideapplication java at org eclipse equinox internal app eclipseapphandle run eclipseapphandle java at org eclipse core runtime internal adaptor eclipseapplauncher runapplication eclipseapplauncher java at org eclipse core runtime internal adaptor eclipseapplauncher start eclipseapplauncher java at org eclipse core runtime adaptor eclipsestarter run eclipsestarter java at org eclipse core runtime adaptor eclipsestarter run eclipsestarter java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke unknown source at java lang reflect method invoke unknown source at org eclipse equinox launcher main invokeframework main java at org eclipse equinox launcher main basicrun main java at org eclipse equinox launcher main run main java from this issue has been automatically imported from our old issue tracker
| 0
|
18,187
| 24,236,165,089
|
IssuesEvent
|
2022-09-26 23:36:58
|
opensearch-project/data-prepper
|
https://api.github.com/repos/opensearch-project/data-prepper
|
closed
|
Support parsing JSON
|
enhancement plugin - processor
|
**Is your feature request related to a problem? Please describe.**
Data Prepper events may have JSON values inside Event fields. Data Prepper should be able to parse these JSON strings and create fields directly in the Event from the JSON.
**Describe the solution you'd like**
Provide a JSON parsing processor - `parse_json`.
It should be able to parse a JSON string from a field and set the values in the Event object. This processor will automatically support nesting.
**Example**
Given the following configuration:
```
processor:
parse_json:
source: my_field
```
Given this input event:
```
"my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}"
```
The input event is changed to:
```
"my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}"
"key1" : "value1"
"key2" : "value2"
```
**Example with Nesting**
Given this input event:
```
"my_field" : "{\"key1\" : \"value1\", \"key2\" : { \"key2child\" : \"innerValue\" }}"
```
The input event is changed to:
```
"my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}"
"key1" : "value1"
"key2" : {
"key2child" : "innerValue"
}
```
**Configurations**
`source` - the field with JSON
`target` - the field to set the values in; by default this is the root object
|
1.0
|
Support parsing JSON - **Is your feature request related to a problem? Please describe.**
Data Prepper events may have JSON values inside Event fields. Data Prepper should be able to parse these JSON strings and create fields directly in the Event from the JSON.
**Describe the solution you'd like**
Provide a JSON parsing processor - `parse_json`.
It should be able to parse a JSON string from a field and set the values in the Event object. This processor will automatically support nesting.
**Example**
Given the following configuration:
```
processor:
parse_json:
source: my_field
```
Given this input event:
```
"my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}"
```
The input event is changed to:
```
"my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}"
"key1" : "value1"
"key2" : "value2"
```
**Example with Nesting**
Given this input event:
```
"my_field" : "{\"key1\" : \"value1\", \"key2\" : { \"key2child\" : \"innerValue\" }}"
```
The input event is changed to:
```
"my_field" : "{\"key1\" : \"value1\", \"key2\" : \"value2\"}"
"key1" : "value1"
"key2" : {
"key2child" : "innerValue"
}
```
**Configurations**
`source` - the field with JSON
`target` - the field to set the values in; by default this is the root object
|
process
|
support parsing json is your feature request related to a problem please describe data prepper events may have json values inside event fields data prepper should be able to parse these json strings and create fields directly in the event from the json describe the solution you d like provide a json parsing processor parse json it should be able to parse a json string from a field and set the values in the event object this processor will automatically support nesting example given the following configuration processor parse json source my field given this input event my field the input event is changed to my field example with nesting given this input event my field innervalue the input event is changed to my field innervalue configurations source the field with json target the field to set the values in by default this is the root object
| 1
|
134,882
| 18,513,051,764
|
IssuesEvent
|
2021-10-20 06:56:12
|
mgh3326/createandread
|
https://api.github.com/repos/mgh3326/createandread
|
opened
|
CVE-2021-37713 (High) detected in tar-4.4.13.tgz
|
security vulnerability
|
## CVE-2021-37713 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.13.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p>
<p>Path to dependency file: createandread/package.json</p>
<p>Path to vulnerable library: createandread/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.9.0.tgz (Root Library)
- chokidar-2.1.8.tgz
- fsevents-1.2.9.tgz
- node-pre-gyp-0.12.0.tgz
- :x: **tar-4.4.13.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mgh3326/createandread/commit/da66d826bfaeb3ccfb4d58ebf3cd7d5c1af76a48">da66d826bfaeb3ccfb4d58ebf3cd7d5c1af76a48</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted. This is, in part, accomplished by sanitizing absolute paths of entries within the archive, skipping archive entries that contain `..` path portions, and resolving the sanitized paths against the extraction target directory. This logic was insufficient on Windows systems when extracting tar files that contained a path that was not an absolute path, but specified a drive letter different from the extraction target, such as `C:some\path`. If the drive letter does not match the extraction target, for example `D:\extraction\dir`, then the result of `path.resolve(extractionDirectory, entryPath)` would resolve against the current working directory on the `C:` drive, rather than the extraction target directory. Additionally, a `..` portion of the path could occur immediately after the drive letter, such as `C:../foo`, and was not properly sanitized by the logic that checked for `..` within the normalized and split portions of the path. This only affects users of `node-tar` on Windows systems. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. There is no reasonable way to work around this issue without performing the same path normalization procedures that node-tar now does. Users are encouraged to upgrade to the latest patched versions of node-tar, rather than attempt to sanitize paths themselves.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37713>CVE-2021-37713</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh">https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution: tar - 4.4.18, 5.0.10, 6.1.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-37713 (High) detected in tar-4.4.13.tgz - ## CVE-2021-37713 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.13.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.13.tgz">https://registry.npmjs.org/tar/-/tar-4.4.13.tgz</a></p>
<p>Path to dependency file: createandread/package.json</p>
<p>Path to vulnerable library: createandread/node_modules/tar/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.9.0.tgz (Root Library)
- chokidar-2.1.8.tgz
- fsevents-1.2.9.tgz
- node-pre-gyp-0.12.0.tgz
- :x: **tar-4.4.13.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mgh3326/createandread/commit/da66d826bfaeb3ccfb4d58ebf3cd7d5c1af76a48">da66d826bfaeb3ccfb4d58ebf3cd7d5c1af76a48</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted. This is, in part, accomplished by sanitizing absolute paths of entries within the archive, skipping archive entries that contain `..` path portions, and resolving the sanitized paths against the extraction target directory. This logic was insufficient on Windows systems when extracting tar files that contained a path that was not an absolute path, but specified a drive letter different from the extraction target, such as `C:some\path`. If the drive letter does not match the extraction target, for example `D:\extraction\dir`, then the result of `path.resolve(extractionDirectory, entryPath)` would resolve against the current working directory on the `C:` drive, rather than the extraction target directory. Additionally, a `..` portion of the path could occur immediately after the drive letter, such as `C:../foo`, and was not properly sanitized by the logic that checked for `..` within the normalized and split portions of the path. This only affects users of `node-tar` on Windows systems. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. There is no reasonable way to work around this issue without performing the same path normalization procedures that node-tar now does. Users are encouraged to upgrade to the latest patched versions of node-tar, rather than attempt to sanitize paths themselves.
<p>Publish Date: 2021-08-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37713>CVE-2021-37713</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh">https://github.com/npm/node-tar/security/advisories/GHSA-5955-9wpr-37jh</a></p>
<p>Release Date: 2021-08-31</p>
<p>Fix Resolution: tar - 4.4.18, 5.0.10, 6.1.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file createandread package json path to vulnerable library createandread node modules tar package json dependency hierarchy webpack dev server tgz root library chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in head commit a href vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be outside of the extraction target directory is not extracted this is in part accomplished by sanitizing absolute paths of entries within the archive skipping archive entries that contain path portions and resolving the sanitized paths against the extraction target directory this logic was insufficient on windows systems when extracting tar files that contained a path that was not an absolute path but specified a drive letter different from the extraction target such as c some path if the drive letter does not match the extraction target for example d extraction dir then the result of path resolve extractiondirectory entrypath would resolve against the current working directory on the c drive rather than the extraction target directory additionally a portion of the path could occur immediately after the drive letter such as c foo and was not properly sanitized by the logic that checked for within the normalized and split portions of the path this only affects users of node tar on windows systems these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar there is no reasonable way to work around this issue without performing the same path normalization procedures that node tar now does users are encouraged to upgrade to the latest patched versions of node tar rather than attempt to sanitize paths themselves publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource
| 0
|
13,584
| 16,135,484,256
|
IssuesEvent
|
2021-04-29 11:15:39
|
sanzcarlos/CiscoAXL
|
https://api.github.com/repos/sanzcarlos/CiscoAXL
|
closed
|
Add new field in CSV
|
enhancement process
|
**Is your feature request related to a problem? Please describe.**
We need add the main DDI of site and create a Translation Pattern to Hunt Pilot
**Describe the solution you'd like**
N/A
**Additional context**
N/A
|
1.0
|
Add new field in CSV - **Is your feature request related to a problem? Please describe.**
We need add the main DDI of site and create a Translation Pattern to Hunt Pilot
**Describe the solution you'd like**
N/A
**Additional context**
N/A
|
process
|
add new field in csv is your feature request related to a problem please describe we need add the main ddi of site and create a translation pattern to hunt pilot describe the solution you d like n a additional context n a
| 1
|
15,881
| 20,070,639,759
|
IssuesEvent
|
2022-02-04 06:06:50
|
codeanit/til
|
https://api.github.com/repos/codeanit/til
|
opened
|
Vanguard Method - Understanding variation in demand
|
wip process
|
The Vanguard Method is one framework for understanding variation in demand. This takes a data driven approach of understanding what the customer wanted to achieve.
Here are the steps:
1. Define the purpose of the system from customer perspective.
2. Study the nature of demand coming towards the system
3. Learn how the system responds to the demands.
4. Understand why this happens.
5. Identify what policies or measures cause problems in the flow of work.
6. Recognize the thinking behind the design and management of the system
|
1.0
|
Vanguard Method - Understanding variation in demand - The Vanguard Method is one framework for understanding variation in demand. This takes a data driven approach of understanding what the customer wanted to achieve.
Here are the steps:
1. Define the purpose of the system from customer perspective.
2. Study the nature of demand coming towards the system
3. Learn how the system responds to the demands.
4. Understand why this happens.
5. Identify what policies or measures cause problems in the flow of work.
6. Recognize the thinking behind the design and management of the system
|
process
|
vanguard method understanding variation in demand the vanguard method is one framework for understanding variation in demand this takes a data driven approach of understanding what the customer wanted to achieve here are the steps define the purpose of the system from customer perspective study the nature of demand coming towards the system learn how the system responds to the demands understand why this happens identify what policies or measures cause problems in the flow of work recognize the thinking behind the design and management of the system
| 1
|
639,809
| 20,766,675,465
|
IssuesEvent
|
2022-03-15 21:23:23
|
SIMPLE-AstroDB/SIMPLE-db
|
https://api.github.com/repos/SIMPLE-AstroDB/SIMPLE-db
|
closed
|
Add tests for Gaia and 2MASS data
|
enhancement low priority
|
I'm not quite sure how to write these, but they would be useful to have:
- If 2MASS designation in Names, 2MASS photometry should exist
- If 2MASS photometry, 2MASS designation should be in Names
- If Gaia designation in Names, Gaia phot and astrometry should exist
- If Gaia phot, Gaia designation should be in Names
- If Gaia pm, Gaia designation should be in Names
- If Gaia parallax, Gaia designation should be in Names
|
1.0
|
Add tests for Gaia and 2MASS data - I'm not quite sure how to write these, but they would be useful to have:
- If 2MASS designation in Names, 2MASS photometry should exist
- If 2MASS photometry, 2MASS designation should be in Names
- If Gaia designation in Names, Gaia phot and astrometry should exist
- If Gaia phot, Gaia designation should be in Names
- If Gaia pm, Gaia designation should be in Names
- If Gaia parallax, Gaia designation should be in Names
|
non_process
|
add tests for gaia and data i m not quite sure how to write these but they would be useful to have if designation in names photometry should exist if photometry designation should be in names if gaia designation in names gaia phot and astrometry should exist if gaia phot gaia designation should be in names if gaia pm gaia designation should be in names if gaia parallax gaia designation should be in names
| 0
|
4,451
| 7,319,320,758
|
IssuesEvent
|
2018-03-02 00:11:45
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Outdated documentation
|
active-directory cxp in-process triaged
|
There is no such option as Single Sign On in azure ad now. Atleast not in my directory.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f97b43d9-4a25-1552-c719-f0f0b418b0a8
* Version Independent ID: 13445782-bdb8-b9fa-b130-630bb770187c
* [Content](https://docs.microsoft.com/en-us/azure/active-directory/active-directory-enterprise-apps-manage-sso)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/active-directory-enterprise-apps-manage-sso.md)
* Service: active-directory
|
1.0
|
Outdated documentation - There is no such option as Single Sign On in azure ad now. Atleast not in my directory.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f97b43d9-4a25-1552-c719-f0f0b418b0a8
* Version Independent ID: 13445782-bdb8-b9fa-b130-630bb770187c
* [Content](https://docs.microsoft.com/en-us/azure/active-directory/active-directory-enterprise-apps-manage-sso)
* [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/active-directory/active-directory-enterprise-apps-manage-sso.md)
* Service: active-directory
|
process
|
outdated documentation there is no such option as single sign on in azure ad now atleast not in my directory document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id service active directory
| 1
|
211,980
| 23,856,895,146
|
IssuesEvent
|
2022-09-07 01:14:24
|
CartoonFan/node-jscs
|
https://api.github.com/repos/CartoonFan/node-jscs
|
opened
|
WS-2021-0638 (High) detected in mocha-2.5.3.tgz
|
security vulnerability
|
## WS-2021-0638 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mocha-2.5.3.tgz</b></p></summary>
<p>simple, flexible, fun test framework</p>
<p>Library home page: <a href="https://registry.npmjs.org/mocha/-/mocha-2.5.3.tgz">https://registry.npmjs.org/mocha/-/mocha-2.5.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/mocha/package.json</p>
<p>
Dependency Hierarchy:
- :x: **mocha-2.5.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/CartoonFan/node-jscs/commit/d7995bed5d8b00cdb4811d5cb4631eb63ac6f70f">d7995bed5d8b00cdb4811d5cb4631eb63ac6f70f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There is regular Expression Denial of Service (ReDoS) vulnerability in mocha.
It allows cause a denial of service when stripping crafted invalid function definition from strs.
<p>Publish Date: 2021-09-18
<p>URL: <a href=https://github.com/mochajs/mocha/commit/61b4b9209c2c64b32c8d48b1761c3b9384d411ea>WS-2021-0638</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/1d8a3d95-d199-4129-a6ad-8eafe5e77b9e/">https://huntr.dev/bounties/1d8a3d95-d199-4129-a6ad-8eafe5e77b9e/</a></p>
<p>Release Date: 2021-09-18</p>
<p>Fix Resolution: 3.0.0-0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2021-0638 (High) detected in mocha-2.5.3.tgz - ## WS-2021-0638 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mocha-2.5.3.tgz</b></p></summary>
<p>simple, flexible, fun test framework</p>
<p>Library home page: <a href="https://registry.npmjs.org/mocha/-/mocha-2.5.3.tgz">https://registry.npmjs.org/mocha/-/mocha-2.5.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/mocha/package.json</p>
<p>
Dependency Hierarchy:
- :x: **mocha-2.5.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/CartoonFan/node-jscs/commit/d7995bed5d8b00cdb4811d5cb4631eb63ac6f70f">d7995bed5d8b00cdb4811d5cb4631eb63ac6f70f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There is regular Expression Denial of Service (ReDoS) vulnerability in mocha.
It allows cause a denial of service when stripping crafted invalid function definition from strs.
<p>Publish Date: 2021-09-18
<p>URL: <a href=https://github.com/mochajs/mocha/commit/61b4b9209c2c64b32c8d48b1761c3b9384d411ea>WS-2021-0638</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/1d8a3d95-d199-4129-a6ad-8eafe5e77b9e/">https://huntr.dev/bounties/1d8a3d95-d199-4129-a6ad-8eafe5e77b9e/</a></p>
<p>Release Date: 2021-09-18</p>
<p>Fix Resolution: 3.0.0-0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
ws high detected in mocha tgz ws high severity vulnerability vulnerable library mocha tgz simple flexible fun test framework library home page a href path to dependency file package json path to vulnerable library node modules mocha package json dependency hierarchy x mocha tgz vulnerable library found in head commit a href found in base branch master vulnerability details there is regular expression denial of service redos vulnerability in mocha it allows cause a denial of service when stripping crafted invalid function definition from strs publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
19,582
| 25,906,137,007
|
IssuesEvent
|
2022-12-15 10:21:17
|
CSE201-project/PaperFriend-desktop-app
|
https://api.github.com/repos/CSE201-project/PaperFriend-desktop-app
|
closed
|
Work on EntryRecap : Best/Worst day (Entry), Text (String), Average mood (Double)
|
file processing frontend analysis
|
Class used to show the recap over week/month/year to the user
Also add another parameter (int type) that will tell the frontend team if it's week/month/year to show it in different colors.
|
1.0
|
Work on EntryRecap : Best/Worst day (Entry), Text (String), Average mood (Double) - Class used to show the recap over week/month/year to the user
Also add another parameter (int type) that will tell the frontend team if it's week/month/year to show it in different colors.
|
process
|
work on entryrecap best worst day entry text string average mood double class used to show the recap over week month year to the user also add another parameter int type that will tell the frontend team if it s week month year to show it in different colors
| 1
|
509,177
| 14,723,296,470
|
IssuesEvent
|
2021-01-06 00:09:23
|
ArctosDB/arctos
|
https://api.github.com/repos/ArctosDB/arctos
|
opened
|
"Creator" button needed in Tools-Customize to display at search results
|
Bug Collection Type - Cultural Collections Function-SearchOrDownload Priority-Critical
|
Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html
**Describe the bug**
Maker --> Creator resulted in customized search results showing no data for Maker. There is no creator field to add in the customize box. Please change maker to creator in the customize box so we can add this field to search results and see who made our stuff!
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://arctos.database.museum/saved/Firearms
2.
<img width="1196" alt="Screen Shot 2021-01-05 at 2 47 16 PM" src="https://user-images.githubusercontent.com/17605945/103712898-6b93b480-4f67-11eb-9736-157f792ae3b8.png">
3. Open record and see there is a creator listed:
<img width="1173" alt="Screen Shot 2021-01-05 at 2 51 33 PM" src="https://user-images.githubusercontent.com/17605945/103712925-7bab9400-4f67-11eb-89ba-859ddcfafcd8.png">
4. See that there is no creator in customize tool box:
<img width="726" alt="Screen Shot 2021-01-05 at 2 47 02 PM" src="https://user-images.githubusercontent.com/17605945/103712958-8f56fa80-4f67-11eb-8690-49b1aba72cc7.png">
**Expected behavior**
The information that previously displayed in Maker column should be accessible via Creator, which needs to be added to the customize box.
**Priority**
Super high priority! Makes us look like we don't know our stuff!
|
1.0
|
"Creator" button needed in Tools-Customize to display at search results - Issue Documentation is http://handbook.arctosdb.org/how_to/How-to-Use-Issues-in-Arctos.html
**Describe the bug**
Maker --> Creator resulted in customized search results showing no data for Maker. There is no creator field to add in the customize box. Please change maker to creator in the customize box so we can add this field to search results and see who made our stuff!
**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://arctos.database.museum/saved/Firearms
2.
<img width="1196" alt="Screen Shot 2021-01-05 at 2 47 16 PM" src="https://user-images.githubusercontent.com/17605945/103712898-6b93b480-4f67-11eb-9736-157f792ae3b8.png">
3. Open record and see there is a creator listed:
<img width="1173" alt="Screen Shot 2021-01-05 at 2 51 33 PM" src="https://user-images.githubusercontent.com/17605945/103712925-7bab9400-4f67-11eb-89ba-859ddcfafcd8.png">
4. See that there is no creator in customize tool box:
<img width="726" alt="Screen Shot 2021-01-05 at 2 47 02 PM" src="https://user-images.githubusercontent.com/17605945/103712958-8f56fa80-4f67-11eb-8690-49b1aba72cc7.png">
**Expected behavior**
The information that previously displayed in Maker column should be accessible via Creator, which needs to be added to the customize box.
**Priority**
Super high priority! Makes us look like we don't know our stuff!
|
non_process
|
creator button needed in tools customize to display at search results issue documentation is describe the bug maker creator resulted in customized search results showing no data for maker there is no creator field to add in the customize box please change maker to creator in the customize box so we can add this field to search results and see who made our stuff to reproduce steps to reproduce the behavior go to img width alt screen shot at pm src open record and see there is a creator listed img width alt screen shot at pm src see that there is no creator in customize tool box img width alt screen shot at pm src expected behavior the information that previously displayed in maker column should be accessible via creator which needs to be added to the customize box priority super high priority makes us look like we don t know our stuff
| 0
|
8,879
| 11,980,553,695
|
IssuesEvent
|
2020-04-07 09:32:08
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
Investigate origin of `AWS Lambda timed out` (#1754)
|
bug/2-confirmed kind/bug process/candidate topic: deployment-platforms
|
I found a solution to #1754 [reproduction is here](https://github.com/UWellFail/prisma2--1754)
https://github.com/prisma/prisma2/issues/1754#issuecomment-601271228
```
// Set to false to send the response right away when the callback executes, instead of waiting for the Node.js event loop to be empty.
context.callbackWaitsForEmptyEventLoop = false;
```
But this means that the Node.js event loop is not empty and it should be empty.
### Without `context.callbackWaitsForEmptyEventLoop = false`
A POST request works perfectly. Only the GET requests in this example have a non empty event loop.
We could start investigating with:
- Try out pure Prisma Client in lambda without the apollo server overhead, get the GET and POST event running
- Instead of the callback based approach, try using a promise and play with async: true as described here https://github.com/serverless/serverless/issues/3171#issuecomment-540101244
- Try to get serverless offline running with a simple prisma client example https://serverless.com/plugins/serverless-offline/
|
1.0
|
Investigate origin of `AWS Lambda timed out` (#1754) - I found a solution to #1754 [reproduction is here](https://github.com/UWellFail/prisma2--1754)
https://github.com/prisma/prisma2/issues/1754#issuecomment-601271228
```
// Set to false to send the response right away when the callback executes, instead of waiting for the Node.js event loop to be empty.
context.callbackWaitsForEmptyEventLoop = false;
```
But this means that the Node.js event loop is not empty and it should be empty.
### Without `context.callbackWaitsForEmptyEventLoop = false`
A POST request works perfectly. Only the GET requests in this example have a non empty event loop.
We could start investigating with:
- Try out pure Prisma Client in lambda without the apollo server overhead, get the GET and POST event running
- Instead of the callback based approach, try using a promise and play with async: true as described here https://github.com/serverless/serverless/issues/3171#issuecomment-540101244
- Try to get serverless offline running with a simple prisma client example https://serverless.com/plugins/serverless-offline/
|
process
|
investigate origin of aws lambda timed out i found a solution to set to false to send the response right away when the callback executes instead of waiting for the node js event loop to be empty context callbackwaitsforemptyeventloop false but this means that the node js event loop is not empty and it should be empty without context callbackwaitsforemptyeventloop false a post request works perfectly only the get requests in this example have a non empty event loop we could start investigating with try out pure prisma client in lambda without the apollo server overhead get the get and post event running instead of the callback based approach try using a promise and play with async true as described here try to get serverless offline running with a simple prisma client example
| 1
|
1,510
| 4,103,283,526
|
IssuesEvent
|
2016-06-04 15:32:37
|
ProgrammingLife2016/PL2-2016
|
https://api.github.com/repos/ProgrammingLife2016/PL2-2016
|
opened
|
Edge drawing optimization
|
data processing development enhancement gui optimization
|
Investigate if it is possible to draw the edges more efficiently, or otherwise draw fewer edges.
The application zooms/scrolls very fast when the edges are not drawn. When the edges are drawn this is not always the case (especially for larger datasets), because there can be 100's of thousands of edges at the same time and currently a separate JavaFX Line object is created for each edge. This can possibly be done more efficient if something like a Canvas is used, where it is not necessary to create a new object for every drawn line.
|
1.0
|
Edge drawing optimization - Investigate if it is possible to draw the edges more efficiently, or otherwise draw fewer edges.
The application zooms/scrolls very fast when the edges are not drawn. When the edges are drawn this is not always the case (especially for larger datasets), because there can be 100's of thousands of edges at the same time and currently a separate JavaFX Line object is created for each edge. This can possibly be done more efficient if something like a Canvas is used, where it is not necessary to create a new object for every drawn line.
|
process
|
edge drawing optimization investigate if it is possible to draw the edges more efficiently or otherwise draw fewer edges the application zooms scrolls very fast when the edges are not drawn when the edges are drawn this is not always the case especially for larger datasets because there can be s of thousands of edges at the same time and currently a separate javafx line object is created for each edge this can possibly be done more efficient if something like a canvas is used where it is not necessary to create a new object for every drawn line
| 1
|
550,977
| 16,135,506,786
|
IssuesEvent
|
2021-04-29 11:17:28
|
bounswe/2021SpringGroup9
|
https://api.github.com/repos/bounswe/2021SpringGroup9
|
opened
|
Research about UML Use Case, Class and Sequence Diagrams
|
priority: high research status: in progress
|
I start doing research about UML diagrams and in particular how to use the arrows in specific cases such as inheritance and so on.
|
1.0
|
Research about UML Use Case, Class and Sequence Diagrams - I start doing research about UML diagrams and in particular how to use the arrows in specific cases such as inheritance and so on.
|
non_process
|
research about uml use case class and sequence diagrams i start doing research about uml diagrams and in particular how to use the arrows in specific cases such as inheritance and so on
| 0
|
17,903
| 23,877,390,484
|
IssuesEvent
|
2022-09-07 20:30:16
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
Requested files vs new visitors a day, discrepency in hits and visitors
|
question log-processing
|
Hi, I am unsure if it is a bug, but when I see the panel of unquie visitors a day including spiders it shows around 1400 hits and 220 visitors. But if I look at requested URLS, it shows what I believe to be more accurate of almost the same number of visitors and hits.

So not sure since the domain ends in .to if there is a regex issue?
The command I use to make this report is `cd /var/log/nginx && zcat access.log.*.gz | goaccess --4xx-to-unique-count access.log access.log.1 - -o report.html --log-format=COMBINED`
Any clarification this would help
|
1.0
|
Requested files vs new visitors a day, discrepency in hits and visitors - Hi, I am unsure if it is a bug, but when I see the panel of unquie visitors a day including spiders it shows around 1400 hits and 220 visitors. But if I look at requested URLS, it shows what I believe to be more accurate of almost the same number of visitors and hits.

So not sure since the domain ends in .to if there is a regex issue?
The command I use to make this report is `cd /var/log/nginx && zcat access.log.*.gz | goaccess --4xx-to-unique-count access.log access.log.1 - -o report.html --log-format=COMBINED`
Any clarification this would help
|
process
|
requested files vs new visitors a day discrepency in hits and visitors hi i am unsure if it is a bug but when i see the panel of unquie visitors a day including spiders it shows around hits and visitors but if i look at requested urls it shows what i believe to be more accurate of almost the same number of visitors and hits so not sure since the domain ends in to if there is a regex issue the command i use to make this report is cd var log nginx zcat access log gz goaccess to unique count access log access log o report html log format combined any clarification this would help
| 1
|
238,446
| 26,112,059,459
|
IssuesEvent
|
2022-12-27 21:43:33
|
microsoft/ebpf-for-windows
|
https://api.github.com/repos/microsoft/ebpf-for-windows
|
closed
|
AddressSanitizer: heap-buffer-overflow in btf_parse_line_information
|
bug triaged security
|
https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129
```
=================================================================
==5[976](https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129#step:16:977)==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x126a01034074 at pc 0x7ff72c0cbe6a bp 0x00f6bbb6cc80 sp 0x00f6bbb6cc80
READ of size 4 at 0x126a01034074 thread T0
#0 0x7ff72c0cbe69 in btf_parse_line_information(class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::function<(class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, unsigned int)>) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14021be69)
#1 0x7ff72c041eeb in read_elf(class std::basic_istream<char, struct std::char_traits<char>> &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, struct ebpf_verifier_options_t const *, struct ebpf_platform_t const *) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140191eeb)
#2 0x7ff72bec2fc6 in _ebpf_api_elf_verify_section_from_stream D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:516
#3 0x7ff72bec4227 in _verify_section_from_string D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:605
#4 0x7ff72bec4631 in ebpf_api_elf_verify_section_from_memory D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:637
#5 0x7ff72bf9c302 in LLVMFuzzerTestOneInput D:\a\ebpf-for-windows\ebpf-for-windows\tests\libfuzzer\verifier\libfuzz_harness.cpp:18
#6 0x7ff72bff5f9f in fuzzer::Fuzzer::ExecuteCallback(unsigned char const *, unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:559
#7 0x7ff72bff80fe in fuzzer::Fuzzer::RunOne(unsigned char const *, unsigned __int64, bool, struct fuzzer::InputInfo *, bool *) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:471
#8 0x7ff72bff6bbd in fuzzer::Fuzzer::MutateAndTestOne(void) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:702
#9 0x7ff72bff6683 in fuzzer::Fuzzer::Loop(class std::vector<struct fuzzer::SizedFile, class fuzzer::fuzzer_allocator<struct fuzzer::SizedFile>> &) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:838
#10 0x7ff72bfeafe9 in fuzzer::FuzzerDriver(int *, char ***, int (__cdecl *)(unsigned char const *, unsigned __int64)) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerDriver.cpp:853
#11 0x7ff72bfdfe82 in main D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerMain.cpp:20
#12 0x7ff72bfdac7b in invoke_main D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78
#13 0x7ff72bfdac7b in __scrt_common_main_seh D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288
#14 0x7ffc1a1a4ecf (C:\Windows\System32\KERNEL32.DLL+0x180014ecf)
#15 0x7ffc1a7be40a (C:\Windows\SYSTEM32\ntdll.dll+0x18007e40a)
0x126a01034076 is located 0 bytes to the right of 54-byte region [0x126a01034040,0x126a01034076)
allocated by thread T0 here:
#0 0x7ff72bfd407a in operator new(unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\asan\asan_win_new_scalar_thunk.cpp:41
#1 0x7ff72bf3065e in std::allocator<unsigned char>::allocate C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\xmemory:835
#2 0x7ff72bf3065e in std::vector<unsigned char, class std::allocator<unsigned char>>::_Buy_raw(unsigned __int64) C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector:2052
#3 0x7ff72c02e050 in std::vector<unsigned char, class std::allocator<unsigned char>>::_Construct_n<unsigned char *, unsigned char *>(unsigned __int64, unsigned char *&&, unsigned char *&&) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14017e050)
#4 0x7ff72c02bf3d in std::vector<unsigned char, class std::allocator<unsigned char>>::vector<unsigned char, class std::allocator<unsigned char>><unsigned char *, 0>(unsigned char *, unsigned char *, class std::allocator<unsigned char> const &) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14017bf3d)
#5 0x7ff72c034705 in std::_Tree<class std::_Tset_traits<unsigned short, struct std::less<unsigned short>, class std::allocator<unsigned short>, 0>>::insert<0, 0>(unsigned short &&) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140184705)
#6 0x7ff72c041ec2 in read_elf(class std::basic_istream<char, struct std::char_traits<char>> &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, struct ebpf_verifier_options_t const *, struct ebpf_platform_t const *) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140191ec2)
#7 0x7ff72bec2fc6 in _ebpf_api_elf_verify_section_from_stream D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:516
#8 0x7ff72bec4227 in _verify_section_from_string D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:605
#9 0x7ff72bec4631 in ebpf_api_elf_verify_section_from_memory D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:637
#10 0x7ff72bf9c302 in LLVMFuzzerTestOneInput D:\a\ebpf-for-windows\ebpf-for-windows\tests\libfuzzer\verifier\libfuzz_harness.cpp:18
#11 0x7ff72bff5f9f in fuzzer::Fuzzer::ExecuteCallback(unsigned char const *, unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:559
#12 0x7ff72bff80fe in fuzzer::Fuzzer::RunOne(unsigned char const *, unsigned __int64, bool, struct fuzzer::InputInfo *, bool *) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:471
#13 0x7ff72bff6bbd in fuzzer::Fuzzer::MutateAndTestOne(void) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:702
#14 0x7ff72bff6683 in fuzzer::Fuzzer::Loop(class std::vector<struct fuzzer::SizedFile, class fuzzer::fuzzer_allocator<struct fuzzer::SizedFile>> &) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:838
#15 0x7ff72bfeafe9 in fuzzer::FuzzerDriver(int *, char ***, int (__cdecl *)(unsigned char const *, unsigned __int64)) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerDriver.cpp:853
#16 0x7ff72bfdfe82 in main D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerMain.cpp:20
#17 0x7ff72bfdac7b in invoke_main D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78
#18 0x7ff72bfdac7b in __scrt_common_main_seh D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288
#19 0x7ffc1a1a4ecf (C:\Windows\System32\KERNEL32.DLL+0x180014ecf)
#20 0x7ffc1a7be40a (C:\Windows\SYSTEM32\ntdll.dll+0x18007e40a)
SUMMARY: AddressSanitizer: heap-buffer-overflow (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14021be69) in btf_parse_line_information(class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::function<(class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, unsigned int)>)
Shadow bytes around the buggy address:
0x04ab410067b0: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa
0x04ab410067c0: fd fd fd fd fd fd fd fa fa fa fa fa fd fd fd fd
0x04ab410067d0: fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd
0x04ab410067e0: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa
0x04ab410067f0: fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd
=>0x04ab41006800: fd fd fd fd fa fa fa fa 00 00 00 00 00 00[06]fa
0x04ab41006810: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa
0x04ab41006820: fa fa fa fa fa fa fa fa fa fa fa fa fd fd fd fd
0x04ab41006830: fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd
0x04ab41006840: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa
0x04ab41006850: fd fd fd fd fd fd fd fa fa fa fa fa fa fa fa fa
Shadow byte legend (one shadow byte represents 8 application bytes):
Addressable: 00
Partially addressable: 01 02 03 04 05 06 07
Heap left redzone: fa
Freed heap region: fd
Stack left redzone: f1
Stack mid redzone: f2
Stack right redzone: f3
Stack after return: f5
Stack use after scope: f8
Global redzone: f9
Global init order: f6
Poisoned by user: f7
Container overflow: fc
Array cookie: ac
Intra object redzone: bb
ASan internal: fe
Left alloca redzone: ca
Right alloca redzone: cb
Shadow gap: cc
==5976==ABORTING
MS: 3 ChangeASCIIInt-ChangeByte-CMP- DE: "6\x00\x00\x00\x00\x00\x00\x00"-; base unit: 8e6ec735553a53dafa51712487c621a5e42321b6
artifact_prefix='Artifacts\'; Test unit written to Artifacts\crash-b7220d69a1e24129[984](https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129#step:16:985)c0a1d855e5cd1146f6938
```
|
True
|
AddressSanitizer: heap-buffer-overflow in btf_parse_line_information - https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129
```
=================================================================
==5[976](https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129#step:16:977)==ERROR: AddressSanitizer: heap-buffer-overflow on address 0x126a01034074 at pc 0x7ff72c0cbe6a bp 0x00f6bbb6cc80 sp 0x00f6bbb6cc80
READ of size 4 at 0x126a01034074 thread T0
#0 0x7ff72c0cbe69 in btf_parse_line_information(class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::function<(class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, unsigned int)>) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14021be69)
#1 0x7ff72c041eeb in read_elf(class std::basic_istream<char, struct std::char_traits<char>> &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, struct ebpf_verifier_options_t const *, struct ebpf_platform_t const *) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140191eeb)
#2 0x7ff72bec2fc6 in _ebpf_api_elf_verify_section_from_stream D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:516
#3 0x7ff72bec4227 in _verify_section_from_string D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:605
#4 0x7ff72bec4631 in ebpf_api_elf_verify_section_from_memory D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:637
#5 0x7ff72bf9c302 in LLVMFuzzerTestOneInput D:\a\ebpf-for-windows\ebpf-for-windows\tests\libfuzzer\verifier\libfuzz_harness.cpp:18
#6 0x7ff72bff5f9f in fuzzer::Fuzzer::ExecuteCallback(unsigned char const *, unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:559
#7 0x7ff72bff80fe in fuzzer::Fuzzer::RunOne(unsigned char const *, unsigned __int64, bool, struct fuzzer::InputInfo *, bool *) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:471
#8 0x7ff72bff6bbd in fuzzer::Fuzzer::MutateAndTestOne(void) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:702
#9 0x7ff72bff6683 in fuzzer::Fuzzer::Loop(class std::vector<struct fuzzer::SizedFile, class fuzzer::fuzzer_allocator<struct fuzzer::SizedFile>> &) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:838
#10 0x7ff72bfeafe9 in fuzzer::FuzzerDriver(int *, char ***, int (__cdecl *)(unsigned char const *, unsigned __int64)) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerDriver.cpp:853
#11 0x7ff72bfdfe82 in main D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerMain.cpp:20
#12 0x7ff72bfdac7b in invoke_main D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78
#13 0x7ff72bfdac7b in __scrt_common_main_seh D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288
#14 0x7ffc1a1a4ecf (C:\Windows\System32\KERNEL32.DLL+0x180014ecf)
#15 0x7ffc1a7be40a (C:\Windows\SYSTEM32\ntdll.dll+0x18007e40a)
0x126a01034076 is located 0 bytes to the right of 54-byte region [0x126a01034040,0x126a01034076)
allocated by thread T0 here:
#0 0x7ff72bfd407a in operator new(unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\asan\asan_win_new_scalar_thunk.cpp:41
#1 0x7ff72bf3065e in std::allocator<unsigned char>::allocate C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\xmemory:835
#2 0x7ff72bf3065e in std::vector<unsigned char, class std::allocator<unsigned char>>::_Buy_raw(unsigned __int64) C:\Program Files\Microsoft Visual Studio\2022\Enterprise\VC\Tools\MSVC\14.34.31933\include\vector:2052
#3 0x7ff72c02e050 in std::vector<unsigned char, class std::allocator<unsigned char>>::_Construct_n<unsigned char *, unsigned char *>(unsigned __int64, unsigned char *&&, unsigned char *&&) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14017e050)
#4 0x7ff72c02bf3d in std::vector<unsigned char, class std::allocator<unsigned char>>::vector<unsigned char, class std::allocator<unsigned char>><unsigned char *, 0>(unsigned char *, unsigned char *, class std::allocator<unsigned char> const &) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14017bf3d)
#5 0x7ff72c034705 in std::_Tree<class std::_Tset_traits<unsigned short, struct std::less<unsigned short>, class std::allocator<unsigned short>, 0>>::insert<0, 0>(unsigned short &&) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140184705)
#6 0x7ff72c041ec2 in read_elf(class std::basic_istream<char, struct std::char_traits<char>> &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, struct ebpf_verifier_options_t const *, struct ebpf_platform_t const *) (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x140191ec2)
#7 0x7ff72bec2fc6 in _ebpf_api_elf_verify_section_from_stream D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:516
#8 0x7ff72bec4227 in _verify_section_from_string D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:605
#9 0x7ff72bec4631 in ebpf_api_elf_verify_section_from_memory D:\a\ebpf-for-windows\ebpf-for-windows\libs\api\Verifier.cpp:637
#10 0x7ff72bf9c302 in LLVMFuzzerTestOneInput D:\a\ebpf-for-windows\ebpf-for-windows\tests\libfuzzer\verifier\libfuzz_harness.cpp:18
#11 0x7ff72bff5f9f in fuzzer::Fuzzer::ExecuteCallback(unsigned char const *, unsigned __int64) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:559
#12 0x7ff72bff80fe in fuzzer::Fuzzer::RunOne(unsigned char const *, unsigned __int64, bool, struct fuzzer::InputInfo *, bool *) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:471
#13 0x7ff72bff6bbd in fuzzer::Fuzzer::MutateAndTestOne(void) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:702
#14 0x7ff72bff6683 in fuzzer::Fuzzer::Loop(class std::vector<struct fuzzer::SizedFile, class fuzzer::fuzzer_allocator<struct fuzzer::SizedFile>> &) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerLoop.cpp:838
#15 0x7ff72bfeafe9 in fuzzer::FuzzerDriver(int *, char ***, int (__cdecl *)(unsigned char const *, unsigned __int64)) D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerDriver.cpp:853
#16 0x7ff72bfdfe82 in main D:\a\_work\1\s\src\vctools\asan\llvm\compiler-rt\lib\fuzzer\FuzzerMain.cpp:20
#17 0x7ff72bfdac7b in invoke_main D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:78
#18 0x7ff72bfdac7b in __scrt_common_main_seh D:\a\_work\1\s\src\vctools\crt\vcstartup\src\startup\exe_common.inl:288
#19 0x7ffc1a1a4ecf (C:\Windows\System32\KERNEL32.DLL+0x180014ecf)
#20 0x7ffc1a7be40a (C:\Windows\SYSTEM32\ntdll.dll+0x18007e40a)
SUMMARY: AddressSanitizer: heap-buffer-overflow (D:\a\ebpf-for-windows\ebpf-for-windows\x64\Release\verifier_fuzzer.exe+0x14021be69) in btf_parse_line_information(class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::vector<unsigned char, class std::allocator<unsigned char>> const &, class std::function<(class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, class std::basic_string<char, struct std::char_traits<char>, class std::allocator<char>> const &, unsigned int, unsigned int)>)
Shadow bytes around the buggy address:
0x04ab410067b0: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa
0x04ab410067c0: fd fd fd fd fd fd fd fa fa fa fa fa fd fd fd fd
0x04ab410067d0: fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd
0x04ab410067e0: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa
0x04ab410067f0: fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd
=>0x04ab41006800: fd fd fd fd fa fa fa fa 00 00 00 00 00 00[06]fa
0x04ab41006810: fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa
0x04ab41006820: fa fa fa fa fa fa fa fa fa fa fa fa fd fd fd fd
0x04ab41006830: fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd
0x04ab41006840: fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa
0x04ab41006850: fd fd fd fd fd fd fd fa fa fa fa fa fa fa fa fa
Shadow byte legend (one shadow byte represents 8 application bytes):
Addressable: 00
Partially addressable: 01 02 03 04 05 06 07
Heap left redzone: fa
Freed heap region: fd
Stack left redzone: f1
Stack mid redzone: f2
Stack right redzone: f3
Stack after return: f5
Stack use after scope: f8
Global redzone: f9
Global init order: f6
Poisoned by user: f7
Container overflow: fc
Array cookie: ac
Intra object redzone: bb
ASan internal: fe
Left alloca redzone: ca
Right alloca redzone: cb
Shadow gap: cc
==5976==ABORTING
MS: 3 ChangeASCIIInt-ChangeByte-CMP- DE: "6\x00\x00\x00\x00\x00\x00\x00"-; base unit: 8e6ec735553a53dafa51712487c621a5e42321b6
artifact_prefix='Artifacts\'; Test unit written to Artifacts\crash-b7220d69a1e24129[984](https://github.com/microsoft/ebpf-for-windows/actions/runs/3666222777/jobs/6197891129#step:16:985)c0a1d855e5cd1146f6938
```
|
non_process
|
addresssanitizer heap buffer overflow in btf parse line information addresssanitizer heap buffer overflow on address at pc bp sp read of size at thread in btf parse line information class std vector const class std vector const class std function class std allocator const unsigned int class std basic string class std allocator const class std basic string class std allocator const unsigned int unsigned int d a ebpf for windows ebpf for windows release verifier fuzzer exe in read elf class std basic istream class std basic string class std allocator const class std basic string class std allocator const struct ebpf verifier options t const struct ebpf platform t const d a ebpf for windows ebpf for windows release verifier fuzzer exe in ebpf api elf verify section from stream d a ebpf for windows ebpf for windows libs api verifier cpp in verify section from string d a ebpf for windows ebpf for windows libs api verifier cpp in ebpf api elf verify section from memory d a ebpf for windows ebpf for windows libs api verifier cpp in llvmfuzzertestoneinput d a ebpf for windows ebpf for windows tests libfuzzer verifier libfuzz harness cpp in fuzzer fuzzer executecallback unsigned char const unsigned d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer runone unsigned char const unsigned bool struct fuzzer inputinfo bool d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer mutateandtestone void d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer loop class std vector d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzerdriver int char int cdecl unsigned char const unsigned d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerdriver cpp in main d a work s src vctools asan llvm compiler rt lib fuzzer fuzzermain cpp in invoke main d a work s src vctools crt vcstartup src startup exe common inl in scrt common main seh d a work s src vctools crt vcstartup src startup exe common inl c windows dll c windows ntdll dll is located bytes to the right of byte region allocated by thread here in operator new unsigned d a work s src vctools asan llvm compiler rt lib asan asan win new scalar thunk cpp in std allocator allocate c program files microsoft visual studio enterprise vc tools msvc include xmemory in std vector buy raw unsigned c program files microsoft visual studio enterprise vc tools msvc include vector in std vector construct n unsigned unsigned char unsigned char d a ebpf for windows ebpf for windows release verifier fuzzer exe in std vector vector unsigned char unsigned char class std allocator const d a ebpf for windows ebpf for windows release verifier fuzzer exe in std tree class std allocator insert unsigned short d a ebpf for windows ebpf for windows release verifier fuzzer exe in read elf class std basic istream class std basic string class std allocator const class std basic string class std allocator const struct ebpf verifier options t const struct ebpf platform t const d a ebpf for windows ebpf for windows release verifier fuzzer exe in ebpf api elf verify section from stream d a ebpf for windows ebpf for windows libs api verifier cpp in verify section from string d a ebpf for windows ebpf for windows libs api verifier cpp in ebpf api elf verify section from memory d a ebpf for windows ebpf for windows libs api verifier cpp in llvmfuzzertestoneinput d a ebpf for windows ebpf for windows tests libfuzzer verifier libfuzz harness cpp in fuzzer fuzzer executecallback unsigned char const unsigned d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer runone unsigned char const unsigned bool struct fuzzer inputinfo bool d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer mutateandtestone void d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzer loop class std vector d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerloop cpp in fuzzer fuzzerdriver int char int cdecl unsigned char const unsigned d a work s src vctools asan llvm compiler rt lib fuzzer fuzzerdriver cpp in main d a work s src vctools asan llvm compiler rt lib fuzzer fuzzermain cpp in invoke main d a work s src vctools crt vcstartup src startup exe common inl in scrt common main seh d a work s src vctools crt vcstartup src startup exe common inl c windows dll c windows ntdll dll summary addresssanitizer heap buffer overflow d a ebpf for windows ebpf for windows release verifier fuzzer exe in btf parse line information class std vector const class std vector const class std function class std allocator const unsigned int class std basic string class std allocator const class std basic string class std allocator const unsigned int unsigned int shadow bytes around the buggy address fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fa fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fd fa fa fa fa fd fd fd fd fd fd fd fa fa fa fa fa fa fa fa fa shadow byte legend one shadow byte represents application bytes addressable partially addressable heap left redzone fa freed heap region fd stack left redzone stack mid redzone stack right redzone stack after return stack use after scope global redzone global init order poisoned by user container overflow fc array cookie ac intra object redzone bb asan internal fe left alloca redzone ca right alloca redzone cb shadow gap cc aborting ms changeasciiint changebyte cmp de base unit artifact prefix artifacts test unit written to artifacts crash
| 0
|
231,524
| 18,772,411,328
|
IssuesEvent
|
2021-11-07 03:30:02
|
ProjectStarlight/StarlightRiver
|
https://api.github.com/repos/ProjectStarlight/StarlightRiver
|
closed
|
scrapshot grapple not working as intended when the snek enemy teleports mid-grapple
|
demo Checked - Test again
|
## Issue Description
Expected behavior: If a grappled Snek were to teleport, I would expect the grapple to break.
Actual behavior: The player is dragged to the top of the map by the Snek.
Context: I was grappling Sneks with the Scrapshot, and suddenly I was on top of the map.
Error message (if applicable): The first time this happened I was given a System.OutOfMemoryException, however on subsequent attempts to reproduce the game did not crash.
## Reproduction
Mod list: Starlight River
Config settings: No custom mod config.
Game settings: All settings default with the exception of custom keybinds and volume.
Gameplay events happening: Fighting a Snek with the Scrapshot and a set of Breacher armor.
Your inputs: All you need to do to recreate this is to time the grapple of the Scrapshot so that the Snek teleports before you reach it.
OS: Windows 10 Pro 64 bit
## Additional context and log files
**[client.log](https://github.com/ProjectStarlight/StarlightRiver/files/7341039/client.log)**
|
1.0
|
scrapshot grapple not working as intended when the snek enemy teleports mid-grapple - ## Issue Description
Expected behavior: If a grappled Snek were to teleport, I would expect the grapple to break.
Actual behavior: The player is dragged to the top of the map by the Snek.
Context: I was grappling Sneks with the Scrapshot, and suddenly I was on top of the map.
Error message (if applicable): The first time this happened I was given a System.OutOfMemoryException, however on subsequent attempts to reproduce the game did not crash.
## Reproduction
Mod list: Starlight River
Config settings: No custom mod config.
Game settings: All settings default with the exception of custom keybinds and volume.
Gameplay events happening: Fighting a Snek with the Scrapshot and a set of Breacher armor.
Your inputs: All you need to do to recreate this is to time the grapple of the Scrapshot so that the Snek teleports before you reach it.
OS: Windows 10 Pro 64 bit
## Additional context and log files
**[client.log](https://github.com/ProjectStarlight/StarlightRiver/files/7341039/client.log)**
|
non_process
|
scrapshot grapple not working as intended when the snek enemy teleports mid grapple issue description expected behavior if a grappled snek were to teleport i would expect the grapple to break actual behavior the player is dragged to the top of the map by the snek context i was grappling sneks with the scrapshot and suddenly i was on top of the map error message if applicable the first time this happened i was given a system outofmemoryexception however on subsequent attempts to reproduce the game did not crash reproduction mod list starlight river config settings no custom mod config game settings all settings default with the exception of custom keybinds and volume gameplay events happening fighting a snek with the scrapshot and a set of breacher armor your inputs all you need to do to recreate this is to time the grapple of the scrapshot so that the snek teleports before you reach it os windows pro bit additional context and log files
| 0
|
4,581
| 23,793,802,947
|
IssuesEvent
|
2022-09-02 17:08:43
|
Vivelin/SMZ3Randomizer
|
https://api.github.com/repos/Vivelin/SMZ3Randomizer
|
opened
|
Split out configs into a separate project
|
:wrench: maintainability
|
The current config system can't be accessed in the randomizer project. Because of that, the configs should be added to a unique project so that they can be accessed globally.
|
True
|
Split out configs into a separate project - The current config system can't be accessed in the randomizer project. Because of that, the configs should be added to a unique project so that they can be accessed globally.
|
non_process
|
split out configs into a separate project the current config system can t be accessed in the randomizer project because of that the configs should be added to a unique project so that they can be accessed globally
| 0
|
18,882
| 24,820,798,387
|
IssuesEvent
|
2022-10-25 16:16:19
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[processor/resourcedetection] Flaky test - TestDetectResource_Parallel
|
bug help wanted good first issue priority:p3 flaky test processor/resourcedetection
|
Seen in https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6580882988?check_suite_focus=true
```
=== RUN TestDetectResource_Parallel
==================
WARNING: DATA RACE
Read at 0x00c000216ba0 by goroutine 20:
go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0x78
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Previous write at 0x00c000216ba0 by goroutine 19:
reflect.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0
internal/reflectlite.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Goroutine 20 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
Goroutine 19 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
==================
==================
WARNING: DATA RACE
Read at 0x00c000216b80 by goroutine 21:
go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0xd4
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:113 +0x1f35
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:127 +0xf79
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.DeepEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:237 +0x3f9
github.com/stretchr/testify/assert.ObjectsAreEqual()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:65 +0x184
github.com/stretchr/testify/assert.Equal()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:343 +0x212
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:232 +0x16a
Previous write at 0x00c000216b90 by goroutine 19:
reflect.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0
internal/reflectlite.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Goroutine 20 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
Goroutine 19 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
==================
testing.go:1312: race detected during execution of test
--- FAIL: TestDetectResource_Parallel (0.01s)=== RUN TestDetectResource_Parallel
==================
WARNING: DATA RACE
Read at 0x00c000216ba0 by goroutine 20:
go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0x78
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Previous write at 0x00c000216ba0 by goroutine 19:
reflect.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0
internal/reflectlite.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Goroutine 20 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
Goroutine 19 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
==================
==================
WARNING: DATA RACE
Read at 0x00c000216b80 by goroutine 21:
go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0xd4
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:113 +0x1f35
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:127 +0xf79
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.DeepEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:237 +0x3f9
github.com/stretchr/testify/assert.ObjectsAreEqual()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:65 +0x184
github.com/stretchr/testify/assert.Equal()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:343 +0x212
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:232 +0x16a
Previous write at 0x00c000216b90 by goroutine 19:
reflect.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0
internal/reflectlite.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Goroutine 20 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
Goroutine 19 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
==================
testing.go:1312: race detected during execution of test
--- FAIL: TestDetectResource_Parallel (0.01s)
```
|
1.0
|
[processor/resourcedetection] Flaky test - TestDetectResource_Parallel - Seen in https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6580882988?check_suite_focus=true
```
=== RUN TestDetectResource_Parallel
==================
WARNING: DATA RACE
Read at 0x00c000216ba0 by goroutine 20:
go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0x78
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Previous write at 0x00c000216ba0 by goroutine 19:
reflect.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0
internal/reflectlite.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Goroutine 20 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
Goroutine 19 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
==================
==================
WARNING: DATA RACE
Read at 0x00c000216b80 by goroutine 21:
go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0xd4
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:113 +0x1f35
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:127 +0xf79
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.DeepEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:237 +0x3f9
github.com/stretchr/testify/assert.ObjectsAreEqual()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:65 +0x184
github.com/stretchr/testify/assert.Equal()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:343 +0x212
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:232 +0x16a
Previous write at 0x00c000216b90 by goroutine 19:
reflect.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0
internal/reflectlite.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Goroutine 20 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
Goroutine 19 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
==================
testing.go:1312: race detected during execution of test
--- FAIL: TestDetectResource_Parallel (0.01s)=== RUN TestDetectResource_Parallel
==================
WARNING: DATA RACE
Read at 0x00c000216ba0 by goroutine 20:
go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0x78
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Previous write at 0x00c000216ba0 by goroutine 19:
reflect.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0
internal/reflectlite.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Goroutine 20 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
Goroutine 19 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
==================
==================
WARNING: DATA RACE
Read at 0x00c000216b80 by goroutine 21:
go.opentelemetry.io/collector/pdata/internal.Map.Sort.func1()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:852 +0xd4
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:12 +0xd7
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:113 +0x1f35
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:127 +0xf79
reflect.deepValueEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:130 +0x209b
reflect.DeepEqual()
/opt/hostedtoolcache/go/1.18.2/x64/src/reflect/deepequal.go:237 +0x3f9
github.com/stretchr/testify/assert.ObjectsAreEqual()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:65 +0x184
github.com/stretchr/testify/assert.Equal()
/home/runner/go/pkg/mod/github.com/stretchr/testify@v1.7.1/assert/assertions.go:343 +0x212
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:232 +0x16a
Previous write at 0x00c000216b90 by goroutine 19:
reflect.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:178 +0x0
internal/reflectlite.typedmemmove()
/opt/hostedtoolcache/go/1.18.2/x64/src/runtime/mbarrier.go:196 +0x18
sort.insertionSort_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:13 +0x94
sort.stable_func()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/zfuncversion.go:171 +0x7c
sort.SliceStable()
/opt/hostedtoolcache/go/1.18.2/x64/src/sort/slice.go:32 +0xae
go.opentelemetry.io/collector/pdata/internal.Map.Sort()
/home/runner/go/pkg/mod/go.opentelemetry.io/collector/pdata@v0.51.1-0.20220519211145-c56d20e9e0af/internal/common.go:851 +0x7a
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel.func1()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:231 +0x133
Goroutine 20 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
Goroutine 19 (running) created at:
github.com/open-telemetry/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal.TestDetectResource_Parallel()
/home/runner/work/opentelemetry-collector-contrib/opentelemetry-collector-contrib/processor/resourcedetectionprocessor/internal/resourcedetection_test.go:227 +0xdf1
testing.tRunner()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1439 +0x213
testing.(*T).Run.func1()
/opt/hostedtoolcache/go/1.18.2/x64/src/testing/testing.go:1486 +0x47
==================
testing.go:1312: race detected during execution of test
--- FAIL: TestDetectResource_Parallel (0.01s)
```
|
process
|
flaky test testdetectresource parallel seen in run testdetectresource parallel warning data race read at by goroutine go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go previous write at by goroutine reflect typedmemmove opt hostedtoolcache go src runtime mbarrier go internal reflectlite typedmemmove opt hostedtoolcache go src runtime mbarrier go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go warning data race read at by goroutine go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepequal opt hostedtoolcache go src reflect deepequal go github com stretchr testify assert objectsareequal home runner go pkg mod github com stretchr testify assert assertions go github com stretchr testify assert equal home runner go pkg mod github com stretchr testify assert assertions go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go previous write at by goroutine reflect typedmemmove opt hostedtoolcache go src runtime mbarrier go internal reflectlite typedmemmove opt hostedtoolcache go src runtime mbarrier go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go testing go race detected during execution of test fail testdetectresource parallel run testdetectresource parallel warning data race read at by goroutine go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go previous write at by goroutine reflect typedmemmove opt hostedtoolcache go src runtime mbarrier go internal reflectlite typedmemmove opt hostedtoolcache go src runtime mbarrier go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go warning data race read at by goroutine go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepvalueequal opt hostedtoolcache go src reflect deepequal go reflect deepequal opt hostedtoolcache go src reflect deepequal go github com stretchr testify assert objectsareequal home runner go pkg mod github com stretchr testify assert assertions go github com stretchr testify assert equal home runner go pkg mod github com stretchr testify assert assertions go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go previous write at by goroutine reflect typedmemmove opt hostedtoolcache go src runtime mbarrier go internal reflectlite typedmemmove opt hostedtoolcache go src runtime mbarrier go sort insertionsort func opt hostedtoolcache go src sort zfuncversion go sort stable func opt hostedtoolcache go src sort zfuncversion go sort slicestable opt hostedtoolcache go src sort slice go go opentelemetry io collector pdata internal map sort home runner go pkg mod go opentelemetry io collector pdata internal common go github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go goroutine running created at github com open telemetry opentelemetry collector contrib processor resourcedetectionprocessor internal testdetectresource parallel home runner work opentelemetry collector contrib opentelemetry collector contrib processor resourcedetectionprocessor internal resourcedetection test go testing trunner opt hostedtoolcache go src testing testing go testing t run opt hostedtoolcache go src testing testing go testing go race detected during execution of test fail testdetectresource parallel
| 1
|
18,391
| 24,522,416,044
|
IssuesEvent
|
2022-10-11 10:31:34
|
streamnative/flink
|
https://api.github.com/repos/streamnative/flink
|
closed
|
[Enhancement][Stream] Remove pulsar.consumer.autoUpdatePartitionsIntervalSeconds
|
compute/data-processing type/enhancement
|
This config option is useless, we should remove it now.
|
1.0
|
[Enhancement][Stream] Remove pulsar.consumer.autoUpdatePartitionsIntervalSeconds - This config option is useless, we should remove it now.
|
process
|
remove pulsar consumer autoupdatepartitionsintervalseconds this config option is useless we should remove it now
| 1
|
15,753
| 19,911,722,699
|
IssuesEvent
|
2022-01-25 17:50:11
|
input-output-hk/high-assurance-legacy
|
https://api.github.com/repos/input-output-hk/high-assurance-legacy
|
closed
|
Generalize the proofs of the basic bisimilarity core laws
|
type: enhancement language: isabelle topic: process calculus
|
We have implemented a number of core laws about the basic strong bisimilarity relation, namely compatibility (preservation) laws, scope extension laws, and neutrality, associativity, and commutativity of parallel composition. These laws have counterparts for the basic _weak_ bisimilarity relation, but these haven’t been proved yet.
Incidentally, the proofs of the core laws for the strong case can be easily generalized to also cover the weak case. This is because the weak analogs of all basic transition rules hold (see #76). The generalization works by replacing all applications of basic transition rules with applications of fact parameters that can be specialized to either the basic transition rules or their weak analogs. Our goal is to generalize the proofs in this way.
|
1.0
|
Generalize the proofs of the basic bisimilarity core laws - We have implemented a number of core laws about the basic strong bisimilarity relation, namely compatibility (preservation) laws, scope extension laws, and neutrality, associativity, and commutativity of parallel composition. These laws have counterparts for the basic _weak_ bisimilarity relation, but these haven’t been proved yet.
Incidentally, the proofs of the core laws for the strong case can be easily generalized to also cover the weak case. This is because the weak analogs of all basic transition rules hold (see #76). The generalization works by replacing all applications of basic transition rules with applications of fact parameters that can be specialized to either the basic transition rules or their weak analogs. Our goal is to generalize the proofs in this way.
|
process
|
generalize the proofs of the basic bisimilarity core laws we have implemented a number of core laws about the basic strong bisimilarity relation namely compatibility preservation laws scope extension laws and neutrality associativity and commutativity of parallel composition these laws have counterparts for the basic weak bisimilarity relation but these haven’t been proved yet incidentally the proofs of the core laws for the strong case can be easily generalized to also cover the weak case this is because the weak analogs of all basic transition rules hold see the generalization works by replacing all applications of basic transition rules with applications of fact parameters that can be specialized to either the basic transition rules or their weak analogs our goal is to generalize the proofs in this way
| 1
|
244,667
| 18,764,946,547
|
IssuesEvent
|
2021-11-05 21:51:28
|
vmware-tanzu/carvel-ytt
|
https://api.github.com/repos/vmware-tanzu/carvel-ytt
|
closed
|
Clarify what "data values" are in the docs
|
enhancement documentation
|
The doc page for https://carvel.dev/ytt/docs/latest/ytt-data-values/ is self referential :)
Also the https://carvel.dev/ytt/docs/latest/how-to-use-data-values/ page seems to reference "Data Values" as a proper qouted term, but doesnt seem to really explain it either.
```
Overview ¶
A ytt run can be configured by supplying custom Data Values.
```
Can we have `Data Values` definition on that page, where we sorta dumb it down... i.e.
It can be a little confusing if you dont use YTT alot, to try to differentiate the data, template, and value...
|
1.0
|
Clarify what "data values" are in the docs - The doc page for https://carvel.dev/ytt/docs/latest/ytt-data-values/ is self referential :)
Also the https://carvel.dev/ytt/docs/latest/how-to-use-data-values/ page seems to reference "Data Values" as a proper qouted term, but doesnt seem to really explain it either.
```
Overview ¶
A ytt run can be configured by supplying custom Data Values.
```
Can we have `Data Values` definition on that page, where we sorta dumb it down... i.e.
It can be a little confusing if you dont use YTT alot, to try to differentiate the data, template, and value...
|
non_process
|
clarify what data values are in the docs the doc page for is self referential also the page seems to reference data values as a proper qouted term but doesnt seem to really explain it either overview ¶ a ytt run can be configured by supplying custom data values can we have data values definition on that page where we sorta dumb it down i e it can be a little confusing if you dont use ytt alot to try to differentiate the data template and value
| 0
|
11,711
| 14,546,460,665
|
IssuesEvent
|
2020-12-15 21:16:20
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Extends template does not render parameters
|
Pri2 devops-cicd-process/tech devops/prod product-feedback
|
I have tried to insert runtime param from simple-param.yml file same as [MS DOC](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#passing-parameters),
The **"yesNo"** parameter in simple-param.yml file not render/view() into AzureDevOps

```
# File: simple-param.yml
parameters:
- name: yesNo # name of the parameter; required
type: boolean # data type of the parameter; required
default: false
steps:
- script: echo ${{ parameters.yesNo }}
```
```
# File: azure-popelines.yml
name: DEMO
parameters:
- name: demo-name# name of the parameter; required
type: string # data type of the parameter; required
default: xxx
extends:
template: simple-param.yml
parameters:
yesNo: false # set to a non-boolean value to have the build fail
```
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Extends template does not render parameters - I have tried to insert runtime param from simple-param.yml file same as [MS DOC](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#passing-parameters),
The **"yesNo"** parameter in simple-param.yml file not render/view() into AzureDevOps

```
# File: simple-param.yml
parameters:
- name: yesNo # name of the parameter; required
type: boolean # data type of the parameter; required
default: false
steps:
- script: echo ${{ parameters.yesNo }}
```
```
# File: azure-popelines.yml
name: DEMO
parameters:
- name: demo-name# name of the parameter; required
type: string # data type of the parameter; required
default: xxx
extends:
template: simple-param.yml
parameters:
yesNo: false # set to a non-boolean value to have the build fail
```
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
extends template does not render parameters i have tried to insert runtime param from simple param yml file same as the yesno parameter in simple param yml file not render view into azuredevops file simple param yml parameters name yesno name of the parameter required type boolean data type of the parameter required default false steps script echo parameters yesno file azure popelines yml name demo parameters name demo name name of the parameter required type string data type of the parameter required default xxx extends template simple param yml parameters yesno false set to a non boolean value to have the build fail document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
315,168
| 27,051,499,129
|
IssuesEvent
|
2023-02-13 13:33:46
|
aquaaerobicsystem/aerationandmixing
|
https://api.github.com/repos/aquaaerobicsystem/aerationandmixing
|
closed
|
One slight change to the report for “sloped basin.” I just want to have this look cleaner on the report.
|
testing
|

|
1.0
|
One slight change to the report for “sloped basin.” I just want to have this look cleaner on the report. - 
|
non_process
|
one slight change to the report for “sloped basin ” i just want to have this look cleaner on the report
| 0
|
94,378
| 11,866,630,990
|
IssuesEvent
|
2020-03-26 04:23:26
|
openstates/openstates.org
|
https://api.github.com/repos/openstates/openstates.org
|
opened
|
voter boxes are different heights based on matched or not
|
design web
|
<img width="1214" alt="Screen Shot 2020-03-26 at 12 23 05 AM" src="https://user-images.githubusercontent.com/8034/77610044-fdaa5f00-6ef7-11ea-9cbc-f468663c5f1a.png">
|
1.0
|
voter boxes are different heights based on matched or not -
<img width="1214" alt="Screen Shot 2020-03-26 at 12 23 05 AM" src="https://user-images.githubusercontent.com/8034/77610044-fdaa5f00-6ef7-11ea-9cbc-f468663c5f1a.png">
|
non_process
|
voter boxes are different heights based on matched or not img width alt screen shot at am src
| 0
|
83,907
| 10,342,548,029
|
IssuesEvent
|
2019-09-04 06:46:42
|
wavespectra/wavespectra
|
https://api.github.com/repos/wavespectra/wavespectra
|
opened
|
Automated doc
|
documentation
|
shpynx stuff have not been updated since generated a few years back. The main issue with automated building was due to the fortran compilation.
|
1.0
|
Automated doc - shpynx stuff have not been updated since generated a few years back. The main issue with automated building was due to the fortran compilation.
|
non_process
|
automated doc shpynx stuff have not been updated since generated a few years back the main issue with automated building was due to the fortran compilation
| 0
|
10,334
| 13,163,456,879
|
IssuesEvent
|
2020-08-11 00:29:03
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Flaky parallel/test-child-process-fork-args on Windows
|
CI / flaky test child_process windows
|
Saw this test crashing in `node-test-binary-windows-js-suites` on https://ci.nodejs.org/job/node-test-pull-request/30753/:
```
12:59:01 not ok 70 parallel/test-child-process-fork-args
12:59:01 ---
12:59:01 duration_ms: 0.305
12:59:01 severity: crashed
12:59:01 exitcode: -1073741819
12:59:01 stack: |-
12:59:01 ...
```
|
1.0
|
Flaky parallel/test-child-process-fork-args on Windows - Saw this test crashing in `node-test-binary-windows-js-suites` on https://ci.nodejs.org/job/node-test-pull-request/30753/:
```
12:59:01 not ok 70 parallel/test-child-process-fork-args
12:59:01 ---
12:59:01 duration_ms: 0.305
12:59:01 severity: crashed
12:59:01 exitcode: -1073741819
12:59:01 stack: |-
12:59:01 ...
```
|
process
|
flaky parallel test child process fork args on windows saw this test crashing in node test binary windows js suites on not ok parallel test child process fork args duration ms severity crashed exitcode stack
| 1
|
71,311
| 7,241,738,759
|
IssuesEvent
|
2018-02-14 03:03:11
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
closed
|
We need a test case for standalone kubelet
|
area/kubelet area/reliability area/test lifecycle/rotten priority/awaiting-more-evidence sig/node
|
We have a scenario of kubelet being used in standalone mode without a master which is the containervm.
In that case, kubelet is started with a `--manifest_url` flag pointing to the URL of a YAML file with instructions of which containers to start.
This case is getting broken often, particularly as we seem to be testing for `nil` and in some cases the pointer vs. interface nil breaks our checks which are not being caught by any of the tests...
I'd say we need to include a test for `--manifest_url` in open source Kubernetes so that we don't get these bugs piling up until we try to build a new containervm with a version of Kubelet.
@dchen1107 @jbeda @brendanburns @satnam6502 @lavalamp
|
1.0
|
We need a test case for standalone kubelet - We have a scenario of kubelet being used in standalone mode without a master which is the containervm.
In that case, kubelet is started with a `--manifest_url` flag pointing to the URL of a YAML file with instructions of which containers to start.
This case is getting broken often, particularly as we seem to be testing for `nil` and in some cases the pointer vs. interface nil breaks our checks which are not being caught by any of the tests...
I'd say we need to include a test for `--manifest_url` in open source Kubernetes so that we don't get these bugs piling up until we try to build a new containervm with a version of Kubelet.
@dchen1107 @jbeda @brendanburns @satnam6502 @lavalamp
|
non_process
|
we need a test case for standalone kubelet we have a scenario of kubelet being used in standalone mode without a master which is the containervm in that case kubelet is started with a manifest url flag pointing to the url of a yaml file with instructions of which containers to start this case is getting broken often particularly as we seem to be testing for nil and in some cases the pointer vs interface nil breaks our checks which are not being caught by any of the tests i d say we need to include a test for manifest url in open source kubernetes so that we don t get these bugs piling up until we try to build a new containervm with a version of kubelet jbeda brendanburns lavalamp
| 0
|
247,320
| 18,857,598,727
|
IssuesEvent
|
2021-11-12 08:48:23
|
DystoriaX/pe
|
https://api.github.com/repos/DystoriaX/pe
|
opened
|
Help Message: Inconsistent Formatting
|
severity.VeryLow type.DocumentationBug
|
Some noticable inconsistent formatting:
- The parameter for `find` and `task find`, the additional `KEYWORD`.
- Parameter of user-keyed-in input is inconsistent with that of UG (i.e. should be capitalized according to UG)
<!--session: 1636702907050-1a839069-12c6-4666-ae30-5fe0011d2628-->
<!--Version: Web v3.4.1-->
|
1.0
|
Help Message: Inconsistent Formatting - Some noticable inconsistent formatting:
- The parameter for `find` and `task find`, the additional `KEYWORD`.
- Parameter of user-keyed-in input is inconsistent with that of UG (i.e. should be capitalized according to UG)
<!--session: 1636702907050-1a839069-12c6-4666-ae30-5fe0011d2628-->
<!--Version: Web v3.4.1-->
|
non_process
|
help message inconsistent formatting some noticable inconsistent formatting the parameter for find and task find the additional keyword parameter of user keyed in input is inconsistent with that of ug i e should be capitalized according to ug
| 0
|
29,484
| 14,141,820,732
|
IssuesEvent
|
2020-11-10 13:15:44
|
microsoft/vscode-pull-request-github
|
https://api.github.com/repos/microsoft/vscode-pull-request-github
|
closed
|
Checkout is slow on Windows
|
performance
|
Checkout is taking like 5+ seconds, here's a log:
```
Review> swtich to Pull Requet #55776
> git merge-base 1e45ce6add2ac86c2a6e214aa84c85281df1814b 44837ff8323d2a9e769a4157f5a62068518132f8
> git diff --no-ext-diff --exit-code
> git symbolic-ref --short HEAD
Review> there is already one local branch origin/tyriar/auto_switch_renderer associated with Pull Request #55776
> git fetch origin
> git rev-parse pr/segevfiner/55194
> git rev-parse --symbolic-full-name pr/segevfiner/55194@{u}
> git rev-list --left-right pr/segevfiner/55194...refs/remotes/segevfiner/cli-env-inheritance-fix
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git config --local --get remote.njkevlani.github-pr-remote
> git config --local --get remote.origin.github-pr-remote
> git config --local --get remote.segevfiner.github-pr-remote
> git config --local --get branch.pr/segevfiner/55194.github-pr-owner-number
> git rev-parse tyriar/auto_switch_renderer
> git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u}
> git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer
> git checkout tyriar/auto_switch_renderer
> git symbolic-ref --short HEAD
> git rev-parse pr/segevfiner/55194
> git rev-parse --symbolic-full-name pr/segevfiner/55194@{u}
> git rev-list --left-right pr/segevfiner/55194...refs/remotes/segevfiner/cli-env-inheritance-fix
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git config --local --get remote.njkevlani.github-pr-remote
GitHelper> associate tyriar/auto_switch_renderer with Pull Request #55776
> git config --local branch.tyriar/auto_switch_renderer.github-pr-owner-number Microsoft#vscode#55776
> git config --local --get remote.origin.github-pr-remote
> git symbolic-ref --short HEAD
> git config --local --get remote.segevfiner.github-pr-remote
> git rev-parse tyriar/auto_switch_renderer
> git config --local --get branch.pr/segevfiner/55194.github-pr-owner-number
> git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u}
> git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git config --local --get remote.njkevlani.github-pr-remote
> git config --local --get remote.origin.github-pr-remote
> git config --local --get remote.segevfiner.github-pr-remote
> git symbolic-ref --short HEAD
> git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number
> git rev-parse tyriar/auto_switch_renderer
Review> current branch tyriar/auto_switch_renderer is associated with pull request #55776
> git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u}
> git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git merge-base 1e45ce6add2ac86c2a6e214aa84c85281df1814b 44837ff8323d2a9e769a4157f5a62068518132f8
> git cat-file -e 7f5beafe6e7bedec2fa309b6429c293523dee6b2:src/vs/workbench/parts/terminal/electron-browser/terminalInstance.ts
Review> register comments provider
Review> display pull request status bar indicator and refresh pull request tree view.
> git config --local --get remote.origin.github-pr-remote
> git symbolic-ref --short HEAD
> git rev-parse tyriar/auto_switch_renderer
> git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u}
> git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git config --local --get remote.njkevlani.github-pr-remote
> git config --local --get remote.origin.github-pr-remote
> git config --local --get remote.segevfiner.github-pr-remote
> git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number
> git merge-base a957654db2cbf8df37f7d0139e103a7ecffdf926 487b6f745a93c9a344676ef2203e643421dbdf89
> git cat-file -e a957654db2cbf8df37f7d0139e103a7ecffdf926:src/vs/workbench/parts/terminal/common/terminalColorRegistry.ts
> git cat-file -e a957654db2cbf8df37f7d0139e103a7ecffdf926:src/vs/workbench/parts/terminal/electron-browser/terminalPanel.ts
> git merge-base 957a16690d156774c918fb447ff5d63b120f503e 945e6f81a9cbe01f12568dd973b8e33cbecfac53
> git cat-file -e 7b7c7dad6f740d73d685b3622b4fe528de7d5804:src/vs/workbench/parts/terminal/electron-browser/terminal.contribution.ts
> git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number
```
|
True
|
Checkout is slow on Windows - Checkout is taking like 5+ seconds, here's a log:
```
Review> swtich to Pull Requet #55776
> git merge-base 1e45ce6add2ac86c2a6e214aa84c85281df1814b 44837ff8323d2a9e769a4157f5a62068518132f8
> git diff --no-ext-diff --exit-code
> git symbolic-ref --short HEAD
Review> there is already one local branch origin/tyriar/auto_switch_renderer associated with Pull Request #55776
> git fetch origin
> git rev-parse pr/segevfiner/55194
> git rev-parse --symbolic-full-name pr/segevfiner/55194@{u}
> git rev-list --left-right pr/segevfiner/55194...refs/remotes/segevfiner/cli-env-inheritance-fix
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git config --local --get remote.njkevlani.github-pr-remote
> git config --local --get remote.origin.github-pr-remote
> git config --local --get remote.segevfiner.github-pr-remote
> git config --local --get branch.pr/segevfiner/55194.github-pr-owner-number
> git rev-parse tyriar/auto_switch_renderer
> git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u}
> git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer
> git checkout tyriar/auto_switch_renderer
> git symbolic-ref --short HEAD
> git rev-parse pr/segevfiner/55194
> git rev-parse --symbolic-full-name pr/segevfiner/55194@{u}
> git rev-list --left-right pr/segevfiner/55194...refs/remotes/segevfiner/cli-env-inheritance-fix
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git config --local --get remote.njkevlani.github-pr-remote
GitHelper> associate tyriar/auto_switch_renderer with Pull Request #55776
> git config --local branch.tyriar/auto_switch_renderer.github-pr-owner-number Microsoft#vscode#55776
> git config --local --get remote.origin.github-pr-remote
> git symbolic-ref --short HEAD
> git config --local --get remote.segevfiner.github-pr-remote
> git rev-parse tyriar/auto_switch_renderer
> git config --local --get branch.pr/segevfiner/55194.github-pr-owner-number
> git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u}
> git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git config --local --get remote.njkevlani.github-pr-remote
> git config --local --get remote.origin.github-pr-remote
> git config --local --get remote.segevfiner.github-pr-remote
> git symbolic-ref --short HEAD
> git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number
> git rev-parse tyriar/auto_switch_renderer
Review> current branch tyriar/auto_switch_renderer is associated with pull request #55776
> git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u}
> git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git merge-base 1e45ce6add2ac86c2a6e214aa84c85281df1814b 44837ff8323d2a9e769a4157f5a62068518132f8
> git cat-file -e 7f5beafe6e7bedec2fa309b6429c293523dee6b2:src/vs/workbench/parts/terminal/electron-browser/terminalInstance.ts
Review> register comments provider
Review> display pull request status bar indicator and refresh pull request tree view.
> git config --local --get remote.origin.github-pr-remote
> git symbolic-ref --short HEAD
> git rev-parse tyriar/auto_switch_renderer
> git rev-parse --symbolic-full-name tyriar/auto_switch_renderer@{u}
> git rev-list --left-right tyriar/auto_switch_renderer...refs/remotes/origin/tyriar/auto_switch_renderer
> git for-each-ref --format %(refname) %(objectname) --sort -committerdate
> git remote --verbose
> git config --local --get remote.njkevlani.github-pr-remote
> git config --local --get remote.origin.github-pr-remote
> git config --local --get remote.segevfiner.github-pr-remote
> git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number
> git merge-base a957654db2cbf8df37f7d0139e103a7ecffdf926 487b6f745a93c9a344676ef2203e643421dbdf89
> git cat-file -e a957654db2cbf8df37f7d0139e103a7ecffdf926:src/vs/workbench/parts/terminal/common/terminalColorRegistry.ts
> git cat-file -e a957654db2cbf8df37f7d0139e103a7ecffdf926:src/vs/workbench/parts/terminal/electron-browser/terminalPanel.ts
> git merge-base 957a16690d156774c918fb447ff5d63b120f503e 945e6f81a9cbe01f12568dd973b8e33cbecfac53
> git cat-file -e 7b7c7dad6f740d73d685b3622b4fe528de7d5804:src/vs/workbench/parts/terminal/electron-browser/terminal.contribution.ts
> git config --local --get branch.tyriar/auto_switch_renderer.github-pr-owner-number
```
|
non_process
|
checkout is slow on windows checkout is taking like seconds here s a log review swtich to pull requet git merge base git diff no ext diff exit code git symbolic ref short head review there is already one local branch origin tyriar auto switch renderer associated with pull request git fetch origin git rev parse pr segevfiner git rev parse symbolic full name pr segevfiner u git rev list left right pr segevfiner refs remotes segevfiner cli env inheritance fix git for each ref format refname objectname sort committerdate git remote verbose git config local get remote njkevlani github pr remote git config local get remote origin github pr remote git config local get remote segevfiner github pr remote git config local get branch pr segevfiner github pr owner number git rev parse tyriar auto switch renderer git rev parse symbolic full name tyriar auto switch renderer u git rev list left right tyriar auto switch renderer refs remotes origin tyriar auto switch renderer git checkout tyriar auto switch renderer git symbolic ref short head git rev parse pr segevfiner git rev parse symbolic full name pr segevfiner u git rev list left right pr segevfiner refs remotes segevfiner cli env inheritance fix git for each ref format refname objectname sort committerdate git remote verbose git config local get remote njkevlani github pr remote githelper associate tyriar auto switch renderer with pull request git config local branch tyriar auto switch renderer github pr owner number microsoft vscode git config local get remote origin github pr remote git symbolic ref short head git config local get remote segevfiner github pr remote git rev parse tyriar auto switch renderer git config local get branch pr segevfiner github pr owner number git rev parse symbolic full name tyriar auto switch renderer u git rev list left right tyriar auto switch renderer refs remotes origin tyriar auto switch renderer git for each ref format refname objectname sort committerdate git remote verbose git config local get remote njkevlani github pr remote git config local get remote origin github pr remote git config local get remote segevfiner github pr remote git symbolic ref short head git config local get branch tyriar auto switch renderer github pr owner number git rev parse tyriar auto switch renderer review current branch tyriar auto switch renderer is associated with pull request git rev parse symbolic full name tyriar auto switch renderer u git rev list left right tyriar auto switch renderer refs remotes origin tyriar auto switch renderer git for each ref format refname objectname sort committerdate git remote verbose git merge base git cat file e src vs workbench parts terminal electron browser terminalinstance ts review register comments provider review display pull request status bar indicator and refresh pull request tree view git config local get remote origin github pr remote git symbolic ref short head git rev parse tyriar auto switch renderer git rev parse symbolic full name tyriar auto switch renderer u git rev list left right tyriar auto switch renderer refs remotes origin tyriar auto switch renderer git for each ref format refname objectname sort committerdate git remote verbose git config local get remote njkevlani github pr remote git config local get remote origin github pr remote git config local get remote segevfiner github pr remote git config local get branch tyriar auto switch renderer github pr owner number git merge base git cat file e src vs workbench parts terminal common terminalcolorregistry ts git cat file e src vs workbench parts terminal electron browser terminalpanel ts git merge base git cat file e src vs workbench parts terminal electron browser terminal contribution ts git config local get branch tyriar auto switch renderer github pr owner number
| 0
|
603,543
| 18,668,845,316
|
IssuesEvent
|
2021-10-30 10:01:49
|
AY2122S1-CS2103T-W13-2/tp
|
https://api.github.com/repos/AY2122S1-CS2103T-W13-2/tp
|
closed
|
[PE-D] Unable to view long fields
|
type.Bug priority.High severity.Medium
|

If any of the fields are very long for some reason (name, email, address, etc), we may not be able to see it, especially if it is longer than the width of our entire screen. Maybe you can add a horizontal scroll bar so that even though it may be long, we are able to scroll horizontally to see it.
<!--session: 1635494539368-6e4ca0fb-c787-43c4-a16e-48a16480861f-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.Low` `type.FeatureFlaw`
original: s7u4rt99/ped#2
|
1.0
|
[PE-D] Unable to view long fields - 
If any of the fields are very long for some reason (name, email, address, etc), we may not be able to see it, especially if it is longer than the width of our entire screen. Maybe you can add a horizontal scroll bar so that even though it may be long, we are able to scroll horizontally to see it.
<!--session: 1635494539368-6e4ca0fb-c787-43c4-a16e-48a16480861f-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.Low` `type.FeatureFlaw`
original: s7u4rt99/ped#2
|
non_process
|
unable to view long fields if any of the fields are very long for some reason name email address etc we may not be able to see it especially if it is longer than the width of our entire screen maybe you can add a horizontal scroll bar so that even though it may be long we are able to scroll horizontally to see it labels severity low type featureflaw original ped
| 0
|
20,999
| 27,866,169,100
|
IssuesEvent
|
2023-03-21 10:23:56
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Ability to replace current Node process with another
|
child_process feature request stale
|
**Edit:** If someone can come up with a better shim for `execve` for Windows, that'd be *far* better. The form below is *very* expensive and *very* horrible.
**Edit 2:** Linked relevant [SO question](https://stackoverflow.com/questions/51185115/what-is-the-ideal-way-to-emulate-process-replacement-on-windows).
**Edit 3:** Clarify FS changes
**Edit 4:** Here's the text from that SO question as of July 6, 2018 (so you don't have to search for it), where I asked about how to do the Windows part.
<details>
<summary>Click to show (warning: lots of text)</summary>
So, in a [feature request I filed against Node.js](https://github.com/nodejs/node/issues/21664), I was looking for a way to replace the current Node process with another. In Linux and friends (really, any POSIX-compliant system), this is easy: use [`execve`](http://man7.org/linux/man-pages/man2/execve.2.html) and friends and call it a day. But obviously, that won't work on Windows, since it only has `CreateProcess` (which `execve` and friends delegate to, [complete with async behavior](https://stackoverflow.com/questions/49736973/blocking-version-of-execvp-windows)). And it's not like [people](https://stackoverflow.com/questions/35111313/windows-exec-equivalent) [haven't](https://stackoverflow.com/questions/6743567/replace-current-process-with-invocation-of-subprocess) [wanted](https://stackoverflow.com/questions/7198666/strategies-for-replacing-program-executable-in-windows) [to](https://stackoverflow.com/questions/198122/how-can-i-replace-the-current-java-process-like-a-unix-style-exec) [do](https://stackoverflow.com/questions/5450147/how-to-replace-the-current-java-process-in-windows-using-jna-jni) [similar](https://stackoverflow.com/questions/45607959/restart-windows-process-inplace-preserving-process-id-and-handles), leading to [numerous duplicate questions on this site](https://www.google.com/search?q=windows+replace+current+process+site:stackoverflow.com). (This isn't a duplicate because it's explicitly seeking a workaround given certain constraints, not just asking for direct replacement.)
Process replacement has several facets that have to addressed:
1. All console I/O streams have to be forwarded to the new process.
1. All signals need transparently forwarded to the new process.
1. The data from the old process have to be destroyed, with as many resources reclaimed as possible.
1. All pre-existing threads and child processes should be destroyed.
1. All pre-existing handles should be destroyed apart from open file descriptors and named pipes/etc.
1. Optimally, the old process's memory should be kept to a minimum after the process is created.
1. For my particular use case, retaining the process ID is not important.
And for my particular case, there are a few constraints:
1. I can control the initial process's startup as well as the location of my "process replacement" function.
1. I could load arbitrary native code via add-ons at potentially any stack offset.
- Implication: I can't even dream of tracking `malloc` calls, handles, thread manipulation, or process manipulation to track and free them all, since DLL rewriting isn't exactly practical.
1. I have no control over *when* my "process replacement" is called. It could be called through an add-on, which could've been called through either interpreted code via FFI or even another add-on recursively. It could even be called during add-on initialization.
- Implication: I would have no ability to know what's in the stack, even if I perfectly instrumented my side. And rewriting all their `call`s and `push`es is far from practical, and would just be all-around slow for obvious reasons.
So, here's the gist of what I was thinking: use something similar to a pseudo-trampoline.
1. Statically allocate the following:
1. A single pointer for the stack pointer.
1. `MAX_PATH + 1` chars for the application path + `'\0'`.
1. `MAX_PATH + 1` chars for the current working directory path + `'\0'`.
1. 32768 chars for the arguments + `'\0'`.
1. 32768 chars for the environment + `'\0'`.
1. On entry, set the global stack pointer reference to the stack pointer.
1. On "replacement":
1. Do relevant process cleanup and lock/release everything you can.
1. Set the stack pointer to the stored original global one.
1. Terminate each child thread.
1. Kill each child process.
1. Free [each open handle](https://stackoverflow.com/questions/733384/how-to-enumerate-process-handles).
1. If possible (i.e. not in a UWP program), [For each heap](https://docs.microsoft.com/en-us/windows/desktop/api/heapapi/nf-heapapi-getprocessheaps), [destroy it](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapdestroy) if it's not the [default heap](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-getprocessheap) or the temporary heap (if it exists).
1. If possible, close [each open handle](https://stackoverflow.com/questions/733384/how-to-enumerate-process-handles).
1. If possible, [walk](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapwalk) the default heap and [free](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapfree) each segment associated with it.
1. Create a new process with the statically allocated file/arguments/environment/etc. with no new window created.
1. Proxy all future received signals, exceptions, etc. without modification to this process somehow. [The standard signals are easy](https://docs.microsoft.com/en-us/windows/console/setconsolectrlhandler), but not so much with the exceptions.
1. Wait for the process to end.
1. Return with [the process's exit code](https://docs.microsoft.com/en-us/windows/desktop/api/processthreadsapi/nf-processthreadsapi-getexitcodeprocess).
The idea here is to use a process-based trampoline and drop the current process size to an absolute minimum while the newly created one is started.
But where I'm not very familiar with Windows, I probably made quite a few mistakes here. Also, the above seems *extremely* inefficient and to an extent it just feels horribly wrong for something a kernel could just release a few memory pages, deallocate a bunch of memory handles, and move some memory around for the next process.
So, to summarize, what's the ideal way to emulate process replacement on Windows with the fewest limitations?
</details>
-----
I would like a means to "replace" the current Node process with another, keeping the same process ID. It would be something morally similar to [this function](https://github.com/isiahmeadows/thallium/blob/master/lib/cli/util.js#L92-L111), but it wouldn't return. This would be most useful for conditionally replacing Node flags in a startup script - for example, if someone wants to enable modules and your behavior needs to change non-trivially in the presence of them (like if you need to install a default loader), you'll want to respawn the process with `--experimental-modules --loader <file>` so you can install the loader.
This is also for scenarios when you want to run a module as a `main` module. If you want to do logic after the process ends, you should be using `child_process.spawn` regardless - you shouldn't be attempting to "replace" it in any capacity.
Here's what I propose:
- `child_process.replaceSpawn(command [ , args] [ , options ])`
- `command` is the path to the new command.
- `args` is the args to replace the arguments with. This defaults to the empty array.
- `options` is for the various options for replacing the process. This defaults to an empty object.
- `options.cwd` is the new cwd to use. (Default: `process.cwd()`)
- `options.env` is the new environment to use. (Default: `process.env`)
- `options.argv0` is the binary to spawn as. (Default: `command`)
- `child_process.replaceFork(mainPath [ , args] [ , options ])` works similarly to above.
- `mainPath` is the path to the new `require.main`.
- `options.execPath` is the new binary to spawn as. (Default: `process.execPath`)
- `options.execArgv` are the new Node flags to spawn with. (Default: `process.execArgv`)
- `options.argv0` is the binary to spawn as. (Default: `process.argv0`)
- The command is the original binary itself.
- Add a `napi_terminating` member for `napi_status` to represent `try_catch.HasTerminated()` and the result of each call after replacement termination.
- Add a `napi_set_terminate_hook(napi_env env, void (*fun)(void*), void* data)` function to register a callback called on termination, to make it easier to clean up resources.
Internally, there are two cases you need to cover, and the simulated part for Windows is where it gets really hairy due to all the edge cases. Here's pseudocode for the basic algorithm (I'm not really familiar with Node internals, so take this as a rough guideline):
1. Stop the main event loop.
1. Go through the standard shutdown routine.
1. Destroy any open libuv handles and cancel any remaining event loop tasks.
1. If we're on a platform that supports process replacement (like Linux or Mac):
1. Invoke `execve` or equivalent with the new process path, arguments, and environment.
1. Else, if we're on Windows (the only supported OS that doesn't), we have to simulate it entirely:
1. Terminate execution via `v8::V8::TerminateExecution()`. All N-API callbacks should return `napi_terminated` during this step.
1. For each loaded native module:
1. If the native module has a terminate hook, call it.
1. Unload the native module's DLL.
1. Close the event loop.
1. Dispose the isolate.
1. Do the rest according to whatever happens to [this SO question](https://stackoverflow.com/questions/51185115/what-is-the-ideal-way-to-emulate-process-replacement-on-windows).
1. Else, on other OSs without a process replacement function, it'd look similar to Windows.
In addition, file system requests will have to generally create each file descriptor with `O_CLOEXEC`.
As for precedent where this could be used immediately:
- [Liftoff](https://www.npmjs.com/package/liftoff) works very similarly, just with a little extra opinionated sugar, and that's used natively in Gulp. This kind of thing would speed that up quite a bit.
- I do [very similar](https://github.com/isiahmeadows/thallium/blob/master/cli.js#L126-L150) to transparently pass through unknown Node flags.
- Babel [attempts to use `kexec`](https://github.com/babel/babel/blob/master/packages/babel-node/src/babel-node.js#L87-L88) where available, which [is a POSIX-only module that replaces the process literally](https://www.npmjs.com/package/kexec). Absent that, it falls back to [its own implementation](https://github.com/babel/babel/blob/master/packages/babel-node/src/babel-node.js#L90-L109) that works like the other two examples.
|
1.0
|
Ability to replace current Node process with another - **Edit:** If someone can come up with a better shim for `execve` for Windows, that'd be *far* better. The form below is *very* expensive and *very* horrible.
**Edit 2:** Linked relevant [SO question](https://stackoverflow.com/questions/51185115/what-is-the-ideal-way-to-emulate-process-replacement-on-windows).
**Edit 3:** Clarify FS changes
**Edit 4:** Here's the text from that SO question as of July 6, 2018 (so you don't have to search for it), where I asked about how to do the Windows part.
<details>
<summary>Click to show (warning: lots of text)</summary>
So, in a [feature request I filed against Node.js](https://github.com/nodejs/node/issues/21664), I was looking for a way to replace the current Node process with another. In Linux and friends (really, any POSIX-compliant system), this is easy: use [`execve`](http://man7.org/linux/man-pages/man2/execve.2.html) and friends and call it a day. But obviously, that won't work on Windows, since it only has `CreateProcess` (which `execve` and friends delegate to, [complete with async behavior](https://stackoverflow.com/questions/49736973/blocking-version-of-execvp-windows)). And it's not like [people](https://stackoverflow.com/questions/35111313/windows-exec-equivalent) [haven't](https://stackoverflow.com/questions/6743567/replace-current-process-with-invocation-of-subprocess) [wanted](https://stackoverflow.com/questions/7198666/strategies-for-replacing-program-executable-in-windows) [to](https://stackoverflow.com/questions/198122/how-can-i-replace-the-current-java-process-like-a-unix-style-exec) [do](https://stackoverflow.com/questions/5450147/how-to-replace-the-current-java-process-in-windows-using-jna-jni) [similar](https://stackoverflow.com/questions/45607959/restart-windows-process-inplace-preserving-process-id-and-handles), leading to [numerous duplicate questions on this site](https://www.google.com/search?q=windows+replace+current+process+site:stackoverflow.com). (This isn't a duplicate because it's explicitly seeking a workaround given certain constraints, not just asking for direct replacement.)
Process replacement has several facets that have to addressed:
1. All console I/O streams have to be forwarded to the new process.
1. All signals need transparently forwarded to the new process.
1. The data from the old process have to be destroyed, with as many resources reclaimed as possible.
1. All pre-existing threads and child processes should be destroyed.
1. All pre-existing handles should be destroyed apart from open file descriptors and named pipes/etc.
1. Optimally, the old process's memory should be kept to a minimum after the process is created.
1. For my particular use case, retaining the process ID is not important.
And for my particular case, there are a few constraints:
1. I can control the initial process's startup as well as the location of my "process replacement" function.
1. I could load arbitrary native code via add-ons at potentially any stack offset.
- Implication: I can't even dream of tracking `malloc` calls, handles, thread manipulation, or process manipulation to track and free them all, since DLL rewriting isn't exactly practical.
1. I have no control over *when* my "process replacement" is called. It could be called through an add-on, which could've been called through either interpreted code via FFI or even another add-on recursively. It could even be called during add-on initialization.
- Implication: I would have no ability to know what's in the stack, even if I perfectly instrumented my side. And rewriting all their `call`s and `push`es is far from practical, and would just be all-around slow for obvious reasons.
So, here's the gist of what I was thinking: use something similar to a pseudo-trampoline.
1. Statically allocate the following:
1. A single pointer for the stack pointer.
1. `MAX_PATH + 1` chars for the application path + `'\0'`.
1. `MAX_PATH + 1` chars for the current working directory path + `'\0'`.
1. 32768 chars for the arguments + `'\0'`.
1. 32768 chars for the environment + `'\0'`.
1. On entry, set the global stack pointer reference to the stack pointer.
1. On "replacement":
1. Do relevant process cleanup and lock/release everything you can.
1. Set the stack pointer to the stored original global one.
1. Terminate each child thread.
1. Kill each child process.
1. Free [each open handle](https://stackoverflow.com/questions/733384/how-to-enumerate-process-handles).
1. If possible (i.e. not in a UWP program), [For each heap](https://docs.microsoft.com/en-us/windows/desktop/api/heapapi/nf-heapapi-getprocessheaps), [destroy it](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapdestroy) if it's not the [default heap](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-getprocessheap) or the temporary heap (if it exists).
1. If possible, close [each open handle](https://stackoverflow.com/questions/733384/how-to-enumerate-process-handles).
1. If possible, [walk](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapwalk) the default heap and [free](https://docs.microsoft.com/en-us/windows/desktop/api/HeapApi/nf-heapapi-heapfree) each segment associated with it.
1. Create a new process with the statically allocated file/arguments/environment/etc. with no new window created.
1. Proxy all future received signals, exceptions, etc. without modification to this process somehow. [The standard signals are easy](https://docs.microsoft.com/en-us/windows/console/setconsolectrlhandler), but not so much with the exceptions.
1. Wait for the process to end.
1. Return with [the process's exit code](https://docs.microsoft.com/en-us/windows/desktop/api/processthreadsapi/nf-processthreadsapi-getexitcodeprocess).
The idea here is to use a process-based trampoline and drop the current process size to an absolute minimum while the newly created one is started.
But where I'm not very familiar with Windows, I probably made quite a few mistakes here. Also, the above seems *extremely* inefficient and to an extent it just feels horribly wrong for something a kernel could just release a few memory pages, deallocate a bunch of memory handles, and move some memory around for the next process.
So, to summarize, what's the ideal way to emulate process replacement on Windows with the fewest limitations?
</details>
-----
I would like a means to "replace" the current Node process with another, keeping the same process ID. It would be something morally similar to [this function](https://github.com/isiahmeadows/thallium/blob/master/lib/cli/util.js#L92-L111), but it wouldn't return. This would be most useful for conditionally replacing Node flags in a startup script - for example, if someone wants to enable modules and your behavior needs to change non-trivially in the presence of them (like if you need to install a default loader), you'll want to respawn the process with `--experimental-modules --loader <file>` so you can install the loader.
This is also for scenarios when you want to run a module as a `main` module. If you want to do logic after the process ends, you should be using `child_process.spawn` regardless - you shouldn't be attempting to "replace" it in any capacity.
Here's what I propose:
- `child_process.replaceSpawn(command [ , args] [ , options ])`
- `command` is the path to the new command.
- `args` is the args to replace the arguments with. This defaults to the empty array.
- `options` is for the various options for replacing the process. This defaults to an empty object.
- `options.cwd` is the new cwd to use. (Default: `process.cwd()`)
- `options.env` is the new environment to use. (Default: `process.env`)
- `options.argv0` is the binary to spawn as. (Default: `command`)
- `child_process.replaceFork(mainPath [ , args] [ , options ])` works similarly to above.
- `mainPath` is the path to the new `require.main`.
- `options.execPath` is the new binary to spawn as. (Default: `process.execPath`)
- `options.execArgv` are the new Node flags to spawn with. (Default: `process.execArgv`)
- `options.argv0` is the binary to spawn as. (Default: `process.argv0`)
- The command is the original binary itself.
- Add a `napi_terminating` member for `napi_status` to represent `try_catch.HasTerminated()` and the result of each call after replacement termination.
- Add a `napi_set_terminate_hook(napi_env env, void (*fun)(void*), void* data)` function to register a callback called on termination, to make it easier to clean up resources.
Internally, there are two cases you need to cover, and the simulated part for Windows is where it gets really hairy due to all the edge cases. Here's pseudocode for the basic algorithm (I'm not really familiar with Node internals, so take this as a rough guideline):
1. Stop the main event loop.
1. Go through the standard shutdown routine.
1. Destroy any open libuv handles and cancel any remaining event loop tasks.
1. If we're on a platform that supports process replacement (like Linux or Mac):
1. Invoke `execve` or equivalent with the new process path, arguments, and environment.
1. Else, if we're on Windows (the only supported OS that doesn't), we have to simulate it entirely:
1. Terminate execution via `v8::V8::TerminateExecution()`. All N-API callbacks should return `napi_terminated` during this step.
1. For each loaded native module:
1. If the native module has a terminate hook, call it.
1. Unload the native module's DLL.
1. Close the event loop.
1. Dispose the isolate.
1. Do the rest according to whatever happens to [this SO question](https://stackoverflow.com/questions/51185115/what-is-the-ideal-way-to-emulate-process-replacement-on-windows).
1. Else, on other OSs without a process replacement function, it'd look similar to Windows.
In addition, file system requests will have to generally create each file descriptor with `O_CLOEXEC`.
As for precedent where this could be used immediately:
- [Liftoff](https://www.npmjs.com/package/liftoff) works very similarly, just with a little extra opinionated sugar, and that's used natively in Gulp. This kind of thing would speed that up quite a bit.
- I do [very similar](https://github.com/isiahmeadows/thallium/blob/master/cli.js#L126-L150) to transparently pass through unknown Node flags.
- Babel [attempts to use `kexec`](https://github.com/babel/babel/blob/master/packages/babel-node/src/babel-node.js#L87-L88) where available, which [is a POSIX-only module that replaces the process literally](https://www.npmjs.com/package/kexec). Absent that, it falls back to [its own implementation](https://github.com/babel/babel/blob/master/packages/babel-node/src/babel-node.js#L90-L109) that works like the other two examples.
|
process
|
ability to replace current node process with another edit if someone can come up with a better shim for execve for windows that d be far better the form below is very expensive and very horrible edit linked relevant edit clarify fs changes edit here s the text from that so question as of july so you don t have to search for it where i asked about how to do the windows part click to show warning lots of text so in a i was looking for a way to replace the current node process with another in linux and friends really any posix compliant system this is easy use and friends and call it a day but obviously that won t work on windows since it only has createprocess which execve and friends delegate to and it s not like leading to this isn t a duplicate because it s explicitly seeking a workaround given certain constraints not just asking for direct replacement process replacement has several facets that have to addressed all console i o streams have to be forwarded to the new process all signals need transparently forwarded to the new process the data from the old process have to be destroyed with as many resources reclaimed as possible all pre existing threads and child processes should be destroyed all pre existing handles should be destroyed apart from open file descriptors and named pipes etc optimally the old process s memory should be kept to a minimum after the process is created for my particular use case retaining the process id is not important and for my particular case there are a few constraints i can control the initial process s startup as well as the location of my process replacement function i could load arbitrary native code via add ons at potentially any stack offset implication i can t even dream of tracking malloc calls handles thread manipulation or process manipulation to track and free them all since dll rewriting isn t exactly practical i have no control over when my process replacement is called it could be called through an add on which could ve been called through either interpreted code via ffi or even another add on recursively it could even be called during add on initialization implication i would have no ability to know what s in the stack even if i perfectly instrumented my side and rewriting all their call s and push es is far from practical and would just be all around slow for obvious reasons so here s the gist of what i was thinking use something similar to a pseudo trampoline statically allocate the following a single pointer for the stack pointer max path chars for the application path max path chars for the current working directory path chars for the arguments chars for the environment on entry set the global stack pointer reference to the stack pointer on replacement do relevant process cleanup and lock release everything you can set the stack pointer to the stored original global one terminate each child thread kill each child process free if possible i e not in a uwp program if it s not the or the temporary heap if it exists if possible close if possible the default heap and each segment associated with it create a new process with the statically allocated file arguments environment etc with no new window created proxy all future received signals exceptions etc without modification to this process somehow but not so much with the exceptions wait for the process to end return with the idea here is to use a process based trampoline and drop the current process size to an absolute minimum while the newly created one is started but where i m not very familiar with windows i probably made quite a few mistakes here also the above seems extremely inefficient and to an extent it just feels horribly wrong for something a kernel could just release a few memory pages deallocate a bunch of memory handles and move some memory around for the next process so to summarize what s the ideal way to emulate process replacement on windows with the fewest limitations i would like a means to replace the current node process with another keeping the same process id it would be something morally similar to but it wouldn t return this would be most useful for conditionally replacing node flags in a startup script for example if someone wants to enable modules and your behavior needs to change non trivially in the presence of them like if you need to install a default loader you ll want to respawn the process with experimental modules loader so you can install the loader this is also for scenarios when you want to run a module as a main module if you want to do logic after the process ends you should be using child process spawn regardless you shouldn t be attempting to replace it in any capacity here s what i propose child process replacespawn command command is the path to the new command args is the args to replace the arguments with this defaults to the empty array options is for the various options for replacing the process this defaults to an empty object options cwd is the new cwd to use default process cwd options env is the new environment to use default process env options is the binary to spawn as default command child process replacefork mainpath works similarly to above mainpath is the path to the new require main options execpath is the new binary to spawn as default process execpath options execargv are the new node flags to spawn with default process execargv options is the binary to spawn as default process the command is the original binary itself add a napi terminating member for napi status to represent try catch hasterminated and the result of each call after replacement termination add a napi set terminate hook napi env env void fun void void data function to register a callback called on termination to make it easier to clean up resources internally there are two cases you need to cover and the simulated part for windows is where it gets really hairy due to all the edge cases here s pseudocode for the basic algorithm i m not really familiar with node internals so take this as a rough guideline stop the main event loop go through the standard shutdown routine destroy any open libuv handles and cancel any remaining event loop tasks if we re on a platform that supports process replacement like linux or mac invoke execve or equivalent with the new process path arguments and environment else if we re on windows the only supported os that doesn t we have to simulate it entirely terminate execution via terminateexecution all n api callbacks should return napi terminated during this step for each loaded native module if the native module has a terminate hook call it unload the native module s dll close the event loop dispose the isolate do the rest according to whatever happens to else on other oss without a process replacement function it d look similar to windows in addition file system requests will have to generally create each file descriptor with o cloexec as for precedent where this could be used immediately works very similarly just with a little extra opinionated sugar and that s used natively in gulp this kind of thing would speed that up quite a bit i do to transparently pass through unknown node flags babel where available which absent that it falls back to that works like the other two examples
| 1
|
17,034
| 22,409,207,055
|
IssuesEvent
|
2022-06-18 13:05:51
|
RobertCraigie/prisma-client-py
|
https://api.github.com/repos/RobertCraigie/prisma-client-py
|
opened
|
Do not override already set env variables from `.env`
|
bug/2-confirmed kind/bug process/candidate topic: client level/beginner priority/high
|
## Problem
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
Currently the `.env` variables take precedence over the system environment variables, this can cause issues as the Prisma CLI will use the system environment variables instead which could lead to migrations being applied to a different database if you have two different connection strings set.
## Suggested solution
<!-- A clear and concise description of what you want to happen. -->
System environment variables should take priority.
|
1.0
|
Do not override already set env variables from `.env` - ## Problem
<!-- A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] -->
Currently the `.env` variables take precedence over the system environment variables, this can cause issues as the Prisma CLI will use the system environment variables instead which could lead to migrations being applied to a different database if you have two different connection strings set.
## Suggested solution
<!-- A clear and concise description of what you want to happen. -->
System environment variables should take priority.
|
process
|
do not override already set env variables from env problem currently the env variables take precedence over the system environment variables this can cause issues as the prisma cli will use the system environment variables instead which could lead to migrations being applied to a different database if you have two different connection strings set suggested solution system environment variables should take priority
| 1
|
11,492
| 14,366,340,810
|
IssuesEvent
|
2020-12-01 04:08:07
|
aodn/imos-toolbox
|
https://api.github.com/repos/aodn/imos-toolbox
|
opened
|
New PP/QC routines from CSIRO
|
Unit:Processing Unit:QC workshop day
|
This is just a placeholder for the merge of the PP/QC routines from https://github.com/aodn/imos-toolbox/pull/683
|
1.0
|
New PP/QC routines from CSIRO - This is just a placeholder for the merge of the PP/QC routines from https://github.com/aodn/imos-toolbox/pull/683
|
process
|
new pp qc routines from csiro this is just a placeholder for the merge of the pp qc routines from
| 1
|
22,466
| 31,273,166,196
|
IssuesEvent
|
2023-08-22 02:52:34
|
q191201771/lal
|
https://api.github.com/repos/q191201771/lal
|
closed
|
重复的Access-Control-Allow-Origin
|
*In process #Opt
|
通过hls拉流m3u8 接口会默认添加Access-Control-Allow-Origin:*,
但是我通过ngingx转发也会给这个接口加上自己的Access-Control-Allow-Origin。会导致头部重复,请问如何去掉lal自带的Access-Control-Allow-Origin。
<img width="979" alt="image" src="https://github.com/q191201771/lal/assets/9177693/8078f246-e722-4b57-8ec5-801f9d011d02">

|
1.0
|
重复的Access-Control-Allow-Origin - 通过hls拉流m3u8 接口会默认添加Access-Control-Allow-Origin:*,
但是我通过ngingx转发也会给这个接口加上自己的Access-Control-Allow-Origin。会导致头部重复,请问如何去掉lal自带的Access-Control-Allow-Origin。
<img width="979" alt="image" src="https://github.com/q191201771/lal/assets/9177693/8078f246-e722-4b57-8ec5-801f9d011d02">

|
process
|
重复的access control allow origin 接口会默认添加access control allow origin: , 但是我通过ngingx转发也会给这个接口加上自己的access control allow origin。会导致头部重复,请问如何去掉lal自带的access control allow origin。 img width alt image src
| 1
|
56,009
| 31,475,376,779
|
IssuesEvent
|
2023-08-30 10:19:00
|
GeopJr/Tuba
|
https://api.github.com/repos/GeopJr/Tuba
|
closed
|
[Feature request]: Status widget needs to be refactored
|
bug status performance
|
### Describe the bug
The more I think about `Widgets.Status`, the more I agree that its architecture has become questionable.
Currently, each `Widgets.Status` has a pre-instanced `Widgets.VoteBox`, `Widgets.AttachmentBox`, and emoji `Gtk.FlowBox`. These widgets stay even if their status entity doesn't really need them, so they just hang there being invisible. There really is no need to create so many useless widgets.
Ideally, the entire content box should be generated dynamically on demand (when bound to a `API.Status`) to avoid creating a larger memory footprint. Some widgets can be skipped altogether if the account backend doesn't support some features (like emoji reactions).
To make things worse, Notification widgets derive from `Widget.Status`, so they all succumb to this behavior as well (even though it makes no sense for a follow request to have a `VoteBox` at its disposal).
### Steps To Reproduce
Inspect https://github.com/GeopJr/Tooth/blob/main/data/ui/widgets/status.ui#L220-L226 and https://github.com/GeopJr/Tooth/blob/main/data/ui/widgets/status.ui#L257-L264
### Logs and/or Screenshots
_No response_
### Instance Backend
Mastodon
### Operating System
Pop!_OS 22.04 LTS
### Package
Flatpak
### Troubleshooting information
flatpak: false
version: main-15b8f4a (development)
gtk: 4.8.4 (4.8.4)
libadwaita: 1.2.1 (1.2.1)
libsoup: 2.74.2 (2.74.2)
### Additional Context
_No response_
|
True
|
[Feature request]: Status widget needs to be refactored - ### Describe the bug
The more I think about `Widgets.Status`, the more I agree that its architecture has become questionable.
Currently, each `Widgets.Status` has a pre-instanced `Widgets.VoteBox`, `Widgets.AttachmentBox`, and emoji `Gtk.FlowBox`. These widgets stay even if their status entity doesn't really need them, so they just hang there being invisible. There really is no need to create so many useless widgets.
Ideally, the entire content box should be generated dynamically on demand (when bound to a `API.Status`) to avoid creating a larger memory footprint. Some widgets can be skipped altogether if the account backend doesn't support some features (like emoji reactions).
To make things worse, Notification widgets derive from `Widget.Status`, so they all succumb to this behavior as well (even though it makes no sense for a follow request to have a `VoteBox` at its disposal).
### Steps To Reproduce
Inspect https://github.com/GeopJr/Tooth/blob/main/data/ui/widgets/status.ui#L220-L226 and https://github.com/GeopJr/Tooth/blob/main/data/ui/widgets/status.ui#L257-L264
### Logs and/or Screenshots
_No response_
### Instance Backend
Mastodon
### Operating System
Pop!_OS 22.04 LTS
### Package
Flatpak
### Troubleshooting information
flatpak: false
version: main-15b8f4a (development)
gtk: 4.8.4 (4.8.4)
libadwaita: 1.2.1 (1.2.1)
libsoup: 2.74.2 (2.74.2)
### Additional Context
_No response_
|
non_process
|
status widget needs to be refactored describe the bug the more i think about widgets status the more i agree that its architecture has become questionable currently each widgets status has a pre instanced widgets votebox widgets attachmentbox and emoji gtk flowbox these widgets stay even if their status entity doesn t really need them so they just hang there being invisible there really is no need to create so many useless widgets ideally the entire content box should be generated dynamically on demand when bound to a api status to avoid creating a larger memory footprint some widgets can be skipped altogether if the account backend doesn t support some features like emoji reactions to make things worse notification widgets derive from widget status so they all succumb to this behavior as well even though it makes no sense for a follow request to have a votebox at its disposal steps to reproduce inspect and logs and or screenshots no response instance backend mastodon operating system pop os lts package flatpak troubleshooting information flatpak false version main development gtk libadwaita libsoup additional context no response
| 0
|
6,955
| 10,113,956,276
|
IssuesEvent
|
2019-07-30 18:00:10
|
material-components/material-components-ios
|
https://api.github.com/repos/material-components/material-components-ios
|
closed
|
Internal issue: b/114705309
|
[schemes] type:Process
|
This is an internal issue. If you are a Googler, please visit [b/114705309](http://b/114705309) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/114705309](http://b/114705309)
|
1.0
|
Internal issue: b/114705309 - This is an internal issue. If you are a Googler, please visit [b/114705309](http://b/114705309) for more details.
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/114705309](http://b/114705309)
|
process
|
internal issue b this is an internal issue if you are a googler please visit for more details internal data associated internal bug
| 1
|
51,965
| 3,016,118,709
|
IssuesEvent
|
2015-07-29 23:31:29
|
archon810/androidpolice
|
https://api.github.com/repos/archon810/androidpolice
|
closed
|
no pushbullet subscriptions buttons are visible in new designed page
|
auto-migrated Priority-Medium Type-Enhancement
|
```
follow the summary
```
Original issue reported on code.google.com by `gopinathms2012` on 1 Feb 2015 at 7:15
|
1.0
|
no pushbullet subscriptions buttons are visible in new designed page - ```
follow the summary
```
Original issue reported on code.google.com by `gopinathms2012` on 1 Feb 2015 at 7:15
|
non_process
|
no pushbullet subscriptions buttons are visible in new designed page follow the summary original issue reported on code google com by on feb at
| 0
|
130,374
| 12,427,465,055
|
IssuesEvent
|
2020-05-25 02:31:30
|
guilherme-adesouza/fretz-freedom
|
https://api.github.com/repos/guilherme-adesouza/fretz-freedom
|
closed
|
Criar apresentação de testes de segurança
|
documentation
|
Criar apresentação de slides explicando os testes de segurança e a(s) ferramenta(s) e medidas utilizadas.
|
1.0
|
Criar apresentação de testes de segurança - Criar apresentação de slides explicando os testes de segurança e a(s) ferramenta(s) e medidas utilizadas.
|
non_process
|
criar apresentação de testes de segurança criar apresentação de slides explicando os testes de segurança e a s ferramenta s e medidas utilizadas
| 0
|
81,769
| 15,800,929,544
|
IssuesEvent
|
2021-04-03 02:01:18
|
PyCQA/flake8
|
https://api.github.com/repos/PyCQA/flake8
|
closed
|
noqa is not respected for E241
|
bug:invalid component:docs component:pycodestyle
|
In GitLab by @gforcada on Jun 3, 2015, 08:53
If you have something like this:
```
dictionary = {
'key1' : 'value1', # noqa
'long_key2': 'value2',
}
```
The ``# noqa`` comment is ignored and a E241 is reported.
|
1.0
|
noqa is not respected for E241 - In GitLab by @gforcada on Jun 3, 2015, 08:53
If you have something like this:
```
dictionary = {
'key1' : 'value1', # noqa
'long_key2': 'value2',
}
```
The ``# noqa`` comment is ignored and a E241 is reported.
|
non_process
|
noqa is not respected for in gitlab by gforcada on jun if you have something like this dictionary noqa long the noqa comment is ignored and a is reported
| 0
|
14,418
| 17,467,088,028
|
IssuesEvent
|
2021-08-06 18:32:11
|
2i2c-org/pilot-hubs
|
https://api.github.com/repos/2i2c-org/pilot-hubs
|
opened
|
Improve our understanding of expected workloads per hub
|
type: enhancement :label: team-process :label: CI/CD
|
# Summary
IN https://github.com/2i2c-org/pilot-hubs/issues/524 we found that the hub's workload scaled to a higher degree than we were expecting when setting up the hub. We should collect better information from Hub Representatives about the expected usage for their hubs, so that we can better estimate what resources to plan ahead for them.
# User Stories
- As a Hub Representative and as a Hub Engineer, I want to know that the hub's environment/resources have been provisioned according to what I'll actually use, so that nothing unexpected happens when we scale.
# Tasks to complete
- [ ] Understand what information we are missing with our current "New Hub Template"
- [ ] Update the template so that we capture this extra information
|
1.0
|
Improve our understanding of expected workloads per hub - # Summary
IN https://github.com/2i2c-org/pilot-hubs/issues/524 we found that the hub's workload scaled to a higher degree than we were expecting when setting up the hub. We should collect better information from Hub Representatives about the expected usage for their hubs, so that we can better estimate what resources to plan ahead for them.
# User Stories
- As a Hub Representative and as a Hub Engineer, I want to know that the hub's environment/resources have been provisioned according to what I'll actually use, so that nothing unexpected happens when we scale.
# Tasks to complete
- [ ] Understand what information we are missing with our current "New Hub Template"
- [ ] Update the template so that we capture this extra information
|
process
|
improve our understanding of expected workloads per hub summary in we found that the hub s workload scaled to a higher degree than we were expecting when setting up the hub we should collect better information from hub representatives about the expected usage for their hubs so that we can better estimate what resources to plan ahead for them user stories as a hub representative and as a hub engineer i want to know that the hub s environment resources have been provisioned according to what i ll actually use so that nothing unexpected happens when we scale tasks to complete understand what information we are missing with our current new hub template update the template so that we capture this extra information
| 1
|
20,871
| 27,658,688,896
|
IssuesEvent
|
2023-03-12 09:09:00
|
polarismesh/polaris
|
https://api.github.com/repos/polarismesh/polaris
|
closed
|
web管理端无法删除openapi接口发布的配置文件
|
bug config in processed
|
**Describe the bug**
web管理端无法删除openapi接口发布的配置文件

**To Reproduce**
```shell
# 创建新的配置文件
curl -X "POST" "http://localhost:8090/config/v1/configfiles" \
-H 'X-Polaris-Token: nu/0WRA4EqSR1FagrjRj0fZwPXuGlMpX+zCuWu4uMqy8xr1vRjisSbA25aAC3mtU8MeeRsKhQiDAynUR09I=' \
-H 'Content-Type: application/json; charset=utf-8' \
-d $'{
"group": "test",
"namespace": "default",
"content": "redis.cache.age=10",
"name": "app/properties.yml"
}'
# 发布该配置文件
curl -X "POST" "http://localhost:8090/config/v1/configfiles/release" \
-H 'X-Polaris-Token: nu/0WRA4EqSR1FagrjRj0fZwPXuGlMpX+zCuWu4uMqy8xr1vRjisSbA25aAC3mtU8MeeRsKhQiDAynUR09I=' \
-H 'Content-Type: application/json; charset=utf-8' \
-d $'{
"namespace": "default",
"group": "test",
"fileName": "app/properties.yml"
}'
```
**Expected behavior**
```text
// polaris-default-error.log
2023-03-09T11:42:11.582147Z error http/handler.go:167 code:<value:500001 > info:<value:"store layer exception" > {"request-id": ""}
```
**Environment**
- Version: polarismesh/polaris-server-standalone:latest
- OS: docker.
**Additional context**
Add any other context about the problem here.
|
1.0
|
web管理端无法删除openapi接口发布的配置文件 - **Describe the bug**
web管理端无法删除openapi接口发布的配置文件

**To Reproduce**
```shell
# 创建新的配置文件
curl -X "POST" "http://localhost:8090/config/v1/configfiles" \
-H 'X-Polaris-Token: nu/0WRA4EqSR1FagrjRj0fZwPXuGlMpX+zCuWu4uMqy8xr1vRjisSbA25aAC3mtU8MeeRsKhQiDAynUR09I=' \
-H 'Content-Type: application/json; charset=utf-8' \
-d $'{
"group": "test",
"namespace": "default",
"content": "redis.cache.age=10",
"name": "app/properties.yml"
}'
# 发布该配置文件
curl -X "POST" "http://localhost:8090/config/v1/configfiles/release" \
-H 'X-Polaris-Token: nu/0WRA4EqSR1FagrjRj0fZwPXuGlMpX+zCuWu4uMqy8xr1vRjisSbA25aAC3mtU8MeeRsKhQiDAynUR09I=' \
-H 'Content-Type: application/json; charset=utf-8' \
-d $'{
"namespace": "default",
"group": "test",
"fileName": "app/properties.yml"
}'
```
**Expected behavior**
```text
// polaris-default-error.log
2023-03-09T11:42:11.582147Z error http/handler.go:167 code:<value:500001 > info:<value:"store layer exception" > {"request-id": ""}
```
**Environment**
- Version: polarismesh/polaris-server-standalone:latest
- OS: docker.
**Additional context**
Add any other context about the problem here.
|
process
|
web管理端无法删除openapi接口发布的配置文件 describe the bug web管理端无法删除openapi接口发布的配置文件 to reproduce shell 创建新的配置文件 curl x post h x polaris token nu h content type application json charset utf d group test namespace default content redis cache age name app properties yml 发布该配置文件 curl x post h x polaris token nu h content type application json charset utf d namespace default group test filename app properties yml expected behavior text polaris default error log error http handler go code info request id environment version polarismesh polaris server standalone latest os docker additional context add any other context about the problem here
| 1
|
259,622
| 27,700,065,207
|
IssuesEvent
|
2023-03-14 07:12:06
|
UrielProd/big-data-plugin
|
https://api.github.com/repos/UrielProd/big-data-plugin
|
opened
|
commons-beanutils-1.9.3.jar: 2 vulnerabilities (highest severity is: 7.3)
|
Mend: dependency security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p></summary>
<p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p>
<p>Path to dependency file: /impl/cluster/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (commons-beanutils version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2019-10086](https://www.mend.io/vulnerability-database/CVE-2019-10086) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.9.3.jar | Direct | 1.9.4 | ✅ |
| [CVE-2014-0114](https://www.mend.io/vulnerability-database/CVE-2014-0114) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.9.3.jar | Direct | 1.9.4 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10086</summary>
### Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p>
<p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p>
<p>Path to dependency file: /impl/cluster/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean.
<p>Publish Date: 2019-08-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10086>CVE-2019-10086</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-08-20</p>
<p>Fix Resolution: 1.9.4</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2014-0114</summary>
### Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p>
<p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p>
<p>Path to dependency file: /impl/cluster/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.
<p>Publish Date: 2014-04-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-0114>CVE-2014-0114</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p>
<p>Release Date: 2014-04-30</p>
<p>Fix Resolution: 1.9.4</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
True
|
commons-beanutils-1.9.3.jar: 2 vulnerabilities (highest severity is: 7.3) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p></summary>
<p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p>
<p>Path to dependency file: /impl/cluster/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (commons-beanutils version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2019-10086](https://www.mend.io/vulnerability-database/CVE-2019-10086) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.9.3.jar | Direct | 1.9.4 | ✅ |
| [CVE-2014-0114](https://www.mend.io/vulnerability-database/CVE-2014-0114) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | commons-beanutils-1.9.3.jar | Direct | 1.9.4 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10086</summary>
### Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p>
<p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p>
<p>Path to dependency file: /impl/cluster/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean.
<p>Publish Date: 2019-08-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-10086>CVE-2019-10086</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2019-08-20</p>
<p>Fix Resolution: 1.9.4</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2014-0114</summary>
### Vulnerable Library - <b>commons-beanutils-1.9.3.jar</b></p>
<p>Apache Commons BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-beanutils/">https://commons.apache.org/proper/commons-beanutils/</a></p>
<p>Path to dependency file: /impl/cluster/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-beanutils/commons-beanutils/1.9.3/commons-beanutils-1.9.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **commons-beanutils-1.9.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/UrielProd/big-data-plugin/commit/62b1abe76efb4b5a5b583beaac54243fa17cfab0">62b1abe76efb4b5a5b583beaac54243fa17cfab0</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Apache Commons BeanUtils, as distributed in lib/commons-beanutils-1.8.0.jar in Apache Struts 1.x through 1.3.10 and in other products requiring commons-beanutils through 1.9.2, does not suppress the class property, which allows remote attackers to "manipulate" the ClassLoader and execute arbitrary code via the class parameter, as demonstrated by the passing of this parameter to the getClass method of the ActionForm object in Struts 1.
<p>Publish Date: 2014-04-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2014-0114>CVE-2014-0114</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-0114</a></p>
<p>Release Date: 2014-04-30</p>
<p>Fix Resolution: 1.9.4</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
non_process
|
commons beanutils jar vulnerabilities highest severity is vulnerable library commons beanutils jar apache commons beanutils provides an easy to use but flexible wrapper around reflection and introspection library home page a href path to dependency file impl cluster pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in commons beanutils version remediation available high commons beanutils jar direct high commons beanutils jar direct details cve vulnerable library commons beanutils jar apache commons beanutils provides an easy to use but flexible wrapper around reflection and introspection library home page a href path to dependency file impl cluster pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy x commons beanutils jar vulnerable library found in head commit a href found in base branch master vulnerability details in apache commons beanutils a special beanintrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all java objects we however were not using this by default characteristic of the propertyutilsbean publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version release date fix resolution rescue worker helmet automatic remediation is available for this issue cve vulnerable library commons beanutils jar apache commons beanutils provides an easy to use but flexible wrapper around reflection and introspection library home page a href path to dependency file impl cluster pom xml path to vulnerable library home wss scanner repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy x commons beanutils jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons beanutils as distributed in lib commons beanutils jar in apache struts x through and in other products requiring commons beanutils through does not suppress the class property which allows remote attackers to manipulate the classloader and execute arbitrary code via the class parameter as demonstrated by the passing of this parameter to the getclass method of the actionform object in struts publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
| 0
|
435,620
| 30,510,349,185
|
IssuesEvent
|
2023-07-18 20:18:13
|
cilium/cilium
|
https://api.github.com/repos/cilium/cilium
|
opened
|
Documentation style: follow K8s guidance on capitalizing API objects
|
area/documentation
|
Opened after reviewing #25945.
I think Cilium docs should follow the [Kubernetes style guide](https://kubernetes.io/docs/contribute/style/style-guide/#use-upper-camel-case-for-api-objects) for when to (not) capitalize API objects. Specifically:
> When you refer specifically to interacting with an API object, use [UpperCamelCase](https://en.wikipedia.org/wiki/Camel_case), also known as Pascal case.
And:
> When you are generally discussing an API object, use [sentence-style capitalization](https://docs.microsoft.com/en-us/style-guide/text-formatting/using-type/use-sentence-style-capitalization).
In the case of #25945, I'd like to see "Gateway API" capitalized, "Gateway" when talking about an API object as an entity, and "gateway" when talking about a specific instance. For example, this line would read:
```rst
Cilium only programs gateways with the class ``cilium``.
```
This issue is non-blocking for #25945. I mention it specifically because it's a good example of a larger phenomenon I see in Cilium API docs in general.
@qmonnet 👋🏻 I'd love your feedback here
cc @learnitall for visibility
|
1.0
|
Documentation style: follow K8s guidance on capitalizing API objects - Opened after reviewing #25945.
I think Cilium docs should follow the [Kubernetes style guide](https://kubernetes.io/docs/contribute/style/style-guide/#use-upper-camel-case-for-api-objects) for when to (not) capitalize API objects. Specifically:
> When you refer specifically to interacting with an API object, use [UpperCamelCase](https://en.wikipedia.org/wiki/Camel_case), also known as Pascal case.
And:
> When you are generally discussing an API object, use [sentence-style capitalization](https://docs.microsoft.com/en-us/style-guide/text-formatting/using-type/use-sentence-style-capitalization).
In the case of #25945, I'd like to see "Gateway API" capitalized, "Gateway" when talking about an API object as an entity, and "gateway" when talking about a specific instance. For example, this line would read:
```rst
Cilium only programs gateways with the class ``cilium``.
```
This issue is non-blocking for #25945. I mention it specifically because it's a good example of a larger phenomenon I see in Cilium API docs in general.
@qmonnet 👋🏻 I'd love your feedback here
cc @learnitall for visibility
|
non_process
|
documentation style follow guidance on capitalizing api objects opened after reviewing i think cilium docs should follow the for when to not capitalize api objects specifically when you refer specifically to interacting with an api object use also known as pascal case and when you are generally discussing an api object use in the case of i d like to see gateway api capitalized gateway when talking about an api object as an entity and gateway when talking about a specific instance for example this line would read rst cilium only programs gateways with the class cilium this issue is non blocking for i mention it specifically because it s a good example of a larger phenomenon i see in cilium api docs in general qmonnet 👋🏻 i d love your feedback here cc learnitall for visibility
| 0
|
23,172
| 10,852,343,693
|
IssuesEvent
|
2019-11-13 12:37:07
|
ChenLuigi/TestingPOM
|
https://api.github.com/repos/ChenLuigi/TestingPOM
|
opened
|
CVE-2017-5648 (High) detected in tomcat-catalina-7.0.42.jar, tomcat-embed-core-8.0.18.jar
|
security vulnerability
|
## CVE-2017-5648 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-catalina-7.0.42.jar</b>, <b>tomcat-embed-core-8.0.18.jar</b></p></summary>
<p>
<details><summary><b>tomcat-catalina-7.0.42.jar</b></p></summary>
<p>Tomcat Servlet Engine Core Classes and Standard implementations</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-embed-core-8.0.18.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-embed-core-8.0.18.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
While investigating bug 60718, it was noticed that some calls to application listeners in Apache Tomcat 9.0.0.M1 to 9.0.0.M17, 8.5.0 to 8.5.11, 8.0.0.RC1 to 8.0.41, and 7.0.0 to 7.0.75 did not use the appropriate facade object. When running an untrusted application under a SecurityManager, it was therefore possible for that untrusted application to retain a reference to the request or response object and thereby access and/or modify information associated with another web application.
<p>Publish Date: 2017-04-17
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648>CVE-2017-5648</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648</a></p>
<p>Release Date: 2017-04-17</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M18,8.5.12,8.0.42,7.0.76,org.apache.tomcat:tomcat-catalina:9.0.0.M18,8.5.12,8.0.42,7.0.76</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2017-5648 (High) detected in tomcat-catalina-7.0.42.jar, tomcat-embed-core-8.0.18.jar - ## CVE-2017-5648 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tomcat-catalina-7.0.42.jar</b>, <b>tomcat-embed-core-8.0.18.jar</b></p></summary>
<p>
<details><summary><b>tomcat-catalina-7.0.42.jar</b></p></summary>
<p>Tomcat Servlet Engine Core Classes and Standard implementations</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-catalina-7.0.42.jar** (Vulnerable Library)
</details>
<details><summary><b>tomcat-embed-core-8.0.18.jar</b></p></summary>
<p>Core Tomcat implementation</p>
<p>
Dependency Hierarchy:
- :x: **tomcat-embed-core-8.0.18.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/ChenLuigi/TestingPOM/commit/503cb25bceffc9eeb6aa89180177d9d618d2e5b8">503cb25bceffc9eeb6aa89180177d9d618d2e5b8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
While investigating bug 60718, it was noticed that some calls to application listeners in Apache Tomcat 9.0.0.M1 to 9.0.0.M17, 8.5.0 to 8.5.11, 8.0.0.RC1 to 8.0.41, and 7.0.0 to 7.0.75 did not use the appropriate facade object. When running an untrusted application under a SecurityManager, it was therefore possible for that untrusted application to retain a reference to the request or response object and thereby access and/or modify information associated with another web application.
<p>Publish Date: 2017-04-17
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648>CVE-2017-5648</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-5648</a></p>
<p>Release Date: 2017-04-17</p>
<p>Fix Resolution: org.apache.tomcat.embed:tomcat-embed-core:9.0.0.M18,8.5.12,8.0.42,7.0.76,org.apache.tomcat:tomcat-catalina:9.0.0.M18,8.5.12,8.0.42,7.0.76</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in tomcat catalina jar tomcat embed core jar cve high severity vulnerability vulnerable libraries tomcat catalina jar tomcat embed core jar tomcat catalina jar tomcat servlet engine core classes and standard implementations dependency hierarchy x tomcat catalina jar vulnerable library tomcat embed core jar core tomcat implementation dependency hierarchy x tomcat embed core jar vulnerable library found in head commit a href vulnerability details while investigating bug it was noticed that some calls to application listeners in apache tomcat to to to and to did not use the appropriate facade object when running an untrusted application under a securitymanager it was therefore possible for that untrusted application to retain a reference to the request or response object and thereby access and or modify information associated with another web application publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache tomcat embed tomcat embed core org apache tomcat tomcat catalina step up your open source security game with whitesource
| 0
|
19,334
| 25,472,602,873
|
IssuesEvent
|
2022-11-25 11:31:27
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[IDP][PM][MFA Enabled] Admin is not authenticated with 2 factor authentication even though MFA is enabled
|
Bug Blocker P0 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
**Pre-condition:** IDP and MFA should be enabled in the PM
**Steps:**
1. Add organizational user in the participant manager
2. Complete set up your account process
3. Sign in with registered credentials and Verify
**AR:** Admin is not authenticated with 2-factor authentication even though MFA is enabled
**ER:** Admin should be authenticated with 2-factor authentication if MFA is enabled
|
3.0
|
[IDP][PM][MFA Enabled] Admin is not authenticated with 2 factor authentication even though MFA is enabled - **Pre-condition:** IDP and MFA should be enabled in the PM
**Steps:**
1. Add organizational user in the participant manager
2. Complete set up your account process
3. Sign in with registered credentials and Verify
**AR:** Admin is not authenticated with 2-factor authentication even though MFA is enabled
**ER:** Admin should be authenticated with 2-factor authentication if MFA is enabled
|
process
|
admin is not authenticated with factor authentication even though mfa is enabled pre condition idp and mfa should be enabled in the pm steps add organizational user in the participant manager complete set up your account process sign in with registered credentials and verify ar admin is not authenticated with factor authentication even though mfa is enabled er admin should be authenticated with factor authentication if mfa is enabled
| 1
|
42,868
| 5,478,204,975
|
IssuesEvent
|
2017-03-12 16:02:58
|
mattstone22133/JavaNetworkLibrary
|
https://api.github.com/repos/mattstone22133/JavaNetworkLibrary
|
opened
|
Network Black Box Test - Test Mode Changing
|
medium length story test
|
testModeChanging - these tests will take some time because multiple disonnects are required
startClient_changeToServer_Listen_ChangeToClient_Join
startServer_changeToClient_Join_Disconnect_ChangeToServer_listenAndAccept_changeBackToClient_repeat
multiple changes (including duplicates to test for realistic behavior
Server -> Server -> client -> client -> server -> client
client -> server -> client -> client -> server -> server
|
1.0
|
Network Black Box Test - Test Mode Changing - testModeChanging - these tests will take some time because multiple disonnects are required
startClient_changeToServer_Listen_ChangeToClient_Join
startServer_changeToClient_Join_Disconnect_ChangeToServer_listenAndAccept_changeBackToClient_repeat
multiple changes (including duplicates to test for realistic behavior
Server -> Server -> client -> client -> server -> client
client -> server -> client -> client -> server -> server
|
non_process
|
network black box test test mode changing testmodechanging these tests will take some time because multiple disonnects are required startclient changetoserver listen changetoclient join startserver changetoclient join disconnect changetoserver listenandaccept changebacktoclient repeat multiple changes including duplicates to test for realistic behavior server server client client server client client server client client server server
| 0
|
295,866
| 22,275,213,890
|
IssuesEvent
|
2022-06-10 15:56:52
|
aws/aws-cdk
|
https://api.github.com/repos/aws/aws-cdk
|
closed
|
(aws-autoscaling): Lambda autoscaling example uses an alias to latestVersion, which doesn't work
|
good first issue p2 @aws-cdk/aws-autoscaling effort/small documentation
|
### link to reference doc page
https://docs.aws.amazon.com/cdk/api/v2/docs/aws-cdk-lib.aws_lambda.AutoScalingOptions.html
### Describe your issue?
When following the example in the docs to setup auto-scaling provisioned concurrency on a lambda function, the docs explicitly state in the example to first add an alias on the lambda function's `latestVersion`. I followed the example, setup an alias to `latestVersion`, added a basic `scaleOnUtilization` target, and then deployed the stack. The deploy went through without errors.
After deploy, I checked the lambda in the AWS Console and noticed there was no provisioned concurrency setup for it. I drilled down into the alias associated with the lambda that I created from the example and promptly saw:

Did I misunderstand something in the docs or is this a bad example? Looking at the docs for `Version` I see it's example uses the `Function` method `addVersion` - and the `addVersion` method docs tell me is deprecated and points me to use `function.currentVersion`. If this is not a misunderstanding on my part, can these be updated to whatever the current best practice for setting up basic provisioned concurrency auto-scaling on the most recently deployed lambda version is?
|
1.0
|
(aws-autoscaling): Lambda autoscaling example uses an alias to latestVersion, which doesn't work - ### link to reference doc page
https://docs.aws.amazon.com/cdk/api/v2/docs/aws-cdk-lib.aws_lambda.AutoScalingOptions.html
### Describe your issue?
When following the example in the docs to setup auto-scaling provisioned concurrency on a lambda function, the docs explicitly state in the example to first add an alias on the lambda function's `latestVersion`. I followed the example, setup an alias to `latestVersion`, added a basic `scaleOnUtilization` target, and then deployed the stack. The deploy went through without errors.
After deploy, I checked the lambda in the AWS Console and noticed there was no provisioned concurrency setup for it. I drilled down into the alias associated with the lambda that I created from the example and promptly saw:

Did I misunderstand something in the docs or is this a bad example? Looking at the docs for `Version` I see it's example uses the `Function` method `addVersion` - and the `addVersion` method docs tell me is deprecated and points me to use `function.currentVersion`. If this is not a misunderstanding on my part, can these be updated to whatever the current best practice for setting up basic provisioned concurrency auto-scaling on the most recently deployed lambda version is?
|
non_process
|
aws autoscaling lambda autoscaling example uses an alias to latestversion which doesn t work link to reference doc page describe your issue when following the example in the docs to setup auto scaling provisioned concurrency on a lambda function the docs explicitly state in the example to first add an alias on the lambda function s latestversion i followed the example setup an alias to latestversion added a basic scaleonutilization target and then deployed the stack the deploy went through without errors after deploy i checked the lambda in the aws console and noticed there was no provisioned concurrency setup for it i drilled down into the alias associated with the lambda that i created from the example and promptly saw did i misunderstand something in the docs or is this a bad example looking at the docs for version i see it s example uses the function method addversion and the addversion method docs tell me is deprecated and points me to use function currentversion if this is not a misunderstanding on my part can these be updated to whatever the current best practice for setting up basic provisioned concurrency auto scaling on the most recently deployed lambda version is
| 0
|
322,875
| 27,640,896,897
|
IssuesEvent
|
2023-03-10 17:57:40
|
Roukys/HHauto
|
https://api.github.com/repos/Roukys/HHauto
|
closed
|
5.6.136 = Testing auto boss bang option.
|
new feature to be tested
|

no autofill girls here - we can do it by myself
just an option start fight, end fight, start new and etc for the end - the end of HP girls or victory and winning event girl here.
|
1.0
|
5.6.136 = Testing auto boss bang option. - 
no autofill girls here - we can do it by myself
just an option start fight, end fight, start new and etc for the end - the end of HP girls or victory and winning event girl here.
|
non_process
|
testing auto boss bang option no autofill girls here we can do it by myself just an option start fight end fight start new and etc for the end the end of hp girls or victory and winning event girl here
| 0
|
174,143
| 14,448,182,588
|
IssuesEvent
|
2020-12-08 05:41:35
|
zmkfirmware/zmk
|
https://api.github.com/repos/zmkfirmware/zmk
|
closed
|
CMake issue when using 3.19+
|
bug core documentation good first issue
|
The build docs specifically mention that our cmake installs shouldn't be "old", but it's only a lower limit at 3.15. But after I installed a fresh cmake v3.19.1, it failed the build saying something along the lines of:
`in function 'exit': exit.c:(.text.exit+0x60): undefined reference to '_exit'`
Tracked this to a [related bug report](https://github.com/zephyrproject-rtos/zephyr/issues/22060) where the consensus seemed to be that this is caused by cmake 3.19 (either 3.19.0 or 3.19.1). Downgrading to 3.18.5 did indeed help. This might be worth noting in the docs @mcrosson...
|
1.0
|
CMake issue when using 3.19+ - The build docs specifically mention that our cmake installs shouldn't be "old", but it's only a lower limit at 3.15. But after I installed a fresh cmake v3.19.1, it failed the build saying something along the lines of:
`in function 'exit': exit.c:(.text.exit+0x60): undefined reference to '_exit'`
Tracked this to a [related bug report](https://github.com/zephyrproject-rtos/zephyr/issues/22060) where the consensus seemed to be that this is caused by cmake 3.19 (either 3.19.0 or 3.19.1). Downgrading to 3.18.5 did indeed help. This might be worth noting in the docs @mcrosson...
|
non_process
|
cmake issue when using the build docs specifically mention that our cmake installs shouldn t be old but it s only a lower limit at but after i installed a fresh cmake it failed the build saying something along the lines of in function exit exit c text exit undefined reference to exit tracked this to a where the consensus seemed to be that this is caused by cmake either or downgrading to did indeed help this might be worth noting in the docs mcrosson
| 0
|
176,851
| 6,565,797,706
|
IssuesEvent
|
2017-09-08 09:47:26
|
hobbit-project/platform
|
https://api.github.com/repos/hobbit-project/platform
|
closed
|
Queue status not shown when controller creates new experiment
|
component: controller priority: high type: bug
|
## Problem
It is possible that the queue status is not shown if the controller is currently starting a new experiment.

This small window in which the controller is not responding is even extended, when the controller starts to prefetch docker images.
## Cause
The problem might be caused by the controller trying to access the currently running experiment data which is secured by a mutex. When the controller is creating a new experiment, this mutex might be owned by a different thread.
|
1.0
|
Queue status not shown when controller creates new experiment - ## Problem
It is possible that the queue status is not shown if the controller is currently starting a new experiment.

This small window in which the controller is not responding is even extended, when the controller starts to prefetch docker images.
## Cause
The problem might be caused by the controller trying to access the currently running experiment data which is secured by a mutex. When the controller is creating a new experiment, this mutex might be owned by a different thread.
|
non_process
|
queue status not shown when controller creates new experiment problem it is possible that the queue status is not shown if the controller is currently starting a new experiment this small window in which the controller is not responding is even extended when the controller starts to prefetch docker images cause the problem might be caused by the controller trying to access the currently running experiment data which is secured by a mutex when the controller is creating a new experiment this mutex might be owned by a different thread
| 0
|
132,577
| 18,751,850,972
|
IssuesEvent
|
2021-11-05 03:46:17
|
hackforla/food-oasis
|
https://api.github.com/repos/hackforla/food-oasis
|
reopened
|
Experiment with creating a widget to embed Food Oasis in a web site
|
Release Note: Enhancement Role: Design UI/UX Role: Front-end 3pt
|
### Overview
We have had several requests to provide a Food Oasis widget that could be embedded in an external organization's web site. Explore the options and implications
### Action Items
- [ ] Look into the functionality provided by an IFrame or other implementation.
- Can the behavior be tailored to the site by customizing the query, allowing some styling, etc.?
- How can this be linked to attribute Food Oasis and re-direct to Food Oasis properly
- [ ] How can we design to work with the Mapbox license agreement to avoid absorbing the charges incurred on the host site by mapbox requests?
- [ ] Can we limit the number of map calls (which cost money) per embedded site so that we don't go broke hosting maps on other sites.
- [ ] Can we generate a static snapshot (ideally hyperlinked) map that links to our site for cost efficiency that can be put on external sites.
### Resources/Instructions
Requests have been made by
Julia Blanton (Food Oasis Santa Barbara)
Nancy Beyda (Food Cycle LA)
|
1.0
|
Experiment with creating a widget to embed Food Oasis in a web site - ### Overview
We have had several requests to provide a Food Oasis widget that could be embedded in an external organization's web site. Explore the options and implications
### Action Items
- [ ] Look into the functionality provided by an IFrame or other implementation.
- Can the behavior be tailored to the site by customizing the query, allowing some styling, etc.?
- How can this be linked to attribute Food Oasis and re-direct to Food Oasis properly
- [ ] How can we design to work with the Mapbox license agreement to avoid absorbing the charges incurred on the host site by mapbox requests?
- [ ] Can we limit the number of map calls (which cost money) per embedded site so that we don't go broke hosting maps on other sites.
- [ ] Can we generate a static snapshot (ideally hyperlinked) map that links to our site for cost efficiency that can be put on external sites.
### Resources/Instructions
Requests have been made by
Julia Blanton (Food Oasis Santa Barbara)
Nancy Beyda (Food Cycle LA)
|
non_process
|
experiment with creating a widget to embed food oasis in a web site overview we have had several requests to provide a food oasis widget that could be embedded in an external organization s web site explore the options and implications action items look into the functionality provided by an iframe or other implementation can the behavior be tailored to the site by customizing the query allowing some styling etc how can this be linked to attribute food oasis and re direct to food oasis properly how can we design to work with the mapbox license agreement to avoid absorbing the charges incurred on the host site by mapbox requests can we limit the number of map calls which cost money per embedded site so that we don t go broke hosting maps on other sites can we generate a static snapshot ideally hyperlinked map that links to our site for cost efficiency that can be put on external sites resources instructions requests have been made by julia blanton food oasis santa barbara nancy beyda food cycle la
| 0
|
43,165
| 2,885,717,061
|
IssuesEvent
|
2015-06-12 00:32:32
|
parallaxinc/ChromeIDE
|
https://api.github.com/repos/parallaxinc/ChromeIDE
|
closed
|
File dialog should set focus on filename field automatically
|
Priority 1
|
The New File dialog should put focus on filename field automatically once opened.
Request: A very slight delay (after dialog appearance) followed by a focus-event animation (perhaps an expanding and translucent-fading circle on the field) to simulate the appearance of a user click/touch on the field, followed by the existing "filename" label slide-shrink animation.
If it's possible for a previous entry to appear (like what happens after an Esc key press on a previous dialog, then the cursor should appear at the end of the field's filename text.
|
1.0
|
File dialog should set focus on filename field automatically - The New File dialog should put focus on filename field automatically once opened.
Request: A very slight delay (after dialog appearance) followed by a focus-event animation (perhaps an expanding and translucent-fading circle on the field) to simulate the appearance of a user click/touch on the field, followed by the existing "filename" label slide-shrink animation.
If it's possible for a previous entry to appear (like what happens after an Esc key press on a previous dialog, then the cursor should appear at the end of the field's filename text.
|
non_process
|
file dialog should set focus on filename field automatically the new file dialog should put focus on filename field automatically once opened request a very slight delay after dialog appearance followed by a focus event animation perhaps an expanding and translucent fading circle on the field to simulate the appearance of a user click touch on the field followed by the existing filename label slide shrink animation if it s possible for a previous entry to appear like what happens after an esc key press on a previous dialog then the cursor should appear at the end of the field s filename text
| 0
|
42,479
| 2,870,709,729
|
IssuesEvent
|
2015-06-07 12:50:43
|
bobbybee/uPresent
|
https://api.github.com/repos/bobbybee/uPresent
|
closed
|
Combining bold and italics _**like this**_ won't compile
|
bug medium-priority
|
It's worth noting `_ **text** _` works
|
1.0
|
Combining bold and italics _**like this**_ won't compile - It's worth noting `_ **text** _` works
|
non_process
|
combining bold and italics like this won t compile it s worth noting text works
| 0
|
25,255
| 12,231,359,090
|
IssuesEvent
|
2020-05-04 07:35:45
|
hashicorp/consul
|
https://api.github.com/repos/hashicorp/consul
|
closed
|
Support server selection priorities / weights
|
theme/service-metadata type/enhancement
|
If we support priories, then you can in support cases like a server in a remote region (for backup / quorum purposes). The priorities would disable client routing through them unless necessary (higher priority servers have failed / are unreachable).
|
1.0
|
Support server selection priorities / weights - If we support priories, then you can in support cases like a server in a remote region (for backup / quorum purposes). The priorities would disable client routing through them unless necessary (higher priority servers have failed / are unreachable).
|
non_process
|
support server selection priorities weights if we support priories then you can in support cases like a server in a remote region for backup quorum purposes the priorities would disable client routing through them unless necessary higher priority servers have failed are unreachable
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.