Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
665
labels
stringlengths
4
554
body
stringlengths
3
235k
index
stringclasses
6 values
text_combine
stringlengths
96
235k
label
stringclasses
2 values
text
stringlengths
96
196k
binary_label
int64
0
1
85,671
15,755,128,957
IssuesEvent
2021-03-31 01:14:09
phytomichael/KSA1
https://api.github.com/repos/phytomichael/KSA1
opened
CVE-2021-23358 (High) detected in underscore-1.9.1.tgz
security vulnerability
## CVE-2021-23358 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore-1.9.1.tgz</b></p></summary> <p>JavaScript's functional programming helper library.</p> <p>Library home page: <a href="https://registry.npmjs.org/underscore/-/underscore-1.9.1.tgz">https://registry.npmjs.org/underscore/-/underscore-1.9.1.tgz</a></p> <p>Path to dependency file: /KSA1/ksa/ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json</p> <p>Path to vulnerable library: KSA1/ksa/ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/node_modules/underscore/package.json</p> <p> Dependency Hierarchy: - recess-1.0.3.tgz (Root Library) - :x: **underscore-1.9.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package underscore from 1.13.0-0 and before 1.13.0-2, from 1.3.2 and before 1.12.1 are vulnerable to Arbitrary Code Execution via the template function, particularly when a variable property is passed as an argument as it is not sanitized. <p>Publish Date: 2021-03-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23358>CVE-2021-23358</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358</a></p> <p>Release Date: 2021-03-29</p> <p>Fix Resolution: underscore - 1.12.1,1.13.0-2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"underscore","packageVersion":"1.9.1","packageFilePaths":["/KSA1/ksa/ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json"],"isTransitiveDependency":true,"dependencyTree":"recess:1.0.3;underscore:1.9.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"underscore - 1.12.1,1.13.0-2"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2021-23358","vulnerabilityDetails":"The package underscore from 1.13.0-0 and before 1.13.0-2, from 1.3.2 and before 1.12.1 are vulnerable to Arbitrary Code Execution via the template function, particularly when a variable property is passed as an argument as it is not sanitized.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23358","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-23358 (High) detected in underscore-1.9.1.tgz - ## CVE-2021-23358 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore-1.9.1.tgz</b></p></summary> <p>JavaScript's functional programming helper library.</p> <p>Library home page: <a href="https://registry.npmjs.org/underscore/-/underscore-1.9.1.tgz">https://registry.npmjs.org/underscore/-/underscore-1.9.1.tgz</a></p> <p>Path to dependency file: /KSA1/ksa/ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json</p> <p>Path to vulnerable library: KSA1/ksa/ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/node_modules/underscore/package.json</p> <p> Dependency Hierarchy: - recess-1.0.3.tgz (Root Library) - :x: **underscore-1.9.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package underscore from 1.13.0-0 and before 1.13.0-2, from 1.3.2 and before 1.12.1 are vulnerable to Arbitrary Code Execution via the template function, particularly when a variable property is passed as an argument as it is not sanitized. <p>Publish Date: 2021-03-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23358>CVE-2021-23358</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23358</a></p> <p>Release Date: 2021-03-29</p> <p>Fix Resolution: underscore - 1.12.1,1.13.0-2</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"underscore","packageVersion":"1.9.1","packageFilePaths":["/KSA1/ksa/ksa/ksa-web-root/ksa-web/src/main/webapp/rs/bootstrap/package.json"],"isTransitiveDependency":true,"dependencyTree":"recess:1.0.3;underscore:1.9.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"underscore - 1.12.1,1.13.0-2"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2021-23358","vulnerabilityDetails":"The package underscore from 1.13.0-0 and before 1.13.0-2, from 1.3.2 and before 1.12.1 are vulnerable to Arbitrary Code Execution via the template function, particularly when a variable property is passed as an argument as it is not sanitized.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23358","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_infrastructure
cve high detected in underscore tgz cve high severity vulnerability vulnerable library underscore tgz javascript s functional programming helper library library home page a href path to dependency file ksa ksa ksa web root ksa web src main webapp rs bootstrap package json path to vulnerable library ksa ksa ksa web root ksa web src main webapp rs bootstrap node modules underscore package json dependency hierarchy recess tgz root library x underscore tgz vulnerable library vulnerability details the package underscore from and before from and before are vulnerable to arbitrary code execution via the template function particularly when a variable property is passed as an argument as it is not sanitized publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution underscore isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree recess underscore isminimumfixversionavailable true minimumfixversion underscore basebranches vulnerabilityidentifier cve vulnerabilitydetails the package underscore from and before from and before are vulnerable to arbitrary code execution via the template function particularly when a variable property is passed as an argument as it is not sanitized vulnerabilityurl
0
12,278
9,667,806,060
IssuesEvent
2019-05-21 13:58:23
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
[InfraUI] Add object nested property accessor function / utility
:Infra UI :infrastructure [zube]: Backlog
In this [PR](https://github.com/elastic/kibana/pull/30826#pullrequestreview-202762767) it was discussed that accessing nested object properties (which may or may not exist) can be a bit of a pain. After further discussion it was decided that we'll [copy APM's solution](https://github.com/elastic/kibana/blob/master/x-pack/plugins/apm/common/idx.ts) for the most part, but with the small enhancement that you can offer a default value. There is no direct urgency for this issue, whoever has the need to access deeply nested properties next should implement this.
1.0
[InfraUI] Add object nested property accessor function / utility - In this [PR](https://github.com/elastic/kibana/pull/30826#pullrequestreview-202762767) it was discussed that accessing nested object properties (which may or may not exist) can be a bit of a pain. After further discussion it was decided that we'll [copy APM's solution](https://github.com/elastic/kibana/blob/master/x-pack/plugins/apm/common/idx.ts) for the most part, but with the small enhancement that you can offer a default value. There is no direct urgency for this issue, whoever has the need to access deeply nested properties next should implement this.
infrastructure
add object nested property accessor function utility in this it was discussed that accessing nested object properties which may or may not exist can be a bit of a pain after further discussion it was decided that we ll for the most part but with the small enhancement that you can offer a default value there is no direct urgency for this issue whoever has the need to access deeply nested properties next should implement this
1
34,603
6,353,503,507
IssuesEvent
2017-07-29 00:04:49
kalabox/lando
https://api.github.com/repos/kalabox/lando
closed
Refactor/reorg docs
documentation
Let's do a pass on the docs so we can update appropriate sections with the new `lando init` command. Particularly we should look at... * Getting Started Section * Add a table of contents to help people pick/choose starting point * Starting w/existing codebase * Pulling codebase from Github * Pulling codebase from Pantheon These rely upon the creation of other docs that we don't have yet: * Configuring lando.yml * Using init command * Manually creating lando.yml * Links to configuration docs * Configuring... * Recipes * Tooling * Services * Advanced Topics/Follow-up * Relevant tutorials * Add helpful tutorials to README/index.html
1.0
Refactor/reorg docs - Let's do a pass on the docs so we can update appropriate sections with the new `lando init` command. Particularly we should look at... * Getting Started Section * Add a table of contents to help people pick/choose starting point * Starting w/existing codebase * Pulling codebase from Github * Pulling codebase from Pantheon These rely upon the creation of other docs that we don't have yet: * Configuring lando.yml * Using init command * Manually creating lando.yml * Links to configuration docs * Configuring... * Recipes * Tooling * Services * Advanced Topics/Follow-up * Relevant tutorials * Add helpful tutorials to README/index.html
non_infrastructure
refactor reorg docs let s do a pass on the docs so we can update appropriate sections with the new lando init command particularly we should look at getting started section add a table of contents to help people pick choose starting point starting w existing codebase pulling codebase from github pulling codebase from pantheon these rely upon the creation of other docs that we don t have yet configuring lando yml using init command manually creating lando yml links to configuration docs configuring recipes tooling services advanced topics follow up relevant tutorials add helpful tutorials to readme index html
0
13,714
10,430,380,754
IssuesEvent
2019-09-17 06:34:50
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
closed
Persistent cache system [Blocked: #6]
Priority: Essential Status: Blocked Type: Improvement Where: Infrastructure
A replacement is needed to SharedPreferences for storing settings and generally caching data since SharedPreferences executes on the main thread which can block the UI rendering and cause ANRs. It's instead recommended to use a proto-backed storage mechanism that relies on background threads (likely via the threading library described below) to perform the I/O. Using proto allows the data structure to be trivially updated to a new structure upon app update without requiring an upgrade flow. Such a system should be strict-mode compliant (that is, does not perform any I/O or long-running operations on the main thread).
1.0
Persistent cache system [Blocked: #6] - A replacement is needed to SharedPreferences for storing settings and generally caching data since SharedPreferences executes on the main thread which can block the UI rendering and cause ANRs. It's instead recommended to use a proto-backed storage mechanism that relies on background threads (likely via the threading library described below) to perform the I/O. Using proto allows the data structure to be trivially updated to a new structure upon app update without requiring an upgrade flow. Such a system should be strict-mode compliant (that is, does not perform any I/O or long-running operations on the main thread).
infrastructure
persistent cache system a replacement is needed to sharedpreferences for storing settings and generally caching data since sharedpreferences executes on the main thread which can block the ui rendering and cause anrs it s instead recommended to use a proto backed storage mechanism that relies on background threads likely via the threading library described below to perform the i o using proto allows the data structure to be trivially updated to a new structure upon app update without requiring an upgrade flow such a system should be strict mode compliant that is does not perform any i o or long running operations on the main thread
1
143,460
11,566,006,439
IssuesEvent
2020-02-20 11:36:58
dgraph-io/badger
https://api.github.com/repos/dgraph-io/badger
closed
ExampleDB_Subscribe test is flaky
area/testing exp/intermediate good first issue kind/maintenance platform/windows priority/P3 status/stale
The test `ExampleDB_Subscribe` https://github.com/dgraph-io/badger/blob/00039ea84781c1c5a219dd22a2e420d082bf4f0f/db_test.go#L1904-L1963 doesn't always work. See https://ci.appveyor.com/project/manishrjain/badger/builds/28179570 for failure. ``` === RUN ExampleDB_Subscribe badger 2019/10/17 13:35:32 INFO: All 0 tables opened in 0s 2019/10/17 13:35:32 stopping subscription 2019/10/17 13:35:32 waiting for subscription to close 2019/10/17 13:35:32 subscription closed badger 2019/10/17 13:35:32 DEBUG: Storing value log head: {Fid:0 Len:29 Offset:107} badger 2019/10/17 13:35:32 INFO: Got compaction priority: {level:0 score:1.73 dropPrefix:[]} badger 2019/10/17 13:35:32 INFO: Running for level: 0 badger 2019/10/17 13:35:32 DEBUG: LOG Compact. Added 3 keys. Skipped 0 keys. Iteration took: 1.0012ms badger 2019/10/17 13:35:32 DEBUG: Discard stats: map[] badger 2019/10/17 13:35:32 INFO: LOG Compact 0->1, del 1 tables, add 1 tables, took 3.0026ms badger 2019/10/17 13:35:32 INFO: Compaction for level: 0 DONE badger 2019/10/17 13:35:32 INFO: Force compaction on level 0 done --- FAIL: ExampleDB_Subscribe (0.02s) got: want: a-key is now set to a-value FAIL ```
1.0
ExampleDB_Subscribe test is flaky - The test `ExampleDB_Subscribe` https://github.com/dgraph-io/badger/blob/00039ea84781c1c5a219dd22a2e420d082bf4f0f/db_test.go#L1904-L1963 doesn't always work. See https://ci.appveyor.com/project/manishrjain/badger/builds/28179570 for failure. ``` === RUN ExampleDB_Subscribe badger 2019/10/17 13:35:32 INFO: All 0 tables opened in 0s 2019/10/17 13:35:32 stopping subscription 2019/10/17 13:35:32 waiting for subscription to close 2019/10/17 13:35:32 subscription closed badger 2019/10/17 13:35:32 DEBUG: Storing value log head: {Fid:0 Len:29 Offset:107} badger 2019/10/17 13:35:32 INFO: Got compaction priority: {level:0 score:1.73 dropPrefix:[]} badger 2019/10/17 13:35:32 INFO: Running for level: 0 badger 2019/10/17 13:35:32 DEBUG: LOG Compact. Added 3 keys. Skipped 0 keys. Iteration took: 1.0012ms badger 2019/10/17 13:35:32 DEBUG: Discard stats: map[] badger 2019/10/17 13:35:32 INFO: LOG Compact 0->1, del 1 tables, add 1 tables, took 3.0026ms badger 2019/10/17 13:35:32 INFO: Compaction for level: 0 DONE badger 2019/10/17 13:35:32 INFO: Force compaction on level 0 done --- FAIL: ExampleDB_Subscribe (0.02s) got: want: a-key is now set to a-value FAIL ```
non_infrastructure
exampledb subscribe test is flaky the test exampledb subscribe doesn t always work see for failure run exampledb subscribe badger info all tables opened in stopping subscription waiting for subscription to close subscription closed badger debug storing value log head fid len offset badger info got compaction priority level score dropprefix badger info running for level badger debug log compact added keys skipped keys iteration took badger debug discard stats map badger info log compact del tables add tables took badger info compaction for level done badger info force compaction on level done fail exampledb subscribe got want a key is now set to a value fail
0
13,061
21,795,593,953
IssuesEvent
2022-05-15 15:27:41
coders-camp-2021-best-team/FitaTAM
https://api.github.com/repos/coders-camp-2021-best-team/FitaTAM
opened
requirement/feed-tracking
requirement
**Base description** As a user I want to track all of my added products at specific time/date, also add products to specific period of time (breakfast etc.) **User action steps** - User is able to check all of his previous "days" - User is able to add product to the breakfast etc. - User can see his overall eaten calories, carbons etc. for each day - User is not able to scroll forward (the limit is a current day) **Hints (optional)** We shouldn't bother with timezones etc. just stick to the ours (Poland ~ GMT + 2)
1.0
requirement/feed-tracking - **Base description** As a user I want to track all of my added products at specific time/date, also add products to specific period of time (breakfast etc.) **User action steps** - User is able to check all of his previous "days" - User is able to add product to the breakfast etc. - User can see his overall eaten calories, carbons etc. for each day - User is not able to scroll forward (the limit is a current day) **Hints (optional)** We shouldn't bother with timezones etc. just stick to the ours (Poland ~ GMT + 2)
non_infrastructure
requirement feed tracking base description as a user i want to track all of my added products at specific time date also add products to specific period of time breakfast etc user action steps user is able to check all of his previous days user is able to add product to the breakfast etc user can see his overall eaten calories carbons etc for each day user is not able to scroll forward the limit is a current day hints optional we shouldn t bother with timezones etc just stick to the ours poland gmt
0
200,333
22,763,093,801
IssuesEvent
2022-07-07 23:37:02
bcgov/cloud-pathfinder
https://api.github.com/repos/bcgov/cloud-pathfinder
closed
SecOps-Exchange log offload in Azure - Review for possible onboarding
Security CSP: Azure User: SecOps CPF Member Onboarding
**Describe the issue** A proposed project by the SecOps and NCCS Exchange team as a solution to the problem of required 13 month log retention. Exchange online keeps logs for approximately 30 days and SecOps' Azure Sentinel SIEM service can keep them for up to 90 days, which may either leave around 8 or 9 months of retention that needs to be addressed **Additional context** - Per Dale L in a recent meeting with Julian S, he expressed that 13 months retention for cloud logs minimum - Further details will be had at initial project meeting - Will discuss the tech involved, caveats and possible options **Definition of done** - PO would like to ride along in these meetings - Surmise options after initial email-based review of situation and send feedback to the SecOps+NCCS teams - Have a meeting to discuss this and any particular caveats around what could be done to accomodate this extended log retention - Discuss with SecOps if there will be other projects that would need this type of offload from Sential or other services, and if creating a SecOps log repo with various storage tiers organized within and possible automation\auto-tiering the data the primary (hot or cool - allowing for quick ref to this data without IO extra cost, for a period of time), then the secondary (cool) and possibly an Archive layer - Review any automation services MS may suggest as easiest for these tasks or have them inquire if not yet available - Post-analysis Technical Review meeting to be held seperately and discuss any remaining questions\decisions - Depending on option chosen, the either config of Sentinel for extended storage or the IaC deploy of the RG and storage needed
True
SecOps-Exchange log offload in Azure - Review for possible onboarding - **Describe the issue** A proposed project by the SecOps and NCCS Exchange team as a solution to the problem of required 13 month log retention. Exchange online keeps logs for approximately 30 days and SecOps' Azure Sentinel SIEM service can keep them for up to 90 days, which may either leave around 8 or 9 months of retention that needs to be addressed **Additional context** - Per Dale L in a recent meeting with Julian S, he expressed that 13 months retention for cloud logs minimum - Further details will be had at initial project meeting - Will discuss the tech involved, caveats and possible options **Definition of done** - PO would like to ride along in these meetings - Surmise options after initial email-based review of situation and send feedback to the SecOps+NCCS teams - Have a meeting to discuss this and any particular caveats around what could be done to accomodate this extended log retention - Discuss with SecOps if there will be other projects that would need this type of offload from Sential or other services, and if creating a SecOps log repo with various storage tiers organized within and possible automation\auto-tiering the data the primary (hot or cool - allowing for quick ref to this data without IO extra cost, for a period of time), then the secondary (cool) and possibly an Archive layer - Review any automation services MS may suggest as easiest for these tasks or have them inquire if not yet available - Post-analysis Technical Review meeting to be held seperately and discuss any remaining questions\decisions - Depending on option chosen, the either config of Sentinel for extended storage or the IaC deploy of the RG and storage needed
non_infrastructure
secops exchange log offload in azure review for possible onboarding describe the issue a proposed project by the secops and nccs exchange team as a solution to the problem of required month log retention exchange online keeps logs for approximately days and secops azure sentinel siem service can keep them for up to days which may either leave around or months of retention that needs to be addressed additional context per dale l in a recent meeting with julian s he expressed that months retention for cloud logs minimum further details will be had at initial project meeting will discuss the tech involved caveats and possible options definition of done po would like to ride along in these meetings surmise options after initial email based review of situation and send feedback to the secops nccs teams have a meeting to discuss this and any particular caveats around what could be done to accomodate this extended log retention discuss with secops if there will be other projects that would need this type of offload from sential or other services and if creating a secops log repo with various storage tiers organized within and possible automation auto tiering the data the primary hot or cool allowing for quick ref to this data without io extra cost for a period of time then the secondary cool and possibly an archive layer review any automation services ms may suggest as easiest for these tasks or have them inquire if not yet available post analysis technical review meeting to be held seperately and discuss any remaining questions decisions depending on option chosen the either config of sentinel for extended storage or the iac deploy of the rg and storage needed
0
49,758
13,462,578,327
IssuesEvent
2020-09-09 16:16:01
LevyForchh/symphony-java-api
https://api.github.com/repos/LevyForchh/symphony-java-api
opened
CVE-2019-16942 (High) detected in jackson-databind-2.8.4.jar
security vulnerability
## CVE-2019-16942 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/symphony-java-api/authenticator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.4/jackson-databind-2.8.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.4/jackson-databind-2.8.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.4/jackson-databind-2.8.4.jar</p> <p> Dependency Hierarchy: - jackson-datatype-jsr310-2.8.4.jar (Root Library) - :x: **jackson-databind-2.8.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LevyForchh/symphony-java-api/commit/0ff7c2c88343428195277225d1875a1c606c1c4b">0ff7c2c88343428195277225d1875a1c606c1c4b</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942>CVE-2019-16942</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.4","isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.datatype:jackson-datatype-jsr310:2.8.4;com.fasterxml.jackson.core:jackson-databind:2.8.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1"}],"vulnerabilityIdentifier":"CVE-2019-16942","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-16942 (High) detected in jackson-databind-2.8.4.jar - ## CVE-2019-16942 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/symphony-java-api/authenticator/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.4/jackson-databind-2.8.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.4/jackson-databind-2.8.4.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.4/jackson-databind-2.8.4.jar</p> <p> Dependency Hierarchy: - jackson-datatype-jsr310-2.8.4.jar (Root Library) - :x: **jackson-databind-2.8.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/LevyForchh/symphony-java-api/commit/0ff7c2c88343428195277225d1875a1c606c1c4b">0ff7c2c88343428195277225d1875a1c606c1c4b</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling. <p>Publish Date: 2019-10-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942>CVE-2019-16942</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-16942</a></p> <p>Release Date: 2019-10-01</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.4","isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.datatype:jackson-datatype-jsr310:2.8.4;com.fasterxml.jackson.core:jackson-databind:2.8.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.6.7.3,2.7.9.7,2.8.11.5,2.9.10.1"}],"vulnerabilityIdentifier":"CVE-2019-16942","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the commons-dbcp (1.4) jar in the classpath, and an attacker can find an RMI service endpoint to access, it is possible to make the service execute a malicious payload. This issue exists because of org.apache.commons.dbcp.datasources.SharedPoolDataSource and org.apache.commons.dbcp.datasources.PerUserPoolDataSource mishandling.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-16942","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_infrastructure
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm symphony java api authenticator pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy jackson datatype jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the commons dbcp jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of org apache commons dbcp datasources sharedpooldatasource and org apache commons dbcp datasources peruserpooldatasource mishandling publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the commons dbcp jar in the classpath and an attacker can find an rmi service endpoint to access it is possible to make the service execute a malicious payload this issue exists because of org apache commons dbcp datasources sharedpooldatasource and org apache commons dbcp datasources peruserpooldatasource mishandling vulnerabilityurl
0
339,020
10,240,673,413
IssuesEvent
2019-08-19 21:26:52
coding-eval-platform/evaluations-service
https://api.github.com/repos/coding-eval-platform/evaluations-service
closed
Should we allow deleting an exam (or its exercises and test cases) only by its owner?
Priority: Mid Question
Similar to #89 , but for deleting an exam Continues #18
1.0
Should we allow deleting an exam (or its exercises and test cases) only by its owner? - Similar to #89 , but for deleting an exam Continues #18
non_infrastructure
should we allow deleting an exam or its exercises and test cases only by its owner similar to but for deleting an exam continues
0
13,403
10,239,202,139
IssuesEvent
2019-08-19 17:40:23
algorand/go-algorand
https://api.github.com/repos/algorand/go-algorand
opened
Add PureStake Algorand Relays into tooling (algorelay) so they're automatically controlled
infrastructure
(originally reported @Karmastic) May want to change to a data-driven tool where we combine hand-curated entries with afdb1-sourced entries. Latest from Tim: ```# Points to relay-amsterdam-mainnet-algorand.alg-945.net via CF - 'r1.algorand-mainnet.network' # Points to relay-dublin-mainnet-algorand.alg-945.net via CF - 'relay-dublin-mainnet-algorand.algorand-mainnet.network' # Points to relay-hongkong-mainnet-algorand.alg-945.net via CF - 'relay-hongkong-mainnet-algorand.algorand-mainnet.network' # Points to relay-london-mainnet-algorand.alg-945.net via CF - 'r13.algorand-mainnet.network' # Points to relay-montreal-mainnet-algorand.alg-945.net via CF - 'relay-montreal-mainnet-algorand.algorand-mainnet.network' # Points to relay-montreal2-mainnet-algorand.alg-945.net via CF - 'relay-montreal2-mainnet-algorand.algorand-mainnet.network' # Points to relay-mumbai-mainnet-algorand.alg-945.net via CF - 'relay-mumbai-mainnet-algorand.algorand-mainnet.network' # Points to relay-saopaulo-mainnet-algorand.alg-945.net via CF - 'r7.algorand-mainnet.network' # Points to relay-singapore-mainnet-algorand.alg-945.net via CF - 'relay-singapore-mainnet-algorand.algorand-mainnet.network' # Points to relay-singaporea-mainnet-algorand.alg-945.net via CF - 'relay-singaporea-mainnet-algorand.algorand-mainnet.network' # Points to relay-sydney-mainnet-algorand.alg-945.net via CF - 'r10.algorand-mainnet.network' # Points to relay-tokyo-mainnet-algorand.alg-945.net via CF - 'relay-tokyo-mainnet-algorand.algorand-mainnet.network' # Points to relay-toronto-mainnet-algorand.alg-945.net via CF - 'r12.algorand-mainnet.network'```
1.0
Add PureStake Algorand Relays into tooling (algorelay) so they're automatically controlled - (originally reported @Karmastic) May want to change to a data-driven tool where we combine hand-curated entries with afdb1-sourced entries. Latest from Tim: ```# Points to relay-amsterdam-mainnet-algorand.alg-945.net via CF - 'r1.algorand-mainnet.network' # Points to relay-dublin-mainnet-algorand.alg-945.net via CF - 'relay-dublin-mainnet-algorand.algorand-mainnet.network' # Points to relay-hongkong-mainnet-algorand.alg-945.net via CF - 'relay-hongkong-mainnet-algorand.algorand-mainnet.network' # Points to relay-london-mainnet-algorand.alg-945.net via CF - 'r13.algorand-mainnet.network' # Points to relay-montreal-mainnet-algorand.alg-945.net via CF - 'relay-montreal-mainnet-algorand.algorand-mainnet.network' # Points to relay-montreal2-mainnet-algorand.alg-945.net via CF - 'relay-montreal2-mainnet-algorand.algorand-mainnet.network' # Points to relay-mumbai-mainnet-algorand.alg-945.net via CF - 'relay-mumbai-mainnet-algorand.algorand-mainnet.network' # Points to relay-saopaulo-mainnet-algorand.alg-945.net via CF - 'r7.algorand-mainnet.network' # Points to relay-singapore-mainnet-algorand.alg-945.net via CF - 'relay-singapore-mainnet-algorand.algorand-mainnet.network' # Points to relay-singaporea-mainnet-algorand.alg-945.net via CF - 'relay-singaporea-mainnet-algorand.algorand-mainnet.network' # Points to relay-sydney-mainnet-algorand.alg-945.net via CF - 'r10.algorand-mainnet.network' # Points to relay-tokyo-mainnet-algorand.alg-945.net via CF - 'relay-tokyo-mainnet-algorand.algorand-mainnet.network' # Points to relay-toronto-mainnet-algorand.alg-945.net via CF - 'r12.algorand-mainnet.network'```
infrastructure
add purestake algorand relays into tooling algorelay so they re automatically controlled originally reported karmastic may want to change to a data driven tool where we combine hand curated entries with sourced entries latest from tim points to relay amsterdam mainnet algorand alg net via cf algorand mainnet network points to relay dublin mainnet algorand alg net via cf relay dublin mainnet algorand algorand mainnet network points to relay hongkong mainnet algorand alg net via cf relay hongkong mainnet algorand algorand mainnet network points to relay london mainnet algorand alg net via cf algorand mainnet network points to relay montreal mainnet algorand alg net via cf relay montreal mainnet algorand algorand mainnet network points to relay mainnet algorand alg net via cf relay mainnet algorand algorand mainnet network points to relay mumbai mainnet algorand alg net via cf relay mumbai mainnet algorand algorand mainnet network points to relay saopaulo mainnet algorand alg net via cf algorand mainnet network points to relay singapore mainnet algorand alg net via cf relay singapore mainnet algorand algorand mainnet network points to relay singaporea mainnet algorand alg net via cf relay singaporea mainnet algorand algorand mainnet network points to relay sydney mainnet algorand alg net via cf algorand mainnet network points to relay tokyo mainnet algorand alg net via cf relay tokyo mainnet algorand algorand mainnet network points to relay toronto mainnet algorand alg net via cf algorand mainnet network
1
256,047
22,033,437,534
IssuesEvent
2022-05-28 07:31:13
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: activerecord failed
C-test-failure O-robot O-roachtest release-blocker branch-release-22.1
roachtest.activerecord [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=5299863&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=5299863&tab=artifacts#/activerecord) on release-22.1 @ [4e5b7aef63f7594f7d56eb24362b9bf78b39e9a2](https://github.com/cockroachdb/cockroach/commits/4e5b7aef63f7594f7d56eb24362b9bf78b39e9a2): ``` The test failed on branch=release-22.1, cloud=gce: test artifacts and logs in: /artifacts/activerecord/run_1 orm_helpers.go:193,orm_helpers.go:119,activerecord.go:251,test_runner.go:883: Tests run on Cockroach v22.1.0-beta.4-361-g4e5b7aef63 Tests run against activerecord 6.1.6 6666 Total Tests Run 6665 tests passed 1 test failed 19 tests skipped 5 tests ignored 0 tests passed unexpectedly 1 test failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- PASS: SerializedAttributeTest#test_unexpected_serialized_type (expected) --- FAIL: BasicsTest#test_default_values_are_deeply_dupped (unexpected) For a full summary look at the activerecord artifacts An updated blocklist (activeRecordBlockList22_1) is available in the artifacts' activerecord log ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #81088 roachtest: activerecord failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.2] </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*activerecord.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: activerecord failed - roachtest.activerecord [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=5299863&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=5299863&tab=artifacts#/activerecord) on release-22.1 @ [4e5b7aef63f7594f7d56eb24362b9bf78b39e9a2](https://github.com/cockroachdb/cockroach/commits/4e5b7aef63f7594f7d56eb24362b9bf78b39e9a2): ``` The test failed on branch=release-22.1, cloud=gce: test artifacts and logs in: /artifacts/activerecord/run_1 orm_helpers.go:193,orm_helpers.go:119,activerecord.go:251,test_runner.go:883: Tests run on Cockroach v22.1.0-beta.4-361-g4e5b7aef63 Tests run against activerecord 6.1.6 6666 Total Tests Run 6665 tests passed 1 test failed 19 tests skipped 5 tests ignored 0 tests passed unexpectedly 1 test failed unexpectedly 0 tests expected failed but skipped 0 tests expected failed but not run --- --- PASS: SerializedAttributeTest#test_unexpected_serialized_type (expected) --- FAIL: BasicsTest#test_default_values_are_deeply_dupped (unexpected) For a full summary look at the activerecord artifacts An updated blocklist (activeRecordBlockList22_1) is available in the artifacts' activerecord log ``` <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #81088 roachtest: activerecord failed [C-test-failure O-roachtest O-robot T-sql-experience branch-release-21.2] </p> </details> /cc @cockroachdb/sql-experience <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*activerecord.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_infrastructure
roachtest activerecord failed roachtest activerecord with on release the test failed on branch release cloud gce test artifacts and logs in artifacts activerecord run orm helpers go orm helpers go activerecord go test runner go tests run on cockroach beta tests run against activerecord total tests run tests passed test failed tests skipped tests ignored tests passed unexpectedly test failed unexpectedly tests expected failed but skipped tests expected failed but not run pass serializedattributetest test unexpected serialized type expected fail basicstest test default values are deeply dupped unexpected for a full summary look at the activerecord artifacts an updated blocklist is available in the artifacts activerecord log help see see same failure on other branches roachtest activerecord failed cc cockroachdb sql experience
0
808,147
30,035,644,727
IssuesEvent
2023-06-27 12:35:55
Laravel-Backpack/CRUD
https://api.github.com/repos/Laravel-Backpack/CRUD
opened
Can we remove the Gravatar dependency in v6?
triage Priority: MUST Priority: SHOULD Size: S
In Backpack v5, we're using `creativeorange/gravatar` to show the admin's profile image, in the top-right corner. This is a good feature, because MOST people identify with their profile image - there are studies showing that alone makes the interface more pleasurable (seeing their face or familiar image). However... there are a number of problems with using this package: - (1) it makes a call to a third-party endpoint, on every pageload; so even though we have removed all CDNs, we still have one URL that pings 3rd parties; - (2) this adds `creativeorange/gravatar` as a dependency for CRUD; we've had no issues with them so far, the package has been SUPER well maintained; every time there was a new Laravel version they were quick to add support; they did not introduce any bugs or breaking changes; but it's one dependency to our core package... that we can _probably_ do without; So here's a crazy idea... What if we bring in [the `Gravatar.php` file](https://github.com/creativeorange/gravatar/blob/1.0/src/Gravatar.php) into Backpack? 👀 That would allow us to: - use Basset in it; Basset gets the profile image ONCE, then caches it, so we no longer do a 3rd party ping on each pageload (fixes problem no 1); - remove the `creativeorange/gravatar` dependency; It feel a little off doing this... but after all the whole package is just this one class. And rather than extending it... I think we get more benefit out of copy-pasting it... What do you think @pxpm ? Do you think we should do this? Or am I missing something here?
2.0
Can we remove the Gravatar dependency in v6? - In Backpack v5, we're using `creativeorange/gravatar` to show the admin's profile image, in the top-right corner. This is a good feature, because MOST people identify with their profile image - there are studies showing that alone makes the interface more pleasurable (seeing their face or familiar image). However... there are a number of problems with using this package: - (1) it makes a call to a third-party endpoint, on every pageload; so even though we have removed all CDNs, we still have one URL that pings 3rd parties; - (2) this adds `creativeorange/gravatar` as a dependency for CRUD; we've had no issues with them so far, the package has been SUPER well maintained; every time there was a new Laravel version they were quick to add support; they did not introduce any bugs or breaking changes; but it's one dependency to our core package... that we can _probably_ do without; So here's a crazy idea... What if we bring in [the `Gravatar.php` file](https://github.com/creativeorange/gravatar/blob/1.0/src/Gravatar.php) into Backpack? 👀 That would allow us to: - use Basset in it; Basset gets the profile image ONCE, then caches it, so we no longer do a 3rd party ping on each pageload (fixes problem no 1); - remove the `creativeorange/gravatar` dependency; It feel a little off doing this... but after all the whole package is just this one class. And rather than extending it... I think we get more benefit out of copy-pasting it... What do you think @pxpm ? Do you think we should do this? Or am I missing something here?
non_infrastructure
can we remove the gravatar dependency in in backpack we re using creativeorange gravatar to show the admin s profile image in the top right corner this is a good feature because most people identify with their profile image there are studies showing that alone makes the interface more pleasurable seeing their face or familiar image however there are a number of problems with using this package it makes a call to a third party endpoint on every pageload so even though we have removed all cdns we still have one url that pings parties this adds creativeorange gravatar as a dependency for crud we ve had no issues with them so far the package has been super well maintained every time there was a new laravel version they were quick to add support they did not introduce any bugs or breaking changes but it s one dependency to our core package that we can probably do without so here s a crazy idea what if we bring in into backpack 👀 that would allow us to use basset in it basset gets the profile image once then caches it so we no longer do a party ping on each pageload fixes problem no remove the creativeorange gravatar dependency it feel a little off doing this but after all the whole package is just this one class and rather than extending it i think we get more benefit out of copy pasting it what do you think pxpm do you think we should do this or am i missing something here
0
243,095
20,361,906,406
IssuesEvent
2022-02-20 20:08:16
dagster-io/dagster
https://api.github.com/repos/dagster-io/dagster
closed
Create release pipeline
test systems good second issue platform
Right now, the pipenv test will fail on any new release commit (since the commit will hit GitHub before the packages are available on PyPI). We really should formalize the release process in CI/CD, and add explicit post-release tests for installability, etc., of which this would be one. Include: * post release smoke tests #1818 * (optional) support for non-prod pypi #1958 * Test/validate K8s release deliverables #3179
1.0
Create release pipeline - Right now, the pipenv test will fail on any new release commit (since the commit will hit GitHub before the packages are available on PyPI). We really should formalize the release process in CI/CD, and add explicit post-release tests for installability, etc., of which this would be one. Include: * post release smoke tests #1818 * (optional) support for non-prod pypi #1958 * Test/validate K8s release deliverables #3179
non_infrastructure
create release pipeline right now the pipenv test will fail on any new release commit since the commit will hit github before the packages are available on pypi we really should formalize the release process in ci cd and add explicit post release tests for installability etc of which this would be one include post release smoke tests optional support for non prod pypi test validate release deliverables
0
23,673
16,515,135,850
IssuesEvent
2021-05-26 09:06:05
crystal-lang/crystal
https://api.github.com/repos/crystal-lang/crystal
closed
The rpm has missing dependencies
kind:feature topic:infrastructure
It seems that the official RPM is missing a ton of dependencies, at least in Fedora 33. Check these out: https://download.copr.fedorainfracloud.org/results/zawertun/crystal/fedora-33-x86_64/01906375-crystal/crystal.spec The dependencies are far more that what gets installed when I use the crystal repo. AFAICS, you're missing: gc-devel gpm-devel libxml2-devel libyaml-devel openssl-devel xz-devel My test was trying to build amber framework with the installed crystal version. I've installed v0.35.1 (because 0.36.0 breaks amber). My system's info: <details> ``` * OS Release (cat /etc/*-release | uniq): Fedora release 33 (Thirty Three) NAME=Fedora VERSION="33 (Workstation Edition)" ID=fedora VERSION_ID=33 VERSION_CODENAME="" PLATFORM_ID="platform:f33" PRETTY_NAME="Fedora 33 (Workstation Edition)" ANSI_COLOR="0;38;2;60;110;180" LOGO=fedora-logo-icon CPE_NAME="cpe:/o:fedoraproject:fedora:33" HOME_URL="https://fedoraproject.org/" DOCUMENTATION_URL="https://docs.fedoraproject.org/en-US/fedora/f33/system-administrators-guide/" SUPPORT_URL="https://fedoraproject.org/wiki/Communicating_and_getting_help" BUG_REPORT_URL="https://bugzilla.redhat.com/" REDHAT_BUGZILLA_PRODUCT="Fedora" REDHAT_BUGZILLA_PRODUCT_VERSION=33 REDHAT_SUPPORT_PRODUCT="Fedora" REDHAT_SUPPORT_PRODUCT_VERSION=33 PRIVACY_POLICY_URL="https://fedoraproject.org/wiki/Legal:PrivacyPolicy" VARIANT="Workstation Edition" VARIANT_ID=workstation Fedora release 33 (Thirty Three) * Kernel (uname -r ; cat /proc/cmdline): 5.10.9-201.fc33.x86_64 BOOT_IMAGE=(hd1,gpt2)/vmlinuz-5.10.9-201.fc33.x86_64 root=UUID=934e5565-cbc7-47c1-b488-ed4dfead07cb ro rootflags=subvol=root rhgb quiet processor.max_cstate=5 rcu_nocbs=0-11 * Desktop(s) Running (ps -eo comm= | grep -E '(gnome-session|startkde|startactive|xfce.?-session|fluxbox|blackbox|hackedbox|ratpoison|enlightenment|icewm-session|od-session|wmaker|wmx|openbox-lxde|openbox-gnome-session|openbox-kde-session|mwm|e16|fvwm|xmonad|sugar-session|mate-session|lxqt-session|cinnamon)' ): gnome-session-b gnome-session-c gnome-session-b * Desktop(s) Installed (ls -m /usr/share/xsessions/ | sed 's/\.desktop//g' ): budgie-desktop, gnome-classic, gnome, gnome-xorg * SELinux Status (sestatus): SELinux status: enabled SELinuxfs mount: /sys/fs/selinux SELinux root directory: /etc/selinux Loaded policy name: targeted Current mode: enforcing Mode from config file: enforcing Policy MLS status: enabled Policy deny_unknown status: allowed Memory protection checking: actual (secure) Max kernel policy version: 33 * SELinux Errors (selinuxenabled && journalctl --since yesterday |grep avc: | grep -Eo comm="[^ ]+" | sort |uniq -c |sort -rn): 2 comm="systemd-tmpfile" * CPU Model (grep 'model name' /proc/cpuinfo | awk -F: '{print $2}' | uniq -c | sed -re 's/^ +//' ): 16 AMD Ryzen 7 1800X Eight-Core Processor * 64-bit Support (grep -q ' lm ' /proc/cpuinfo && echo Yes || echo No): Yes * Hardware Virtualization Support (grep -Eq '(vmx|svm)' /proc/cpuinfo && echo Yes || echo No): Yes * Load average (uptime): 17:34:16 up 2 days, 12:16, 1 user, load average: 0.67, 1.18, 1.30 * Memory usage (free -m): total used free shared buff/cache available Mem: 64301 12574 1430 207 50297 50956 Swap: 8191 3 8188 * Top 5 CPU hogs (ps axuScnh | awk '$2!=436183' | sort -rnk3 | head -5): 107 88678 27.8 9.2 21782828 6095088 ? Sl Jan25 701:04 qemu-system-x86 1000 9830 22.9 0.0 26140 15692 ? Ss Jan25 750:25 systemd 1000 418425 21.6 1.1 4657628 776676 ? Sl 17:08 5:34 firefox 1000 423177 14.6 0.0 233804 6704 pts/0 Ss 17:17 2:27 bash 1000 418649 8.4 0.6 3601236 445680 ? Sl 17:08 2:10 Web Content * Top 5 Memory hogs (ps axuScnh | sort -rnk4 | head -5): 107 88678 27.8 9.2 21782828 6095088 ? Sl Jan25 701:04 qemu-system-x86 1000 418425 21.6 1.1 4657628 776676 ? Sl 17:08 5:34 firefox 1000 10426 0.0 1.0 1452188 702940 ? Sl Jan25 0:48 gnome-software 0 1689 0.0 0.9 1184620 612796 ? Ssl Jan25 1:09 packagekitd 1000 418656 2.7 0.6 3439884 451992 ? Sl 17:08 0:42 Web Content * Disk space usage (df): Filesystem 1K-blocks Used Available Use% Mounted on devtmpfs 32864392 0 32864392 0% /dev tmpfs 32922604 35576 32887028 1% /dev/shm tmpfs 13169044 2456 13166588 1% /run /dev/nvme0n1p3 498443264 320413504 176661440 65% / /dev/nvme0n1p3 498443264 320413504 176661440 65% /srv /dev/nvme0n1p3 498443264 320413504 176661440 65% /home /dev/nvme0n1p2 999320 288000 642508 31% /boot /dev/nvme0n1p1 613184 20772 592412 4% /boot/efi tmpfs 32922608 8072 32914536 1% /tmp /dev/sda1 3907016704 601673476 3303447612 16% /var/lib/libvirt/images /dev/nvme0n1p3 498443264 320413504 176661440 65% /var/lib/libvirt/images/fast /dev/sda1 3907016704 601673476 3303447612 16% /home/renich/Downloads /dev/sda1 3907016704 601673476 3303447612 16% /home/renich/VirtualMachines/isos /dev/sda1 3907016704 601673476 3303447612 16% /home/renich/Videos/movies /dev/sda1 3907016704 601673476 3303447612 16% /home/renich/xzone tmpfs 6584520 240 6584280 1% /run/user/1000 * Block devices (blkid): /dev/nvme0n1p3: LABEL="fedora" UUID="934e5565-cbc7-47c1-b488-ed4dfead07cb" UUID_SUB="d07aecb3-ed70-4fea-bce7-f568c999b165" BLOCK_SIZE="4096" TYPE="btrfs" PARTUUID="5a6076d9-174e-46a0-a6f3-38f05ea33d89" /dev/nvme0n1p2: UUID="2919b7e6-0ca5-494e-9ca2-f9502faf6d01" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="55884fcb-1fc5-4157-a3ba-9677fc0dd2d2" /dev/nvme0n1p1: UUID="866B-15BA" BLOCK_SIZE="512" TYPE="vfat" PARTLABEL="EFI System Partition" PARTUUID="1c6865f5-cd08-411d-aaf6-c53c3fca0c57" * PCI devices (lspci -nn): 00:00.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Root Complex [1022:1450] 00:00.2 IOMMU [0806]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) I/O Memory Management Unit [1022:1451] 00:01.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:01.1 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) PCIe GPP Bridge [1022:1453] 00:01.3 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) PCIe GPP Bridge [1022:1453] 00:02.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:03.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:03.1 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) PCIe GPP Bridge [1022:1453] 00:04.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:07.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:07.1 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Internal PCIe GPP Bridge 0 to Bus B [1022:1454] 00:08.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:08.1 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Internal PCIe GPP Bridge 0 to Bus B [1022:1454] 00:14.0 SMBus [0c05]: Advanced Micro Devices, Inc. [AMD] FCH SMBus Controller [1022:790b] (rev 59) 00:14.3 ISA bridge [0601]: Advanced Micro Devices, Inc. [AMD] FCH LPC Bridge [1022:790e] (rev 51) 00:18.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 0 [1022:1460] 00:18.1 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 1 [1022:1461] 00:18.2 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 2 [1022:1462] 00:18.3 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 3 [1022:1463] 00:18.4 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 4 [1022:1464] 00:18.5 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 5 [1022:1465] 00:18.6 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 6 [1022:1466] 00:18.7 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 7 [1022:1467] 01:00.0 Non-Volatile memory controller [0108]: Samsung Electronics Co Ltd NVMe SSD Controller SM961/PM961/SM963 [144d:a804] 02:00.0 USB controller [0c03]: Advanced Micro Devices, Inc. [AMD] X370 Series Chipset USB 3.1 xHCI Controller [1022:43b9] (rev 02) 02:00.1 SATA controller [0106]: Advanced Micro Devices, Inc. [AMD] X370 Series Chipset SATA Controller [1022:43b5] (rev 02) 02:00.2 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] X370 Series Chipset PCIe Upstream Port [1022:43b0] (rev 02) 03:00.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:02.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:03.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:04.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:05.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:06.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:07.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 04:00.0 USB controller [0c03]: ASMedia Technology Inc. ASM1142 USB 3.1 Host Controller [1b21:1242] 05:00.0 Ethernet controller [0200]: Intel Corporation I211 Gigabit Network Connection [8086:1539] (rev 03) 06:00.0 Network controller [0280]: Qualcomm Atheros QCA6174 802.11ac Wireless Network Adapter [168c:003e] (rev 32) 08:00.0 RAID bus controller [0104]: Silicon Image, Inc. SiI 3132 Serial ATA Raid II Controller [1095:3132] (rev 01) 0b:00.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD/ATI] Device [1002:14a0] (rev c1) 0c:00.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD/ATI] Device [1002:14a1] 0d:00.0 VGA compatible controller [0300]: Advanced Micro Devices, Inc. [AMD/ATI] Vega 20 [Radeon VII] [1002:66af] (rev c1) 0d:00.1 Audio device [0403]: Advanced Micro Devices, Inc. [AMD/ATI] Vega 20 HDMI Audio [Radeon VII] [1002:ab20] 0e:00.0 Non-Essential Instrumentation [1300]: Advanced Micro Devices, Inc. [AMD] Zeppelin/Raven/Raven2 PCIe Dummy Function [1022:145a] 0e:00.2 Encryption controller [1080]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Platform Security Processor [1022:1456] 0e:00.3 USB controller [0c03]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) USB 3.0 Host Controller [1022:145c] 0f:00.0 Non-Essential Instrumentation [1300]: Advanced Micro Devices, Inc. [AMD] Zeppelin/Renoir PCIe Dummy Function [1022:1455] 0f:00.2 SATA controller [0106]: Advanced Micro Devices, Inc. [AMD] FCH SATA Controller [AHCI mode] [1022:7901] (rev 51) 0f:00.3 Audio device [0403]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) HD Audio Controller [1022:1457] * USB devices (lsusb): Bus 006 Device 002: ID 2109:0817 VIA Labs, Inc. USB3.0 Hub Bus 006 Device 001: ID 1d6b:0003 Linux Foundation 3.0 root hub Bus 005 Device 004: ID 05e3:0745 Genesys Logic, Inc. Logilink CR0012 Bus 005 Device 003: ID 05e3:0608 Genesys Logic, Inc. Hub Bus 005 Device 002: ID 2109:2817 VIA Labs, Inc. USB2.0 Hub Bus 005 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub Bus 004 Device 001: ID 1d6b:0003 Linux Foundation 3.0 root hub Bus 003 Device 002: ID 051d:0002 American Power Conversion Uninterruptible Power Supply Bus 003 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub Bus 002 Device 001: ID 1d6b:0003 Linux Foundation 3.0 root hub Bus 001 Device 004: ID 046d:085e Logitech, Inc. Logitech BRIO Bus 001 Device 003: ID 09e8:0045 AKAI Professional M.I. Corp. MPK Mini Mk II MIDI Controller Bus 001 Device 002: ID 0b05:1825 ASUSTek Computer, Inc. Qualcomm Bluetooth 4.1 Bus 001 Device 006: ID 145f:024b Trust Trust Wired Keyboard Bus 001 Device 007: ID 1d57:ad17 Xenta ZELOTES GAME MOUSE Bus 001 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub * DRM Information (journalctl -k -b | grep -o 'kernel:.*drm.*$' | cut -d ' ' -f 2- ): [drm] amdgpu kernel modesetting enabled. fb0: switching to amdgpudrmfb from EFI VGA [drm] initializing kernel modesetting (VEGA20 0x1002:0x66AF 0x1002:0x081E 0xC1). [drm] register mmio base: 0xFCB00000 [drm] register mmio size: 524288 [drm] add ip block number 0 <soc15_common> [drm] add ip block number 1 <gmc_v9_0> [drm] add ip block number 2 <vega10_ih> [drm] add ip block number 3 <psp> [drm] add ip block number 4 <gfx_v9_0> [drm] add ip block number 5 <sdma_v4_0> [drm] add ip block number 6 <powerplay> [drm] add ip block number 7 <dm> [drm] add ip block number 8 <uvd_v7_0> [drm] add ip block number 9 <vce_v4_0> [drm] UVD(0) is enabled in VM mode [drm] UVD(1) is enabled in VM mode [drm] UVD(0) ENC is enabled in VM mode [drm] UVD(1) ENC is enabled in VM mode [drm] VCE enabled in VM mode [drm] vm size is 262144 GB, 4 levels, block size is 9-bit, fragment size is 9-bit [drm] Detected VRAM RAM=16368M, BAR=256M [drm] RAM width 4096bits HBM [drm] amdgpu: 16368M of VRAM memory ready [drm] amdgpu: 16368M of GTT memory ready. [drm] GART: num cpu pages 131072, num gpu pages 131072 [drm] PCIE GART of 512M enabled (table at 0x0000008000300000). [drm] Found UVD firmware ENC: 1.2 DEC: .43 Family ID: 19 [drm] PSP loading UVD firmware [drm] Found VCE firmware Version: 57.6 Binary ID: 4 [drm] PSP loading VCE firmware [drm] reserve 0x400000 from 0x83fec00000 for PSP TMR [drm] kiq ring mec 2 pipe 1 q 0 [drm] Display Core initialized with v3.2.104! [drm] UVD and UVD ENC initialized successfully. [drm] VCE initialized successfully. [drm] fb mappable at 0xE0959000 [drm] vram apper at 0xE0000000 [drm] size 19906560 [drm] fb depth is 24 [drm] pitch is 13824 fbcon: amdgpudrmfb (fb0) is primary device amdgpu 0000:0d:00.0: [drm] fb0: amdgpudrmfb frame buffer device [drm] Initialized amdgpu 3.40.0 20150101 for 0000:0d:00.0 on minor 0 * Xorg modules (grep LoadModule /var/log/Xorg.0.log ~/.local/share/xorg/Xorg.0.log | cut -d \" -f 2 | xargs): glx amdgpu ati radeon modesetting fbdev vesa fbdevhw fb dri2 glamoregl ramdac libinput * GL Support (glxinfo | grep -E "OpenGL version|OpenGL renderer"): OpenGL renderer string: AMD Radeon VII (VEGA20, DRM 3.40.0, 5.10.9-201.fc33.x86_64, LLVM 11.0.0) OpenGL version string: 4.6 (Compatibility Profile) Mesa 20.3.3 * Xorg errors (grep '^\[.*(EE)' /var/log/Xorg.0.log ~/.local/share/xorg/Xorg.0.log | cut -d ':' -f 2- ): [ 23882.071] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 23ms, your system is too slow [ 26328.538] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 14ms, your system is too slow [ 26703.167] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 29ms, your system is too slow [ 27254.938] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 33ms, your system is too slow [ 28592.084] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 21ms, your system is too slow [ 29737.977] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 22ms, your system is too slow [ 30311.456] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 32ms, your system is too slow [ 31027.544] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [ 65641.903] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 23ms, your system is too slow [102414.603] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 12ms, your system is too slow [103826.055] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 24ms, your system is too slow [104286.367] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 31ms, your system is too slow [104374.050] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [111278.504] (EE) event7 - ZLY ZELOTES GAME MOUSE: client bug: event processing lagging behind by 124ms, your system is too slow [114794.283] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 20ms, your system is too slow [116422.089] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [117879.555] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 23ms, your system is too slow [119364.104] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 28ms, your system is too slow [119915.987] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 29ms, your system is too slow [120554.073] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 32ms, your system is too slow [126507.594] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 24ms, your system is too slow [127504.971] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 25ms, your system is too slow [128269.945] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [193225.271] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 14ms, your system is too slow [197628.944] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [197634.041] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 33ms, your system is too slow [197788.651] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 19ms, your system is too slow [216002.069] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 31ms, your system is too slow * Kernel buffer tail (dmesg | tail): [214756.158881] wlp6s0: authenticate with b6:c4:40:89:8b:f4 [214756.212074] wlp6s0: send auth to b6:c4:40:89:8b:f4 (try 1/3) [214756.320436] wlp6s0: send auth to b6:c4:40:89:8b:f4 (try 2/3) [214756.326458] wlp6s0: authenticated [214756.327410] wlp6s0: associate with b6:c4:40:89:8b:f4 (try 1/3) [214756.341557] wlp6s0: RX AssocResp from b6:c4:40:89:8b:f4 (capab=0x1431 status=0 aid=5) [214756.343713] wlp6s0: associated [214756.361502] IPv6: ADDRCONF(NETDEV_CHANGE): wlp6s0: link becomes ready [214756.393780] wlp6s0: Limiting TX power to 30 (30 - 0) dBm as advertised by b6:c4:40:89:8b:f4 [215317.939882] wlp6s0: deauthenticating from b6:c4:40:89:8b:f4 by local choice (Reason: 3=DEAUTH_LEAVING) * Last few reboots (last -x -n10 reboot runlevel): runlevel (to lvl 5) 5.10.9-201.fc33. Mon Jan 25 05:18 still running reboot system boot 5.10.9-201.fc33. Mon Jan 25 05:17 still running runlevel (to lvl 5) 5.10.9-201.fc33. Sun Jan 24 12:09 - 05:18 (17:08) reboot system boot 5.10.9-201.fc33. Sun Jan 24 12:08 still running runlevel (to lvl 5) 5.10.8-200.fc33. Fri Jan 22 08:57 - 12:08 (2+03:10) reboot system boot 5.10.8-200.fc33. Fri Jan 22 08:56 - 12:08 (2+03:11) reboot system boot 5.10.8-200.fc33. Fri Jan 22 00:14 - 00:14 (00:00) runlevel (to lvl 5) 5.10.8-200.fc33. Wed Jan 20 09:24 - 00:13 (1+14:49) reboot system boot 5.10.8-200.fc33. Wed Jan 20 09:24 - 00:13 (1+14:49) reboot system boot 5.10.8-200.fc33. Wed Jan 20 09:15 - 00:13 (1+14:58) wtmp begins Wed Oct 28 01:12:12 2020 * DNF Repositories (dnf repolist): repo id repo name copr:copr.fedorainfracloud.org:elxreno:go-ipfs Copr repo for go-ipfs owned by elxreno copr:copr.fedorainfracloud.org:rodrigofarias77:budgie-desktop Copr repo for budgie-desktop owned by rodrigofarias77 copr:copr.fedorainfracloud.org:zawertun:crystal Copr repo for crystal owned by zawertun crystal Crystal fedora Fedora 33 - x86_64 fedora-cisco-openh264 Fedora 33 openh264 (From Cisco) - x86_64 fedora-modular Fedora Modular 33 - x86_64 google-chrome google-chrome google-cloud-sdk Google Cloud SDK hashicorp Hashicorp Stable - x86_64 kubernetes Kubernetes rpmfusion-free RPM Fusion for Fedora 33 - Free rpmfusion-free-updates RPM Fusion for Fedora 33 - Free - Updates rpmfusion-nonfree RPM Fusion for Fedora 33 - Nonfree rpmfusion-nonfree-updates RPM Fusion for Fedora 33 - Nonfree - Updates updates Fedora 33 - x86_64 - Updates updates-modular Fedora Modular 33 - x86_64 - Updates * DNF Extras (dnf -C list extras): Crystal 11 kB/s | 3.1 kB 00:00 Is this ok [y/N]: Google Cloud SDK 11 kB/s | 3.6 kB 00:00 Is this ok [y/N]: Is this ok [y/N]: Is this ok [y/N]: Google Cloud SDK 3.1 kB/s | 975 B 00:00 Is this ok [y/N]: Kubernetes 13 kB/s | 3.6 kB 00:00 Is this ok [y/N]: Is this ok [y/N]: Is this ok [y/N]: Kubernetes 3.2 kB/s | 975 B 00:00 Is this ok [y/N]: Last metadata expiration check: 0:12:54 ago on Wed 27 Jan 2021 05:21:26 PM CST. Extra Packages google-cloud-sdk.x86_64 325.0.0-1 @google-cloud-sdk kubectl.x86_64 1.20.2-0 @kubernetes minikube.x86_64 1.17.0-0 @@commandline trezor-bridge.x86_64 2.0.27-1 @@commandline * Last 20 packages installed (rpm -qa --nodigest --nosignature --last | head -20): openssl-devel-1.1.1i-1.fc33.x86_64 Wed 27 Jan 2021 05:31:02 PM CST libyaml-devel-0.2.5-3.fc33.x86_64 Wed 27 Jan 2021 05:31:01 PM CST libxml2-devel-2.9.10-8.fc33.x86_64 Wed 27 Jan 2021 05:31:01 PM CST gpm-devel-1.20.7-24.fc33.x86_64 Wed 27 Jan 2021 05:31:01 PM CST gc-devel-8.0.4-4.fc33.x86_64 Wed 27 Jan 2021 05:31:01 PM CST xz-devel-5.2.5-4.fc33.x86_64 Wed 27 Jan 2021 05:31:00 PM CST crystal-0.35.1-1.x86_64 Wed 27 Jan 2021 05:15:10 PM CST pcre-utf32-8.44-2.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST pcre-utf16-8.44-2.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST pcre-devel-8.44-2.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST pcre-cpp-8.44-2.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST libevent-devel-2.1.8-10.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST kernel-devel-5.10.10-200.fc33.x86_64 Wed 27 Jan 2021 06:50:23 AM CST pango-1.48.1-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST mtr-0.94-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST libpskc-2.6.6-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST libaom-2.0.1-3.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST fuse-overlayfs-1.4.0-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST crun-0.17-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST urw-base35-fonts-legacy-20200910-2.fc33.noarch Wed 27 Jan 2021 06:50:11 AM CST * EFI boot manager output (efibootmgr -v): BootCurrent: 0000 Timeout: 1 seconds BootOrder: 0000,0022,0018,0001,0002,0003 Boot0000* Fedora HD(1,GPT,1c6865f5-cd08-411d-aaf6-c53c3fca0c57,0x800,0x12c000)/File(\EFI\FEDORA\SHIMX64.EFI) Boot0001* UEFI:CD/DVD Drive BBS(129,,0x0) Boot0002* UEFI:Removable Device BBS(130,,0x0) Boot0003* UEFI:Network Device BBS(131,,0x0) Boot0018 Hard Drive BBS(HD,,0x0)..GO..NO........q.S.a.m.s.u.n.g. .S.S.D. .9.6.0. .P.R.O. .5.1.2.G.B....................A...........................%8Vq..a.....4..Gd-.;.A..MQ..L.S.3.E.W.N.X.0.J.6.0.1.3.7.7.T........BO..NO........u.S.T.4.0.0.0.D.M.0.0.5.-.2.D.P.1.6.6....................A.................................>..Gd-.;.A..MQ..L. . . . . . . . . . . . .G.Z.0.Y.F.J.Z.9........BO Boot0022* Fedora HD(1,GPT,1c6865f5-cd08-411d-aaf6-c53c3fca0c57,0x800,0x12c000)/File(\EFI\FEDORA\SHIM.EFI)..BO ```
1.0
The rpm has missing dependencies - It seems that the official RPM is missing a ton of dependencies, at least in Fedora 33. Check these out: https://download.copr.fedorainfracloud.org/results/zawertun/crystal/fedora-33-x86_64/01906375-crystal/crystal.spec The dependencies are far more that what gets installed when I use the crystal repo. AFAICS, you're missing: gc-devel gpm-devel libxml2-devel libyaml-devel openssl-devel xz-devel My test was trying to build amber framework with the installed crystal version. I've installed v0.35.1 (because 0.36.0 breaks amber). My system's info: <details> ``` * OS Release (cat /etc/*-release | uniq): Fedora release 33 (Thirty Three) NAME=Fedora VERSION="33 (Workstation Edition)" ID=fedora VERSION_ID=33 VERSION_CODENAME="" PLATFORM_ID="platform:f33" PRETTY_NAME="Fedora 33 (Workstation Edition)" ANSI_COLOR="0;38;2;60;110;180" LOGO=fedora-logo-icon CPE_NAME="cpe:/o:fedoraproject:fedora:33" HOME_URL="https://fedoraproject.org/" DOCUMENTATION_URL="https://docs.fedoraproject.org/en-US/fedora/f33/system-administrators-guide/" SUPPORT_URL="https://fedoraproject.org/wiki/Communicating_and_getting_help" BUG_REPORT_URL="https://bugzilla.redhat.com/" REDHAT_BUGZILLA_PRODUCT="Fedora" REDHAT_BUGZILLA_PRODUCT_VERSION=33 REDHAT_SUPPORT_PRODUCT="Fedora" REDHAT_SUPPORT_PRODUCT_VERSION=33 PRIVACY_POLICY_URL="https://fedoraproject.org/wiki/Legal:PrivacyPolicy" VARIANT="Workstation Edition" VARIANT_ID=workstation Fedora release 33 (Thirty Three) * Kernel (uname -r ; cat /proc/cmdline): 5.10.9-201.fc33.x86_64 BOOT_IMAGE=(hd1,gpt2)/vmlinuz-5.10.9-201.fc33.x86_64 root=UUID=934e5565-cbc7-47c1-b488-ed4dfead07cb ro rootflags=subvol=root rhgb quiet processor.max_cstate=5 rcu_nocbs=0-11 * Desktop(s) Running (ps -eo comm= | grep -E '(gnome-session|startkde|startactive|xfce.?-session|fluxbox|blackbox|hackedbox|ratpoison|enlightenment|icewm-session|od-session|wmaker|wmx|openbox-lxde|openbox-gnome-session|openbox-kde-session|mwm|e16|fvwm|xmonad|sugar-session|mate-session|lxqt-session|cinnamon)' ): gnome-session-b gnome-session-c gnome-session-b * Desktop(s) Installed (ls -m /usr/share/xsessions/ | sed 's/\.desktop//g' ): budgie-desktop, gnome-classic, gnome, gnome-xorg * SELinux Status (sestatus): SELinux status: enabled SELinuxfs mount: /sys/fs/selinux SELinux root directory: /etc/selinux Loaded policy name: targeted Current mode: enforcing Mode from config file: enforcing Policy MLS status: enabled Policy deny_unknown status: allowed Memory protection checking: actual (secure) Max kernel policy version: 33 * SELinux Errors (selinuxenabled && journalctl --since yesterday |grep avc: | grep -Eo comm="[^ ]+" | sort |uniq -c |sort -rn): 2 comm="systemd-tmpfile" * CPU Model (grep 'model name' /proc/cpuinfo | awk -F: '{print $2}' | uniq -c | sed -re 's/^ +//' ): 16 AMD Ryzen 7 1800X Eight-Core Processor * 64-bit Support (grep -q ' lm ' /proc/cpuinfo && echo Yes || echo No): Yes * Hardware Virtualization Support (grep -Eq '(vmx|svm)' /proc/cpuinfo && echo Yes || echo No): Yes * Load average (uptime): 17:34:16 up 2 days, 12:16, 1 user, load average: 0.67, 1.18, 1.30 * Memory usage (free -m): total used free shared buff/cache available Mem: 64301 12574 1430 207 50297 50956 Swap: 8191 3 8188 * Top 5 CPU hogs (ps axuScnh | awk '$2!=436183' | sort -rnk3 | head -5): 107 88678 27.8 9.2 21782828 6095088 ? Sl Jan25 701:04 qemu-system-x86 1000 9830 22.9 0.0 26140 15692 ? Ss Jan25 750:25 systemd 1000 418425 21.6 1.1 4657628 776676 ? Sl 17:08 5:34 firefox 1000 423177 14.6 0.0 233804 6704 pts/0 Ss 17:17 2:27 bash 1000 418649 8.4 0.6 3601236 445680 ? Sl 17:08 2:10 Web Content * Top 5 Memory hogs (ps axuScnh | sort -rnk4 | head -5): 107 88678 27.8 9.2 21782828 6095088 ? Sl Jan25 701:04 qemu-system-x86 1000 418425 21.6 1.1 4657628 776676 ? Sl 17:08 5:34 firefox 1000 10426 0.0 1.0 1452188 702940 ? Sl Jan25 0:48 gnome-software 0 1689 0.0 0.9 1184620 612796 ? Ssl Jan25 1:09 packagekitd 1000 418656 2.7 0.6 3439884 451992 ? Sl 17:08 0:42 Web Content * Disk space usage (df): Filesystem 1K-blocks Used Available Use% Mounted on devtmpfs 32864392 0 32864392 0% /dev tmpfs 32922604 35576 32887028 1% /dev/shm tmpfs 13169044 2456 13166588 1% /run /dev/nvme0n1p3 498443264 320413504 176661440 65% / /dev/nvme0n1p3 498443264 320413504 176661440 65% /srv /dev/nvme0n1p3 498443264 320413504 176661440 65% /home /dev/nvme0n1p2 999320 288000 642508 31% /boot /dev/nvme0n1p1 613184 20772 592412 4% /boot/efi tmpfs 32922608 8072 32914536 1% /tmp /dev/sda1 3907016704 601673476 3303447612 16% /var/lib/libvirt/images /dev/nvme0n1p3 498443264 320413504 176661440 65% /var/lib/libvirt/images/fast /dev/sda1 3907016704 601673476 3303447612 16% /home/renich/Downloads /dev/sda1 3907016704 601673476 3303447612 16% /home/renich/VirtualMachines/isos /dev/sda1 3907016704 601673476 3303447612 16% /home/renich/Videos/movies /dev/sda1 3907016704 601673476 3303447612 16% /home/renich/xzone tmpfs 6584520 240 6584280 1% /run/user/1000 * Block devices (blkid): /dev/nvme0n1p3: LABEL="fedora" UUID="934e5565-cbc7-47c1-b488-ed4dfead07cb" UUID_SUB="d07aecb3-ed70-4fea-bce7-f568c999b165" BLOCK_SIZE="4096" TYPE="btrfs" PARTUUID="5a6076d9-174e-46a0-a6f3-38f05ea33d89" /dev/nvme0n1p2: UUID="2919b7e6-0ca5-494e-9ca2-f9502faf6d01" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="55884fcb-1fc5-4157-a3ba-9677fc0dd2d2" /dev/nvme0n1p1: UUID="866B-15BA" BLOCK_SIZE="512" TYPE="vfat" PARTLABEL="EFI System Partition" PARTUUID="1c6865f5-cd08-411d-aaf6-c53c3fca0c57" * PCI devices (lspci -nn): 00:00.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Root Complex [1022:1450] 00:00.2 IOMMU [0806]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) I/O Memory Management Unit [1022:1451] 00:01.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:01.1 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) PCIe GPP Bridge [1022:1453] 00:01.3 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) PCIe GPP Bridge [1022:1453] 00:02.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:03.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:03.1 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) PCIe GPP Bridge [1022:1453] 00:04.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:07.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:07.1 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Internal PCIe GPP Bridge 0 to Bus B [1022:1454] 00:08.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-1fh) PCIe Dummy Host Bridge [1022:1452] 00:08.1 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Internal PCIe GPP Bridge 0 to Bus B [1022:1454] 00:14.0 SMBus [0c05]: Advanced Micro Devices, Inc. [AMD] FCH SMBus Controller [1022:790b] (rev 59) 00:14.3 ISA bridge [0601]: Advanced Micro Devices, Inc. [AMD] FCH LPC Bridge [1022:790e] (rev 51) 00:18.0 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 0 [1022:1460] 00:18.1 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 1 [1022:1461] 00:18.2 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 2 [1022:1462] 00:18.3 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 3 [1022:1463] 00:18.4 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 4 [1022:1464] 00:18.5 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 5 [1022:1465] 00:18.6 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 6 [1022:1466] 00:18.7 Host bridge [0600]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Data Fabric: Device 18h; Function 7 [1022:1467] 01:00.0 Non-Volatile memory controller [0108]: Samsung Electronics Co Ltd NVMe SSD Controller SM961/PM961/SM963 [144d:a804] 02:00.0 USB controller [0c03]: Advanced Micro Devices, Inc. [AMD] X370 Series Chipset USB 3.1 xHCI Controller [1022:43b9] (rev 02) 02:00.1 SATA controller [0106]: Advanced Micro Devices, Inc. [AMD] X370 Series Chipset SATA Controller [1022:43b5] (rev 02) 02:00.2 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] X370 Series Chipset PCIe Upstream Port [1022:43b0] (rev 02) 03:00.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:02.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:03.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:04.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:05.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:06.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 03:07.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD] 300 Series Chipset PCIe Port [1022:43b4] (rev 02) 04:00.0 USB controller [0c03]: ASMedia Technology Inc. ASM1142 USB 3.1 Host Controller [1b21:1242] 05:00.0 Ethernet controller [0200]: Intel Corporation I211 Gigabit Network Connection [8086:1539] (rev 03) 06:00.0 Network controller [0280]: Qualcomm Atheros QCA6174 802.11ac Wireless Network Adapter [168c:003e] (rev 32) 08:00.0 RAID bus controller [0104]: Silicon Image, Inc. SiI 3132 Serial ATA Raid II Controller [1095:3132] (rev 01) 0b:00.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD/ATI] Device [1002:14a0] (rev c1) 0c:00.0 PCI bridge [0604]: Advanced Micro Devices, Inc. [AMD/ATI] Device [1002:14a1] 0d:00.0 VGA compatible controller [0300]: Advanced Micro Devices, Inc. [AMD/ATI] Vega 20 [Radeon VII] [1002:66af] (rev c1) 0d:00.1 Audio device [0403]: Advanced Micro Devices, Inc. [AMD/ATI] Vega 20 HDMI Audio [Radeon VII] [1002:ab20] 0e:00.0 Non-Essential Instrumentation [1300]: Advanced Micro Devices, Inc. [AMD] Zeppelin/Raven/Raven2 PCIe Dummy Function [1022:145a] 0e:00.2 Encryption controller [1080]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) Platform Security Processor [1022:1456] 0e:00.3 USB controller [0c03]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) USB 3.0 Host Controller [1022:145c] 0f:00.0 Non-Essential Instrumentation [1300]: Advanced Micro Devices, Inc. [AMD] Zeppelin/Renoir PCIe Dummy Function [1022:1455] 0f:00.2 SATA controller [0106]: Advanced Micro Devices, Inc. [AMD] FCH SATA Controller [AHCI mode] [1022:7901] (rev 51) 0f:00.3 Audio device [0403]: Advanced Micro Devices, Inc. [AMD] Family 17h (Models 00h-0fh) HD Audio Controller [1022:1457] * USB devices (lsusb): Bus 006 Device 002: ID 2109:0817 VIA Labs, Inc. USB3.0 Hub Bus 006 Device 001: ID 1d6b:0003 Linux Foundation 3.0 root hub Bus 005 Device 004: ID 05e3:0745 Genesys Logic, Inc. Logilink CR0012 Bus 005 Device 003: ID 05e3:0608 Genesys Logic, Inc. Hub Bus 005 Device 002: ID 2109:2817 VIA Labs, Inc. USB2.0 Hub Bus 005 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub Bus 004 Device 001: ID 1d6b:0003 Linux Foundation 3.0 root hub Bus 003 Device 002: ID 051d:0002 American Power Conversion Uninterruptible Power Supply Bus 003 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub Bus 002 Device 001: ID 1d6b:0003 Linux Foundation 3.0 root hub Bus 001 Device 004: ID 046d:085e Logitech, Inc. Logitech BRIO Bus 001 Device 003: ID 09e8:0045 AKAI Professional M.I. Corp. MPK Mini Mk II MIDI Controller Bus 001 Device 002: ID 0b05:1825 ASUSTek Computer, Inc. Qualcomm Bluetooth 4.1 Bus 001 Device 006: ID 145f:024b Trust Trust Wired Keyboard Bus 001 Device 007: ID 1d57:ad17 Xenta ZELOTES GAME MOUSE Bus 001 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub * DRM Information (journalctl -k -b | grep -o 'kernel:.*drm.*$' | cut -d ' ' -f 2- ): [drm] amdgpu kernel modesetting enabled. fb0: switching to amdgpudrmfb from EFI VGA [drm] initializing kernel modesetting (VEGA20 0x1002:0x66AF 0x1002:0x081E 0xC1). [drm] register mmio base: 0xFCB00000 [drm] register mmio size: 524288 [drm] add ip block number 0 <soc15_common> [drm] add ip block number 1 <gmc_v9_0> [drm] add ip block number 2 <vega10_ih> [drm] add ip block number 3 <psp> [drm] add ip block number 4 <gfx_v9_0> [drm] add ip block number 5 <sdma_v4_0> [drm] add ip block number 6 <powerplay> [drm] add ip block number 7 <dm> [drm] add ip block number 8 <uvd_v7_0> [drm] add ip block number 9 <vce_v4_0> [drm] UVD(0) is enabled in VM mode [drm] UVD(1) is enabled in VM mode [drm] UVD(0) ENC is enabled in VM mode [drm] UVD(1) ENC is enabled in VM mode [drm] VCE enabled in VM mode [drm] vm size is 262144 GB, 4 levels, block size is 9-bit, fragment size is 9-bit [drm] Detected VRAM RAM=16368M, BAR=256M [drm] RAM width 4096bits HBM [drm] amdgpu: 16368M of VRAM memory ready [drm] amdgpu: 16368M of GTT memory ready. [drm] GART: num cpu pages 131072, num gpu pages 131072 [drm] PCIE GART of 512M enabled (table at 0x0000008000300000). [drm] Found UVD firmware ENC: 1.2 DEC: .43 Family ID: 19 [drm] PSP loading UVD firmware [drm] Found VCE firmware Version: 57.6 Binary ID: 4 [drm] PSP loading VCE firmware [drm] reserve 0x400000 from 0x83fec00000 for PSP TMR [drm] kiq ring mec 2 pipe 1 q 0 [drm] Display Core initialized with v3.2.104! [drm] UVD and UVD ENC initialized successfully. [drm] VCE initialized successfully. [drm] fb mappable at 0xE0959000 [drm] vram apper at 0xE0000000 [drm] size 19906560 [drm] fb depth is 24 [drm] pitch is 13824 fbcon: amdgpudrmfb (fb0) is primary device amdgpu 0000:0d:00.0: [drm] fb0: amdgpudrmfb frame buffer device [drm] Initialized amdgpu 3.40.0 20150101 for 0000:0d:00.0 on minor 0 * Xorg modules (grep LoadModule /var/log/Xorg.0.log ~/.local/share/xorg/Xorg.0.log | cut -d \" -f 2 | xargs): glx amdgpu ati radeon modesetting fbdev vesa fbdevhw fb dri2 glamoregl ramdac libinput * GL Support (glxinfo | grep -E "OpenGL version|OpenGL renderer"): OpenGL renderer string: AMD Radeon VII (VEGA20, DRM 3.40.0, 5.10.9-201.fc33.x86_64, LLVM 11.0.0) OpenGL version string: 4.6 (Compatibility Profile) Mesa 20.3.3 * Xorg errors (grep '^\[.*(EE)' /var/log/Xorg.0.log ~/.local/share/xorg/Xorg.0.log | cut -d ':' -f 2- ): [ 23882.071] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 23ms, your system is too slow [ 26328.538] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 14ms, your system is too slow [ 26703.167] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 29ms, your system is too slow [ 27254.938] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 33ms, your system is too slow [ 28592.084] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 21ms, your system is too slow [ 29737.977] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 22ms, your system is too slow [ 30311.456] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 32ms, your system is too slow [ 31027.544] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [ 65641.903] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 23ms, your system is too slow [102414.603] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 12ms, your system is too slow [103826.055] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 24ms, your system is too slow [104286.367] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 31ms, your system is too slow [104374.050] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [111278.504] (EE) event7 - ZLY ZELOTES GAME MOUSE: client bug: event processing lagging behind by 124ms, your system is too slow [114794.283] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 20ms, your system is too slow [116422.089] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [117879.555] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 23ms, your system is too slow [119364.104] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 28ms, your system is too slow [119915.987] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 29ms, your system is too slow [120554.073] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 32ms, your system is too slow [126507.594] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 24ms, your system is too slow [127504.971] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 25ms, your system is too slow [128269.945] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [193225.271] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 14ms, your system is too slow [197628.944] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 17ms, your system is too slow [197634.041] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 33ms, your system is too slow [197788.651] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 19ms, your system is too slow [216002.069] (EE) event3 - SEMICO Trust Wired Keyboard: client bug: event processing lagging behind by 31ms, your system is too slow * Kernel buffer tail (dmesg | tail): [214756.158881] wlp6s0: authenticate with b6:c4:40:89:8b:f4 [214756.212074] wlp6s0: send auth to b6:c4:40:89:8b:f4 (try 1/3) [214756.320436] wlp6s0: send auth to b6:c4:40:89:8b:f4 (try 2/3) [214756.326458] wlp6s0: authenticated [214756.327410] wlp6s0: associate with b6:c4:40:89:8b:f4 (try 1/3) [214756.341557] wlp6s0: RX AssocResp from b6:c4:40:89:8b:f4 (capab=0x1431 status=0 aid=5) [214756.343713] wlp6s0: associated [214756.361502] IPv6: ADDRCONF(NETDEV_CHANGE): wlp6s0: link becomes ready [214756.393780] wlp6s0: Limiting TX power to 30 (30 - 0) dBm as advertised by b6:c4:40:89:8b:f4 [215317.939882] wlp6s0: deauthenticating from b6:c4:40:89:8b:f4 by local choice (Reason: 3=DEAUTH_LEAVING) * Last few reboots (last -x -n10 reboot runlevel): runlevel (to lvl 5) 5.10.9-201.fc33. Mon Jan 25 05:18 still running reboot system boot 5.10.9-201.fc33. Mon Jan 25 05:17 still running runlevel (to lvl 5) 5.10.9-201.fc33. Sun Jan 24 12:09 - 05:18 (17:08) reboot system boot 5.10.9-201.fc33. Sun Jan 24 12:08 still running runlevel (to lvl 5) 5.10.8-200.fc33. Fri Jan 22 08:57 - 12:08 (2+03:10) reboot system boot 5.10.8-200.fc33. Fri Jan 22 08:56 - 12:08 (2+03:11) reboot system boot 5.10.8-200.fc33. Fri Jan 22 00:14 - 00:14 (00:00) runlevel (to lvl 5) 5.10.8-200.fc33. Wed Jan 20 09:24 - 00:13 (1+14:49) reboot system boot 5.10.8-200.fc33. Wed Jan 20 09:24 - 00:13 (1+14:49) reboot system boot 5.10.8-200.fc33. Wed Jan 20 09:15 - 00:13 (1+14:58) wtmp begins Wed Oct 28 01:12:12 2020 * DNF Repositories (dnf repolist): repo id repo name copr:copr.fedorainfracloud.org:elxreno:go-ipfs Copr repo for go-ipfs owned by elxreno copr:copr.fedorainfracloud.org:rodrigofarias77:budgie-desktop Copr repo for budgie-desktop owned by rodrigofarias77 copr:copr.fedorainfracloud.org:zawertun:crystal Copr repo for crystal owned by zawertun crystal Crystal fedora Fedora 33 - x86_64 fedora-cisco-openh264 Fedora 33 openh264 (From Cisco) - x86_64 fedora-modular Fedora Modular 33 - x86_64 google-chrome google-chrome google-cloud-sdk Google Cloud SDK hashicorp Hashicorp Stable - x86_64 kubernetes Kubernetes rpmfusion-free RPM Fusion for Fedora 33 - Free rpmfusion-free-updates RPM Fusion for Fedora 33 - Free - Updates rpmfusion-nonfree RPM Fusion for Fedora 33 - Nonfree rpmfusion-nonfree-updates RPM Fusion for Fedora 33 - Nonfree - Updates updates Fedora 33 - x86_64 - Updates updates-modular Fedora Modular 33 - x86_64 - Updates * DNF Extras (dnf -C list extras): Crystal 11 kB/s | 3.1 kB 00:00 Is this ok [y/N]: Google Cloud SDK 11 kB/s | 3.6 kB 00:00 Is this ok [y/N]: Is this ok [y/N]: Is this ok [y/N]: Google Cloud SDK 3.1 kB/s | 975 B 00:00 Is this ok [y/N]: Kubernetes 13 kB/s | 3.6 kB 00:00 Is this ok [y/N]: Is this ok [y/N]: Is this ok [y/N]: Kubernetes 3.2 kB/s | 975 B 00:00 Is this ok [y/N]: Last metadata expiration check: 0:12:54 ago on Wed 27 Jan 2021 05:21:26 PM CST. Extra Packages google-cloud-sdk.x86_64 325.0.0-1 @google-cloud-sdk kubectl.x86_64 1.20.2-0 @kubernetes minikube.x86_64 1.17.0-0 @@commandline trezor-bridge.x86_64 2.0.27-1 @@commandline * Last 20 packages installed (rpm -qa --nodigest --nosignature --last | head -20): openssl-devel-1.1.1i-1.fc33.x86_64 Wed 27 Jan 2021 05:31:02 PM CST libyaml-devel-0.2.5-3.fc33.x86_64 Wed 27 Jan 2021 05:31:01 PM CST libxml2-devel-2.9.10-8.fc33.x86_64 Wed 27 Jan 2021 05:31:01 PM CST gpm-devel-1.20.7-24.fc33.x86_64 Wed 27 Jan 2021 05:31:01 PM CST gc-devel-8.0.4-4.fc33.x86_64 Wed 27 Jan 2021 05:31:01 PM CST xz-devel-5.2.5-4.fc33.x86_64 Wed 27 Jan 2021 05:31:00 PM CST crystal-0.35.1-1.x86_64 Wed 27 Jan 2021 05:15:10 PM CST pcre-utf32-8.44-2.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST pcre-utf16-8.44-2.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST pcre-devel-8.44-2.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST pcre-cpp-8.44-2.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST libevent-devel-2.1.8-10.fc33.x86_64 Wed 27 Jan 2021 05:15:09 PM CST kernel-devel-5.10.10-200.fc33.x86_64 Wed 27 Jan 2021 06:50:23 AM CST pango-1.48.1-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST mtr-0.94-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST libpskc-2.6.6-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST libaom-2.0.1-3.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST fuse-overlayfs-1.4.0-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST crun-0.17-1.fc33.x86_64 Wed 27 Jan 2021 06:50:12 AM CST urw-base35-fonts-legacy-20200910-2.fc33.noarch Wed 27 Jan 2021 06:50:11 AM CST * EFI boot manager output (efibootmgr -v): BootCurrent: 0000 Timeout: 1 seconds BootOrder: 0000,0022,0018,0001,0002,0003 Boot0000* Fedora HD(1,GPT,1c6865f5-cd08-411d-aaf6-c53c3fca0c57,0x800,0x12c000)/File(\EFI\FEDORA\SHIMX64.EFI) Boot0001* UEFI:CD/DVD Drive BBS(129,,0x0) Boot0002* UEFI:Removable Device BBS(130,,0x0) Boot0003* UEFI:Network Device BBS(131,,0x0) Boot0018 Hard Drive BBS(HD,,0x0)..GO..NO........q.S.a.m.s.u.n.g. .S.S.D. .9.6.0. .P.R.O. .5.1.2.G.B....................A...........................%8Vq..a.....4..Gd-.;.A..MQ..L.S.3.E.W.N.X.0.J.6.0.1.3.7.7.T........BO..NO........u.S.T.4.0.0.0.D.M.0.0.5.-.2.D.P.1.6.6....................A.................................>..Gd-.;.A..MQ..L. . . . . . . . . . . . .G.Z.0.Y.F.J.Z.9........BO Boot0022* Fedora HD(1,GPT,1c6865f5-cd08-411d-aaf6-c53c3fca0c57,0x800,0x12c000)/File(\EFI\FEDORA\SHIM.EFI)..BO ```
infrastructure
the rpm has missing dependencies it seems that the official rpm is missing a ton of dependencies at least in fedora check these out the dependencies are far more that what gets installed when i use the crystal repo afaics you re missing gc devel gpm devel devel libyaml devel openssl devel xz devel my test was trying to build amber framework with the installed crystal version i ve installed because breaks amber my system s info os release cat etc release uniq fedora release thirty three name fedora version workstation edition id fedora version id version codename platform id platform pretty name fedora workstation edition ansi color logo fedora logo icon cpe name cpe o fedoraproject fedora home url documentation url support url bug report url redhat bugzilla product fedora redhat bugzilla product version redhat support product fedora redhat support product version privacy policy url variant workstation edition variant id workstation fedora release thirty three kernel uname r cat proc cmdline boot image vmlinuz root uuid ro rootflags subvol root rhgb quiet processor max cstate rcu nocbs desktop s running ps eo comm grep e gnome session startkde startactive xfce session fluxbox blackbox hackedbox ratpoison enlightenment icewm session od session wmaker wmx openbox lxde openbox gnome session openbox kde session mwm fvwm xmonad sugar session mate session lxqt session cinnamon gnome session b gnome session c gnome session b desktop s installed ls m usr share xsessions sed s desktop g budgie desktop gnome classic gnome gnome xorg selinux status sestatus selinux status enabled selinuxfs mount sys fs selinux selinux root directory etc selinux loaded policy name targeted current mode enforcing mode from config file enforcing policy mls status enabled policy deny unknown status allowed memory protection checking actual secure max kernel policy version selinux errors selinuxenabled journalctl since yesterday grep avc grep eo comm sort uniq c sort rn comm systemd tmpfile cpu model grep model name proc cpuinfo awk f print uniq c sed re s amd ryzen eight core processor bit support grep q lm proc cpuinfo echo yes echo no yes hardware virtualization support grep eq vmx svm proc cpuinfo echo yes echo no yes load average uptime up days user load average memory usage free m total used free shared buff cache available mem swap top cpu hogs ps axuscnh awk sort head sl qemu system ss systemd sl firefox pts ss bash sl web content top memory hogs ps axuscnh sort head sl qemu system sl firefox sl gnome software ssl packagekitd sl web content disk space usage df filesystem blocks used available use mounted on devtmpfs dev tmpfs dev shm tmpfs run dev dev srv dev home dev boot dev boot efi tmpfs tmp dev var lib libvirt images dev var lib libvirt images fast dev home renich downloads dev home renich virtualmachines isos dev home renich videos movies dev home renich xzone tmpfs run user block devices blkid dev label fedora uuid uuid sub block size type btrfs partuuid dev uuid block size type partuuid dev uuid block size type vfat partlabel efi system partition partuuid pci devices lspci nn host bridge advanced micro devices inc family models root complex iommu advanced micro devices inc family models i o memory management unit host bridge advanced micro devices inc family models pcie dummy host bridge pci bridge advanced micro devices inc family models pcie gpp bridge pci bridge advanced micro devices inc family models pcie gpp bridge host bridge advanced micro devices inc family models pcie dummy host bridge host bridge advanced micro devices inc family models pcie dummy host bridge pci bridge advanced micro devices inc family models pcie gpp bridge host bridge advanced micro devices inc family models pcie dummy host bridge host bridge advanced micro devices inc family models pcie dummy host bridge pci bridge advanced micro devices inc family models internal pcie gpp bridge to bus b host bridge advanced micro devices inc family models pcie dummy host bridge pci bridge advanced micro devices inc family models internal pcie gpp bridge to bus b smbus advanced micro devices inc fch smbus controller rev isa bridge advanced micro devices inc fch lpc bridge rev host bridge advanced micro devices inc family models data fabric device function host bridge advanced micro devices inc family models data fabric device function host bridge advanced micro devices inc family models data fabric device function host bridge advanced micro devices inc family models data fabric device function host bridge advanced micro devices inc family models data fabric device function host bridge advanced micro devices inc family models data fabric device function host bridge advanced micro devices inc family models data fabric device function host bridge advanced micro devices inc family models data fabric device function non volatile memory controller samsung electronics co ltd nvme ssd controller usb controller advanced micro devices inc series chipset usb xhci controller rev sata controller advanced micro devices inc series chipset sata controller rev pci bridge advanced micro devices inc series chipset pcie upstream port rev pci bridge advanced micro devices inc series chipset pcie port rev pci bridge advanced micro devices inc series chipset pcie port rev pci bridge advanced micro devices inc series chipset pcie port rev pci bridge advanced micro devices inc series chipset pcie port rev pci bridge advanced micro devices inc series chipset pcie port rev pci bridge advanced micro devices inc series chipset pcie port rev pci bridge advanced micro devices inc series chipset pcie port rev usb controller asmedia technology inc usb host controller ethernet controller intel corporation gigabit network connection rev network controller qualcomm atheros wireless network adapter rev raid bus controller silicon image inc sii serial ata raid ii controller rev pci bridge advanced micro devices inc device rev pci bridge advanced micro devices inc device vga compatible controller advanced micro devices inc vega rev audio device advanced micro devices inc vega hdmi audio non essential instrumentation advanced micro devices inc zeppelin raven pcie dummy function encryption controller advanced micro devices inc family models platform security processor usb controller advanced micro devices inc family models usb host controller non essential instrumentation advanced micro devices inc zeppelin renoir pcie dummy function sata controller advanced micro devices inc fch sata controller rev audio device advanced micro devices inc family models hd audio controller usb devices lsusb bus device id via labs inc hub bus device id linux foundation root hub bus device id genesys logic inc logilink bus device id genesys logic inc hub bus device id via labs inc hub bus device id linux foundation root hub bus device id linux foundation root hub bus device id american power conversion uninterruptible power supply bus device id linux foundation root hub bus device id linux foundation root hub bus device id logitech inc logitech brio bus device id akai professional m i corp mpk mini mk ii midi controller bus device id asustek computer inc qualcomm bluetooth bus device id trust trust wired keyboard bus device id xenta zelotes game mouse bus device id linux foundation root hub drm information journalctl k b grep o kernel drm cut d f amdgpu kernel modesetting enabled switching to amdgpudrmfb from efi vga initializing kernel modesetting register mmio base register mmio size add ip block number add ip block number add ip block number add ip block number add ip block number add ip block number add ip block number add ip block number add ip block number add ip block number uvd is enabled in vm mode uvd is enabled in vm mode uvd enc is enabled in vm mode uvd enc is enabled in vm mode vce enabled in vm mode vm size is gb levels block size is bit fragment size is bit detected vram ram bar ram width hbm amdgpu of vram memory ready amdgpu of gtt memory ready gart num cpu pages num gpu pages pcie gart of enabled table at found uvd firmware enc dec family id psp loading uvd firmware found vce firmware version binary id psp loading vce firmware reserve from for psp tmr kiq ring mec pipe q display core initialized with uvd and uvd enc initialized successfully vce initialized successfully fb mappable at vram apper at size fb depth is pitch is fbcon amdgpudrmfb is primary device amdgpu amdgpudrmfb frame buffer device initialized amdgpu for on minor xorg modules grep loadmodule var log xorg log local share xorg xorg log cut d f xargs glx amdgpu ati radeon modesetting fbdev vesa fbdevhw fb glamoregl ramdac libinput gl support glxinfo grep e opengl version opengl renderer opengl renderer string amd radeon vii drm llvm opengl version string compatibility profile mesa xorg errors grep ee var log xorg log local share xorg xorg log cut d f ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee zly zelotes game mouse client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow ee semico trust wired keyboard client bug event processing lagging behind by your system is too slow kernel buffer tail dmesg tail authenticate with send auth to try send auth to try authenticated associate with try rx assocresp from capab status aid associated addrconf netdev change link becomes ready limiting tx power to dbm as advertised by deauthenticating from by local choice reason deauth leaving last few reboots last x reboot runlevel runlevel to lvl mon jan still running reboot system boot mon jan still running runlevel to lvl sun jan reboot system boot sun jan still running runlevel to lvl fri jan reboot system boot fri jan reboot system boot fri jan runlevel to lvl wed jan reboot system boot wed jan reboot system boot wed jan wtmp begins wed oct dnf repositories dnf repolist repo id repo name copr copr fedorainfracloud org elxreno go ipfs copr repo for go ipfs owned by elxreno copr copr fedorainfracloud org budgie desktop copr repo for budgie desktop owned by copr copr fedorainfracloud org zawertun crystal copr repo for crystal owned by zawertun crystal crystal fedora fedora fedora cisco fedora from cisco fedora modular fedora modular google chrome google chrome google cloud sdk google cloud sdk hashicorp hashicorp stable kubernetes kubernetes rpmfusion free rpm fusion for fedora free rpmfusion free updates rpm fusion for fedora free updates rpmfusion nonfree rpm fusion for fedora nonfree rpmfusion nonfree updates rpm fusion for fedora nonfree updates updates fedora updates updates modular fedora modular updates dnf extras dnf c list extras crystal kb s kb is this ok google cloud sdk kb s kb is this ok is this ok is this ok google cloud sdk kb s b is this ok kubernetes kb s kb is this ok is this ok is this ok kubernetes kb s b is this ok last metadata expiration check ago on wed jan pm cst extra packages google cloud sdk google cloud sdk kubectl kubernetes minikube commandline trezor bridge commandline last packages installed rpm qa nodigest nosignature last head openssl devel wed jan pm cst libyaml devel wed jan pm cst devel wed jan pm cst gpm devel wed jan pm cst gc devel wed jan pm cst xz devel wed jan pm cst crystal wed jan pm cst pcre wed jan pm cst pcre wed jan pm cst pcre devel wed jan pm cst pcre cpp wed jan pm cst libevent devel wed jan pm cst kernel devel wed jan am cst pango wed jan am cst mtr wed jan am cst libpskc wed jan am cst libaom wed jan am cst fuse overlayfs wed jan am cst crun wed jan am cst urw fonts legacy noarch wed jan am cst efi boot manager output efibootmgr v bootcurrent timeout seconds bootorder fedora hd gpt file efi fedora efi uefi cd dvd drive bbs uefi removable device bbs uefi network device bbs hard drive bbs hd go no q s a m s u n g s s d p r o g b a a gd a mq l s e w n x j t bo no u s t d m d p a gd a mq l g z y f j z bo fedora hd gpt file efi fedora shim efi bo
1
8,898
7,724,180,611
IssuesEvent
2018-05-24 14:27:41
Microsoft/WindowsTemplateStudio
https://api.github.com/repos/Microsoft/WindowsTemplateStudio
closed
Update localization docs (in wiki)
Infrastructure localization meta
The information in https://github.com/Microsoft/WindowsTemplateStudio/wiki/Localization-Tool and in [the help output from the app](https://github.com/Microsoft/WindowsTemplateStudio/blob/dev/code/tools/Localization/Program.cs#L100) is wrong. the `ext` verb actually takes three parameters `-o`, `-a` & `-s` (`-t` & `-c` are not valid args)
1.0
Update localization docs (in wiki) - The information in https://github.com/Microsoft/WindowsTemplateStudio/wiki/Localization-Tool and in [the help output from the app](https://github.com/Microsoft/WindowsTemplateStudio/blob/dev/code/tools/Localization/Program.cs#L100) is wrong. the `ext` verb actually takes three parameters `-o`, `-a` & `-s` (`-t` & `-c` are not valid args)
infrastructure
update localization docs in wiki the information in and in is wrong the ext verb actually takes three parameters o a s t c are not valid args
1
12,208
9,640,105,898
IssuesEvent
2019-05-16 14:51:14
hzi-braunschweig/SORMAS-Project
https://api.github.com/repos/hzi-braunschweig/SORMAS-Project
opened
Fix problems in CI process
Infrastructure
Jenkins promotion process does not work, due to "uncomitted changes" Maybe related to maven goal calling sormas-api/pom.xml with > test > -Dtest=*Generator > -e > -X
1.0
Fix problems in CI process - Jenkins promotion process does not work, due to "uncomitted changes" Maybe related to maven goal calling sormas-api/pom.xml with > test > -Dtest=*Generator > -e > -X
infrastructure
fix problems in ci process jenkins promotion process does not work due to uncomitted changes maybe related to maven goal calling sormas api pom xml with test dtest generator e x
1
2,493
3,703,399,526
IssuesEvent
2016-02-29 20:16:10
phpmyadmin/phpmyadmin
https://api.github.com/repos/phpmyadmin/phpmyadmin
opened
`upload-release` error
infrastructure
`$ ./scripts/upload-release 4.6.0-rc1 release` During the ssh call to `./bin/sync-files-cdn`, after the file transfer I get this: ``` total size is 9,079,773,550 speedup is 130.56 Added 4.6.0-rc1 Traceback (most recent call last): File "./manage.py", line 31, in <module> execute_from_command_line(sys.argv) File "/usr/lib/python2.7/dist-packages/django/core/management/__init__.py", line 385, in execute_from_command_line utility.execute() File "/usr/lib/python2.7/dist-packages/django/core/management/__init__.py", line 377, in execute self.fetch_command(subcommand).run_from_argv(self.argv) File "/usr/lib/python2.7/dist-packages/django/core/management/base.py", line 288, in run_from_argv self.execute(*args, **options.__dict__) File "/usr/lib/python2.7/dist-packages/django/core/management/base.py", line 338, in execute output = self.handle(*args, **options) File "/home/builder/website/files/management/commands/import_files.py", line 85, in handle self.process_releases(os.path.join(settings.FILES_PATH, 'phpMyAdmin')) File "/home/builder/website/files/management/commands/import_files.py", line 81, in process_releases release File "/home/builder/website/files/management/commands/import_files.py", line 49, in process_files release=release, filename=filename File "/usr/lib/python2.7/dist-packages/django/db/models/manager.py", line 92, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/usr/lib/python2.7/dist-packages/django/db/models/query.py", line 424, in get_or_create return self._create_object_from_params(lookup, params) File "/usr/lib/python2.7/dist-packages/django/db/models/query.py", line 457, in _create_object_from_params obj.save(force_insert=True, using=self.db) File "/usr/lib/python2.7/dist-packages/django/db/models/base.py", line 589, in save force_update=force_update, update_fields=update_fields) File "/usr/lib/python2.7/dist-packages/django/db/models/base.py", line 626, in save_base update_fields=update_fields, raw=raw, using=using) File "/usr/lib/python2.7/dist-packages/django/dispatch/dispatcher.py", line 198, in send response = receiver(signal=self, sender=sender, **named) File "/home/builder/website/files/models.py", line 366, in purge_download purge_release(sender, instance.release) File "/home/builder/website/files/models.py", line 360, in purge_release instance.get_absolute_url(), File "/home/builder/website/pmaweb/cdn.py", line 55, in purge_cdn return perform(URL, data) File "/home/builder/website/pmaweb/cdn.py", line 40, in perform raise Exception(decoded) Exception: {u'status': u'error', u'description': u"Purge is NOT successful because purging data didn't succeed."} ```
1.0
`upload-release` error - `$ ./scripts/upload-release 4.6.0-rc1 release` During the ssh call to `./bin/sync-files-cdn`, after the file transfer I get this: ``` total size is 9,079,773,550 speedup is 130.56 Added 4.6.0-rc1 Traceback (most recent call last): File "./manage.py", line 31, in <module> execute_from_command_line(sys.argv) File "/usr/lib/python2.7/dist-packages/django/core/management/__init__.py", line 385, in execute_from_command_line utility.execute() File "/usr/lib/python2.7/dist-packages/django/core/management/__init__.py", line 377, in execute self.fetch_command(subcommand).run_from_argv(self.argv) File "/usr/lib/python2.7/dist-packages/django/core/management/base.py", line 288, in run_from_argv self.execute(*args, **options.__dict__) File "/usr/lib/python2.7/dist-packages/django/core/management/base.py", line 338, in execute output = self.handle(*args, **options) File "/home/builder/website/files/management/commands/import_files.py", line 85, in handle self.process_releases(os.path.join(settings.FILES_PATH, 'phpMyAdmin')) File "/home/builder/website/files/management/commands/import_files.py", line 81, in process_releases release File "/home/builder/website/files/management/commands/import_files.py", line 49, in process_files release=release, filename=filename File "/usr/lib/python2.7/dist-packages/django/db/models/manager.py", line 92, in manager_method return getattr(self.get_queryset(), name)(*args, **kwargs) File "/usr/lib/python2.7/dist-packages/django/db/models/query.py", line 424, in get_or_create return self._create_object_from_params(lookup, params) File "/usr/lib/python2.7/dist-packages/django/db/models/query.py", line 457, in _create_object_from_params obj.save(force_insert=True, using=self.db) File "/usr/lib/python2.7/dist-packages/django/db/models/base.py", line 589, in save force_update=force_update, update_fields=update_fields) File "/usr/lib/python2.7/dist-packages/django/db/models/base.py", line 626, in save_base update_fields=update_fields, raw=raw, using=using) File "/usr/lib/python2.7/dist-packages/django/dispatch/dispatcher.py", line 198, in send response = receiver(signal=self, sender=sender, **named) File "/home/builder/website/files/models.py", line 366, in purge_download purge_release(sender, instance.release) File "/home/builder/website/files/models.py", line 360, in purge_release instance.get_absolute_url(), File "/home/builder/website/pmaweb/cdn.py", line 55, in purge_cdn return perform(URL, data) File "/home/builder/website/pmaweb/cdn.py", line 40, in perform raise Exception(decoded) Exception: {u'status': u'error', u'description': u"Purge is NOT successful because purging data didn't succeed."} ```
infrastructure
upload release error scripts upload release release during the ssh call to bin sync files cdn after the file transfer i get this total size is speedup is added traceback most recent call last file manage py line in execute from command line sys argv file usr lib dist packages django core management init py line in execute from command line utility execute file usr lib dist packages django core management init py line in execute self fetch command subcommand run from argv self argv file usr lib dist packages django core management base py line in run from argv self execute args options dict file usr lib dist packages django core management base py line in execute output self handle args options file home builder website files management commands import files py line in handle self process releases os path join settings files path phpmyadmin file home builder website files management commands import files py line in process releases release file home builder website files management commands import files py line in process files release release filename filename file usr lib dist packages django db models manager py line in manager method return getattr self get queryset name args kwargs file usr lib dist packages django db models query py line in get or create return self create object from params lookup params file usr lib dist packages django db models query py line in create object from params obj save force insert true using self db file usr lib dist packages django db models base py line in save force update force update update fields update fields file usr lib dist packages django db models base py line in save base update fields update fields raw raw using using file usr lib dist packages django dispatch dispatcher py line in send response receiver signal self sender sender named file home builder website files models py line in purge download purge release sender instance release file home builder website files models py line in purge release instance get absolute url file home builder website pmaweb cdn py line in purge cdn return perform url data file home builder website pmaweb cdn py line in perform raise exception decoded exception u status u error u description u purge is not successful because purging data didn t succeed
1
239,091
19,808,589,569
IssuesEvent
2022-01-19 09:45:54
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
sql: TestTelemetry failed
C-test-failure O-robot branch-master
sql.TestTelemetry [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4150708&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4150708&tab=artifacts#/) on master @ [912964e02ddd951c77d4f71981ae18b3894e9084](https://github.com/cockroachdb/cockroach/commits/912964e02ddd951c77d4f71981ae18b3894e9084): ``` === RUN TestTelemetry test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTelemetry1418443230 test_log_scope.go:80: use -show-logs to present logs inline === CONT TestTelemetry telemetry_test.go:45: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTelemetry1418443230 --- FAIL: TestTelemetry (74.29s) === RUN TestTelemetry/planning/server telemetry.go:90: testdata/telemetry/planning:178: SELECT * FROM (VALUES (1), (2)) AS a(x) JOIN LATERAL (SELECT a.x+1 AS x) AS b ON a.x < b.x expected: sql.plan.lateral-join found: sql.plan.lateral-join sql.plan.subquery --- FAIL: TestTelemetry/planning/server (1.97s) === RUN TestTelemetry/planning --- FAIL: TestTelemetry/planning (3.82s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=deadlock - GOFLAGS=-parallel=4 </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTelemetry.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
sql: TestTelemetry failed - sql.TestTelemetry [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4150708&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4150708&tab=artifacts#/) on master @ [912964e02ddd951c77d4f71981ae18b3894e9084](https://github.com/cockroachdb/cockroach/commits/912964e02ddd951c77d4f71981ae18b3894e9084): ``` === RUN TestTelemetry test_log_scope.go:79: test logs captured to: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTelemetry1418443230 test_log_scope.go:80: use -show-logs to present logs inline === CONT TestTelemetry telemetry_test.go:45: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestTelemetry1418443230 --- FAIL: TestTelemetry (74.29s) === RUN TestTelemetry/planning/server telemetry.go:90: testdata/telemetry/planning:178: SELECT * FROM (VALUES (1), (2)) AS a(x) JOIN LATERAL (SELECT a.x+1 AS x) AS b ON a.x < b.x expected: sql.plan.lateral-join found: sql.plan.lateral-join sql.plan.subquery --- FAIL: TestTelemetry/planning/server (1.97s) === RUN TestTelemetry/planning --- FAIL: TestTelemetry/planning (3.82s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - TAGS=deadlock - GOFLAGS=-parallel=4 </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTelemetry.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_infrastructure
sql testtelemetry failed sql testtelemetry with on master run testtelemetry test log scope go test logs captured to go src github com cockroachdb cockroach artifacts test log scope go use show logs to present logs inline cont testtelemetry telemetry test go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts fail testtelemetry run testtelemetry planning server telemetry go testdata telemetry planning select from values as a x join lateral select a x as x as b on a x b x expected sql plan lateral join found sql plan lateral join sql plan subquery fail testtelemetry planning server run testtelemetry planning fail testtelemetry planning help see also parameters in this failure tags deadlock goflags parallel cc cockroachdb sql queries
0
25,993
19,541,591,808
IssuesEvent
2022-01-01 01:57:51
rjust/defects4j
https://api.github.com/repos/rjust/defects4j
closed
Should all modules in `cpanfile` need to be installed manually?
question infrastructure tutorial
Hi, I'm setting up defects4j, but there was a DBI issue. The steps written in `README.md` do not seem to setup perl module dependencies well. First, here is my environment. - Ubuntu 18.04 (on virtualbox) - openjdk version "1.8.0_282" - git version 2.17.1 - svn, version 1.9.7 (r1800392) - This is perl 5, version 26, subversion 1 (v5.26.1) All above satisfy the requirements written in README.md Here are the commands to install `defects4j`, exactly same with what `README.md` is saying on __Steps to set up Defects4J__ section. ```bash git clone https://github.com/rjust/defects4j cd defects4j cpanm --installdeps . ./init.sh echo 'export PATH=/home/vagrant/github/rjust/defects4j/framework/bin:${PATH}' >> ~/.bashrc . ~/.bashrc defects4j info -p Lang ``` There are following messages in terminal output from `cpanm --installdeps .` ```bash ==> Found dependencies: DBI --> Working on DBI Fetching http://www.cpan.org/authors/id/T/TI/TIMB/DBI-1.643.tar.gz ... OK Configuring DBI-1.643 ... OK Building and testing DBI-1.643 ... OK Successfully installed DBI-1.643 ``` It seems that DBI perl module is installed. However, when checking installation with `defects4j info -p Lang`, there is following error. ```bash Can't locate DBI.pm in @INC (you may need to install the DBI module) ... ``` However, after referring #125 and after running `sudo cpan DBI` (`sudo` is required for my case), the output is normal. ``` Summary of configuration for Project: Lang -------------------------------------------------------------------------------- Script dir: /home/vagrant/github/rjust/defects4j/framework Base dir: /home/vagrant/github/rjust/defects4j Major root: /home/vagrant/github/rjust/defects4j/major Repo dir: /home/vagrant/github/rjust/defects4j/project_repos -------------------------------------------------------------------------------- Project ID: Lang Program: commons-lang Build file: /home/vagrant/github/rjust/defects4j/framework/projects/Lang/Lang.build.xml -------------------------------------------------------------------------------- Vcs: Vcs::Git Repository: /home/vagrant/github/rjust/defects4j/project_repos/commons-lang.git Commit db: /home/vagrant/github/rjust/defects4j/framework/projects/Lang/active-bugs.csv Number of bugs: 64 -------------------------------------------------------------------------------- ``` Then, should I run `sudo cpan MODULE` for all written in `cpanfile` ? The [comments for the issues #125](https://github.com/rjust/defects4j/issues/125#issuecomment-338849526) is saying so I think... Sorry for the disturbance with minor question, but hope someone can answer about it, since it's my first time to deal with projects using `perl` modules.
1.0
Should all modules in `cpanfile` need to be installed manually? - Hi, I'm setting up defects4j, but there was a DBI issue. The steps written in `README.md` do not seem to setup perl module dependencies well. First, here is my environment. - Ubuntu 18.04 (on virtualbox) - openjdk version "1.8.0_282" - git version 2.17.1 - svn, version 1.9.7 (r1800392) - This is perl 5, version 26, subversion 1 (v5.26.1) All above satisfy the requirements written in README.md Here are the commands to install `defects4j`, exactly same with what `README.md` is saying on __Steps to set up Defects4J__ section. ```bash git clone https://github.com/rjust/defects4j cd defects4j cpanm --installdeps . ./init.sh echo 'export PATH=/home/vagrant/github/rjust/defects4j/framework/bin:${PATH}' >> ~/.bashrc . ~/.bashrc defects4j info -p Lang ``` There are following messages in terminal output from `cpanm --installdeps .` ```bash ==> Found dependencies: DBI --> Working on DBI Fetching http://www.cpan.org/authors/id/T/TI/TIMB/DBI-1.643.tar.gz ... OK Configuring DBI-1.643 ... OK Building and testing DBI-1.643 ... OK Successfully installed DBI-1.643 ``` It seems that DBI perl module is installed. However, when checking installation with `defects4j info -p Lang`, there is following error. ```bash Can't locate DBI.pm in @INC (you may need to install the DBI module) ... ``` However, after referring #125 and after running `sudo cpan DBI` (`sudo` is required for my case), the output is normal. ``` Summary of configuration for Project: Lang -------------------------------------------------------------------------------- Script dir: /home/vagrant/github/rjust/defects4j/framework Base dir: /home/vagrant/github/rjust/defects4j Major root: /home/vagrant/github/rjust/defects4j/major Repo dir: /home/vagrant/github/rjust/defects4j/project_repos -------------------------------------------------------------------------------- Project ID: Lang Program: commons-lang Build file: /home/vagrant/github/rjust/defects4j/framework/projects/Lang/Lang.build.xml -------------------------------------------------------------------------------- Vcs: Vcs::Git Repository: /home/vagrant/github/rjust/defects4j/project_repos/commons-lang.git Commit db: /home/vagrant/github/rjust/defects4j/framework/projects/Lang/active-bugs.csv Number of bugs: 64 -------------------------------------------------------------------------------- ``` Then, should I run `sudo cpan MODULE` for all written in `cpanfile` ? The [comments for the issues #125](https://github.com/rjust/defects4j/issues/125#issuecomment-338849526) is saying so I think... Sorry for the disturbance with minor question, but hope someone can answer about it, since it's my first time to deal with projects using `perl` modules.
infrastructure
should all modules in cpanfile need to be installed manually hi i m setting up but there was a dbi issue the steps written in readme md do not seem to setup perl module dependencies well first here is my environment ubuntu on virtualbox openjdk version git version svn version this is perl version subversion all above satisfy the requirements written in readme md here are the commands to install exactly same with what readme md is saying on steps to set up section bash git clone cd cpanm installdeps init sh echo export path home vagrant github rjust framework bin path bashrc bashrc info p lang there are following messages in terminal output from cpanm installdeps bash found dependencies dbi working on dbi fetching ok configuring dbi ok building and testing dbi ok successfully installed dbi it seems that dbi perl module is installed however when checking installation with info p lang there is following error bash can t locate dbi pm in inc you may need to install the dbi module however after referring and after running sudo cpan dbi sudo is required for my case the output is normal summary of configuration for project lang script dir home vagrant github rjust framework base dir home vagrant github rjust major root home vagrant github rjust major repo dir home vagrant github rjust project repos project id lang program commons lang build file home vagrant github rjust framework projects lang lang build xml vcs vcs git repository home vagrant github rjust project repos commons lang git commit db home vagrant github rjust framework projects lang active bugs csv number of bugs then should i run sudo cpan module for all written in cpanfile the is saying so i think sorry for the disturbance with minor question but hope someone can answer about it since it s my first time to deal with projects using perl modules
1
4,359
16,137,374,695
IssuesEvent
2021-04-29 13:30:05
Rust-GCC/gccrs
https://api.github.com/repos/Rust-GCC/gccrs
opened
Add automation to run rustc against testcases
automation diagnostic question
When writing test cases it could be useful to have automation that checks that the test case behaves in a similar fashion with rustc. It could be that for compile/execute test cases we see that the compile returns zero, and that the rust compiler returns non zero for xfail test cases. This will save us from adding test cases which act differently against rustc. This could potentially be built upon to have a test-runner such that we can execute rustc testcases against gccrs for a general Rust compiler testing framework. This issue is intended to create a discussion if this seems like an interesting idea or not.
1.0
Add automation to run rustc against testcases - When writing test cases it could be useful to have automation that checks that the test case behaves in a similar fashion with rustc. It could be that for compile/execute test cases we see that the compile returns zero, and that the rust compiler returns non zero for xfail test cases. This will save us from adding test cases which act differently against rustc. This could potentially be built upon to have a test-runner such that we can execute rustc testcases against gccrs for a general Rust compiler testing framework. This issue is intended to create a discussion if this seems like an interesting idea or not.
non_infrastructure
add automation to run rustc against testcases when writing test cases it could be useful to have automation that checks that the test case behaves in a similar fashion with rustc it could be that for compile execute test cases we see that the compile returns zero and that the rust compiler returns non zero for xfail test cases this will save us from adding test cases which act differently against rustc this could potentially be built upon to have a test runner such that we can execute rustc testcases against gccrs for a general rust compiler testing framework this issue is intended to create a discussion if this seems like an interesting idea or not
0
9,957
8,286,196,662
IssuesEvent
2018-09-19 03:18:03
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Factor with one level
bug interface/infrastructure
In a simulation, I have set two factors 1) Cultivar with two levels, 2) Management with one level ![image](https://user-images.githubusercontent.com/1657381/45273083-317b8a80-b4f4-11e8-91ac-06f90983f339.png) However, the Management factor is stored as the column name "Factors" and value as "Management". ![image](https://user-images.githubusercontent.com/1657381/45273119-70a9db80-b4f4-11e8-8d95-1eaed54c1d39.png) I would expect to store as the first Factor, e.g. | Cultivar | Management | | ------------- | ------------- | | P100 | IHN | | P150 | IHN | In this way, I can mixed experiment with multiple levels and only one levels when plotting figures.
1.0
Factor with one level - In a simulation, I have set two factors 1) Cultivar with two levels, 2) Management with one level ![image](https://user-images.githubusercontent.com/1657381/45273083-317b8a80-b4f4-11e8-91ac-06f90983f339.png) However, the Management factor is stored as the column name "Factors" and value as "Management". ![image](https://user-images.githubusercontent.com/1657381/45273119-70a9db80-b4f4-11e8-8d95-1eaed54c1d39.png) I would expect to store as the first Factor, e.g. | Cultivar | Management | | ------------- | ------------- | | P100 | IHN | | P150 | IHN | In this way, I can mixed experiment with multiple levels and only one levels when plotting figures.
infrastructure
factor with one level in a simulation i have set two factors cultivar with two levels management with one level however the management factor is stored as the column name factors and value as management i would expect to store as the first factor e g cultivar management ihn ihn in this way i can mixed experiment with multiple levels and only one levels when plotting figures
1
62,174
6,778,489,392
IssuesEvent
2017-10-28 11:34:24
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
AMP not working on password-protected pages
bug Need Testing
If mobile redirection is ON, then it's giving too many redirect error. Ref: https://secure.helpscout.net/conversation/441028863/16581/?folderId=1622967
1.0
AMP not working on password-protected pages - If mobile redirection is ON, then it's giving too many redirect error. Ref: https://secure.helpscout.net/conversation/441028863/16581/?folderId=1622967
non_infrastructure
amp not working on password protected pages if mobile redirection is on then it s giving too many redirect error ref
0
46,698
13,180,995,805
IssuesEvent
2020-08-12 13:40:56
mibo32/fitbit-api-example-java
https://api.github.com/repos/mibo32/fitbit-api-example-java
opened
CVE-2018-12022 (High) detected in jackson-databind-2.8.1.jar
security vulnerability
## CVE-2018-12022 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/fitbit-api-example-java/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mibo32/fitbit-api-example-java/commit/fdd0855fc8f3b846f83506299759cd9cc820e5d2">fdd0855fc8f3b846f83506299759cd9cc820e5d2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022>CVE-2018-12022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: 2.7.9.4, 2.8.11.2, 2.9.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-12022 (High) detected in jackson-databind-2.8.1.jar - ## CVE-2018-12022 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/fitbit-api-example-java/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/mibo32/fitbit-api-example-java/commit/fdd0855fc8f3b846f83506299759cd9cc820e5d2">fdd0855fc8f3b846f83506299759cd9cc820e5d2</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022>CVE-2018-12022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: 2.7.9.4, 2.8.11.2, 2.9.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm fitbit api example java pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library found in head commit a href vulnerability details an issue was discovered in fasterxml jackson databind prior to and when default typing is enabled either globally or for a specific property the service has the jodd db jar for database access for the jodd framework in the classpath and an attacker can provide an ldap service to access it is possible to make the service execute a malicious payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
69,120
30,083,381,309
IssuesEvent
2023-06-29 06:36:05
vmware/singleton
https://api.github.com/repos/vmware/singleton
opened
[Security][Service] Please upgrade dependency libraries to latest version (jackson-databind)
area/java-service kind/security priority/medium
In Singleton Service build, below dependency library is out of date, please upgrade it to latest version: - jackson-databind (2.15.1)
1.0
[Security][Service] Please upgrade dependency libraries to latest version (jackson-databind) - In Singleton Service build, below dependency library is out of date, please upgrade it to latest version: - jackson-databind (2.15.1)
non_infrastructure
please upgrade dependency libraries to latest version jackson databind in singleton service build below dependency library is out of date please upgrade it to latest version jackson databind
0
178,925
30,035,925,221
IssuesEvent
2023-06-27 12:46:02
code4romania/asistent-medical-comunitar
https://api.github.com/repos/code4romania/asistent-medical-comunitar
opened
[Activități comunitare] Implement 'Adaugă activitate' button
design low-priority
**Describe the bug** The 'Adaugă activitate' button is displayed only in 'Acțiuni rapide' button. **To Reproduce** Steps to reproduce the behavior: 1. Go to 'Activități comunitare' 2. Go to 'Activități de mediu' tab 3. Click on 'Acțiuni rapide' button 4. Notice the 'Adaugă campanie de sănătate', 'Adaugă activitate de mediu' and 'Adaugă activitate administrative' buttons **Expected behavior** The 'Adaugă activitate' button should also be displayed on the top right corner of the table.
1.0
[Activități comunitare] Implement 'Adaugă activitate' button - **Describe the bug** The 'Adaugă activitate' button is displayed only in 'Acțiuni rapide' button. **To Reproduce** Steps to reproduce the behavior: 1. Go to 'Activități comunitare' 2. Go to 'Activități de mediu' tab 3. Click on 'Acțiuni rapide' button 4. Notice the 'Adaugă campanie de sănătate', 'Adaugă activitate de mediu' and 'Adaugă activitate administrative' buttons **Expected behavior** The 'Adaugă activitate' button should also be displayed on the top right corner of the table.
non_infrastructure
implement adaugă activitate button describe the bug the adaugă activitate button is displayed only in acțiuni rapide button to reproduce steps to reproduce the behavior go to activități comunitare go to activități de mediu tab click on acțiuni rapide button notice the adaugă campanie de sănătate adaugă activitate de mediu and adaugă activitate administrative buttons expected behavior the adaugă activitate button should also be displayed on the top right corner of the table
0
5,233
5,535,604,814
IssuesEvent
2017-03-21 17:42:49
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
opened
[MultiVersion] Mechanism for importing versioned models is too cumbersome
Infrastructure
From PR #2526: Original ```Python from azure.mgmt.storage.models import Encryption, EncryptionServices, EncryptionService ``` Current ```Python Encryption, EncryptionServices, EncryptionService = versioned_models(ResourceType.MGMT_STORAGE_STORAGE_ACCOUNTS, 'Encryption', 'EncryptionServices', 'EncryptionService') ``` Before we apply the second pattern wholesale across the CLI, we should investigate how we can approach the ideal of the original, handling the versioning behind this scenes.
1.0
[MultiVersion] Mechanism for importing versioned models is too cumbersome - From PR #2526: Original ```Python from azure.mgmt.storage.models import Encryption, EncryptionServices, EncryptionService ``` Current ```Python Encryption, EncryptionServices, EncryptionService = versioned_models(ResourceType.MGMT_STORAGE_STORAGE_ACCOUNTS, 'Encryption', 'EncryptionServices', 'EncryptionService') ``` Before we apply the second pattern wholesale across the CLI, we should investigate how we can approach the ideal of the original, handling the versioning behind this scenes.
infrastructure
mechanism for importing versioned models is too cumbersome from pr original python from azure mgmt storage models import encryption encryptionservices encryptionservice current python encryption encryptionservices encryptionservice versioned models resourcetype mgmt storage storage accounts encryption encryptionservices encryptionservice before we apply the second pattern wholesale across the cli we should investigate how we can approach the ideal of the original handling the versioning behind this scenes
1
24,651
17,552,243,807
IssuesEvent
2021-08-13 00:16:37
scala-js/scala-js-dom
https://api.github.com/repos/scala-js/scala-js-dom
reopened
Change auto-formatting settings
infrastructure
Once our PR queue is cleared, on the 2.x branch lets: * Add import normalisation and sorting to auto-formatting (I use it everywhere and it's GLORIOUS) * I wouldn't mind increasing the column width from 79 to 120 * Format ScalaDoc in a more concise way - it takes up so much vertical space atm * More?
1.0
Change auto-formatting settings - Once our PR queue is cleared, on the 2.x branch lets: * Add import normalisation and sorting to auto-formatting (I use it everywhere and it's GLORIOUS) * I wouldn't mind increasing the column width from 79 to 120 * Format ScalaDoc in a more concise way - it takes up so much vertical space atm * More?
infrastructure
change auto formatting settings once our pr queue is cleared on the x branch lets add import normalisation and sorting to auto formatting i use it everywhere and it s glorious i wouldn t mind increasing the column width from to format scaladoc in a more concise way it takes up so much vertical space atm more
1
250,846
7,988,443,059
IssuesEvent
2018-07-19 11:04:33
fxi/map-x-mgl
https://api.github.com/repos/fxi/map-x-mgl
opened
MapX V5: list of "Priority 1" bugs
Priority 1 bug
- [ ] Upload data - drag and drop method: doesn't work - [ ] Upload data - upload module: Shapefile & GeoJSON -> OK; other formats fail (GPX, KML...) - [ ] Download data: "Country used to spatially select the data. By default = entire world" for now it's not the case, by default = current project (highlighted countries).
1.0
MapX V5: list of "Priority 1" bugs - - [ ] Upload data - drag and drop method: doesn't work - [ ] Upload data - upload module: Shapefile & GeoJSON -> OK; other formats fail (GPX, KML...) - [ ] Download data: "Country used to spatially select the data. By default = entire world" for now it's not the case, by default = current project (highlighted countries).
non_infrastructure
mapx list of priority bugs upload data drag and drop method doesn t work upload data upload module shapefile geojson ok other formats fail gpx kml download data country used to spatially select the data by default entire world for now it s not the case by default current project highlighted countries
0
289,149
21,770,628,581
IssuesEvent
2022-05-13 08:43:24
so-c/soc-toolbox
https://api.github.com/repos/so-c/soc-toolbox
closed
voicevox-callerインストール手順実施中の-File ""指定誤りが減るようにする
documentation enhancement destructive
動作不良の問い合わせ対応で、ここの指定誤りが原因であるケースが2回あった。予防したい。 手段案 * 間違いの入り込みにくいマニュアルに改善する * rvls を書き換えるcmdバッチを同梱。これを実行する手順にする
1.0
voicevox-callerインストール手順実施中の-File ""指定誤りが減るようにする - 動作不良の問い合わせ対応で、ここの指定誤りが原因であるケースが2回あった。予防したい。 手段案 * 間違いの入り込みにくいマニュアルに改善する * rvls を書き換えるcmdバッチを同梱。これを実行する手順にする
non_infrastructure
voicevox callerインストール手順実施中の file 指定誤りが減るようにする 動作不良の問い合わせ対応で、 。予防したい。 手段案 間違いの入り込みにくいマニュアルに改善する rvls を書き換えるcmdバッチを同梱。これを実行する手順にする
0
452,341
32,057,388,689
IssuesEvent
2023-09-24 08:44:21
AcademySoftwareFoundation/OpenColorIO
https://api.github.com/repos/AcademySoftwareFoundation/OpenColorIO
opened
Update API Docs Sphinx Theme
Documentation
As discussed in the TSC already, our present theme used in the API documentation is [Press Theme](https://github.com/schettino72/sphinx_press_theme), and it has unfortunately been marked as looking for a new maintainer. I'm sure someone will take up that request, but it won't be us. Given the slightly uncertain future of the theme, and some visual oddities it can produce, its a good opportunity to revamp the theme to something like [Furo](https://github.com/pradyunsg/furo) which keeps a clean and modern aesthetic with much more community support for the project its self. We've also seen that fellow ASWF/O*IO project OpenImageIO has adopted it themselves. In theory its a single line change in the Sphinx configuration, so may be a good first issue, but its not immediately clear if it will require larger changes to maintain readability across the docs.
1.0
Update API Docs Sphinx Theme - As discussed in the TSC already, our present theme used in the API documentation is [Press Theme](https://github.com/schettino72/sphinx_press_theme), and it has unfortunately been marked as looking for a new maintainer. I'm sure someone will take up that request, but it won't be us. Given the slightly uncertain future of the theme, and some visual oddities it can produce, its a good opportunity to revamp the theme to something like [Furo](https://github.com/pradyunsg/furo) which keeps a clean and modern aesthetic with much more community support for the project its self. We've also seen that fellow ASWF/O*IO project OpenImageIO has adopted it themselves. In theory its a single line change in the Sphinx configuration, so may be a good first issue, but its not immediately clear if it will require larger changes to maintain readability across the docs.
non_infrastructure
update api docs sphinx theme as discussed in the tsc already our present theme used in the api documentation is and it has unfortunately been marked as looking for a new maintainer i m sure someone will take up that request but it won t be us given the slightly uncertain future of the theme and some visual oddities it can produce its a good opportunity to revamp the theme to something like which keeps a clean and modern aesthetic with much more community support for the project its self we ve also seen that fellow aswf o io project openimageio has adopted it themselves in theory its a single line change in the sphinx configuration so may be a good first issue but its not immediately clear if it will require larger changes to maintain readability across the docs
0
2,302
3,610,185,365
IssuesEvent
2016-02-05 03:17:42
elmsln/elmsln
https://api.github.com/repos/elmsln/elmsln
closed
Kill off passwords printing on drush SI calls
infrastructure security related
We did some minor testing locally and this should do the trick, no reason for drush to claim it did anything because it'll print a password to the screen which would remain in logs forever and is either reset immediately or is to a place holder site that should have no reference anywhere. testing in a branch
1.0
Kill off passwords printing on drush SI calls - We did some minor testing locally and this should do the trick, no reason for drush to claim it did anything because it'll print a password to the screen which would remain in logs forever and is either reset immediately or is to a place holder site that should have no reference anywhere. testing in a branch
infrastructure
kill off passwords printing on drush si calls we did some minor testing locally and this should do the trick no reason for drush to claim it did anything because it ll print a password to the screen which would remain in logs forever and is either reset immediately or is to a place holder site that should have no reference anywhere testing in a branch
1
1,208
3,079,825,831
IssuesEvent
2015-08-21 18:28:59
GoogleChrome/accessibility-developer-tools
https://api.github.com/repos/GoogleChrome/accessibility-developer-tools
closed
Automate style checking
infrastructure
I recommend running a linting tool as part of the grunt build. This would help reduce the overhead of reviewing PRs. In my experience the best by far is ESLint: http://eslint.org/ Closure Linter would be a good fit too but is not as configurable.
1.0
Automate style checking - I recommend running a linting tool as part of the grunt build. This would help reduce the overhead of reviewing PRs. In my experience the best by far is ESLint: http://eslint.org/ Closure Linter would be a good fit too but is not as configurable.
infrastructure
automate style checking i recommend running a linting tool as part of the grunt build this would help reduce the overhead of reviewing prs in my experience the best by far is eslint closure linter would be a good fit too but is not as configurable
1
160,466
13,790,006,035
IssuesEvent
2020-10-09 09:45:13
sButtons/sbuttons
https://api.github.com/repos/sButtons/sbuttons
closed
Make changes to CONTRIBUTING.md
Hacktoberfest Priority: Low documentation enhancement good first issue help wanted up-for-grabs
In CONTRIBUTING.md under Instructions item number 3, we've added the structure of the button to be added to `assets/js/buttons-examples.js`. Currently `shouldHaveBlockType` is the last item. We need to move `variations` and `children` to be at the button of the list.
1.0
Make changes to CONTRIBUTING.md - In CONTRIBUTING.md under Instructions item number 3, we've added the structure of the button to be added to `assets/js/buttons-examples.js`. Currently `shouldHaveBlockType` is the last item. We need to move `variations` and `children` to be at the button of the list.
non_infrastructure
make changes to contributing md in contributing md under instructions item number we ve added the structure of the button to be added to assets js buttons examples js currently shouldhaveblocktype is the last item we need to move variations and children to be at the button of the list
0
27,406
21,698,978,381
IssuesEvent
2022-05-10 00:24:38
celeritas-project/celeritas
https://api.github.com/repos/celeritas-project/celeritas
closed
Prototype performance portability
infrastructure
Do an initial port of enough core Celeritas components to at least run some demo applications on non-CUDA hardware.
1.0
Prototype performance portability - Do an initial port of enough core Celeritas components to at least run some demo applications on non-CUDA hardware.
infrastructure
prototype performance portability do an initial port of enough core celeritas components to at least run some demo applications on non cuda hardware
1
158,739
6,034,339,373
IssuesEvent
2017-06-09 10:48:27
igabriel85/dmon-adp
https://api.github.com/repos/igabriel85/dmon-adp
closed
Use multiprocess instead of threads
High Priority
Look into implementing multiprocess instead of threads in python.
1.0
Use multiprocess instead of threads - Look into implementing multiprocess instead of threads in python.
non_infrastructure
use multiprocess instead of threads look into implementing multiprocess instead of threads in python
0
17,274
12,269,761,114
IssuesEvent
2020-05-07 14:32:04
acts-project/acts
https://api.github.com/repos/acts-project/acts
closed
Add DD4hep to the `EVERYTHING` build flag
Impact - Minor Infrastructure
Requires DD4hep to be present in the Ubuntu image as tracked in acts-project/machines#5.
1.0
Add DD4hep to the `EVERYTHING` build flag - Requires DD4hep to be present in the Ubuntu image as tracked in acts-project/machines#5.
infrastructure
add to the everything build flag requires to be present in the ubuntu image as tracked in acts project machines
1
533
2,771,924,665
IssuesEvent
2015-05-02 05:12:28
letsencrypt/boulder
https://api.github.com/repos/letsencrypt/boulder
closed
Don't ignore random number generation errors
security blocker
In https://github.com/letsencrypt/boulder/blob/81b0d8034a9705eff5de2201a721ce8deb877f49/core/util.go#L71: ``` func RandomString(byteLength int) string { b := make([]byte, byteLength) rand.Read(b) // NOTE: Ignoring errors return B64enc(b) } ``` An error in `rand.Read` might indicate that the system RNG is failing. You probably want to make sure callers hard-fail if there is an error here so that tokens, keys, etc. don't end up with too little entropy. From https://golang.org/pkg/crypto/rand/#Read: ``` On return, n == len(b) if and only if err == nil. ```
True
Don't ignore random number generation errors - In https://github.com/letsencrypt/boulder/blob/81b0d8034a9705eff5de2201a721ce8deb877f49/core/util.go#L71: ``` func RandomString(byteLength int) string { b := make([]byte, byteLength) rand.Read(b) // NOTE: Ignoring errors return B64enc(b) } ``` An error in `rand.Read` might indicate that the system RNG is failing. You probably want to make sure callers hard-fail if there is an error here so that tokens, keys, etc. don't end up with too little entropy. From https://golang.org/pkg/crypto/rand/#Read: ``` On return, n == len(b) if and only if err == nil. ```
non_infrastructure
don t ignore random number generation errors in func randomstring bytelength int string b make byte bytelength rand read b note ignoring errors return b an error in rand read might indicate that the system rng is failing you probably want to make sure callers hard fail if there is an error here so that tokens keys etc don t end up with too little entropy from on return n len b if and only if err nil
0
118,411
25,303,453,078
IssuesEvent
2022-11-17 12:30:38
gleam-lang/gleam
https://api.github.com/repos/gleam-lang/gleam
closed
Javascript nested sequences
bug help wanted good first issue area:codegen
I believe nested sequences within JavaScript is broken. Take this example program (that granted is doing naughty things ignoring the `Result`). ```gleam pub fn main() -> Result(Int, Nil) { let x = 1; { try b = Error(Nil); b } Ok(x) } ``` The result of `main` in Erlang is `Ok(1)` but `Error(Nil)` in JavaScript. Output Erlang: ```erlang -module(main). -compile(no_auto_import). -export([main/0]). -spec main() -> {ok, integer()} | {error, nil}. main() -> X = 1, case {error, nil} of {error, _try} -> {error, _try}; {ok, B} -> B end, {ok, X}. ``` Output JavaScript: ```javascript import { Ok, Error } from "./gleam.mjs"; export function main() { let x = 1; let $ = new Error(undefined); if (!$.isOk()) return $; let b = $[0]; b; return new Ok(x); } ``` I think sequences in JS need to be unconditionally wrapped in a lambda due to `try`. It maybe that if there are no try statements then the lambda may be omitted, but that seems unnecessary.
1.0
Javascript nested sequences - I believe nested sequences within JavaScript is broken. Take this example program (that granted is doing naughty things ignoring the `Result`). ```gleam pub fn main() -> Result(Int, Nil) { let x = 1; { try b = Error(Nil); b } Ok(x) } ``` The result of `main` in Erlang is `Ok(1)` but `Error(Nil)` in JavaScript. Output Erlang: ```erlang -module(main). -compile(no_auto_import). -export([main/0]). -spec main() -> {ok, integer()} | {error, nil}. main() -> X = 1, case {error, nil} of {error, _try} -> {error, _try}; {ok, B} -> B end, {ok, X}. ``` Output JavaScript: ```javascript import { Ok, Error } from "./gleam.mjs"; export function main() { let x = 1; let $ = new Error(undefined); if (!$.isOk()) return $; let b = $[0]; b; return new Ok(x); } ``` I think sequences in JS need to be unconditionally wrapped in a lambda due to `try`. It maybe that if there are no try statements then the lambda may be omitted, but that seems unnecessary.
non_infrastructure
javascript nested sequences i believe nested sequences within javascript is broken take this example program that granted is doing naughty things ignoring the result gleam pub fn main result int nil let x try b error nil b ok x the result of main in erlang is ok but error nil in javascript output erlang erlang module main compile no auto import export spec main ok integer error nil main x case error nil of error try error try ok b b end ok x output javascript javascript import ok error from gleam mjs export function main let x let new error undefined if isok return let b b return new ok x i think sequences in js need to be unconditionally wrapped in a lambda due to try it maybe that if there are no try statements then the lambda may be omitted but that seems unnecessary
0
73,821
24,814,074,473
IssuesEvent
2022-10-25 11:48:53
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
opened
Component: Table
defect
### Describe the bug When using p-table with stateStorage and resizableColumns the table is being extended over the size of container - in my p-card. This happens only after restoring state from local storage, so make sure you will resize any column then refresh the view. I do believe it may be connected with adding style="width: XXXXpx;" property for first <div> under the <p-table> Are there any known workarounds for that issue? Or maybe I've configured something wrongly. Thank you in advance for helping me out. Here is the screenshot of the issue: Actual: ![image](https://user-images.githubusercontent.com/50959052/197764667-b05bc79b-d677-4e5a-8f66-d2263392430a.png) Expected: ![image](https://user-images.githubusercontent.com/50959052/197765155-8b475678-dc5e-4840-8a70-ceadce550541.png) ### Environment Windows 10 or 11 ### Reproducer https://stackblitz.com/edit/primeng-tablestate-demo-qdp7qc?file=src%2Fapp%2Fapp.component.html ### Angular version 11.0.4 ### PrimeNG version 13.0.4 ### Build / Runtime Angular CLI App ### Language ALL ### Node version (for AoT issues node --version) 0.0.0 ### Browser(s) Chrome 106 ### Steps to reproduce the behavior 1. Resize any column to be wider 2. Refresh view ### Expected behavior The table should remain inside the card and should show scrollbar on the bottom
1.0
Component: Table - ### Describe the bug When using p-table with stateStorage and resizableColumns the table is being extended over the size of container - in my p-card. This happens only after restoring state from local storage, so make sure you will resize any column then refresh the view. I do believe it may be connected with adding style="width: XXXXpx;" property for first <div> under the <p-table> Are there any known workarounds for that issue? Or maybe I've configured something wrongly. Thank you in advance for helping me out. Here is the screenshot of the issue: Actual: ![image](https://user-images.githubusercontent.com/50959052/197764667-b05bc79b-d677-4e5a-8f66-d2263392430a.png) Expected: ![image](https://user-images.githubusercontent.com/50959052/197765155-8b475678-dc5e-4840-8a70-ceadce550541.png) ### Environment Windows 10 or 11 ### Reproducer https://stackblitz.com/edit/primeng-tablestate-demo-qdp7qc?file=src%2Fapp%2Fapp.component.html ### Angular version 11.0.4 ### PrimeNG version 13.0.4 ### Build / Runtime Angular CLI App ### Language ALL ### Node version (for AoT issues node --version) 0.0.0 ### Browser(s) Chrome 106 ### Steps to reproduce the behavior 1. Resize any column to be wider 2. Refresh view ### Expected behavior The table should remain inside the card and should show scrollbar on the bottom
non_infrastructure
component table describe the bug when using p table with statestorage and resizablecolumns the table is being extended over the size of container in my p card this happens only after restoring state from local storage so make sure you will resize any column then refresh the view i do believe it may be connected with adding style width xxxxpx property for first under the are there any known workarounds for that issue or maybe i ve configured something wrongly thank you in advance for helping me out here is the screenshot of the issue actual expected environment windows or reproducer angular version primeng version build runtime angular cli app language all node version for aot issues node version browser s chrome steps to reproduce the behavior resize any column to be wider refresh view expected behavior the table should remain inside the card and should show scrollbar on the bottom
0
4,222
4,893,833,367
IssuesEvent
2016-11-19 01:22:13
eriq-augustine/psl
https://api.github.com/repos/eriq-augustine/psl
closed
Continuous Integration
Infrastructure
We need some continuous integration over here. Travis-CI is the go to for github.
1.0
Continuous Integration - We need some continuous integration over here. Travis-CI is the go to for github.
infrastructure
continuous integration we need some continuous integration over here travis ci is the go to for github
1
31,308
11,907,314,145
IssuesEvent
2020-03-30 22:01:59
sammcj/anz-test-2
https://api.github.com/repos/sammcj/anz-test-2
opened
CVE-2020-7608 (High) detected in yargs-parser-13.1.1.tgz
security vulnerability
## CVE-2020-7608 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yargs-parser-13.1.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/anz-test-2/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/anz-test-2/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - mocha-7.0.0.tgz (Root Library) - :x: **yargs-parser-13.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sammcj/anz-test-2/commit/a5f4c263664c76bac168a76b0a24fbeedbedc8db">a5f4c263664c76bac168a76b0a24fbeedbedc8db</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7608 (High) detected in yargs-parser-13.1.1.tgz - ## CVE-2020-7608 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>yargs-parser-13.1.1.tgz</b></p></summary> <p>the mighty option parser used by yargs</p> <p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-13.1.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/anz-test-2/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/anz-test-2/node_modules/yargs-parser/package.json</p> <p> Dependency Hierarchy: - mocha-7.0.0.tgz (Root Library) - :x: **yargs-parser-13.1.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sammcj/anz-test-2/commit/a5f4c263664c76bac168a76b0a24fbeedbedc8db">a5f4c263664c76bac168a76b0a24fbeedbedc8db</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload. <p>Publish Date: 2020-03-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p> <p>Release Date: 2020-03-16</p> <p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in yargs parser tgz cve high severity vulnerability vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file tmp ws scm anz test package json path to vulnerable library tmp ws scm anz test node modules yargs parser package json dependency hierarchy mocha tgz root library x yargs parser tgz vulnerable library found in head commit a href vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
420,553
28,289,902,189
IssuesEvent
2023-04-09 04:00:03
HNToan7202/video-editing
https://api.github.com/repos/HNToan7202/video-editing
closed
Quay video Trình bày cách sử dụng công cụ testing (Selenium WebDriver)
documentation
- [x] Tìm hiểu về WebDriver - [x] Tìm hiểu Chức năng WebDriver trong Selenium - [x] Xây dựng dự án testing - [x] Demo chi tiết các sử dụng công cụ
1.0
Quay video Trình bày cách sử dụng công cụ testing (Selenium WebDriver) - - [x] Tìm hiểu về WebDriver - [x] Tìm hiểu Chức năng WebDriver trong Selenium - [x] Xây dựng dự án testing - [x] Demo chi tiết các sử dụng công cụ
non_infrastructure
quay video trình bày cách sử dụng công cụ testing selenium webdriver tìm hiểu về webdriver tìm hiểu chức năng webdriver trong selenium xây dựng dự án testing demo chi tiết các sử dụng công cụ
0
11,307
9,090,870,419
IssuesEvent
2019-02-19 01:26:53
kevinkjt2000/bowser
https://api.github.com/repos/kevinkjt2000/bowser
closed
Need to investigate what "TLS verification failed" means
Bug Infrastructure
Bowser was not responding to messages, so I tried to connect to the docker-machine this error message was displayed when trying to enable the bowser environment: ``` $ eval $(docker-machine env bowser) Error checking TLS connection: Error checking and/or regenerating the certs: There was an error validating certificates for host "34.228.210.108:2376": dial tcp 34.228.210.108:2376: connect: connection refused You can attempt to regenerate them using 'docker-machine regenerate-certs [name]'. Be advised that this will trigger a Docker daemon restart which might stop running containers. ``` I ran the suggested command `docker-machine regenerate-certs bowser`, which also restarted the docker services. This could be because it has been while since the machine was started. If that is the case, these TLS certs may need to be regenerated on a regular basis.
1.0
Need to investigate what "TLS verification failed" means - Bowser was not responding to messages, so I tried to connect to the docker-machine this error message was displayed when trying to enable the bowser environment: ``` $ eval $(docker-machine env bowser) Error checking TLS connection: Error checking and/or regenerating the certs: There was an error validating certificates for host "34.228.210.108:2376": dial tcp 34.228.210.108:2376: connect: connection refused You can attempt to regenerate them using 'docker-machine regenerate-certs [name]'. Be advised that this will trigger a Docker daemon restart which might stop running containers. ``` I ran the suggested command `docker-machine regenerate-certs bowser`, which also restarted the docker services. This could be because it has been while since the machine was started. If that is the case, these TLS certs may need to be regenerated on a regular basis.
infrastructure
need to investigate what tls verification failed means bowser was not responding to messages so i tried to connect to the docker machine this error message was displayed when trying to enable the bowser environment eval docker machine env bowser error checking tls connection error checking and or regenerating the certs there was an error validating certificates for host dial tcp connect connection refused you can attempt to regenerate them using docker machine regenerate certs be advised that this will trigger a docker daemon restart which might stop running containers i ran the suggested command docker machine regenerate certs bowser which also restarted the docker services this could be because it has been while since the machine was started if that is the case these tls certs may need to be regenerated on a regular basis
1
7,950
11,137,554,528
IssuesEvent
2019-12-20 19:40:56
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Bug: Drawers for experience and education aren't expanding on first click
Apply Process Bug State Dept.
Environment: Test Steps to reproduce: 1) Apply for an internship 2) on the work experience page, click on the plus sign for a work experience (or education) to expand - on the first try the row expands and quickly closes - must click again for it to expand and stay open
1.0
Bug: Drawers for experience and education aren't expanding on first click - Environment: Test Steps to reproduce: 1) Apply for an internship 2) on the work experience page, click on the plus sign for a work experience (or education) to expand - on the first try the row expands and quickly closes - must click again for it to expand and stay open
non_infrastructure
bug drawers for experience and education aren t expanding on first click environment test steps to reproduce apply for an internship on the work experience page click on the plus sign for a work experience or education to expand on the first try the row expands and quickly closes must click again for it to expand and stay open
0
22,389
15,164,303,706
IssuesEvent
2021-02-12 13:33:06
NeuromatchAcademy/course-content
https://api.github.com/repos/NeuromatchAcademy/course-content
opened
CI robot should revise its PR comments instead of making new ones
Infrastructure
There must be an API for this because e.g. the codecov bot works this way ... will take some figuring out, however. It is not urgent but it will increase the SNR of pull request feedback and cut down on notifications.
1.0
CI robot should revise its PR comments instead of making new ones - There must be an API for this because e.g. the codecov bot works this way ... will take some figuring out, however. It is not urgent but it will increase the SNR of pull request feedback and cut down on notifications.
infrastructure
ci robot should revise its pr comments instead of making new ones there must be an api for this because e g the codecov bot works this way will take some figuring out however it is not urgent but it will increase the snr of pull request feedback and cut down on notifications
1
32,700
26,921,841,530
IssuesEvent
2023-02-07 11:01:06
lunakv/academyruins-api
https://api.github.com/repos/lunakv/academyruins-api
opened
Normalize trailing slashes in URLs
infrastructure low priority
Some routes return 307s when sent without a trailing slash. Some return 307 when sent with one. We should make the slash irrelevant and just return the response. FastAPI doesn't fully support this nicely, so the way to do it is probably with a combination of making sure the line without the trailing slash is always the canonical one (this is not true for router roots), then strip the slash in the nginx proxy.
1.0
Normalize trailing slashes in URLs - Some routes return 307s when sent without a trailing slash. Some return 307 when sent with one. We should make the slash irrelevant and just return the response. FastAPI doesn't fully support this nicely, so the way to do it is probably with a combination of making sure the line without the trailing slash is always the canonical one (this is not true for router roots), then strip the slash in the nginx proxy.
infrastructure
normalize trailing slashes in urls some routes return when sent without a trailing slash some return when sent with one we should make the slash irrelevant and just return the response fastapi doesn t fully support this nicely so the way to do it is probably with a combination of making sure the line without the trailing slash is always the canonical one this is not true for router roots then strip the slash in the nginx proxy
1
33,268
27,351,087,074
IssuesEvent
2023-02-27 09:36:34
Altinn/altinn-platform
https://api.github.com/repos/Altinn/altinn-platform
opened
POC: apps AKS with dualstack ipv4 and ipv6
kind/user-story status/draft area/infrastructure
### Description To be compliant with [Forskrift om IT-standarder i offentlig forvaltning](https://lovdata.no/dokument/SF/forskrift/2013-04-05-959) we need to run dualstack with ipv4 and ipv6 in all apps clusters. We need to investigate how to convert existing clusters, and make choices about new clusters. ### Additional Information - [Use dual-stack kubenet networking in Azure Kubernetes Service (AKS) (Preview)](https://learn.microsoft.com/en-us/azure/aks/configure-kubenet-dual-stack) - [Configure Azure CNI networking in Azure Kubernetes Service (AKS)](https://learn.microsoft.com/en-us/azure/aks/configure-azure-cni) # current config - [Configure Azure CNI Powered by Cilium in Azure Kubernetes Service (AKS) (Preview)](https://learn.microsoft.com/en-us/azure/aks/azure-cni-powered-by-cilium) # may be new config - [Bring your own Container Network Interface (CNI) plugin with Azure Kubernetes Service (AKS)](https://learn.microsoft.com/en-us/azure/aks/use-byo-cni?tabs=azure-cli) # should evaluate with cilium ### Tasks Deploy ttd-at25-aks: - Check if converting existing clusters is possible in place - Evaluate what CNI we should set as standard - Test "Azure CNI Powered by Cilium" vs "Bring your own CNI with Cilium" - Update iac to set up dualstack ipv4 and ipv6 ### Acceptance Criterias apps AKS expose ipv6 for inbound and aoutbound traffic as minimum. Preferably ipv6 should be used internal in cluster.
1.0
POC: apps AKS with dualstack ipv4 and ipv6 - ### Description To be compliant with [Forskrift om IT-standarder i offentlig forvaltning](https://lovdata.no/dokument/SF/forskrift/2013-04-05-959) we need to run dualstack with ipv4 and ipv6 in all apps clusters. We need to investigate how to convert existing clusters, and make choices about new clusters. ### Additional Information - [Use dual-stack kubenet networking in Azure Kubernetes Service (AKS) (Preview)](https://learn.microsoft.com/en-us/azure/aks/configure-kubenet-dual-stack) - [Configure Azure CNI networking in Azure Kubernetes Service (AKS)](https://learn.microsoft.com/en-us/azure/aks/configure-azure-cni) # current config - [Configure Azure CNI Powered by Cilium in Azure Kubernetes Service (AKS) (Preview)](https://learn.microsoft.com/en-us/azure/aks/azure-cni-powered-by-cilium) # may be new config - [Bring your own Container Network Interface (CNI) plugin with Azure Kubernetes Service (AKS)](https://learn.microsoft.com/en-us/azure/aks/use-byo-cni?tabs=azure-cli) # should evaluate with cilium ### Tasks Deploy ttd-at25-aks: - Check if converting existing clusters is possible in place - Evaluate what CNI we should set as standard - Test "Azure CNI Powered by Cilium" vs "Bring your own CNI with Cilium" - Update iac to set up dualstack ipv4 and ipv6 ### Acceptance Criterias apps AKS expose ipv6 for inbound and aoutbound traffic as minimum. Preferably ipv6 should be used internal in cluster.
infrastructure
poc apps aks with dualstack and description to be compliant with we need to run dualstack with and in all apps clusters we need to investigate how to convert existing clusters and make choices about new clusters additional information current config may be new config should evaluate with cilium tasks deploy ttd aks check if converting existing clusters is possible in place evaluate what cni we should set as standard test azure cni powered by cilium vs bring your own cni with cilium update iac to set up dualstack and acceptance criterias apps aks expose for inbound and aoutbound traffic as minimum preferably should be used internal in cluster
1
2,965
3,987,990,366
IssuesEvent
2016-05-09 07:50:42
scenarioo/scenarioo
https://api.github.com/repos/scenarioo/scenarioo
closed
Remove bower and always npm instead
topic:infrastructure topic:webapp
Why? Because its simpler to just have one dependency manager. ### Acceptance criteria * bower is not needed anymore for Scenarioo development * bower is removed from the build server * npm is used instead for all dependencies * The wiki is adapted * All developers are informed about the change by e-mail ### Implementation See: https://gofore.com/ohjelmistokehitys/stop-using-bower/
1.0
Remove bower and always npm instead - Why? Because its simpler to just have one dependency manager. ### Acceptance criteria * bower is not needed anymore for Scenarioo development * bower is removed from the build server * npm is used instead for all dependencies * The wiki is adapted * All developers are informed about the change by e-mail ### Implementation See: https://gofore.com/ohjelmistokehitys/stop-using-bower/
infrastructure
remove bower and always npm instead why because its simpler to just have one dependency manager acceptance criteria bower is not needed anymore for scenarioo development bower is removed from the build server npm is used instead for all dependencies the wiki is adapted all developers are informed about the change by e mail implementation see
1
10,789
8,722,284,222
IssuesEvent
2018-12-09 10:50:16
coq/coq
https://api.github.com/repos/coq/coq
closed
Change important CI DOS batch files to CRLF
kind: infrastructure platform: Windows
Currently all Windows batch files used on CI use LF line endings. Unfortunately this can have the effect that jump/call labels are not found (but only if the jump label is more than 512 bytes away from the call and the moon is in a certain constellation to Saturn, otherwise it works fine - oh how we love MS). So I fear I have to change some important CI batch files to CRLF, although this makes diffing hard. See https://web.archive.org/web/20160305063042/help.wugnet.com/windows/system-find-batch-label-ftopict615555.html I actually run into this issue in CI job https://gitlab.com/coq/coq/-/jobs/103241210 At the end it complains "The system cannot find the batch label specified - CopyLogFiles", but the label is definitely there.
1.0
Change important CI DOS batch files to CRLF - Currently all Windows batch files used on CI use LF line endings. Unfortunately this can have the effect that jump/call labels are not found (but only if the jump label is more than 512 bytes away from the call and the moon is in a certain constellation to Saturn, otherwise it works fine - oh how we love MS). So I fear I have to change some important CI batch files to CRLF, although this makes diffing hard. See https://web.archive.org/web/20160305063042/help.wugnet.com/windows/system-find-batch-label-ftopict615555.html I actually run into this issue in CI job https://gitlab.com/coq/coq/-/jobs/103241210 At the end it complains "The system cannot find the batch label specified - CopyLogFiles", but the label is definitely there.
infrastructure
change important ci dos batch files to crlf currently all windows batch files used on ci use lf line endings unfortunately this can have the effect that jump call labels are not found but only if the jump label is more than bytes away from the call and the moon is in a certain constellation to saturn otherwise it works fine oh how we love ms so i fear i have to change some important ci batch files to crlf although this makes diffing hard see i actually run into this issue in ci job at the end it complains the system cannot find the batch label specified copylogfiles but the label is definitely there
1
54,563
23,304,284,919
IssuesEvent
2022-08-07 19:45:33
Fevol/obsidian-translate
https://api.github.com/repos/Fevol/obsidian-translate
closed
Fanyi Youdao integration
translation service
(Request from Discord) Implement Youdao as a translation service ## Tasks - [x] Adapt DummyTranslate to use Youdao's API - [x] Add appropriate options to settings page - [x] Add service info to constants.ts - [x] Add service info to README.md
1.0
Fanyi Youdao integration - (Request from Discord) Implement Youdao as a translation service ## Tasks - [x] Adapt DummyTranslate to use Youdao's API - [x] Add appropriate options to settings page - [x] Add service info to constants.ts - [x] Add service info to README.md
non_infrastructure
fanyi youdao integration request from discord implement youdao as a translation service tasks adapt dummytranslate to use youdao s api add appropriate options to settings page add service info to constants ts add service info to readme md
0
21,673
14,704,799,674
IssuesEvent
2021-01-04 17:03:05
raiden-network/light-client
https://api.github.com/repos/raiden-network/light-client
opened
Upgrade webpack to 5.0
dependencies enhancement infrastructure 🚧
## Description We're still using webpack 4.0, while 5.0 is available. Certain plugins start requiring webpack 5, like https://github.com/webpack-contrib/source-map-loader/releases/tag/v2.0.0 #### Resources - https://webpack.js.org/migrate/5/ ## Acceptance criteria - CI and build work as before ## Tasks - [ ] Update to webpack 5.0
1.0
Upgrade webpack to 5.0 - ## Description We're still using webpack 4.0, while 5.0 is available. Certain plugins start requiring webpack 5, like https://github.com/webpack-contrib/source-map-loader/releases/tag/v2.0.0 #### Resources - https://webpack.js.org/migrate/5/ ## Acceptance criteria - CI and build work as before ## Tasks - [ ] Update to webpack 5.0
infrastructure
upgrade webpack to description we re still using webpack while is available certain plugins start requiring webpack like resources acceptance criteria ci and build work as before tasks update to webpack
1
304,159
9,321,964,356
IssuesEvent
2019-03-27 06:30:17
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.reddit.com - site is not usable
browser-firefox-mobile priority-critical
<!-- @browser: Firefox Mobile 66.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.reddit.com/r/ketoscience/comments/abtjp4/effects_of_resistance_vs_aerobic_training/ **Browser / Version**: Firefox Mobile 66.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: site requests app or chrome **Steps to Reproduce**: site pops up an overlay of the article asking for you to continue using either the reddit mobile app or chrome. No option for firefox and the site seems to work fine (save for the obnoxious overlay) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190304101322</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://stats.redditmedia.com/. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://stats.redditmedia.com/. (Reason: CORS request did not succeed)."]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.reddit.com - site is not usable - <!-- @browser: Firefox Mobile 66.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:66.0) Gecko/66.0 Firefox/66.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.reddit.com/r/ketoscience/comments/abtjp4/effects_of_resistance_vs_aerobic_training/ **Browser / Version**: Firefox Mobile 66.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: site requests app or chrome **Steps to Reproduce**: site pops up an overlay of the article asking for you to continue using either the reddit mobile app or chrome. No option for firefox and the site seems to work fine (save for the obnoxious overlay) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190304101322</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: true</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://stats.redditmedia.com/. (Reason: CORS request did not succeed)."]', u'[JavaScript Warning: "Cross-Origin Request Blocked: The Same Origin Policy disallows reading the remote resource at https://stats.redditmedia.com/. (Reason: CORS request did not succeed)."]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_infrastructure
site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description site requests app or chrome steps to reproduce site pops up an overlay of the article asking for you to continue using either the reddit mobile app or chrome no option for firefox and the site seems to work fine save for the obnoxious overlay browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen true mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u from with ❤️
0
27,736
22,284,423,569
IssuesEvent
2022-06-11 11:47:17
srobo/tasks
https://api.github.com/repos/srobo/tasks
opened
Audit network cables
A: Infrastructure A: Competition Operations
SR has a number of network cables of various sizes, histories and states of repair. At the SR2022 competition it was noted that some of them have damaged clips and/or may not conform to the 568B wiring spec, the latter potentially making it harder to diagnose network issues. At some point it might be worth reviewing the state of the cables we have and potentially rationalising them.
1.0
Audit network cables - SR has a number of network cables of various sizes, histories and states of repair. At the SR2022 competition it was noted that some of them have damaged clips and/or may not conform to the 568B wiring spec, the latter potentially making it harder to diagnose network issues. At some point it might be worth reviewing the state of the cables we have and potentially rationalising them.
infrastructure
audit network cables sr has a number of network cables of various sizes histories and states of repair at the competition it was noted that some of them have damaged clips and or may not conform to the wiring spec the latter potentially making it harder to diagnose network issues at some point it might be worth reviewing the state of the cables we have and potentially rationalising them
1
363,247
25,414,732,590
IssuesEvent
2022-11-22 22:30:58
vaporsoftwares/Vapor
https://api.github.com/repos/vaporsoftwares/Vapor
closed
Known bugs
documentation
### Known bugs * Dev armor renders as diamond armor. * Dev Tools name are messed * F3 is something else... * Missing Textures (Thanks Optifine) * Open to LAN button (Fixed) __PATCHED BUGS__ * ~Glass block (and pane) has texture of 1.13 but renders has the normal texture.~ _PATCHED_ * ~dev_sword has no name set.~ _PATCHED_ * ~dev_pickaxe has no name set.~ _PATCHED_ * ~dev_axe has no name set.~ _PATCHED_ ➟ _These will be patch with version 0.4. If you know any other bugs, please create a new issue._
1.0
Known bugs - ### Known bugs * Dev armor renders as diamond armor. * Dev Tools name are messed * F3 is something else... * Missing Textures (Thanks Optifine) * Open to LAN button (Fixed) __PATCHED BUGS__ * ~Glass block (and pane) has texture of 1.13 but renders has the normal texture.~ _PATCHED_ * ~dev_sword has no name set.~ _PATCHED_ * ~dev_pickaxe has no name set.~ _PATCHED_ * ~dev_axe has no name set.~ _PATCHED_ ➟ _These will be patch with version 0.4. If you know any other bugs, please create a new issue._
non_infrastructure
known bugs known bugs dev armor renders as diamond armor dev tools name are messed is something else missing textures thanks optifine open to lan button fixed patched bugs glass block and pane has texture of but renders has the normal texture patched dev sword has no name set patched dev pickaxe has no name set patched dev axe has no name set patched ➟ these will be patch with version if you know any other bugs please create a new issue
0
22,098
14,986,767,187
IssuesEvent
2021-01-28 21:43:40
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Pasting a model when no model is selected results in an error
bug interface/infrastructure
Pressing ctrl-v to paste a model while no model is selected results in an error being displayed - it would be better to not attempt to perform a paste operation in this case.
1.0
Pasting a model when no model is selected results in an error - Pressing ctrl-v to paste a model while no model is selected results in an error being displayed - it would be better to not attempt to perform a paste operation in this case.
infrastructure
pasting a model when no model is selected results in an error pressing ctrl v to paste a model while no model is selected results in an error being displayed it would be better to not attempt to perform a paste operation in this case
1
538
8,432,632,329
IssuesEvent
2018-10-17 03:06:54
dotnet/project-system
https://api.github.com/repos/dotnet/project-system
closed
NRE in TaskDelayScheduler
Bug Project-System-CPS Tenet-Reliability
Observed debugging 031b0db6d. ![image](https://user-images.githubusercontent.com/350947/46945196-15778580-d06c-11e8-85cb-aac8bdf7a71f.png) `cancel` is `true`, so line 131 must have run, and `PendingUpdateTokenSource` would not have been `null`. To have `cts` equal `null` on line 135, `PendingUpdateTokenSource` must have been cleared between 131 and 133. Presumably a callback on cancellation did on the calling thread which would enter the lock. A simple fix for the NRE would be to copy `PendingUpdateTokenSource` to a local earlier and re-use it for reads. However there's possibly a deeper issue here. ``` System.NullReferenceException HResult=0x80004003 Message=Object reference not set to an instance of an object. Source=Microsoft.VisualStudio.ProjectSystem.Managed StackTrace: at Microsoft.VisualStudio.Threading.Tasks.TaskDelayScheduler.ClearPendingUpdates(Boolean cancel) in D:\repos\project-system\src\Microsoft.VisualStudio.ProjectSystem.Managed\Threading\Tasks\TaskDelayScheduler.cs:line 135 at Microsoft.VisualStudio.Threading.Tasks.TaskDelayScheduler.ScheduleAsyncTask(Func`2 asyncFunctionToCall) in D:\repos\project-system\src\Microsoft.VisualStudio.ProjectSystem.Managed\Threading\Tasks\TaskDelayScheduler.cs:line 63 at Microsoft.VisualStudio.ProjectSystem.VS.NuGet.ProjectAssetFileWatcher.FilesChanged(UInt32 cChanges, String[] rgpszFile, UInt32[] rggrfChange) in D:\repos\project-system\src\Microsoft.VisualStudio.ProjectSystem.Managed.VS\ProjectSystem\VS\NuGet\ProjectAssetFileWatcher.cs:line 306 at Microsoft.VisualStudio.Services.FileChangeSubscription.<>c__DisplayClass33_0.<NotifyCore>b__0() ```
True
NRE in TaskDelayScheduler - Observed debugging 031b0db6d. ![image](https://user-images.githubusercontent.com/350947/46945196-15778580-d06c-11e8-85cb-aac8bdf7a71f.png) `cancel` is `true`, so line 131 must have run, and `PendingUpdateTokenSource` would not have been `null`. To have `cts` equal `null` on line 135, `PendingUpdateTokenSource` must have been cleared between 131 and 133. Presumably a callback on cancellation did on the calling thread which would enter the lock. A simple fix for the NRE would be to copy `PendingUpdateTokenSource` to a local earlier and re-use it for reads. However there's possibly a deeper issue here. ``` System.NullReferenceException HResult=0x80004003 Message=Object reference not set to an instance of an object. Source=Microsoft.VisualStudio.ProjectSystem.Managed StackTrace: at Microsoft.VisualStudio.Threading.Tasks.TaskDelayScheduler.ClearPendingUpdates(Boolean cancel) in D:\repos\project-system\src\Microsoft.VisualStudio.ProjectSystem.Managed\Threading\Tasks\TaskDelayScheduler.cs:line 135 at Microsoft.VisualStudio.Threading.Tasks.TaskDelayScheduler.ScheduleAsyncTask(Func`2 asyncFunctionToCall) in D:\repos\project-system\src\Microsoft.VisualStudio.ProjectSystem.Managed\Threading\Tasks\TaskDelayScheduler.cs:line 63 at Microsoft.VisualStudio.ProjectSystem.VS.NuGet.ProjectAssetFileWatcher.FilesChanged(UInt32 cChanges, String[] rgpszFile, UInt32[] rggrfChange) in D:\repos\project-system\src\Microsoft.VisualStudio.ProjectSystem.Managed.VS\ProjectSystem\VS\NuGet\ProjectAssetFileWatcher.cs:line 306 at Microsoft.VisualStudio.Services.FileChangeSubscription.<>c__DisplayClass33_0.<NotifyCore>b__0() ```
non_infrastructure
nre in taskdelayscheduler observed debugging cancel is true so line must have run and pendingupdatetokensource would not have been null to have cts equal null on line pendingupdatetokensource must have been cleared between and presumably a callback on cancellation did on the calling thread which would enter the lock a simple fix for the nre would be to copy pendingupdatetokensource to a local earlier and re use it for reads however there s possibly a deeper issue here system nullreferenceexception hresult message object reference not set to an instance of an object source microsoft visualstudio projectsystem managed stacktrace at microsoft visualstudio threading tasks taskdelayscheduler clearpendingupdates boolean cancel in d repos project system src microsoft visualstudio projectsystem managed threading tasks taskdelayscheduler cs line at microsoft visualstudio threading tasks taskdelayscheduler scheduleasynctask func asyncfunctiontocall in d repos project system src microsoft visualstudio projectsystem managed threading tasks taskdelayscheduler cs line at microsoft visualstudio projectsystem vs nuget projectassetfilewatcher fileschanged cchanges string rgpszfile rggrfchange in d repos project system src microsoft visualstudio projectsystem managed vs projectsystem vs nuget projectassetfilewatcher cs line at microsoft visualstudio services filechangesubscription c b
0
13,800
16,523,997,783
IssuesEvent
2021-05-26 17:36:10
opensearch-project/opensearch-plugins
https://api.github.com/repos/opensearch-project/opensearch-plugins
closed
Document RestAPI backward compatibility support for routes on 9600 port
backwards-compatibility enhancement
The RestAPIs design documented for backward compatibility takes care of routes in 9200. For other routes which run on 9600 we have a different design.
True
Document RestAPI backward compatibility support for routes on 9600 port - The RestAPIs design documented for backward compatibility takes care of routes in 9200. For other routes which run on 9600 we have a different design.
non_infrastructure
document restapi backward compatibility support for routes on port the restapis design documented for backward compatibility takes care of routes in for other routes which run on we have a different design
0
25,367
18,666,711,080
IssuesEvent
2021-10-30 00:38:55
Mudlet/Mudlet
https://api.github.com/repos/Mudlet/Mudlet
closed
PRs made by automation don't run checks
Public infrastructure
#### Brief summary of issue / Description of requested feature: PRs made by automation don't run checks - this has been an issue for a while. And because the checks are mandatory, the PRs can't be merged either because the checks aren't passing. #### Steps to reproduce the issue / Reasons for adding feature: 1. See https://github.com/Mudlet/Mudlet/pull/5447 for example #### Error output / Expected result of feature All of the usual checks are run as needed. #### Extra information, such as Mudlet version, operating system and ideas for how to solve / implement: Reached out to Github [community here](https://github.community/t/workflow-not-running-for-pull-requests-made-by-bots/170245?u=vadi2), but there's no real responses.
1.0
PRs made by automation don't run checks - #### Brief summary of issue / Description of requested feature: PRs made by automation don't run checks - this has been an issue for a while. And because the checks are mandatory, the PRs can't be merged either because the checks aren't passing. #### Steps to reproduce the issue / Reasons for adding feature: 1. See https://github.com/Mudlet/Mudlet/pull/5447 for example #### Error output / Expected result of feature All of the usual checks are run as needed. #### Extra information, such as Mudlet version, operating system and ideas for how to solve / implement: Reached out to Github [community here](https://github.community/t/workflow-not-running-for-pull-requests-made-by-bots/170245?u=vadi2), but there's no real responses.
infrastructure
prs made by automation don t run checks brief summary of issue description of requested feature prs made by automation don t run checks this has been an issue for a while and because the checks are mandatory the prs can t be merged either because the checks aren t passing steps to reproduce the issue reasons for adding feature see for example error output expected result of feature all of the usual checks are run as needed extra information such as mudlet version operating system and ideas for how to solve implement reached out to github but there s no real responses
1
24,178
16,989,257,538
IssuesEvent
2021-06-30 18:08:29
dotnet/dotnet-docker
https://api.github.com/repos/dotnet/dotnet-docker
closed
Unable to run published files into container with latest .net 6 preview
area-infrastructure bug untriaged
[Dockerfile.txt](https://github.com/dotnet/dotnet-docker/files/6729636/Dockerfile.txt) When I try the commands to copy the published files to docker and run, it always fails. Gives some type of access file violation. Tried the same thing under .net 5 and no problems. An earlier preview of .net 6 did not have the problem at all. Would be disappointed if a person was not allowed to publish files and copying those files to docker. There are cases where I need libraries not in the folders and not on nuget and don't want to be forced to publish all my libraries to nuget even though it may access files from my personal computer.
1.0
Unable to run published files into container with latest .net 6 preview - [Dockerfile.txt](https://github.com/dotnet/dotnet-docker/files/6729636/Dockerfile.txt) When I try the commands to copy the published files to docker and run, it always fails. Gives some type of access file violation. Tried the same thing under .net 5 and no problems. An earlier preview of .net 6 did not have the problem at all. Would be disappointed if a person was not allowed to publish files and copying those files to docker. There are cases where I need libraries not in the folders and not on nuget and don't want to be forced to publish all my libraries to nuget even though it may access files from my personal computer.
infrastructure
unable to run published files into container with latest net preview when i try the commands to copy the published files to docker and run it always fails gives some type of access file violation tried the same thing under net and no problems an earlier preview of net did not have the problem at all would be disappointed if a person was not allowed to publish files and copying those files to docker there are cases where i need libraries not in the folders and not on nuget and don t want to be forced to publish all my libraries to nuget even though it may access files from my personal computer
1
11,572
9,299,608,315
IssuesEvent
2019-03-23 05:15:34
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
opened
few issues with tests/setup-stress-dependencies.sh
GCStress area-Infrastructure os-linux
We have warnings like: `/__w/3/s/tests/../init-tools.sh: line 20: return: can only `return' from a function or sourced script` because `init-tools.sh` was supposed to be called via `source`. and we have `grep: write error: Broken pipe` issue that we need to fix.
1.0
few issues with tests/setup-stress-dependencies.sh - We have warnings like: `/__w/3/s/tests/../init-tools.sh: line 20: return: can only `return' from a function or sourced script` because `init-tools.sh` was supposed to be called via `source`. and we have `grep: write error: Broken pipe` issue that we need to fix.
infrastructure
few issues with tests setup stress dependencies sh we have warnings like w s tests init tools sh line return can only return from a function or sourced script because init tools sh was supposed to be called via source and we have grep write error broken pipe issue that we need to fix
1
33,206
12,194,095,292
IssuesEvent
2020-04-29 15:19:06
kenferrara/atlasdb
https://api.github.com/repos/kenferrara/atlasdb
opened
CVE-2018-1320 (High) detected in libthrift-0.11.0.jar
security vulnerability
## CVE-2018-1320 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libthrift-0.11.0.jar</b></p></summary> <p>Thrift is a software framework for scalable cross-language services development.</p> <p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,le/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,le/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar</p> <p> Dependency Hierarchy: - :x: **libthrift-0.11.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kenferrara/atlasdb/commit/9298e273ad7b99d46b6349ab7e65393752bb7e71">9298e273ad7b99d46b6349ab7e65393752bb7e71</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Thrift Java client library versions 0.5.0 through 0.11.0 can bypass SASL negotiation isComplete validation in the org.apache.thrift.transport.TSaslTransport class. An assert used to determine if the SASL handshake had successfully completed could be disabled in production settings making the validation incomplete. <p>Publish Date: 2019-01-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1320>CVE-2018-1320</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1320">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1320</a></p> <p>Release Date: 2019-01-07</p> <p>Fix Resolution: 0.12.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.thrift","packageName":"libthrift","packageVersion":"0.11.0","isTransitiveDependency":false,"dependencyTree":"org.apache.thrift:libthrift:0.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.12.0"}],"vulnerabilityIdentifier":"CVE-2018-1320","vulnerabilityDetails":"Apache Thrift Java client library versions 0.5.0 through 0.11.0 can bypass SASL negotiation isComplete validation in the org.apache.thrift.transport.TSaslTransport class. An assert used to determine if the SASL handshake had successfully completed could be disabled in production settings making the validation incomplete.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1320","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-1320 (High) detected in libthrift-0.11.0.jar - ## CVE-2018-1320 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>libthrift-0.11.0.jar</b></p></summary> <p>Thrift is a software framework for scalable cross-language services development.</p> <p>Path to dependency file: /tmp/ws-scm/atlasdb/atlasdb-cassandra/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,le/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,le/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar,/root/.gradle/caches/modules-2/files-2.1/org.apache.thrift/libthrift/0.11.0/4f4f1c1fbbae63258625dea71007fa41bee7edb3/libthrift-0.11.0.jar</p> <p> Dependency Hierarchy: - :x: **libthrift-0.11.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kenferrara/atlasdb/commit/9298e273ad7b99d46b6349ab7e65393752bb7e71">9298e273ad7b99d46b6349ab7e65393752bb7e71</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Thrift Java client library versions 0.5.0 through 0.11.0 can bypass SASL negotiation isComplete validation in the org.apache.thrift.transport.TSaslTransport class. An assert used to determine if the SASL handshake had successfully completed could be disabled in production settings making the validation incomplete. <p>Publish Date: 2019-01-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1320>CVE-2018-1320</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1320">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1320</a></p> <p>Release Date: 2019-01-07</p> <p>Fix Resolution: 0.12.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.thrift","packageName":"libthrift","packageVersion":"0.11.0","isTransitiveDependency":false,"dependencyTree":"org.apache.thrift:libthrift:0.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.12.0"}],"vulnerabilityIdentifier":"CVE-2018-1320","vulnerabilityDetails":"Apache Thrift Java client library versions 0.5.0 through 0.11.0 can bypass SASL negotiation isComplete validation in the org.apache.thrift.transport.TSaslTransport class. An assert used to determine if the SASL handshake had successfully completed could be disabled in production settings making the validation incomplete.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1320","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_infrastructure
cve high detected in libthrift jar cve high severity vulnerability vulnerable library libthrift jar thrift is a software framework for scalable cross language services development path to dependency file tmp ws scm atlasdb atlasdb cassandra build gradle path to vulnerable library root gradle caches modules files org apache thrift libthrift libthrift jar root gradle caches modules files org apache thrift libthrift libthrift jar le caches modules files org apache thrift libthrift libthrift jar root gradle caches modules files org apache thrift libthrift libthrift jar le caches modules files org apache thrift libthrift libthrift jar root gradle caches modules files org apache thrift libthrift libthrift jar dependency hierarchy x libthrift jar vulnerable library found in head commit a href vulnerability details apache thrift java client library versions through can bypass sasl negotiation iscomplete validation in the org apache thrift transport tsasltransport class an assert used to determine if the sasl handshake had successfully completed could be disabled in production settings making the validation incomplete publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails apache thrift java client library versions through can bypass sasl negotiation iscomplete validation in the org apache thrift transport tsasltransport class an assert used to determine if the sasl handshake had successfully completed could be disabled in production settings making the validation incomplete vulnerabilityurl
0
16,129
11,848,024,015
IssuesEvent
2020-03-24 13:06:49
eclipse/antenna
https://api.github.com/repos/eclipse/antenna
opened
Continuously build Antenna on Eclipse Infrastructure ad deploy artifacts
infrastructure
### Summary of the Improvement Establish builds on Eclipse Jenkins and deploy artifacts on a regular basis to the repositories ### Acceptance Criteria - [ ] Jenkins build runs on Eclipse Jenkins - [ ] Snapshot builds are deployed to Snapshot repo - [ ] Weekly release builds are deployed as Releases ### Definition of Done - Acceptance criteria fulfilled - The ticket describes what has been done and describes how the results can be verified (e.g. new Jenkins step is inserted, the report can be viewed at X) - All known limitations are listed and it is described how they will be addressed (e.g., a ticket for the remaining stuff) - Documentation is updated - The result has been reviewed: Everything works as described and can be verified reading only the documentation
1.0
Continuously build Antenna on Eclipse Infrastructure ad deploy artifacts - ### Summary of the Improvement Establish builds on Eclipse Jenkins and deploy artifacts on a regular basis to the repositories ### Acceptance Criteria - [ ] Jenkins build runs on Eclipse Jenkins - [ ] Snapshot builds are deployed to Snapshot repo - [ ] Weekly release builds are deployed as Releases ### Definition of Done - Acceptance criteria fulfilled - The ticket describes what has been done and describes how the results can be verified (e.g. new Jenkins step is inserted, the report can be viewed at X) - All known limitations are listed and it is described how they will be addressed (e.g., a ticket for the remaining stuff) - Documentation is updated - The result has been reviewed: Everything works as described and can be verified reading only the documentation
infrastructure
continuously build antenna on eclipse infrastructure ad deploy artifacts summary of the improvement establish builds on eclipse jenkins and deploy artifacts on a regular basis to the repositories acceptance criteria jenkins build runs on eclipse jenkins snapshot builds are deployed to snapshot repo weekly release builds are deployed as releases definition of done acceptance criteria fulfilled the ticket describes what has been done and describes how the results can be verified e g new jenkins step is inserted the report can be viewed at x all known limitations are listed and it is described how they will be addressed e g a ticket for the remaining stuff documentation is updated the result has been reviewed everything works as described and can be verified reading only the documentation
1
334,788
24,437,834,683
IssuesEvent
2022-10-06 12:45:54
scikit-learn/scikit-learn
https://api.github.com/repos/scikit-learn/scikit-learn
opened
Possible erroneous use of chi2
Documentation Needs Triage
### Describe the issue linked to the documentation The [documentation page](https://scikit-learn.org/stable/modules/feature_selection.html#univariate-feature-selection) for univariate feature selection suggests that the [`sklearn.feature_selection.chi2`](https://github.com/scikit-learn/scikit-learn/blob/36958fb24/sklearn/feature_selection/_univariate_selection.py#L168) function could be used to rank (and possibly select) features. The problem is that the example given in the documentation is the Iris data set where the features are continuous. The `chi2` function is however supposed to work with contingency matrices of discrete features (i.e., typically counts, frequencies, or booleans). The use of `chi2` is therefore likely to be misleading in this context. See [this discussion](https://stats.stackexchange.com/questions/591202/can-the-χ²-test-be-used-without-a-contingency-table/) on StackExchange for further details. ### Suggest a potential alternative/fix Either find a different use of `chi` that is more aligned with the fact that it works with contingency matrices or explain more clearly how it could be used with continuous features.
1.0
Possible erroneous use of chi2 - ### Describe the issue linked to the documentation The [documentation page](https://scikit-learn.org/stable/modules/feature_selection.html#univariate-feature-selection) for univariate feature selection suggests that the [`sklearn.feature_selection.chi2`](https://github.com/scikit-learn/scikit-learn/blob/36958fb24/sklearn/feature_selection/_univariate_selection.py#L168) function could be used to rank (and possibly select) features. The problem is that the example given in the documentation is the Iris data set where the features are continuous. The `chi2` function is however supposed to work with contingency matrices of discrete features (i.e., typically counts, frequencies, or booleans). The use of `chi2` is therefore likely to be misleading in this context. See [this discussion](https://stats.stackexchange.com/questions/591202/can-the-χ²-test-be-used-without-a-contingency-table/) on StackExchange for further details. ### Suggest a potential alternative/fix Either find a different use of `chi` that is more aligned with the fact that it works with contingency matrices or explain more clearly how it could be used with continuous features.
non_infrastructure
possible erroneous use of describe the issue linked to the documentation the for univariate feature selection suggests that the function could be used to rank and possibly select features the problem is that the example given in the documentation is the iris data set where the features are continuous the function is however supposed to work with contingency matrices of discrete features i e typically counts frequencies or booleans the use of is therefore likely to be misleading in this context see on stackexchange for further details suggest a potential alternative fix either find a different use of chi that is more aligned with the fact that it works with contingency matrices or explain more clearly how it could be used with continuous features
0
189,567
15,191,215,566
IssuesEvent
2021-02-15 19:27:45
udistrital/financiera_documentacion
https://api.github.com/repos/udistrital/financiera_documentacion
opened
Historia de Usuario de Transacciones contables
Documentation
De acuerdo a los análisis planteados en las mesas de trabajo anteriores, se procede a la elaboración de las respectiva historia de usuario.
1.0
Historia de Usuario de Transacciones contables - De acuerdo a los análisis planteados en las mesas de trabajo anteriores, se procede a la elaboración de las respectiva historia de usuario.
non_infrastructure
historia de usuario de transacciones contables de acuerdo a los análisis planteados en las mesas de trabajo anteriores se procede a la elaboración de las respectiva historia de usuario
0
33,288
27,363,880,096
IssuesEvent
2023-02-27 17:39:10
swmcc/swm.cc
https://api.github.com/repos/swmcc/swm.cc
opened
Add the-mcculloughs.org to the project section
infrastructure
*Feature:* Adding an article to my personal website about the-mcculloughs.org project **Scenario:** Detailing technical issues faced and project goals **Given** I have been thinking on the-mcculloughs.org project for several months *And* I will encounter various technical issues during development *And* I have successfully started the project *When* I decide to write an article about the project on my personal website *And* include a detailed overview of the project goals and technical issues faced **Then** I will begin by describing the project's goal to create an online database of all the McCulloughs around the world **And** I will explain the technical challenges I faced, such as data management, web hosting, and software integration **And** I will describe how I overcame these issues to complete the project successfully
1.0
Add the-mcculloughs.org to the project section - *Feature:* Adding an article to my personal website about the-mcculloughs.org project **Scenario:** Detailing technical issues faced and project goals **Given** I have been thinking on the-mcculloughs.org project for several months *And* I will encounter various technical issues during development *And* I have successfully started the project *When* I decide to write an article about the project on my personal website *And* include a detailed overview of the project goals and technical issues faced **Then** I will begin by describing the project's goal to create an online database of all the McCulloughs around the world **And** I will explain the technical challenges I faced, such as data management, web hosting, and software integration **And** I will describe how I overcame these issues to complete the project successfully
infrastructure
add the mcculloughs org to the project section feature adding an article to my personal website about the mcculloughs org project scenario detailing technical issues faced and project goals given i have been thinking on the mcculloughs org project for several months and i will encounter various technical issues during development and i have successfully started the project when i decide to write an article about the project on my personal website and include a detailed overview of the project goals and technical issues faced then i will begin by describing the project s goal to create an online database of all the mcculloughs around the world and i will explain the technical challenges i faced such as data management web hosting and software integration and i will describe how i overcame these issues to complete the project successfully
1
12,246
9,661,184,879
IssuesEvent
2019-05-20 17:19:43
apple/turicreate
https://api.github.com/repos/apple/turicreate
closed
Import broken in 5.4 on linux GPUs for OD
bug infrastructure object detection p1 style transfer
Due to the lazy import of MXNet, on GPUs, object detection is broken with 5.4. https://github.com/apple/turicreate/search?q=_print_neural_compute_device&unscoped_q=_print_neural_compute_device ``` PIP: package/turicreate/toolkits/_internal_utils.py, line 724 should be: from . Import _mxnet And line 728 should be: _mxnet._mxnet_utils... ```
1.0
Import broken in 5.4 on linux GPUs for OD - Due to the lazy import of MXNet, on GPUs, object detection is broken with 5.4. https://github.com/apple/turicreate/search?q=_print_neural_compute_device&unscoped_q=_print_neural_compute_device ``` PIP: package/turicreate/toolkits/_internal_utils.py, line 724 should be: from . Import _mxnet And line 728 should be: _mxnet._mxnet_utils... ```
infrastructure
import broken in on linux gpus for od due to the lazy import of mxnet on gpus object detection is broken with pip package turicreate toolkits internal utils py line should be from import mxnet and line should be mxnet mxnet utils
1
15,110
11,355,647,318
IssuesEvent
2020-01-24 20:33:31
enarx/enarx
https://api.github.com/repos/enarx/enarx
closed
Issue Templates
infrastructure
We need some decent issue templates. This is related, in part, to project infrastructure automation.
1.0
Issue Templates - We need some decent issue templates. This is related, in part, to project infrastructure automation.
infrastructure
issue templates we need some decent issue templates this is related in part to project infrastructure automation
1
184,848
21,785,014,046
IssuesEvent
2022-05-14 02:09:45
ChoeMinji/spring_framework-5.1.8
https://api.github.com/repos/ChoeMinji/spring_framework-5.1.8
reopened
CVE-2022-22965 (High) detected in SpringFrameworkv5.1.8.RELEASE
security vulnerability
## CVE-2022-22965 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>SpringFrameworkv5.1.8.RELEASE</b></p></summary> <p> <p>Library home page: <a href=https://github.com/MeghanaSudhakarB/SpringFramework.git>https://github.com/MeghanaSudhakarB/SpringFramework.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/spring-beans/src/main/java/org/springframework/beans/CachedIntrospectionResults.java</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/spring-beans/src/main/java/org/springframework/beans/CachedIntrospectionResults.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Framework before 5.2.20 and 5.3.x before 5.3.18 are vulnerable due to a vulnerability in Spring-beans which allows attackers under certain circumstances to achieve remote code execution, this vulnerability is also known as ״Spring4Shell״ or ״SpringShell״. The current POC related to the attack is done by creating a specially crafted request which manipulates ClassLoader to successfully achieve RCE (Remote Code Execution). Please note that the ease of exploitation may diverge by the code implementation.Currently, the exploit requires JDK 9 or higher, Apache Tomcat as the Servlet container, the application Packaged as WAR, and dependency on spring-webmvc or spring-webflux. Spring Framework 5.3.18 and 5.2.20 have already been released. WhiteSource's research team is carefully observing developments and researching the case. We will keep updating this page and our WhiteSource resources with updates. <p>Publish Date: 2022-01-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22965>CVE-2022-22965</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://spring.io/blog/2022/03/31/spring-framework-rce-early-announcement">https://spring.io/blog/2022/03/31/spring-framework-rce-early-announcement</a></p> <p>Release Date: 2022-01-11</p> <p>Fix Resolution: org.springframework:spring-beans:5.2.20.RELEASE,5.3.18</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-22965 (High) detected in SpringFrameworkv5.1.8.RELEASE - ## CVE-2022-22965 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>SpringFrameworkv5.1.8.RELEASE</b></p></summary> <p> <p>Library home page: <a href=https://github.com/MeghanaSudhakarB/SpringFramework.git>https://github.com/MeghanaSudhakarB/SpringFramework.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/spring-beans/src/main/java/org/springframework/beans/CachedIntrospectionResults.java</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/spring-beans/src/main/java/org/springframework/beans/CachedIntrospectionResults.java</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Spring Framework before 5.2.20 and 5.3.x before 5.3.18 are vulnerable due to a vulnerability in Spring-beans which allows attackers under certain circumstances to achieve remote code execution, this vulnerability is also known as ״Spring4Shell״ or ״SpringShell״. The current POC related to the attack is done by creating a specially crafted request which manipulates ClassLoader to successfully achieve RCE (Remote Code Execution). Please note that the ease of exploitation may diverge by the code implementation.Currently, the exploit requires JDK 9 or higher, Apache Tomcat as the Servlet container, the application Packaged as WAR, and dependency on spring-webmvc or spring-webflux. Spring Framework 5.3.18 and 5.2.20 have already been released. WhiteSource's research team is carefully observing developments and researching the case. We will keep updating this page and our WhiteSource resources with updates. <p>Publish Date: 2022-01-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22965>CVE-2022-22965</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://spring.io/blog/2022/03/31/spring-framework-rce-early-announcement">https://spring.io/blog/2022/03/31/spring-framework-rce-early-announcement</a></p> <p>Release Date: 2022-01-11</p> <p>Fix Resolution: org.springframework:spring-beans:5.2.20.RELEASE,5.3.18</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in release cve high severity vulnerability vulnerable library release library home page a href found in base branch master vulnerable source files spring beans src main java org springframework beans cachedintrospectionresults java spring beans src main java org springframework beans cachedintrospectionresults java vulnerability details spring framework before and x before are vulnerable due to a vulnerability in spring beans which allows attackers under certain circumstances to achieve remote code execution this vulnerability is also known as ״ ״ or ״springshell״ the current poc related to the attack is done by creating a specially crafted request which manipulates classloader to successfully achieve rce remote code execution please note that the ease of exploitation may diverge by the code implementation currently the exploit requires jdk or higher apache tomcat as the servlet container the application packaged as war and dependency on spring webmvc or spring webflux spring framework and have already been released whitesource s research team is carefully observing developments and researching the case we will keep updating this page and our whitesource resources with updates publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring beans release step up your open source security game with whitesource
0
649,951
21,330,710,738
IssuesEvent
2022-04-18 07:58:16
ZapSquared/Quaver
https://api.github.com/repos/ZapSquared/Quaver
closed
Quaver is unable to find its en-US locale strings.
type:bug released on @next affects:locale affects:functionality priority:p0
**Describe the bug** What isn't working as intended, and what does it affect? Locale Affects functionality. **Affected versions** What versions are affected by this bug? (e.g. >=3.0.1, 2.5.1-2.6.3, >=1.2.0) 3.4.0-next.48 **Steps to reproduce** Steps to reproduce the behavior. (e.g. click on a button, enter a value, etc. and see error) **Expected behavior** What is expected to happen? Locate the right directory to find the en-US strings. **Actual behavior** What actually happens? Attach or add errors or screenshots here as well. Looks at a json file that doesn't exist.
1.0
Quaver is unable to find its en-US locale strings. - **Describe the bug** What isn't working as intended, and what does it affect? Locale Affects functionality. **Affected versions** What versions are affected by this bug? (e.g. >=3.0.1, 2.5.1-2.6.3, >=1.2.0) 3.4.0-next.48 **Steps to reproduce** Steps to reproduce the behavior. (e.g. click on a button, enter a value, etc. and see error) **Expected behavior** What is expected to happen? Locate the right directory to find the en-US strings. **Actual behavior** What actually happens? Attach or add errors or screenshots here as well. Looks at a json file that doesn't exist.
non_infrastructure
quaver is unable to find its en us locale strings describe the bug what isn t working as intended and what does it affect locale affects functionality affected versions what versions are affected by this bug e g next steps to reproduce steps to reproduce the behavior e g click on a button enter a value etc and see error expected behavior what is expected to happen locate the right directory to find the en us strings actual behavior what actually happens attach or add errors or screenshots here as well looks at a json file that doesn t exist
0
21,235
14,446,345,861
IssuesEvent
2020-12-08 01:02:50
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Show non-existing table in replacement
bug interface/infrastructure
If a report model exist in the replacement, but not in the paddock, the data store is not generated a table for it as expected. However, when clicking the `Data` panel under `Replacement`, the `Data` panel is showing the first table in the data store, e.g. the `HarvestReport` in the figure below. ![image](https://user-images.githubusercontent.com/1657381/100832602-efafd380-34b3-11eb-8d27-bed253e9a4d3.png)
1.0
Show non-existing table in replacement - If a report model exist in the replacement, but not in the paddock, the data store is not generated a table for it as expected. However, when clicking the `Data` panel under `Replacement`, the `Data` panel is showing the first table in the data store, e.g. the `HarvestReport` in the figure below. ![image](https://user-images.githubusercontent.com/1657381/100832602-efafd380-34b3-11eb-8d27-bed253e9a4d3.png)
infrastructure
show non existing table in replacement if a report model exist in the replacement but not in the paddock the data store is not generated a table for it as expected however when clicking the data panel under replacement the data panel is showing the first table in the data store e g the harvestreport in the figure below
1
429,756
12,427,010,683
IssuesEvent
2020-05-25 00:28:44
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
Cannot use deploymentParams to specify application name as documentation suggests
Component: embedded ERR: Assignee Priority: Minor Stale Type: Bug community embedded
The maven-embedded-glassfish-plugin documentation clearly prefers we use <deploymentParams> wherever possible. One cannot supply an application name in this manner, however: <deploymentParams> <param>--name=foo</param> </deploymentParams> This is because the Maven plugin property "name" has a default value. The result is that if you omit it, "myapp" is used instead, and (in the example above) both "foo" and "myapp" are passed as name values to the deployment. This is of course an error and you get a nasty message that says that repeated parameter values (in this case it's probably --name) are not allowed. The workaround is to simply use the <name> property and to not use a deployment param for this case. #### Affected Versions [3.1.1]
1.0
Cannot use deploymentParams to specify application name as documentation suggests - The maven-embedded-glassfish-plugin documentation clearly prefers we use <deploymentParams> wherever possible. One cannot supply an application name in this manner, however: <deploymentParams> <param>--name=foo</param> </deploymentParams> This is because the Maven plugin property "name" has a default value. The result is that if you omit it, "myapp" is used instead, and (in the example above) both "foo" and "myapp" are passed as name values to the deployment. This is of course an error and you get a nasty message that says that repeated parameter values (in this case it's probably --name) are not allowed. The workaround is to simply use the <name> property and to not use a deployment param for this case. #### Affected Versions [3.1.1]
non_infrastructure
cannot use deploymentparams to specify application name as documentation suggests the maven embedded glassfish plugin documentation clearly prefers we use wherever possible one cannot supply an application name in this manner however name foo this is because the maven plugin property name has a default value the result is that if you omit it myapp is used instead and in the example above both foo and myapp are passed as name values to the deployment this is of course an error and you get a nasty message that says that repeated parameter values in this case it s probably name are not allowed the workaround is to simply use the property and to not use a deployment param for this case affected versions
0
222,423
17,422,195,002
IssuesEvent
2021-08-04 03:45:39
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Add timestamp to flake rate charts for when the data was updated
area/testing kind/improvement priority/important-longterm
This is just a check to do with caching and stuff.
1.0
Add timestamp to flake rate charts for when the data was updated - This is just a check to do with caching and stuff.
non_infrastructure
add timestamp to flake rate charts for when the data was updated this is just a check to do with caching and stuff
0
47,563
19,675,064,729
IssuesEvent
2022-01-11 11:26:39
GovernIB/ripea
https://api.github.com/repos/GovernIB/ripea
closed
Problema amb respostes incorrectes des de PINBAL
Tipus:Error Prioritat:Alta Lloc:WebServices
Quan RIPEA realitzar una petició de document/justificante a PINBAL espera rebre un document PDF. Moltes vegades pot ocórrer un error (per exemple per dades incorrectes, falta de permisos,...). Faltaria capturar l'error i mostrar un missatge del problema més clar a l'usuari
1.0
Problema amb respostes incorrectes des de PINBAL - Quan RIPEA realitzar una petició de document/justificante a PINBAL espera rebre un document PDF. Moltes vegades pot ocórrer un error (per exemple per dades incorrectes, falta de permisos,...). Faltaria capturar l'error i mostrar un missatge del problema més clar a l'usuari
non_infrastructure
problema amb respostes incorrectes des de pinbal quan ripea realitzar una petició de document justificante a pinbal espera rebre un document pdf moltes vegades pot ocórrer un error per exemple per dades incorrectes falta de permisos faltaria capturar l error i mostrar un missatge del problema més clar a l usuari
0
17,592
6,478,372,706
IssuesEvent
2017-08-18 07:41:29
JabRef/jabref
https://api.github.com/repos/JabRef/jabref
opened
Self-made deb and rpm packages
build-system help-wanted linux
This tracks the state of jabref-issued `deb` and `rpm` packages. We are experimenting using https://github.com/nebula-plugins/gradle-ospackage-plugin on the branch https://github.com/JabRef/jabref/tree/deb-and-rpm. Current state: `jar` is packed into the `deb`. No startup scripts, no desktop integration. This issue becomes obsolete in case JabRef is fully integrated with all features in Debian - see https://github.com/koppor/jabref/issues/135.
1.0
Self-made deb and rpm packages - This tracks the state of jabref-issued `deb` and `rpm` packages. We are experimenting using https://github.com/nebula-plugins/gradle-ospackage-plugin on the branch https://github.com/JabRef/jabref/tree/deb-and-rpm. Current state: `jar` is packed into the `deb`. No startup scripts, no desktop integration. This issue becomes obsolete in case JabRef is fully integrated with all features in Debian - see https://github.com/koppor/jabref/issues/135.
non_infrastructure
self made deb and rpm packages this tracks the state of jabref issued deb and rpm packages we are experimenting using on the branch current state jar is packed into the deb no startup scripts no desktop integration this issue becomes obsolete in case jabref is fully integrated with all features in debian see
0
299,707
9,205,778,323
IssuesEvent
2019-03-08 11:37:30
qissue-bot/QGIS
https://api.github.com/repos/qissue-bot/QGIS
closed
Symbols area scale
Category: Symbology Component: Easy fix? Component: Pull Request or Patch supplied Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Feature request
--- Author Name: **pls -** (pls -) Original Redmine Issue: 1077, https://issues.qgis.org/issues/1077 Original Assignee: nobody - --- Actually, a size can be manually specified for each symbol; it could be helpful to fecth that size from a field of the vector layer's database instead. There could be also an option that allows to scale the symbol's size according to a scale factor; this could be employed to visualize the same symbol associated to cities with different sizes according to their population rank, for example (see attachment).
1.0
Symbols area scale - --- Author Name: **pls -** (pls -) Original Redmine Issue: 1077, https://issues.qgis.org/issues/1077 Original Assignee: nobody - --- Actually, a size can be manually specified for each symbol; it could be helpful to fecth that size from a field of the vector layer's database instead. There could be also an option that allows to scale the symbol's size according to a scale factor; this could be employed to visualize the same symbol associated to cities with different sizes according to their population rank, for example (see attachment).
non_infrastructure
symbols area scale author name pls pls original redmine issue original assignee nobody actually a size can be manually specified for each symbol it could be helpful to fecth that size from a field of the vector layer s database instead there could be also an option that allows to scale the symbol s size according to a scale factor this could be employed to visualize the same symbol associated to cities with different sizes according to their population rank for example see attachment
0
3,434
4,311,445,544
IssuesEvent
2016-07-21 22:47:34
dart-lang/site-www
https://api.github.com/repos/dart-lang/site-www
opened
Automated release testing
Infrastructure MoveToSite-www P2
_From @srawlins on September 2, 2014 20:22_ We should employ some simple sanity tests for releases. Things like: 1. no broken links (including assets: images, js, css) (simple wget with grep script) 2. simple codelab verification with webdriver (webdriver; < 100 lines Ruby script + 1 line in Gemfile OR... some such Dart equivalent + a pubspec.yaml file) 3. maybe markdown formatting tests if jekyll or markdown breaks sometimes? i.e. backticks should be converted to `<code>` (simple curl script + grep, or a more robust Ruby or Dart script) _Copied from original issue: dart-lang/www.dartlang.org#992_
1.0
Automated release testing - _From @srawlins on September 2, 2014 20:22_ We should employ some simple sanity tests for releases. Things like: 1. no broken links (including assets: images, js, css) (simple wget with grep script) 2. simple codelab verification with webdriver (webdriver; < 100 lines Ruby script + 1 line in Gemfile OR... some such Dart equivalent + a pubspec.yaml file) 3. maybe markdown formatting tests if jekyll or markdown breaks sometimes? i.e. backticks should be converted to `<code>` (simple curl script + grep, or a more robust Ruby or Dart script) _Copied from original issue: dart-lang/www.dartlang.org#992_
infrastructure
automated release testing from srawlins on september we should employ some simple sanity tests for releases things like no broken links including assets images js css simple wget with grep script simple codelab verification with webdriver webdriver lines ruby script line in gemfile or some such dart equivalent a pubspec yaml file maybe markdown formatting tests if jekyll or markdown breaks sometimes i e backticks should be converted to simple curl script grep or a more robust ruby or dart script copied from original issue dart lang
1
11,982
9,541,990,311
IssuesEvent
2019-05-01 00:54:33
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
Enable using RabbitMQ in the Zulip backend unit tests
area: testing-infrastructure help wanted
As part of #732, I noticed that the largest concentration of code not tested via the Zulip backend test suite is the event system: - zerver/lib/queue.py - zerver/lib/event_queue.py - zerver/lib/handlers.py - zerver/worker/queue_processors.py A bit part of the reason for this is that we aren't using RabbitMQ at all int the backend test suite. One the big barriers to solving this problem is that we need a way to separate RabbitMQ manual testing traffic from RabbitMQ backend testing traffic (otherwise you'd get all sorts of weird interaction issues). Probably this can be done by e.g. adding "test:" to the names of all the queues when running as part of the backend test suite.
1.0
Enable using RabbitMQ in the Zulip backend unit tests - As part of #732, I noticed that the largest concentration of code not tested via the Zulip backend test suite is the event system: - zerver/lib/queue.py - zerver/lib/event_queue.py - zerver/lib/handlers.py - zerver/worker/queue_processors.py A bit part of the reason for this is that we aren't using RabbitMQ at all int the backend test suite. One the big barriers to solving this problem is that we need a way to separate RabbitMQ manual testing traffic from RabbitMQ backend testing traffic (otherwise you'd get all sorts of weird interaction issues). Probably this can be done by e.g. adding "test:" to the names of all the queues when running as part of the backend test suite.
infrastructure
enable using rabbitmq in the zulip backend unit tests as part of i noticed that the largest concentration of code not tested via the zulip backend test suite is the event system zerver lib queue py zerver lib event queue py zerver lib handlers py zerver worker queue processors py a bit part of the reason for this is that we aren t using rabbitmq at all int the backend test suite one the big barriers to solving this problem is that we need a way to separate rabbitmq manual testing traffic from rabbitmq backend testing traffic otherwise you d get all sorts of weird interaction issues probably this can be done by e g adding test to the names of all the queues when running as part of the backend test suite
1
625,343
19,726,034,803
IssuesEvent
2022-01-13 20:03:44
MatthewSpofford/Multiscale-Statistical-Analysis
https://api.github.com/repos/MatthewSpofford/Multiscale-Statistical-Analysis
closed
In ui_overhaul, creating a new area/complexity graph after clearing it throws an error
bug priority low ui in waiting
Throws error stating C/C++ object has already been destroyed from MatPlotLib. This exists due to #7 being worked on.
1.0
In ui_overhaul, creating a new area/complexity graph after clearing it throws an error - Throws error stating C/C++ object has already been destroyed from MatPlotLib. This exists due to #7 being worked on.
non_infrastructure
in ui overhaul creating a new area complexity graph after clearing it throws an error throws error stating c c object has already been destroyed from matplotlib this exists due to being worked on
0
32,450
26,707,575,043
IssuesEvent
2023-01-27 19:42:18
DickinsonCollege/FarmData2
https://api.github.com/repos/DickinsonCollege/FarmData2
opened
Port conflicts with Development Enviornment
enhancement infrastructure
Consider changing the ports on which each of the FarmData2 containers is exposed. If developers have other applications using a web browser (default port 80) or mysql (default port 3306) or phpmyadmin or express API then there are conflicts when `fd2-up.bash` is run. Of course those services can be stopped on the host machine and everything will be fine. But it might be nice if we could find ports that are less likely to conflict.
1.0
Port conflicts with Development Enviornment - Consider changing the ports on which each of the FarmData2 containers is exposed. If developers have other applications using a web browser (default port 80) or mysql (default port 3306) or phpmyadmin or express API then there are conflicts when `fd2-up.bash` is run. Of course those services can be stopped on the host machine and everything will be fine. But it might be nice if we could find ports that are less likely to conflict.
infrastructure
port conflicts with development enviornment consider changing the ports on which each of the containers is exposed if developers have other applications using a web browser default port or mysql default port or phpmyadmin or express api then there are conflicts when up bash is run of course those services can be stopped on the host machine and everything will be fine but it might be nice if we could find ports that are less likely to conflict
1
21,976
30,468,563,138
IssuesEvent
2023-07-17 12:11:48
q191201771/lal
https://api.github.com/repos/q191201771/lal
closed
多路同时推、拉流,报Broken pipe
#Question *In process
用如下命令推流: ffmpeg -f libk_video -wh 1920x1080 -sensor 3 -i video="test " -f alsa -ac 2 -ar 32000 -i hw:0 -idr_freq 25 -vcodec libk_h264 -acodec aac -f rtsp rtsp://10.20.1.55:5544/xxxx/com28 ffmpeg version 4.4 Copyright (c) 2000-2021 the FFmpeg developers built with gcc 7.3.0 (2019-11-20_nds64le-linux-glibc-v5d-6c120106e03) configuration: --cross-prefix=riscv64-linux- --enable-cross-compile --target-os=linux --cc=riscv64-linux-gcc --arch=riscv64 --extra-ldflags=-L./ --extra-ldflags=-ldl --extra-ldflags='-Wl,-rpath .' --enable-static --enable-libk_video --enable-libk_h264 --enable-libk_jpeg --enable-alsa --disable-autodetect --disable-ffplay --disable-ffprobe --disable-doc --enable-audio3a --enable-indev=v4l2 libavutil 56. 70.100 / 56. 70.100 libavcodec 58.134.100 / 58.134.100 libavformat 58. 76.100 / 58. 76.100 libavdevice 58. 13.100 / 58. 13.100 libavfilter 7.110.100 / 7.110.100 libswscale 5. 9.100 / 5. 9.100 libswresample 3. 9.100 / 3. 9.100 0x970E00f4: from 0x00550000 to 0x00110000 0x970E00f8: from 0x00000000 to 0x00770000 0x970E00fc: from 0x0fffff00 to 0x0fffff00 0x99900290: from 0x00000133 to 0x00000110 0x9990028c: from 0x00000001 to 0x00000000 0x9990038c: from 0x00000003 to 0x00000000 0x99900388: from 0x80000501 to 0x80000707 0x98000504: from 0x0001ffff to 0x00010303 k_video_read_header>w 1920, h 1080, stride = 1920 alloc_memory>phy_addr 0x1aebd000, size 68431872 k_video_read_header>isp_buf_paddr 0x1aebd000, isp_buf_vaddr 0x565000, isp_buf_size 68428800 isp_video ds0 block alloc:0x1aabc000,size:4194304,align 4096 ds0_out_addr =0x1aabc000 twod block alloc:0x182bb000,size:41943040,align 4096 isp_info.ds1_addr is 1aebd000 isp_info.ds1_stride is 780 video_set_mipicsi start! set_vi_params set_isp_params run_video!! Open struct isp_device addr = 0x25c0b0,vi = 0x25c150,mmio_base = 0x92620700 [ 18.682787] mipi_coner_init done, pvt code 0x1ffff [ 19.126175] lcd init ok lcd is 1 ---------------------------------!!! [ 19.133179] mipi_dsi_init done [ 19.136259] mipi_rx_dphy_init done! Call cmd ISP_CMD_MIPI_DSI_INIT s[ 19.139902] isp_act_sensor_rst success uccss system is ready to Release rst_n system rst_n has all release system rst_n has all release i2c_num = 0 imx219_i2c_init config done Isp_f2k_Init start Isp_f2k_Init end video_in_Init start vi_wrap_config start struct isp_device = 0x25c0b0,vi= 0x25c150 vi_wrap_rst struct isp_device 0x25c0b0,VI_WRAP_SWRST_CTL = 0x34f vi_wrap_config end video_in_Init end Imx219_1080p30_init reg is 100 val is 1 reg is 30eb val is 0 reg is 30eb val is 0 reg is 300a val is ff reg is 300b val is ff reg is 30eb val is 0 reg is 30eb val is 0 reg is 114 val is 1 reg is 128 val is 0 reg is 12a val is 18 reg is 12b val is 0 reg is 160 val is 4 reg is 161 val is 8e reg is 162 val is d reg is 163 val is 94 reg is 164 val is 2 reg is 165 val is a8 reg is 166 val is a reg is 167 val is 27 reg is 168 val is 2 reg is 169 val is b4 reg is 16a val is 6 reg is 16b val is eb reg is 16c val is 7 reg is 16d val is 80 reg is 16e val is 4 reg is 16f val is 38 reg is 170 val is 1 reg is 171 val is 1 reg is 174 val is 0 reg is 175 val is 0 reg is 301 val is 5 reg is 303 val is 1 reg is 304 val is 3 reg is 305 val is 3 reg is 306 val is 0 reg is 307 val is 26 reg is 30b val is 1 reg is 30c val is 0 reg is 30d val is 30 reg is 624 val is 7 reg is 625 val is 80 reg is 626 val is 4 reg is 627 val is 38 reg is 455e val is 0 reg is 471e val is 0 reg is 4767 val is 0 reg is 4750 val is 0 reg is 4540 val is 0 reg is 47b4 val is 0 reg is 4713 val is 0 reg is 478b val is 0 reg is 478f val is 0 reg is 4793 val is 0 reg is 4797 val is 0 reg is 479b val is 0 reg is 157 val is 40 reg is 158 val is 1 reg is 159 val is 0 reg is 15a val is 3 reg is 15b val is e8 reg is 100 val is 1 isp_f2k_core_table_init start! Isp2K RGB Gamma TABLE config done! Isp2K YUV Gamma TABLE config done! VO VCoef Config done![ 23.393613] plat->ds1_addr is1aebd000 VO HCoef Config done! VO GA[ 23.397728] plat->ds1_buf_cut is 14 MMA Coef Config done! Call cmd [ 23.404124] ISP_CMD_SET_DS1_SIZE fram_uv_addr is 1fa400 ISP_CMD_DS1_ADDR succss Call cmd ISP_CMD_DS1_BUFF_COUNT succss Call cmd ISP_CMD_SET_DS1_SIZE succss k_video_read_header>fd_isp: 0x7 isp_ouput Input #0, libk_video, from 'video=test': Duration: N/A, start: 0.033333, bitrate: 746496 kb/s Stream #0:0: Video: rawvideo (NV12 / 0x3231564E), nv12, 1920x1080, 746496 kb/s, 30 tbr, 30 tbn, 30 tbc Guessed Channel Layout for Input Stream #1.0 : stereo Input #1, alsa, from 'hw:0': Duration: N/A, start: 1650597527.574041, bitrate: 1024 kb/s Stream #1:0: Audio: pcm_s16le, 32000 Hz, stereo, s16, 1024 kb/s Stream mapping: Stream #0:0 -> #0:0 (rawvideo (native) -> h264 (libk_h264)) Stream #1:0 -> #0:1 (pcm_s16le (native) -> aac (native)) Press [q] to stop, [?] for help alloc_memory>phy_addr 0x17ccb000, size 6221824 k_h264_encode_init>yuv_vAddr 0x2007338000, yuv_phyAddr 0x17ccb000, yuv_size 6220800 Encoder Settings: width : 1920 height : 1080 level : 42 profile : 2 FreqIDR : 25 gopLen : 25 FrameRate : 30 rcMode : 1 SliceQP : 25 bitrate : 4000000 maxbitrate : 4000000 AL_ShareMemAlloc_Create>fd_ddr 0xe, fd_share_memory 0xd ---- FPGA board is ready ---- Board UID : 30AB6E51 Board HW ID : 620000E0 Board rev. : DC4054E7 Board date : 20191115 ----------------------------- Create_OutBuffers>count 4, size 3172352 VideoEncoder_Create>ok, hEnc 0x19c6eb0 pic: format 23, linesize 1920, 1920, 0, pts 0 pic data 0x1d255800, 0x1d44fc00, (nil), (nil), (nil), (nil), (nil), (nil) [alsa @ 0x19aa580] Thread message queue blocking; consider raising the thread_queue_size option (current value: 8) Output #0, rtsp, to 'rtsp://10.20.1.55:5544/xxxx/com28': Metadata: encoder : Lavf58.76.100 Stream #0:0: Video: h264, nv12(progressive), 1920x1080, q=2-31, 30 fps, 90k tbn Metadata: encoder : Lavc58.134.100 libk_h264 Stream #0:1: Audio: aac (LC), 32000 Hz, stereo, fltp, 128 kb/s Metadata: encoder : Lavc58.134.100 aac [libk_video @ 0x19a63d0] Thread message queue blocking; consider raising the thread_queue_size option (current value: 8) av_interleaved_write_frame(): Broken pipe00:00:15.60 bitrate=N/A speed=0.995x Receive NULL pic Last message repeated 1 times Error writing trailer of rtsp://10.20.1.55:5544/xxxx/com28: Broken pipe frame= 482 fps= 30 q=-0.0 Lsize=N/A time=00:00:16.09 bitrate=N/A speed=0.994x video:7870kB audio:253kB subtitle:0kB other streams:0kB global headers:0kB muxing overhead: unknown k_h264_encode_close> VideoEncoder_Destroy>ch 0 ok [aac @ 0x19c1ec0] Qavg: 161.536 QoS restore k_video_read_close> Conversion failed! 然后在vlc里拉流,单独推、拉,还是比较稳定的,如果多路,比如同时推、拉3 路,过一会就会报上面的 Broken pipe 部署是直接用docker启动的。 是不是我哪里做错了?
1.0
多路同时推、拉流,报Broken pipe - 用如下命令推流: ffmpeg -f libk_video -wh 1920x1080 -sensor 3 -i video="test " -f alsa -ac 2 -ar 32000 -i hw:0 -idr_freq 25 -vcodec libk_h264 -acodec aac -f rtsp rtsp://10.20.1.55:5544/xxxx/com28 ffmpeg version 4.4 Copyright (c) 2000-2021 the FFmpeg developers built with gcc 7.3.0 (2019-11-20_nds64le-linux-glibc-v5d-6c120106e03) configuration: --cross-prefix=riscv64-linux- --enable-cross-compile --target-os=linux --cc=riscv64-linux-gcc --arch=riscv64 --extra-ldflags=-L./ --extra-ldflags=-ldl --extra-ldflags='-Wl,-rpath .' --enable-static --enable-libk_video --enable-libk_h264 --enable-libk_jpeg --enable-alsa --disable-autodetect --disable-ffplay --disable-ffprobe --disable-doc --enable-audio3a --enable-indev=v4l2 libavutil 56. 70.100 / 56. 70.100 libavcodec 58.134.100 / 58.134.100 libavformat 58. 76.100 / 58. 76.100 libavdevice 58. 13.100 / 58. 13.100 libavfilter 7.110.100 / 7.110.100 libswscale 5. 9.100 / 5. 9.100 libswresample 3. 9.100 / 3. 9.100 0x970E00f4: from 0x00550000 to 0x00110000 0x970E00f8: from 0x00000000 to 0x00770000 0x970E00fc: from 0x0fffff00 to 0x0fffff00 0x99900290: from 0x00000133 to 0x00000110 0x9990028c: from 0x00000001 to 0x00000000 0x9990038c: from 0x00000003 to 0x00000000 0x99900388: from 0x80000501 to 0x80000707 0x98000504: from 0x0001ffff to 0x00010303 k_video_read_header>w 1920, h 1080, stride = 1920 alloc_memory>phy_addr 0x1aebd000, size 68431872 k_video_read_header>isp_buf_paddr 0x1aebd000, isp_buf_vaddr 0x565000, isp_buf_size 68428800 isp_video ds0 block alloc:0x1aabc000,size:4194304,align 4096 ds0_out_addr =0x1aabc000 twod block alloc:0x182bb000,size:41943040,align 4096 isp_info.ds1_addr is 1aebd000 isp_info.ds1_stride is 780 video_set_mipicsi start! set_vi_params set_isp_params run_video!! Open struct isp_device addr = 0x25c0b0,vi = 0x25c150,mmio_base = 0x92620700 [ 18.682787] mipi_coner_init done, pvt code 0x1ffff [ 19.126175] lcd init ok lcd is 1 ---------------------------------!!! [ 19.133179] mipi_dsi_init done [ 19.136259] mipi_rx_dphy_init done! Call cmd ISP_CMD_MIPI_DSI_INIT s[ 19.139902] isp_act_sensor_rst success uccss system is ready to Release rst_n system rst_n has all release system rst_n has all release i2c_num = 0 imx219_i2c_init config done Isp_f2k_Init start Isp_f2k_Init end video_in_Init start vi_wrap_config start struct isp_device = 0x25c0b0,vi= 0x25c150 vi_wrap_rst struct isp_device 0x25c0b0,VI_WRAP_SWRST_CTL = 0x34f vi_wrap_config end video_in_Init end Imx219_1080p30_init reg is 100 val is 1 reg is 30eb val is 0 reg is 30eb val is 0 reg is 300a val is ff reg is 300b val is ff reg is 30eb val is 0 reg is 30eb val is 0 reg is 114 val is 1 reg is 128 val is 0 reg is 12a val is 18 reg is 12b val is 0 reg is 160 val is 4 reg is 161 val is 8e reg is 162 val is d reg is 163 val is 94 reg is 164 val is 2 reg is 165 val is a8 reg is 166 val is a reg is 167 val is 27 reg is 168 val is 2 reg is 169 val is b4 reg is 16a val is 6 reg is 16b val is eb reg is 16c val is 7 reg is 16d val is 80 reg is 16e val is 4 reg is 16f val is 38 reg is 170 val is 1 reg is 171 val is 1 reg is 174 val is 0 reg is 175 val is 0 reg is 301 val is 5 reg is 303 val is 1 reg is 304 val is 3 reg is 305 val is 3 reg is 306 val is 0 reg is 307 val is 26 reg is 30b val is 1 reg is 30c val is 0 reg is 30d val is 30 reg is 624 val is 7 reg is 625 val is 80 reg is 626 val is 4 reg is 627 val is 38 reg is 455e val is 0 reg is 471e val is 0 reg is 4767 val is 0 reg is 4750 val is 0 reg is 4540 val is 0 reg is 47b4 val is 0 reg is 4713 val is 0 reg is 478b val is 0 reg is 478f val is 0 reg is 4793 val is 0 reg is 4797 val is 0 reg is 479b val is 0 reg is 157 val is 40 reg is 158 val is 1 reg is 159 val is 0 reg is 15a val is 3 reg is 15b val is e8 reg is 100 val is 1 isp_f2k_core_table_init start! Isp2K RGB Gamma TABLE config done! Isp2K YUV Gamma TABLE config done! VO VCoef Config done![ 23.393613] plat->ds1_addr is1aebd000 VO HCoef Config done! VO GA[ 23.397728] plat->ds1_buf_cut is 14 MMA Coef Config done! Call cmd [ 23.404124] ISP_CMD_SET_DS1_SIZE fram_uv_addr is 1fa400 ISP_CMD_DS1_ADDR succss Call cmd ISP_CMD_DS1_BUFF_COUNT succss Call cmd ISP_CMD_SET_DS1_SIZE succss k_video_read_header>fd_isp: 0x7 isp_ouput Input #0, libk_video, from 'video=test': Duration: N/A, start: 0.033333, bitrate: 746496 kb/s Stream #0:0: Video: rawvideo (NV12 / 0x3231564E), nv12, 1920x1080, 746496 kb/s, 30 tbr, 30 tbn, 30 tbc Guessed Channel Layout for Input Stream #1.0 : stereo Input #1, alsa, from 'hw:0': Duration: N/A, start: 1650597527.574041, bitrate: 1024 kb/s Stream #1:0: Audio: pcm_s16le, 32000 Hz, stereo, s16, 1024 kb/s Stream mapping: Stream #0:0 -> #0:0 (rawvideo (native) -> h264 (libk_h264)) Stream #1:0 -> #0:1 (pcm_s16le (native) -> aac (native)) Press [q] to stop, [?] for help alloc_memory>phy_addr 0x17ccb000, size 6221824 k_h264_encode_init>yuv_vAddr 0x2007338000, yuv_phyAddr 0x17ccb000, yuv_size 6220800 Encoder Settings: width : 1920 height : 1080 level : 42 profile : 2 FreqIDR : 25 gopLen : 25 FrameRate : 30 rcMode : 1 SliceQP : 25 bitrate : 4000000 maxbitrate : 4000000 AL_ShareMemAlloc_Create>fd_ddr 0xe, fd_share_memory 0xd ---- FPGA board is ready ---- Board UID : 30AB6E51 Board HW ID : 620000E0 Board rev. : DC4054E7 Board date : 20191115 ----------------------------- Create_OutBuffers>count 4, size 3172352 VideoEncoder_Create>ok, hEnc 0x19c6eb0 pic: format 23, linesize 1920, 1920, 0, pts 0 pic data 0x1d255800, 0x1d44fc00, (nil), (nil), (nil), (nil), (nil), (nil) [alsa @ 0x19aa580] Thread message queue blocking; consider raising the thread_queue_size option (current value: 8) Output #0, rtsp, to 'rtsp://10.20.1.55:5544/xxxx/com28': Metadata: encoder : Lavf58.76.100 Stream #0:0: Video: h264, nv12(progressive), 1920x1080, q=2-31, 30 fps, 90k tbn Metadata: encoder : Lavc58.134.100 libk_h264 Stream #0:1: Audio: aac (LC), 32000 Hz, stereo, fltp, 128 kb/s Metadata: encoder : Lavc58.134.100 aac [libk_video @ 0x19a63d0] Thread message queue blocking; consider raising the thread_queue_size option (current value: 8) av_interleaved_write_frame(): Broken pipe00:00:15.60 bitrate=N/A speed=0.995x Receive NULL pic Last message repeated 1 times Error writing trailer of rtsp://10.20.1.55:5544/xxxx/com28: Broken pipe frame= 482 fps= 30 q=-0.0 Lsize=N/A time=00:00:16.09 bitrate=N/A speed=0.994x video:7870kB audio:253kB subtitle:0kB other streams:0kB global headers:0kB muxing overhead: unknown k_h264_encode_close> VideoEncoder_Destroy>ch 0 ok [aac @ 0x19c1ec0] Qavg: 161.536 QoS restore k_video_read_close> Conversion failed! 然后在vlc里拉流,单独推、拉,还是比较稳定的,如果多路,比如同时推、拉3 路,过一会就会报上面的 Broken pipe 部署是直接用docker启动的。 是不是我哪里做错了?
non_infrastructure
多路同时推、拉流,报broken pipe 用如下命令推流: ffmpeg f libk video wh sensor i video test f alsa ac ar i hw idr freq vcodec libk acodec aac f rtsp rtsp xxxx ffmpeg version copyright c the ffmpeg developers built with gcc linux glibc configuration cross prefix linux enable cross compile target os linux cc linux gcc arch extra ldflags l extra ldflags ldl extra ldflags wl rpath enable static enable libk video enable libk enable libk jpeg enable alsa disable autodetect disable ffplay disable ffprobe disable doc enable enable indev libavutil libavcodec libavformat libavdevice libavfilter libswscale libswresample from to from to from to from to from to from to from to from to k video read header w h stride alloc memory phy addr size k video read header isp buf paddr isp buf vaddr isp buf size isp video block alloc size align out addr twod block alloc size align isp info addr is isp info stride is video set mipicsi start set vi params set isp params run video open struct isp device addr vi mmio base mipi coner init done pvt code lcd init ok lcd is mipi dsi init done mipi rx dphy init done call cmd isp cmd mipi dsi init s isp act sensor rst success uccss system is ready to release rst n system rst n has all release system rst n has all release num init config done isp init start isp init end video in init start vi wrap config start struct isp device vi vi wrap rst struct isp device vi wrap swrst ctl vi wrap config end video in init end init reg is val is reg is val is reg is val is reg is val is ff reg is val is ff reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is d reg is val is reg is val is reg is val is reg is val is a reg is val is reg is val is reg is val is reg is val is reg is val is eb reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is reg is val is isp core table init start! rgb gamma table config done yuv gamma table config done vo vcoef config done plat addr vo hcoef config done vo ga plat buf cut is mma coef config done call cmd isp cmd set size fram uv addr is isp cmd addr succss call cmd isp cmd buff count succss call cmd isp cmd set size succss k video read header fd isp isp ouput input libk video from video test duration n a start bitrate kb s stream video rawvideo kb s tbr tbn tbc guessed channel layout for input stream stereo input alsa from hw duration n a start bitrate kb s stream audio pcm hz stereo kb s stream mapping stream rawvideo native libk stream pcm native aac native press to stop for help alloc memory phy addr size k encode init yuv vaddr yuv phyaddr yuv size encoder settings width height level profile freqidr goplen framerate rcmode sliceqp bitrate maxbitrate al sharememalloc create fd ddr fd share memory fpga board is ready board uid board hw id board rev board date create outbuffers count size videoencoder create ok henc pic format linesize pts pic data nil nil nil nil nil nil thread message queue blocking consider raising the thread queue size option current value output rtsp to rtsp xxxx metadata encoder stream video progressive q fps tbn metadata encoder libk stream audio aac lc hz stereo fltp kb s metadata encoder aac thread message queue blocking consider raising the thread queue size option current value av interleaved write frame broken bitrate n a speed receive null pic last message repeated times error writing trailer of rtsp xxxx broken pipe frame fps q lsize n a time bitrate n a speed video audio subtitle other streams global headers muxing overhead unknown k encode close videoencoder destroy ch ok qavg qos restore k video read close conversion failed 然后在vlc里拉流,单独推、拉,还是比较稳定的,如果多路,比如同时推、 路,过一会就会报上面的 broken pipe 部署是直接用docker启动的。 是不是我哪里做错了?
0
347,144
24,887,601,153
IssuesEvent
2022-10-28 09:07:22
Rachel-AG/ped
https://api.github.com/repos/Rachel-AG/ped
opened
`inspect` in the UG displays the wrong screenshot
severity.VeryLow type.DocumentationBug
`inspect` in the UG displays the help page instead of the result of `inspect`. ![image.png](https://raw.githubusercontent.com/Rachel-AG/ped/main/files/203cc0ff-765f-4e91-a1a1-5d5b65dd24b3.png) <!--session: 1666944915665-071b1e36-98db-452b-9ede-ebf3bf2576bb--> <!--Version: Web v3.4.4-->
1.0
`inspect` in the UG displays the wrong screenshot - `inspect` in the UG displays the help page instead of the result of `inspect`. ![image.png](https://raw.githubusercontent.com/Rachel-AG/ped/main/files/203cc0ff-765f-4e91-a1a1-5d5b65dd24b3.png) <!--session: 1666944915665-071b1e36-98db-452b-9ede-ebf3bf2576bb--> <!--Version: Web v3.4.4-->
non_infrastructure
inspect in the ug displays the wrong screenshot inspect in the ug displays the help page instead of the result of inspect
0
799
2,915,271,968
IssuesEvent
2015-06-23 11:29:01
hackndev/zinc
https://api.github.com/repos/hackndev/zinc
closed
Coverage makes little sense
infrastructure ready for review
[Coverage data](http://ci.zinc.rs/job/zinc/23/PLATFORM=native/cobertura/) has little insight on files that weren't used in testing, giving out silly result of 93% covered when it's closed to 7% covered. One way to fix it is to expand cobertura.xml with all the missing files from relevant subdir. Those should not include lines with comments or empty lines; actual coverage is even more small, but that should give a reasonable approximation.
1.0
Coverage makes little sense - [Coverage data](http://ci.zinc.rs/job/zinc/23/PLATFORM=native/cobertura/) has little insight on files that weren't used in testing, giving out silly result of 93% covered when it's closed to 7% covered. One way to fix it is to expand cobertura.xml with all the missing files from relevant subdir. Those should not include lines with comments or empty lines; actual coverage is even more small, but that should give a reasonable approximation.
infrastructure
coverage makes little sense has little insight on files that weren t used in testing giving out silly result of covered when it s closed to covered one way to fix it is to expand cobertura xml with all the missing files from relevant subdir those should not include lines with comments or empty lines actual coverage is even more small but that should give a reasonable approximation
1
157,331
19,957,099,694
IssuesEvent
2022-01-28 01:24:26
panasalap/linux-4.1.15
https://api.github.com/repos/panasalap/linux-4.1.15
opened
CVE-2016-7915 (Medium) detected in linux-yoctov3.18.26
security vulnerability
## CVE-2016-7915 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov3.18.26</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/hid-core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/hid-core.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The hid_input_field function in drivers/hid/hid-core.c in the Linux kernel before 4.6 allows physically proximate attackers to obtain sensitive information from kernel memory or cause a denial of service (out-of-bounds read) by connecting a device, as demonstrated by a Logitech DJ receiver. <p>Publish Date: 2016-11-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-7915>CVE-2016-7915</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-7915">https://nvd.nist.gov/vuln/detail/CVE-2016-7915</a></p> <p>Release Date: 2016-11-16</p> <p>Fix Resolution: 4.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2016-7915 (Medium) detected in linux-yoctov3.18.26 - ## CVE-2016-7915 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov3.18.26</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/hid-core.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/hid/hid-core.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The hid_input_field function in drivers/hid/hid-core.c in the Linux kernel before 4.6 allows physically proximate attackers to obtain sensitive information from kernel memory or cause a denial of service (out-of-bounds read) by connecting a device, as demonstrated by a Logitech DJ receiver. <p>Publish Date: 2016-11-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-7915>CVE-2016-7915</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-7915">https://nvd.nist.gov/vuln/detail/CVE-2016-7915</a></p> <p>Release Date: 2016-11-16</p> <p>Fix Resolution: 4.6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve medium detected in linux cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in base branch master vulnerable source files drivers hid hid core c drivers hid hid core c vulnerability details the hid input field function in drivers hid hid core c in the linux kernel before allows physically proximate attackers to obtain sensitive information from kernel memory or cause a denial of service out of bounds read by connecting a device as demonstrated by a logitech dj receiver publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
39,183
6,726,669,580
IssuesEvent
2017-10-17 10:44:01
pact-foundation/pact-js
https://api.github.com/repos/pact-foundation/pact-js
opened
Documentation: "Using Pact with Docker"
Documentation good-first-issue help wanted
Given the prevalence of Docker, in particular dev/test, we should have a first class example for people to use. TODO: - [ ] Use of Docker compose to demonstrate both consumer tests and provider verification - [ ] How to separate the test container from the mock server container (could be running `pact-node`, [`pact-mock-service-docker`](https://github.com/pact-foundation/pact-mock-service-docker) or another variant) - [ ] E2E example integrated into CI See https://github.com/pact-foundation/pact-js/pull/115 for inspiration to create this article.
1.0
Documentation: "Using Pact with Docker" - Given the prevalence of Docker, in particular dev/test, we should have a first class example for people to use. TODO: - [ ] Use of Docker compose to demonstrate both consumer tests and provider verification - [ ] How to separate the test container from the mock server container (could be running `pact-node`, [`pact-mock-service-docker`](https://github.com/pact-foundation/pact-mock-service-docker) or another variant) - [ ] E2E example integrated into CI See https://github.com/pact-foundation/pact-js/pull/115 for inspiration to create this article.
non_infrastructure
documentation using pact with docker given the prevalence of docker in particular dev test we should have a first class example for people to use todo use of docker compose to demonstrate both consumer tests and provider verification how to separate the test container from the mock server container could be running pact node or another variant example integrated into ci see for inspiration to create this article
0
22,194
15,035,999,986
IssuesEvent
2021-02-02 14:46:40
SzFMV2021-Tavasz/handout
https://api.github.com/repos/SzFMV2021-Tavasz/handout
opened
Set up infrastructure for 2021 spring
effort: moderate priority: normal status: in progress type: infrastructure
- [ ] Create new GitHub Organization - [ ] Create Teams - [ ] Everyone - [ ] Company-A - Team-A1 - Team-A2 - Team-A3 - Team-A4 - [ ] Company-B - Team-B1 - Team-B2 - Team-B3 - Team-B4 - [ ] Instructors - [ ] Transfer handout repo from the previous organization - [ ] Add labels as repository defaults in organization - effort: high - effort: low - effort: moderate - priority: critical - priority: high - priority: low - priority: normal - status: completed - status: duplicate - status: help wanted - status: in progress - status: invalid - status: pending - status: review needed - status: stalled - status: wontfix - type: bug - type: design - type: documentation - type: enhancement - type: integration - type: question - type: user story - [ ] Create new repos for the project - AutomatedCar-A - AutomatedCar-B - [ ] Add repositories to Teams and set permissions - Match `AutomatedCar-A` to `Company-A` with `write` permissions - Match `AutomatedCar-B` to `Company-B` with `write` permissions - Set admin persmissions to team `Instructors` - [ ] Add milestones - Planning, Release Candidate and Final for every sprints - [ ] Add projects (kanban boards) from `Automated kanban with review` template - one project for every teams for both repositories - [ ] Add first user stories - [ ] Add link to team forums (discussion boards) and user stories in the project descriptions - [ ] Integrate CI - GitHub Actions - [ ] Integrate static code analyzer - CodeFactor.io - [ ] Integrate code coverage tracker - CodeCov.io - [ ] Enable stale bot to mark issues as staled if there is no activity for a while - [ ] Update badges in the README - [ ] Create a tag from the skeleton - `git tag -a skeleton -m 'Tag skeleton'` - and push it: `git push origin --tags` - [ ] Create team branches - `git checkout -b TeamA1` - `git push origin TeamA1` - [ ] Protect branches
1.0
Set up infrastructure for 2021 spring - - [ ] Create new GitHub Organization - [ ] Create Teams - [ ] Everyone - [ ] Company-A - Team-A1 - Team-A2 - Team-A3 - Team-A4 - [ ] Company-B - Team-B1 - Team-B2 - Team-B3 - Team-B4 - [ ] Instructors - [ ] Transfer handout repo from the previous organization - [ ] Add labels as repository defaults in organization - effort: high - effort: low - effort: moderate - priority: critical - priority: high - priority: low - priority: normal - status: completed - status: duplicate - status: help wanted - status: in progress - status: invalid - status: pending - status: review needed - status: stalled - status: wontfix - type: bug - type: design - type: documentation - type: enhancement - type: integration - type: question - type: user story - [ ] Create new repos for the project - AutomatedCar-A - AutomatedCar-B - [ ] Add repositories to Teams and set permissions - Match `AutomatedCar-A` to `Company-A` with `write` permissions - Match `AutomatedCar-B` to `Company-B` with `write` permissions - Set admin persmissions to team `Instructors` - [ ] Add milestones - Planning, Release Candidate and Final for every sprints - [ ] Add projects (kanban boards) from `Automated kanban with review` template - one project for every teams for both repositories - [ ] Add first user stories - [ ] Add link to team forums (discussion boards) and user stories in the project descriptions - [ ] Integrate CI - GitHub Actions - [ ] Integrate static code analyzer - CodeFactor.io - [ ] Integrate code coverage tracker - CodeCov.io - [ ] Enable stale bot to mark issues as staled if there is no activity for a while - [ ] Update badges in the README - [ ] Create a tag from the skeleton - `git tag -a skeleton -m 'Tag skeleton'` - and push it: `git push origin --tags` - [ ] Create team branches - `git checkout -b TeamA1` - `git push origin TeamA1` - [ ] Protect branches
infrastructure
set up infrastructure for spring create new github organization create teams everyone company a team team team team company b team team team team instructors transfer handout repo from the previous organization add labels as repository defaults in organization effort high effort low effort moderate priority critical priority high priority low priority normal status completed status duplicate status help wanted status in progress status invalid status pending status review needed status stalled status wontfix type bug type design type documentation type enhancement type integration type question type user story create new repos for the project automatedcar a automatedcar b add repositories to teams and set permissions match automatedcar a to company a with write permissions match automatedcar b to company b with write permissions set admin persmissions to team instructors add milestones planning release candidate and final for every sprints add projects kanban boards from automated kanban with review template one project for every teams for both repositories add first user stories add link to team forums discussion boards and user stories in the project descriptions integrate ci github actions integrate static code analyzer codefactor io integrate code coverage tracker codecov io enable stale bot to mark issues as staled if there is no activity for a while update badges in the readme create a tag from the skeleton git tag a skeleton m tag skeleton and push it git push origin tags create team branches git checkout b git push origin protect branches
1
27,741
22,289,368,257
IssuesEvent
2022-06-12 05:47:00
cthing/meazure
https://api.github.com/repos/cthing/meazure
closed
Create an XML writer to centralize handling of escaping and pretty printing
infrastructure
Currently, the position log and profile classes handle the details of writing well formed and pretty printed XML. This should be centralized to an XML writer class. This class will handle: * Escaping XML content as appropriate * Normalizing CR+LF to LF * Encoding the output in UTF-8 * Pretty printing the XML
1.0
Create an XML writer to centralize handling of escaping and pretty printing - Currently, the position log and profile classes handle the details of writing well formed and pretty printed XML. This should be centralized to an XML writer class. This class will handle: * Escaping XML content as appropriate * Normalizing CR+LF to LF * Encoding the output in UTF-8 * Pretty printing the XML
infrastructure
create an xml writer to centralize handling of escaping and pretty printing currently the position log and profile classes handle the details of writing well formed and pretty printed xml this should be centralized to an xml writer class this class will handle escaping xml content as appropriate normalizing cr lf to lf encoding the output in utf pretty printing the xml
1
14,547
10,927,875,132
IssuesEvent
2019-11-22 17:40:54
peterhp/ruch
https://api.github.com/repos/peterhp/ruch
closed
Create Ruch Project
infrastructure
Create a new Rust project for Ruch - [x] Create a Rust project - [x] Add basic build script - [x] Add demo code snippets with necessary unit tests
1.0
Create Ruch Project - Create a new Rust project for Ruch - [x] Create a Rust project - [x] Add basic build script - [x] Add demo code snippets with necessary unit tests
infrastructure
create ruch project create a new rust project for ruch create a rust project add basic build script add demo code snippets with necessary unit tests
1
4,706
5,237,718,344
IssuesEvent
2017-01-31 00:51:02
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Microsoft.VisualStudio.Shell.15.0 used in Roslyn build requires strong name skipping
Area-Infrastructure Bug
It's currently not signed properly.
1.0
Microsoft.VisualStudio.Shell.15.0 used in Roslyn build requires strong name skipping - It's currently not signed properly.
infrastructure
microsoft visualstudio shell used in roslyn build requires strong name skipping it s currently not signed properly
1
7,116
6,776,556,154
IssuesEvent
2017-10-27 18:19:59
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
opened
Move compiler test ref assemblies to one of the test resources assemblies
Area-Compilers Area-Infrastructure
Some background here: The compiler tests usually need to compile code against some set of assemblies. Most of those assemblies are stored as resources in a few "TestResources" projects. When the compiler tests need reference assemblies, they pull the resource out of the test resources assembly and use the byte stream to construct a reference assembly in-memory. There's an important exception to the aforementioned: the "netstandard" and similar references when compiling against `netcoreapp2.0`. In this case, we copy the ref assemblies out of a well-known NuGet path to a `ref` folder in the test output folder: https://github.com/dotnet/roslyn/blob/4b1c677e0161df22f12073a1493747532152b115/build/Targets/Settings.props#L39-L40. The reasoning here is that the surface area of `netstandard 2.0`, which is our desired target, was not finalized when we started moving tests to target netcoreapp2.0. Now that netcoreapp2.0 is finalized and shipped, we should take these ref assemblies and add them to one of our TestResources projects (probably ProprietaryTestResources in the roslyn-internal repo to prevent adding too many binary files to the Roslyn git repo) and change the reference code to use those resources like normal. One thing to watch out for when making this change: I believe the version of netstandard.dll we pulled did not include System.ValueTuple, while the newest one does. When upgrading, some tests may fail because they expect to reference their own System.ValueTuple. If these tests depend on constructing specific versions of System.ValueTuple we should probably change those tests to explicitly list their reference assemblies instead of referencing netstandard.dll.
1.0
Move compiler test ref assemblies to one of the test resources assemblies - Some background here: The compiler tests usually need to compile code against some set of assemblies. Most of those assemblies are stored as resources in a few "TestResources" projects. When the compiler tests need reference assemblies, they pull the resource out of the test resources assembly and use the byte stream to construct a reference assembly in-memory. There's an important exception to the aforementioned: the "netstandard" and similar references when compiling against `netcoreapp2.0`. In this case, we copy the ref assemblies out of a well-known NuGet path to a `ref` folder in the test output folder: https://github.com/dotnet/roslyn/blob/4b1c677e0161df22f12073a1493747532152b115/build/Targets/Settings.props#L39-L40. The reasoning here is that the surface area of `netstandard 2.0`, which is our desired target, was not finalized when we started moving tests to target netcoreapp2.0. Now that netcoreapp2.0 is finalized and shipped, we should take these ref assemblies and add them to one of our TestResources projects (probably ProprietaryTestResources in the roslyn-internal repo to prevent adding too many binary files to the Roslyn git repo) and change the reference code to use those resources like normal. One thing to watch out for when making this change: I believe the version of netstandard.dll we pulled did not include System.ValueTuple, while the newest one does. When upgrading, some tests may fail because they expect to reference their own System.ValueTuple. If these tests depend on constructing specific versions of System.ValueTuple we should probably change those tests to explicitly list their reference assemblies instead of referencing netstandard.dll.
infrastructure
move compiler test ref assemblies to one of the test resources assemblies some background here the compiler tests usually need to compile code against some set of assemblies most of those assemblies are stored as resources in a few testresources projects when the compiler tests need reference assemblies they pull the resource out of the test resources assembly and use the byte stream to construct a reference assembly in memory there s an important exception to the aforementioned the netstandard and similar references when compiling against in this case we copy the ref assemblies out of a well known nuget path to a ref folder in the test output folder the reasoning here is that the surface area of netstandard which is our desired target was not finalized when we started moving tests to target now that is finalized and shipped we should take these ref assemblies and add them to one of our testresources projects probably proprietarytestresources in the roslyn internal repo to prevent adding too many binary files to the roslyn git repo and change the reference code to use those resources like normal one thing to watch out for when making this change i believe the version of netstandard dll we pulled did not include system valuetuple while the newest one does when upgrading some tests may fail because they expect to reference their own system valuetuple if these tests depend on constructing specific versions of system valuetuple we should probably change those tests to explicitly list their reference assemblies instead of referencing netstandard dll
1
24,916
17,893,426,096
IssuesEvent
2021-09-08 04:34:38
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Resizing of property presenter widgets with window resize
bug interface/infrastructure
The resizing of property presenter entry boxes cannot keep up with a quick resize of the window under windows. Entry box right hand side resize works if you slowly shrink the window in windows with the mouse, but stops resizing as you start to change the window size quickly, or one rapid shrink.
1.0
Resizing of property presenter widgets with window resize - The resizing of property presenter entry boxes cannot keep up with a quick resize of the window under windows. Entry box right hand side resize works if you slowly shrink the window in windows with the mouse, but stops resizing as you start to change the window size quickly, or one rapid shrink.
infrastructure
resizing of property presenter widgets with window resize the resizing of property presenter entry boxes cannot keep up with a quick resize of the window under windows entry box right hand side resize works if you slowly shrink the window in windows with the mouse but stops resizing as you start to change the window size quickly or one rapid shrink
1
72,763
24,279,630,407
IssuesEvent
2022-09-28 16:15:36
mooltiverse/nyx
https://api.github.com/repos/mooltiverse/nyx
closed
Exception when enabling the resume flag
type::defect:bug type::task
Enabling `resume` throws an exception when unmarshalling the State file. At a first glance the problem is inside the `State.etBump(String bump)` method as it retrieves the configuration object reference from within the State (getConfiguration().getBump()) but object might still be null when unmarshalling. The exception looks like: ``` org.gradle.internal.event.ListenerNotificationException: Failed to notify build listener. at org.gradle.internal.event.AbstractBroadcastDispatch.dispatch(AbstractBroadcastDispatch.java:89) at org.gradle.internal.event.BroadcastDispatch$CompositeDispatch.dispatch(BroadcastDispatch.java:3[46](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:47)) at org.gradle.internal.event.BroadcastDispatch$CompositeDispatch.dispatch(BroadcastDispatch.java:249) at org.gradle.internal.event.ListenerBroadcast.dispatch(ListenerBroadcast.java:141) at org.gradle.internal.event.ListenerBroadcast.dispatch(ListenerBroadcast.java:37) at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94) at com.sun.proxy.$Proxy27.projectsLoaded(Unknown Source) at org.gradle.initialization.NotifyingBuildLoader$3.run(NotifyingBuildLoader.java:74) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:[47](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:48)) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.initialization.NotifyingBuildLoader.load(NotifyingBuildLoader.java:71) at org.gradle.configuration.BuildTreePreparingProjectsPreparer.prepareProjects(BuildTreePreparingProjectsPreparer.java:56) at org.gradle.configuration.BuildOperationFiringProjectsPreparer$ConfigureBuild.run(BuildOperationFiringProjectsPreparer.java:52) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.configuration.BuildOperationFiringProjectsPreparer.prepareProjects(BuildOperationFiringProjectsPreparer.java:40) at org.gradle.initialization.VintageBuildModelController.lambda$prepareProjects$3(VintageBuildModelController.java:89) at org.gradle.internal.model.StateTransitionController.lambda$doTransition$12(StateTransitionController.java:236) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:235) at org.gradle.internal.model.StateTransitionController.lambda$transitionIfNotPreviously$10(StateTransitionController.java:210) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:34) at org.gradle.internal.model.StateTransitionController.transitionIfNotPreviously(StateTransitionController.java:206) at org.gradle.initialization.VintageBuildModelController.prepareProjects(VintageBuildModelController.java:89) at org.gradle.initialization.VintageBuildModelController.prepareToScheduleTasks(VintageBuildModelController.java:71) at org.gradle.internal.build.DefaultBuildLifecycleController.lambda$prepareToScheduleTasks$2(DefaultBuildLifecycleController.java:127) at org.gradle.internal.model.StateTransitionController.lambda$doTransition$12(StateTransitionController.java:236) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:235) at org.gradle.internal.model.StateTransitionController.lambda$maybeTransition$8(StateTransitionController.java:187) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:34) at org.gradle.internal.model.StateTransitionController.maybeTransition(StateTransitionController.java:183) at org.gradle.internal.build.DefaultBuildLifecycleController.prepareToScheduleTasks(DefaultBuildLifecycleController.java:125) at org.gradle.internal.buildtree.DefaultBuildTreeWorkPreparer.scheduleRequestedTasks(DefaultBuildTreeWorkPreparer.java:33) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$doScheduleAndRunTasks$2(DefaultBuildTreeLifecycleController.java:89) at org.gradle.composite.internal.DefaultIncludedBuildTaskGraph.withNewWorkGraph(DefaultIncludedBuildTaskGraph.java:101) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.doScheduleAndRunTasks(DefaultBuildTreeLifecycleController.java:88) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$4(DefaultBuildTreeLifecycleController.java:106) at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44) at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249) at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110) at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64) at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88) at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31) at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75) at org.gradle.util.internal.Swapper.swap(Swapper.java:38) at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52) at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:[48](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:49)) Caused by: com.fasterxml.jackson.databind.JsonMappingException: N/A at [Source: (File); line: 3, column: 12] (through reference chain: com.mooltiverse.oss.nyx.state.State["bump"]) at com.fasterxml.jackson.databind.JsonMappingException.from(JsonMappingException.java:276) at com.fasterxml.jackson.databind.deser.SettableBeanProperty._throwAsIOE(SettableBeanProperty.java:623) at com.fasterxml.jackson.databind.deser.SettableBeanProperty._throwAsIOE(SettableBeanProperty.java:611) at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:143) at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:313) at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:176) at com.fasterxml.jackson.databind.deser.DefaultDeserializationContext.readRootValue(DefaultDeserializationContext.java:322) at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4675) at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3[49](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:50)5) at com.mooltiverse.oss.nyx.io.FileMapper.load(FileMapper.java:108) ... 190 more Caused by: java.lang.NullPointerException at com.mooltiverse.oss.nyx.state.State.setBump(State.java:2[52](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:53)) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:[62](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:63)) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:141) ... 1[96](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:97) more ```
1.0
Exception when enabling the resume flag - Enabling `resume` throws an exception when unmarshalling the State file. At a first glance the problem is inside the `State.etBump(String bump)` method as it retrieves the configuration object reference from within the State (getConfiguration().getBump()) but object might still be null when unmarshalling. The exception looks like: ``` org.gradle.internal.event.ListenerNotificationException: Failed to notify build listener. at org.gradle.internal.event.AbstractBroadcastDispatch.dispatch(AbstractBroadcastDispatch.java:89) at org.gradle.internal.event.BroadcastDispatch$CompositeDispatch.dispatch(BroadcastDispatch.java:3[46](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:47)) at org.gradle.internal.event.BroadcastDispatch$CompositeDispatch.dispatch(BroadcastDispatch.java:249) at org.gradle.internal.event.ListenerBroadcast.dispatch(ListenerBroadcast.java:141) at org.gradle.internal.event.ListenerBroadcast.dispatch(ListenerBroadcast.java:37) at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94) at com.sun.proxy.$Proxy27.projectsLoaded(Unknown Source) at org.gradle.initialization.NotifyingBuildLoader$3.run(NotifyingBuildLoader.java:74) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:[47](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:48)) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.initialization.NotifyingBuildLoader.load(NotifyingBuildLoader.java:71) at org.gradle.configuration.BuildTreePreparingProjectsPreparer.prepareProjects(BuildTreePreparingProjectsPreparer.java:56) at org.gradle.configuration.BuildOperationFiringProjectsPreparer$ConfigureBuild.run(BuildOperationFiringProjectsPreparer.java:52) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.configuration.BuildOperationFiringProjectsPreparer.prepareProjects(BuildOperationFiringProjectsPreparer.java:40) at org.gradle.initialization.VintageBuildModelController.lambda$prepareProjects$3(VintageBuildModelController.java:89) at org.gradle.internal.model.StateTransitionController.lambda$doTransition$12(StateTransitionController.java:236) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:235) at org.gradle.internal.model.StateTransitionController.lambda$transitionIfNotPreviously$10(StateTransitionController.java:210) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:34) at org.gradle.internal.model.StateTransitionController.transitionIfNotPreviously(StateTransitionController.java:206) at org.gradle.initialization.VintageBuildModelController.prepareProjects(VintageBuildModelController.java:89) at org.gradle.initialization.VintageBuildModelController.prepareToScheduleTasks(VintageBuildModelController.java:71) at org.gradle.internal.build.DefaultBuildLifecycleController.lambda$prepareToScheduleTasks$2(DefaultBuildLifecycleController.java:127) at org.gradle.internal.model.StateTransitionController.lambda$doTransition$12(StateTransitionController.java:236) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:235) at org.gradle.internal.model.StateTransitionController.lambda$maybeTransition$8(StateTransitionController.java:187) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:34) at org.gradle.internal.model.StateTransitionController.maybeTransition(StateTransitionController.java:183) at org.gradle.internal.build.DefaultBuildLifecycleController.prepareToScheduleTasks(DefaultBuildLifecycleController.java:125) at org.gradle.internal.buildtree.DefaultBuildTreeWorkPreparer.scheduleRequestedTasks(DefaultBuildTreeWorkPreparer.java:33) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$doScheduleAndRunTasks$2(DefaultBuildTreeLifecycleController.java:89) at org.gradle.composite.internal.DefaultIncludedBuildTaskGraph.withNewWorkGraph(DefaultIncludedBuildTaskGraph.java:101) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.doScheduleAndRunTasks(DefaultBuildTreeLifecycleController.java:88) at org.gradle.internal.buildtree.DefaultBuildTreeLifecycleController.lambda$runBuild$4(DefaultBuildTreeLifecycleController.java:106) at org.gradle.internal.model.StateTransitionController.lambda$transition$5(StateTransitionController.java:166) at org.gradle.internal.model.StateTransitionController.doTransition(StateTransitionController.java:247) at org.gradle.internal.model.StateTransitionController.lambda$transition$6(StateTransitionController.java:166) at org.gradle.internal.work.DefaultSynchronizer.withLock(DefaultSynchronizer.java:44) at org.gradle.internal.model.StateTransitionController.transition(StateTransitionController.java:166) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.lambda$execute$0(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.composite.internal.DefaultRootBuildState.run(DefaultRootBuildState.java:122) at org.gradle.launcher.exec.RootBuildLifecycleBuildActionExecutor.execute(RootBuildLifecycleBuildActionExecutor.java:40) at org.gradle.internal.buildtree.DefaultBuildTreeContext.execute(DefaultBuildTreeContext.java:40) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.lambda$execute$0(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.internal.buildtree.BuildTreeState.run(BuildTreeState.java:53) at org.gradle.launcher.exec.BuildTreeLifecycleBuildActionExecutor.execute(BuildTreeLifecycleBuildActionExecutor.java:65) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:61) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor$3.call(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:204) at org.gradle.internal.operations.DefaultBuildOperationRunner$CallableBuildOperationWorker.execute(DefaultBuildOperationRunner.java:199) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.call(DefaultBuildOperationRunner.java:53) at org.gradle.internal.operations.DefaultBuildOperationExecutor.call(DefaultBuildOperationExecutor.java:73) at org.gradle.launcher.exec.RunAsBuildOperationBuildActionExecutor.execute(RunAsBuildOperationBuildActionExecutor.java:57) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.lambda$execute$0(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.internal.work.DefaultWorkerLeaseService.withLocks(DefaultWorkerLeaseService.java:249) at org.gradle.internal.work.DefaultWorkerLeaseService.runAsWorkerThread(DefaultWorkerLeaseService.java:109) at org.gradle.launcher.exec.RunAsWorkerThreadBuildActionExecutor.execute(RunAsWorkerThreadBuildActionExecutor.java:36) at org.gradle.tooling.internal.provider.continuous.ContinuousBuildActionExecutor.execute(ContinuousBuildActionExecutor.java:110) at org.gradle.tooling.internal.provider.SubscribableBuildActionExecutor.execute(SubscribableBuildActionExecutor.java:64) at org.gradle.internal.session.DefaultBuildSessionContext.execute(DefaultBuildSessionContext.java:46) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:100) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter$ActionImpl.apply(BuildSessionLifecycleBuildActionExecuter.java:88) at org.gradle.internal.session.BuildSessionState.run(BuildSessionState.java:69) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:62) at org.gradle.tooling.internal.provider.BuildSessionLifecycleBuildActionExecuter.execute(BuildSessionLifecycleBuildActionExecuter.java:41) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:63) at org.gradle.tooling.internal.provider.StartParamsValidatingActionExecuter.execute(StartParamsValidatingActionExecuter.java:31) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:52) at org.gradle.tooling.internal.provider.SessionFailureReportingActionExecuter.execute(SessionFailureReportingActionExecuter.java:40) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:47) at org.gradle.tooling.internal.provider.SetupLoggingActionExecuter.execute(SetupLoggingActionExecuter.java:31) at org.gradle.launcher.daemon.server.exec.ExecuteBuild.doBuild(ExecuteBuild.java:65) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.WatchForDisconnection.execute(WatchForDisconnection.java:39) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ResetDeprecationLogger.execute(ResetDeprecationLogger.java:29) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.RequestStopIfSingleUsedDaemon.execute(RequestStopIfSingleUsedDaemon.java:35) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:78) at org.gradle.launcher.daemon.server.exec.ForwardClientInput$2.create(ForwardClientInput.java:75) at org.gradle.util.internal.Swapper.swap(Swapper.java:38) at org.gradle.launcher.daemon.server.exec.ForwardClientInput.execute(ForwardClientInput.java:75) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogAndCheckHealth.execute(LogAndCheckHealth.java:55) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.LogToClient.doBuild(LogToClient.java:63) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.EstablishBuildEnvironment.doBuild(EstablishBuildEnvironment.java:84) at org.gradle.launcher.daemon.server.exec.BuildCommandOnly.execute(BuildCommandOnly.java:37) at org.gradle.launcher.daemon.server.api.DaemonCommandExecution.proceed(DaemonCommandExecution.java:104) at org.gradle.launcher.daemon.server.exec.StartBuildOrRespondWithBusy$1.run(StartBuildOrRespondWithBusy.java:52) at org.gradle.launcher.daemon.server.DaemonStateCoordinator$1.run(DaemonStateCoordinator.java:297) at org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64) at org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:[48](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:49)) Caused by: com.fasterxml.jackson.databind.JsonMappingException: N/A at [Source: (File); line: 3, column: 12] (through reference chain: com.mooltiverse.oss.nyx.state.State["bump"]) at com.fasterxml.jackson.databind.JsonMappingException.from(JsonMappingException.java:276) at com.fasterxml.jackson.databind.deser.SettableBeanProperty._throwAsIOE(SettableBeanProperty.java:623) at com.fasterxml.jackson.databind.deser.SettableBeanProperty._throwAsIOE(SettableBeanProperty.java:611) at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:143) at com.fasterxml.jackson.databind.deser.BeanDeserializer.vanillaDeserialize(BeanDeserializer.java:313) at com.fasterxml.jackson.databind.deser.BeanDeserializer.deserialize(BeanDeserializer.java:176) at com.fasterxml.jackson.databind.deser.DefaultDeserializationContext.readRootValue(DefaultDeserializationContext.java:322) at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4675) at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3[49](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:50)5) at com.mooltiverse.oss.nyx.io.FileMapper.load(FileMapper.java:108) ... 190 more Caused by: java.lang.NullPointerException at com.mooltiverse.oss.nyx.state.State.setBump(State.java:2[52](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:53)) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:[62](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:63)) at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at com.fasterxml.jackson.databind.deser.impl.MethodProperty.deserializeAndSet(MethodProperty.java:141) ... 1[96](https://github.com/mooltiverse/nyx/actions/runs/3136282871/jobs/5093245154#step:7:97) more ```
non_infrastructure
exception when enabling the resume flag enabling resume throws an exception when unmarshalling the state file at a first glance the problem is inside the state etbump string bump method as it retrieves the configuration object reference from within the state getconfiguration getbump but object might still be null when unmarshalling the exception looks like org gradle internal event listenernotificationexception failed to notify build listener at org gradle internal event abstractbroadcastdispatch dispatch abstractbroadcastdispatch java at org gradle internal event broadcastdispatch compositedispatch dispatch broadcastdispatch java at org gradle internal event broadcastdispatch compositedispatch dispatch broadcastdispatch java at org gradle internal event listenerbroadcast dispatch listenerbroadcast java at org gradle internal event listenerbroadcast dispatch listenerbroadcast java at org gradle internal dispatch proxydispatchadapter dispatchinginvocationhandler invoke proxydispatchadapter java at com sun proxy projectsloaded unknown source at org gradle initialization notifyingbuildloader run notifyingbuildloader java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner run defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle initialization notifyingbuildloader load notifyingbuildloader java at org gradle configuration buildtreepreparingprojectspreparer prepareprojects buildtreepreparingprojectspreparer java at org gradle configuration buildoperationfiringprojectspreparer configurebuild run buildoperationfiringprojectspreparer java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner run defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle configuration buildoperationfiringprojectspreparer prepareprojects buildoperationfiringprojectspreparer java at org gradle initialization vintagebuildmodelcontroller lambda prepareprojects vintagebuildmodelcontroller java at org gradle internal model statetransitioncontroller lambda dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller lambda transitionifnotpreviously statetransitioncontroller java at org gradle internal work defaultsynchronizer withlock defaultsynchronizer java at org gradle internal model statetransitioncontroller transitionifnotpreviously statetransitioncontroller java at org gradle initialization vintagebuildmodelcontroller prepareprojects vintagebuildmodelcontroller java at org gradle initialization vintagebuildmodelcontroller preparetoscheduletasks vintagebuildmodelcontroller java at org gradle internal build defaultbuildlifecyclecontroller lambda preparetoscheduletasks defaultbuildlifecyclecontroller java at org gradle internal model statetransitioncontroller lambda dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller lambda maybetransition statetransitioncontroller java at org gradle internal work defaultsynchronizer withlock defaultsynchronizer java at org gradle internal model statetransitioncontroller maybetransition statetransitioncontroller java at org gradle internal build defaultbuildlifecyclecontroller preparetoscheduletasks defaultbuildlifecyclecontroller java at org gradle internal buildtree defaultbuildtreeworkpreparer schedulerequestedtasks defaultbuildtreeworkpreparer java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda doscheduleandruntasks defaultbuildtreelifecyclecontroller java at org gradle composite internal defaultincludedbuildtaskgraph withnewworkgraph defaultincludedbuildtaskgraph java at org gradle internal buildtree defaultbuildtreelifecyclecontroller doscheduleandruntasks defaultbuildtreelifecyclecontroller java at org gradle internal buildtree defaultbuildtreelifecyclecontroller lambda runbuild defaultbuildtreelifecyclecontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal model statetransitioncontroller dotransition statetransitioncontroller java at org gradle internal model statetransitioncontroller lambda transition statetransitioncontroller java at org gradle internal work defaultsynchronizer withlock defaultsynchronizer java at org gradle internal model statetransitioncontroller transition statetransitioncontroller java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor lambda execute rootbuildlifecyclebuildactionexecutor java at org gradle composite internal defaultrootbuildstate run defaultrootbuildstate java at org gradle launcher exec rootbuildlifecyclebuildactionexecutor execute rootbuildlifecyclebuildactionexecutor java at org gradle internal buildtree defaultbuildtreecontext execute defaultbuildtreecontext java at org gradle launcher exec buildtreelifecyclebuildactionexecutor lambda execute buildtreelifecyclebuildactionexecutor java at org gradle internal buildtree buildtreestate run buildtreestate java at org gradle launcher exec buildtreelifecyclebuildactionexecutor execute buildtreelifecyclebuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor call runasbuildoperationbuildactionexecutor java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner callablebuildoperationworker execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner call defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor call defaultbuildoperationexecutor java at org gradle launcher exec runasbuildoperationbuildactionexecutor execute runasbuildoperationbuildactionexecutor java at org gradle launcher exec runasworkerthreadbuildactionexecutor lambda execute runasworkerthreadbuildactionexecutor java at org gradle internal work defaultworkerleaseservice withlocks defaultworkerleaseservice java at org gradle internal work defaultworkerleaseservice runasworkerthread defaultworkerleaseservice java at org gradle launcher exec runasworkerthreadbuildactionexecutor execute runasworkerthreadbuildactionexecutor java at org gradle tooling internal provider continuous continuousbuildactionexecutor execute continuousbuildactionexecutor java at org gradle tooling internal provider subscribablebuildactionexecutor execute subscribablebuildactionexecutor java at org gradle internal session defaultbuildsessioncontext execute defaultbuildsessioncontext java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter actionimpl apply buildsessionlifecyclebuildactionexecuter java at org gradle internal session buildsessionstate run buildsessionstate java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider buildsessionlifecyclebuildactionexecuter execute buildsessionlifecyclebuildactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider startparamsvalidatingactionexecuter execute startparamsvalidatingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider sessionfailurereportingactionexecuter execute sessionfailurereportingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle tooling internal provider setuploggingactionexecuter execute setuploggingactionexecuter java at org gradle launcher daemon server exec executebuild dobuild executebuild java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec watchfordisconnection execute watchfordisconnection java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec resetdeprecationlogger execute resetdeprecationlogger java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec requeststopifsingleuseddaemon execute requeststopifsingleuseddaemon java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle launcher daemon server exec forwardclientinput create forwardclientinput java at org gradle util internal swapper swap swapper java at org gradle launcher daemon server exec forwardclientinput execute forwardclientinput java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logandcheckhealth execute logandcheckhealth java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec logtoclient dobuild logtoclient java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec establishbuildenvironment dobuild establishbuildenvironment java at org gradle launcher daemon server exec buildcommandonly execute buildcommandonly java at org gradle launcher daemon server api daemoncommandexecution proceed daemoncommandexecution java at org gradle launcher daemon server exec startbuildorrespondwithbusy run startbuildorrespondwithbusy java at org gradle launcher daemon server daemonstatecoordinator run daemonstatecoordinator java at org gradle internal concurrent executorpolicy catchandrecordfailures onexecute executorpolicy java at org gradle internal concurrent managedexecutorimpl run managedexecutorimpl java caused by com fasterxml jackson databind jsonmappingexception n a at through reference chain com mooltiverse oss nyx state state at com fasterxml jackson databind jsonmappingexception from jsonmappingexception java at com fasterxml jackson databind deser settablebeanproperty throwasioe settablebeanproperty java at com fasterxml jackson databind deser settablebeanproperty throwasioe settablebeanproperty java at com fasterxml jackson databind deser impl methodproperty deserializeandset methodproperty java at com fasterxml jackson databind deser beandeserializer vanilladeserialize beandeserializer java at com fasterxml jackson databind deser beandeserializer deserialize beandeserializer java at com fasterxml jackson databind deser defaultdeserializationcontext readrootvalue defaultdeserializationcontext java at com fasterxml jackson databind objectmapper readmapandclose objectmapper java at com fasterxml jackson databind objectmapper readvalue objectmapper java at com mooltiverse oss nyx io filemapper load filemapper java more caused by java lang nullpointerexception at com mooltiverse oss nyx state state setbump state java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at com fasterxml jackson databind deser impl methodproperty deserializeandset methodproperty java more
0
100,970
4,104,730,529
IssuesEvent
2016-06-05 15:32:26
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
opened
check that if user calls dmri.nii.gz dwi.nii.gz, there is no overwriting of file
priority: high sct_dmri_moco
because moco is supposed to output a file called dwi.nii.gz
1.0
check that if user calls dmri.nii.gz dwi.nii.gz, there is no overwriting of file - because moco is supposed to output a file called dwi.nii.gz
non_infrastructure
check that if user calls dmri nii gz dwi nii gz there is no overwriting of file because moco is supposed to output a file called dwi nii gz
0
22,467
15,211,005,481
IssuesEvent
2021-02-17 08:26:52
asciidoctor/asciidoctor-documentation-planning
https://api.github.com/repos/asciidoctor/asciidoctor-documentation-planning
closed
Update Release Notes Guide per feedback
content-edit infrastructure
- Fix link style in Credits section and mention to copy and paste template in raw and/or comment them out so they show up - Need to confirm that template is adequately covered by the MIT license or if a CC license should also be applied. - Need to choose between New and Noteworthy or Highlights or What's New - Need to choose between Additional Features and Changes or Minor Features and Changes or Other Features and Changes Once this issue is complete, spawn this issue: Need links to examples once the release notes for 0.1.4 are completed.
1.0
Update Release Notes Guide per feedback - - Fix link style in Credits section and mention to copy and paste template in raw and/or comment them out so they show up - Need to confirm that template is adequately covered by the MIT license or if a CC license should also be applied. - Need to choose between New and Noteworthy or Highlights or What's New - Need to choose between Additional Features and Changes or Minor Features and Changes or Other Features and Changes Once this issue is complete, spawn this issue: Need links to examples once the release notes for 0.1.4 are completed.
infrastructure
update release notes guide per feedback fix link style in credits section and mention to copy and paste template in raw and or comment them out so they show up need to confirm that template is adequately covered by the mit license or if a cc license should also be applied need to choose between new and noteworthy or highlights or what s new need to choose between additional features and changes or minor features and changes or other features and changes once this issue is complete spawn this issue need links to examples once the release notes for are completed
1
5,773
13,152,598,823
IssuesEvent
2020-08-09 23:15:59
docwhite/dreamdrugs
https://api.github.com/repos/docwhite/dreamdrugs
closed
Explicitly route socket.io traffic from nginx so just the correct headers are added.
architecture enhancement
Right now the react application and the sockets share the same location.
1.0
Explicitly route socket.io traffic from nginx so just the correct headers are added. - Right now the react application and the sockets share the same location.
non_infrastructure
explicitly route socket io traffic from nginx so just the correct headers are added right now the react application and the sockets share the same location
0
23,979
16,735,606,370
IssuesEvent
2021-06-11 00:31:49
noahpresler/semesterly
https://api.github.com/repos/noahpresler/semesterly
closed
Refactor OAUTH redirect URIs
Help Wanted Infrastructure Move to New Repo P0
Currently, everytime we add a school we must add the subdomain as a whitelisted URI on both the Facebook and Google OAUTH api. They do not accept wildcards like `*.semester.ly` This causes substantial issues, especially during rapid, open source expansion. ### To mitigate: Make `semester.ly` without a subdomain accept all of the OAUTH related requests and redirect the user. The oauth could pass a string parameter which might redirect the user for instance.
1.0
Refactor OAUTH redirect URIs - Currently, everytime we add a school we must add the subdomain as a whitelisted URI on both the Facebook and Google OAUTH api. They do not accept wildcards like `*.semester.ly` This causes substantial issues, especially during rapid, open source expansion. ### To mitigate: Make `semester.ly` without a subdomain accept all of the OAUTH related requests and redirect the user. The oauth could pass a string parameter which might redirect the user for instance.
infrastructure
refactor oauth redirect uris currently everytime we add a school we must add the subdomain as a whitelisted uri on both the facebook and google oauth api they do not accept wildcards like semester ly this causes substantial issues especially during rapid open source expansion to mitigate make semester ly without a subdomain accept all of the oauth related requests and redirect the user the oauth could pass a string parameter which might redirect the user for instance
1
817,951
30,664,385,377
IssuesEvent
2023-07-25 17:08:52
episphere/connect
https://api.github.com/repos/episphere/connect
closed
Restrict Accession ID Page to Numeric Values Only
enhancement Biospecimen Dashboard Clinical Priority 1 Backend
On the Accession ID linking page in the clinical biospecimen dashboard, make the scanning fields for both the blood and urine Accession ID only accept numeric values; The fields should **not** accept string data.
1.0
Restrict Accession ID Page to Numeric Values Only - On the Accession ID linking page in the clinical biospecimen dashboard, make the scanning fields for both the blood and urine Accession ID only accept numeric values; The fields should **not** accept string data.
non_infrastructure
restrict accession id page to numeric values only on the accession id linking page in the clinical biospecimen dashboard make the scanning fields for both the blood and urine accession id only accept numeric values the fields should not accept string data
0
31,256
25,485,532,648
IssuesEvent
2022-11-26 10:40:07
nf-core/nf-co.re
https://api.github.com/repos/nf-core/nf-co.re
closed
Make listing pages for /usage and /developers
bug infrastructure
Currently https://nf-co.re/usage and https://nf-co.re/developers give a 404 error. It would be nice if these could (automatically?) show the sub-page documentation under those sections.
1.0
Make listing pages for /usage and /developers - Currently https://nf-co.re/usage and https://nf-co.re/developers give a 404 error. It would be nice if these could (automatically?) show the sub-page documentation under those sections.
infrastructure
make listing pages for usage and developers currently and give a error it would be nice if these could automatically show the sub page documentation under those sections
1
5,313
5,560,479,878
IssuesEvent
2017-03-24 19:34:55
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Clean-up binplacing targets
area-Infrastructure enhancement
We need to remove binplacing.targets and merge a more complete bin-placing logic somewhere into the build tools targets (perhaps FrameworkTargeting.targets). We will likely need to have a build task to produce the set of directories to binplace to. I've got a start of some of that work at https://github.com/weshaggard/corefx/tree/BinPlacingTask.
1.0
Clean-up binplacing targets - We need to remove binplacing.targets and merge a more complete bin-placing logic somewhere into the build tools targets (perhaps FrameworkTargeting.targets). We will likely need to have a build task to produce the set of directories to binplace to. I've got a start of some of that work at https://github.com/weshaggard/corefx/tree/BinPlacingTask.
infrastructure
clean up binplacing targets we need to remove binplacing targets and merge a more complete bin placing logic somewhere into the build tools targets perhaps frameworktargeting targets we will likely need to have a build task to produce the set of directories to binplace to i ve got a start of some of that work at
1
12,483
9,805,728,847
IssuesEvent
2019-06-12 09:38:41
maidsafe/safe-build-infrastructure
https://api.github.com/repos/maidsafe/safe-build-infrastructure
closed
SSL Certificate Not Working Correctly on Re-provision
bug infrastructure
We have seen that when an environment is re-provisioned, the SSL certificate doesn't seem to work correctly. On prod it appears like an unsigned SSL certificate, and on staging with the self-signed cert, you don't seem to be able to hit the environment. I did reproduce this on the staging environment, but I don't yet know what causes it.
1.0
SSL Certificate Not Working Correctly on Re-provision - We have seen that when an environment is re-provisioned, the SSL certificate doesn't seem to work correctly. On prod it appears like an unsigned SSL certificate, and on staging with the self-signed cert, you don't seem to be able to hit the environment. I did reproduce this on the staging environment, but I don't yet know what causes it.
infrastructure
ssl certificate not working correctly on re provision we have seen that when an environment is re provisioned the ssl certificate doesn t seem to work correctly on prod it appears like an unsigned ssl certificate and on staging with the self signed cert you don t seem to be able to hit the environment i did reproduce this on the staging environment but i don t yet know what causes it
1
126,013
10,373,320,079
IssuesEvent
2019-09-09 06:55:57
widelands/widelands-issue-migration2
https://api.github.com/repos/widelands/widelands-issue-migration2
opened
input_queues test will often stall
Fix Committed High economy test
We often get this on Travis: test/maps/plain.wmf/scripting/test_inputqueues.lua ... Running Widelands ... done. Loading savegame: inputqueues ... No output has been received in the last 10m0s, this potentially indicates a stalled build or something wrong with the build itself. Check the details on how to adjust your build configuration on: https://docs.travis-ci.com/user/common-build-problems/#Build-times-out-because-no-output-was-received We should investigate.
1.0
input_queues test will often stall - We often get this on Travis: test/maps/plain.wmf/scripting/test_inputqueues.lua ... Running Widelands ... done. Loading savegame: inputqueues ... No output has been received in the last 10m0s, this potentially indicates a stalled build or something wrong with the build itself. Check the details on how to adjust your build configuration on: https://docs.travis-ci.com/user/common-build-problems/#Build-times-out-because-no-output-was-received We should investigate.
non_infrastructure
input queues test will often stall we often get this on travis test maps plain wmf scripting test inputqueues lua running widelands done loading savegame inputqueues no output has been received in the last this potentially indicates a stalled build or something wrong with the build itself check the details on how to adjust your build configuration on we should investigate
0